You can set the settings.llm
with any LangChain ChatModel.
from langchain_openai.chat_models import AzureChatOpenAI
settings.llm = AzureChatOpenAI(...)
You can also set the settings.llm
with a string identifier of a ChatModel including local models.
<provider>/<model_name>:<optional_label>
openai
: OpenAI Chat Modelsllamacpp
: Run local models directly using llamacpp (alias:thebloke
,gguf
)ollama
: Run local models through Ollama (wrapper for llamacpp)azure
: Azure Chat Modelsanthropic
: Anthropic Chat Modelsgoogle
: Google Chat Models
openai/gpt-3.5-turbo
: ChatGPT Classicopenai/gpt-4-1106-preview
: GPT-4-Turboollama/openchat
: OpenChat3.5-1210ollama/openhermes2.5-mistral
: OpenHermes 2.5llamacpp/openchat-3.5-1210
: OpenChat3.5-1210TheBloke/Nous-Hermes-2-SOLAR-10.7B-GGUF
: alias forllamacpp/...
TheBloke/openchat-3.5-0106-GGUF:Q3_K_L
: with Q label
Checkout the file src/funcchain/model/defaults.py
for the code that parses the string identifier.
Feel free to create a PR to add more models to the defaults. Or tell me how wrong I am and create a better system.