Generation/Completion/Chat Completion Models
OpenAI Chat Completion Models
Model Name | Function Call | Required OS Variables |
---|---|---|
gpt-3.5-turbo | completion('gpt-3.5-turbo', messages) |
os.environ['OPENAI_API_KEY'] |
gpt-3.5-turbo-16k | completion('gpt-3.5-turbo-16k', messages) |
os.environ['OPENAI_API_KEY'] |
gpt-3.5-turbo-16k-0613 | completion('gpt-3.5-turbo-16k-0613', messages) |
os.environ['OPENAI_API_KEY'] |
gpt-4 | completion('gpt-4', messages) |
os.environ['OPENAI_API_KEY'] |
Azure OpenAI Chat Completion Models
For Azure calls add the azure/
prefix to model
. If your azure deployment name is gpt-v-2
set model
= azure/gpt-v-2
Model Name | Function Call | Required OS Variables |
---|---|---|
gpt-3.5-turbo | completion('azure/gpt-3.5-turbo-deployment', messages) |
os.environ['AZURE_API_KEY'] ,os.environ['AZURE_API_BASE'] ,os.environ['AZURE_API_VERSION'] |
gpt-4 | completion('azure/gpt-4-deployment', messages) |
os.environ['AZURE_API_KEY'] ,os.environ['AZURE_API_BASE'] ,os.environ['AZURE_API_VERSION'] |
OpenAI Text Completion Models
Model Name | Function Call | Required OS Variables |
---|---|---|
text-davinci-003 | completion('text-davinci-003', messages) |
os.environ['OPENAI_API_KEY'] |
Cohere Models
Model Name | Function Call | Required OS Variables |
---|---|---|
command-nightly | completion('command-nightly', messages) |
os.environ['COHERE_API_KEY'] |
Anthropic Models
Model Name | Function Call | Required OS Variables |
---|---|---|
claude-instant-1 | completion('claude-instant-1', messages) |
os.environ['ANTHROPIC_API_KEY'] |
claude-2 | completion('claude-2', messages) |
os.environ['ANTHROPIC_API_KEY'] |
Hugging Face Inference API
All text2text-generation
and text-generation
models are supported by liteLLM. You can use any text model from Hugging Face with the following steps:
- Copy the
model repo
URL from Hugging Face and set it as themodel
parameter in the completion call. - Set
hugging_face
parameter toTrue
. - Make sure to set the hugging face API key
Here are some examples of supported models: Note that the models mentioned in the table are examples, and you can use any text model available on Hugging Face by following the steps above.
Model Name | Function Call | Required OS Variables |
---|---|---|
stabilityai/stablecode-completion-alpha-3b-4k | completion(model="stabilityai/stablecode-completion-alpha-3b-4k", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
bigcode/starcoder | completion(model="bigcode/starcoder", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
google/flan-t5-xxl | completion(model="google/flan-t5-xxl", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
google/flan-t5-large | completion(model="google/flan-t5-large", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
OpenRouter Completion Models
All the text models from OpenRouter are supported by liteLLM.
Model Name | Function Call | Required OS Variables |
---|---|---|
openai/gpt-3.5-turbo | completion('openai/gpt-3.5-turbo', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
openai/gpt-3.5-turbo-16k | completion('openai/gpt-3.5-turbo-16k', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
openai/gpt-4 | completion('openai/gpt-4', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
openai/gpt-4-32k | completion('openai/gpt-4-32k', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
anthropic/claude-2 | completion('anthropic/claude-2', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
anthropic/claude-instant-v1 | completion('anthropic/claude-instant-v1', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
google/palm-2-chat-bison | completion('google/palm-2-chat-bison', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
google/palm-2-codechat-bison | completion('google/palm-2-codechat-bison', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
meta-llama/llama-2-13b-chat | completion('meta-llama/llama-2-13b-chat', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |
meta-llama/llama-2-70b-chat | completion('meta-llama/llama-2-70b-chat', messages) |
os.environ['OR_SITE_URL'] ,os.environ['OR_APP_NAME'] ,os.environ['OR_API_KEY'] |