reztilop commited on
Commit
0c5567e
·
verified ·
1 Parent(s): f5eb2b6

Update config.yaml

Browse files
Files changed (1) hide show
  1. config.yaml +41 -14
config.yaml CHANGED
@@ -1,17 +1,44 @@
1
- general_settings:
2
- master_key: sk-1234567890
3
-
4
- litellm_settings:
5
- drop_params: True
6
- set_verbose: True
7
-
8
  model_list:
9
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
 
 
 
11
 
12
- - model_name: nousresearch/hermes-3-llama-3.1-405b:free
13
- litellm_params: # all params accepted by litellm.completion() - https://docs.litellm.ai/docs/completion/input
14
- model: nousresearch/hermes-3-llama-3.1-405b:free ### MODEL NAME sent to `litellm.completion()` ###
15
- api_base: https://openrouter.ai/api/v1
16
- api_key: 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 # does os.getenv("AZURE_API_KEY_EU")
17
- rpm: 60 # [OPTIONAL] Rate limit for this deployment: in requests per minute (rpm)
 
 
 
 
 
 
 
 
1
  model_list:
2
+ - model_name: gpt-3.5-turbo ### RECEIVED MODEL NAME ###
3
+ litellm_params: # all params accepted by litellm.completion() - https://docs.litellm.ai/docs/completion/input
4
+ model: azure/gpt-turbo-small-eu ### MODEL NAME sent to `litellm.completion()` ###
5
+ api_base: https://my-endpoint-europe-berri-992.openai.azure.com/
6
+ api_key: "os.environ/AZURE_API_KEY_EU" # does os.getenv("AZURE_API_KEY_EU")
7
+ rpm: 6 # [OPTIONAL] Rate limit for this deployment: in requests per minute (rpm)
8
+ - model_name: bedrock-claude-v1
9
+ litellm_params:
10
+ model: bedrock/anthropic.claude-instant-v1
11
+ - model_name: gpt-3.5-turbo
12
+ litellm_params:
13
+ model: azure/gpt-turbo-small-ca
14
+ api_base: https://my-endpoint-canada-berri992.openai.azure.com/
15
+ api_key: "os.environ/AZURE_API_KEY_CA"
16
+ rpm: 6
17
+ - model_name: anthropic-claude
18
+ litellm_params:
19
+ model: bedrock/anthropic.claude-instant-v1
20
+ ### [OPTIONAL] SET AWS REGION ###
21
+ aws_region_name: us-east-1
22
+ - model_name: vllm-models
23
+ litellm_params:
24
+ model: openai/facebook/opt-125m # the `openai/` prefix tells litellm it's openai compatible
25
+ api_base: http://0.0.0.0:4000/v1
26
+ api_key: none
27
+ rpm: 1440
28
+ model_info:
29
+ version: 2
30
+
31
+ # Use this if you want to make requests to `claude-3-haiku-20240307`,`claude-3-opus-20240229`,`claude-2.1` without defining them on the config.yaml
32
+ # Default models
33
+ # Works for ALL Providers and needs the default provider credentials in .env
34
+ - model_name: "*"
35
+ litellm_params:
36
+ model: "*"
37
 
38
+ litellm_settings: # module level litellm settings - https://github.com/BerriAI/litellm/blob/main/litellm/__init__.py
39
+ drop_params: True
40
+ success_callback: ["langfuse"] # OPTIONAL - if you want to start sending LLM Logs to Langfuse. Make sure to set `LANGFUSE_PUBLIC_KEY` and `LANGFUSE_SECRET_KEY` in your env
41
 
42
+ general_settings:
43
+ master_key: sk-1234 # [OPTIONAL] Only use this if you to require all calls to contain this key (Authorization: Bearer sk-1234)
44
+ alerting: ["slack"] # [OPTIONAL] If you want Slack Alerts for Hanging LLM requests, Slow llm responses, Budget Alerts. Make sure to set `SLACK_WEBHOOK_URL` in your env