Spaces:
Sleeping
Sleeping
Update config.yaml
Browse files- config.yaml +41 -14
config.yaml
CHANGED
@@ -1,17 +1,44 @@
|
|
1 |
-
general_settings:
|
2 |
-
master_key: sk-1234567890
|
3 |
-
|
4 |
-
litellm_settings:
|
5 |
-
drop_params: True
|
6 |
-
set_verbose: True
|
7 |
-
|
8 |
model_list:
|
9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
|
|
|
|
|
|
|
11 |
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
api_base: https://openrouter.ai/api/v1
|
16 |
-
api_key: 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 # does os.getenv("AZURE_API_KEY_EU")
|
17 |
-
rpm: 60 # [OPTIONAL] Rate limit for this deployment: in requests per minute (rpm)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
model_list:
|
2 |
+
- model_name: gpt-3.5-turbo ### RECEIVED MODEL NAME ###
|
3 |
+
litellm_params: # all params accepted by litellm.completion() - https://docs.litellm.ai/docs/completion/input
|
4 |
+
model: azure/gpt-turbo-small-eu ### MODEL NAME sent to `litellm.completion()` ###
|
5 |
+
api_base: https://my-endpoint-europe-berri-992.openai.azure.com/
|
6 |
+
api_key: "os.environ/AZURE_API_KEY_EU" # does os.getenv("AZURE_API_KEY_EU")
|
7 |
+
rpm: 6 # [OPTIONAL] Rate limit for this deployment: in requests per minute (rpm)
|
8 |
+
- model_name: bedrock-claude-v1
|
9 |
+
litellm_params:
|
10 |
+
model: bedrock/anthropic.claude-instant-v1
|
11 |
+
- model_name: gpt-3.5-turbo
|
12 |
+
litellm_params:
|
13 |
+
model: azure/gpt-turbo-small-ca
|
14 |
+
api_base: https://my-endpoint-canada-berri992.openai.azure.com/
|
15 |
+
api_key: "os.environ/AZURE_API_KEY_CA"
|
16 |
+
rpm: 6
|
17 |
+
- model_name: anthropic-claude
|
18 |
+
litellm_params:
|
19 |
+
model: bedrock/anthropic.claude-instant-v1
|
20 |
+
### [OPTIONAL] SET AWS REGION ###
|
21 |
+
aws_region_name: us-east-1
|
22 |
+
- model_name: vllm-models
|
23 |
+
litellm_params:
|
24 |
+
model: openai/facebook/opt-125m # the `openai/` prefix tells litellm it's openai compatible
|
25 |
+
api_base: http://0.0.0.0:4000/v1
|
26 |
+
api_key: none
|
27 |
+
rpm: 1440
|
28 |
+
model_info:
|
29 |
+
version: 2
|
30 |
+
|
31 |
+
# Use this if you want to make requests to `claude-3-haiku-20240307`,`claude-3-opus-20240229`,`claude-2.1` without defining them on the config.yaml
|
32 |
+
# Default models
|
33 |
+
# Works for ALL Providers and needs the default provider credentials in .env
|
34 |
+
- model_name: "*"
|
35 |
+
litellm_params:
|
36 |
+
model: "*"
|
37 |
|
38 |
+
litellm_settings: # module level litellm settings - https://github.com/BerriAI/litellm/blob/main/litellm/__init__.py
|
39 |
+
drop_params: True
|
40 |
+
success_callback: ["langfuse"] # OPTIONAL - if you want to start sending LLM Logs to Langfuse. Make sure to set `LANGFUSE_PUBLIC_KEY` and `LANGFUSE_SECRET_KEY` in your env
|
41 |
|
42 |
+
general_settings:
|
43 |
+
master_key: sk-1234 # [OPTIONAL] Only use this if you to require all calls to contain this key (Authorization: Bearer sk-1234)
|
44 |
+
alerting: ["slack"] # [OPTIONAL] If you want Slack Alerts for Hanging LLM requests, Slow llm responses, Budget Alerts. Make sure to set `SLACK_WEBHOOK_URL` in your env
|
|
|
|
|
|