Spaces:
Running
Running
Bump, pull in generated configs from personal setup
Browse files- Dockerfile +4 -4
- azure-models.txt +0 -27
- generated/anthropic.yaml +24 -0
- generated/github.yaml +56 -0
- generated/groq.yaml +8 -0
- generated/huggingface.yaml +15 -0
- generated/openai.yaml +32 -0
- generated/perplexity.yaml +12 -0
- start.sh +18 -26
Dockerfile
CHANGED
@@ -1,16 +1,16 @@
|
|
1 |
# Building from source is too slow, so let's just customize the existing image
|
2 |
-
FROM ghcr.io/open-webui/open-webui:git-
|
3 |
|
4 |
WORKDIR /app
|
5 |
|
6 |
USER 0:0
|
7 |
|
8 |
# HACK for huggingface.co iframe
|
9 |
-
RUN sed -i "s|samesite=WEBUI_SESSION_COOKIE_SAME_SITE|samesite='none'|g" backend/open_webui/
|
10 |
|
11 |
-
RUN pip install "litellm[proxy]==1.
|
12 |
USER 1000:0
|
13 |
|
14 |
-
COPY ./
|
15 |
COPY ./start.sh /start.sh
|
16 |
CMD [ "bash", "/start.sh" ]
|
|
|
1 |
# Building from source is too slow, so let's just customize the existing image
|
2 |
+
FROM ghcr.io/open-webui/open-webui:git-b72150c
|
3 |
|
4 |
WORKDIR /app
|
5 |
|
6 |
USER 0:0
|
7 |
|
8 |
# HACK for huggingface.co iframe
|
9 |
+
RUN sed -i "s|samesite=WEBUI_SESSION_COOKIE_SAME_SITE|samesite='none'|g" backend/open_webui/routers/auths.py
|
10 |
|
11 |
+
RUN pip install "litellm[proxy]==1.59.9" && chown -R 1000:0 /app
|
12 |
USER 1000:0
|
13 |
|
14 |
+
COPY ./generated /assets/
|
15 |
COPY ./start.sh /start.sh
|
16 |
CMD [ "bash", "/start.sh" ]
|
azure-models.txt
DELETED
@@ -1,27 +0,0 @@
|
|
1 |
-
AI21-Jamba-Instruct
|
2 |
-
Cohere-command-r
|
3 |
-
Cohere-command-r-plus
|
4 |
-
Cohere-embed-v3-english
|
5 |
-
Cohere-embed-v3-multilingual
|
6 |
-
Meta-Llama-3-70B-Instruct
|
7 |
-
Meta-Llama-3-8B-Instruct
|
8 |
-
Meta-Llama-3.1-405B-Instruct
|
9 |
-
Meta-Llama-3.1-70B-Instruct
|
10 |
-
Meta-Llama-3.1-8B-Instruct
|
11 |
-
Mistral-large
|
12 |
-
Mistral-large-2407
|
13 |
-
Mistral-Nemo
|
14 |
-
Mistral-small
|
15 |
-
gpt-4o
|
16 |
-
gpt-4o-mini
|
17 |
-
text-embedding-3-large
|
18 |
-
text-embedding-3-small
|
19 |
-
Phi-3-medium-128k-instruct
|
20 |
-
Phi-3-medium-4k-instruct
|
21 |
-
Phi-3-mini-128k-instruct
|
22 |
-
Phi-3-mini-4k-instruct
|
23 |
-
Phi-3-small-128k-instruct
|
24 |
-
Phi-3-small-8k-instruct
|
25 |
-
Phi-3.5-mini-instruct
|
26 |
-
o1-preview
|
27 |
-
o1-mini
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
generated/anthropic.yaml
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- litellm_params:
|
2 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
3 |
+
model: anthropic/claude-3-5-haiku-latest
|
4 |
+
model_name: anthropic/claude-3-5-haiku
|
5 |
+
- litellm_params:
|
6 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
7 |
+
model: anthropic/claude-3-5-sonnet-latest
|
8 |
+
model_name: anthropic/claude-3-5-sonnet
|
9 |
+
- litellm_params:
|
10 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
11 |
+
model: anthropic/claude-3-5-opus-latest
|
12 |
+
model_name: anthropic/claude-3-5-opus
|
13 |
+
- litellm_params:
|
14 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
15 |
+
model: anthropic/claude-3-haiku-latest
|
16 |
+
model_name: anthropic/claude-3-haiku
|
17 |
+
- litellm_params:
|
18 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
19 |
+
model: anthropic/claude-3-sonnet-latest
|
20 |
+
model_name: anthropic/claude-3-sonnet
|
21 |
+
- litellm_params:
|
22 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
23 |
+
model: anthropic/claude-3-opus-latest
|
24 |
+
model_name: anthropic/claude-3-opus
|
generated/github.yaml
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- litellm_params:
|
2 |
+
api_key: os.environ/GITHUB_API_KEY
|
3 |
+
model: github/DeepSeek-R1
|
4 |
+
model_name: deepseek/r1
|
5 |
+
- litellm_params:
|
6 |
+
api_key: os.environ/GITHUB_API_KEY
|
7 |
+
model: github/gpt-4o
|
8 |
+
model_name: openai/gpt-4o
|
9 |
+
- litellm_params:
|
10 |
+
api_key: os.environ/GITHUB_API_KEY
|
11 |
+
model: github/gpt-4o-mini
|
12 |
+
model_name: openai/gpt-4o-mini
|
13 |
+
- litellm_params:
|
14 |
+
api_key: os.environ/GITHUB_API_KEY
|
15 |
+
model: github/o1
|
16 |
+
model_name: openai/o1
|
17 |
+
- litellm_params:
|
18 |
+
api_key: os.environ/GITHUB_API_KEY
|
19 |
+
model: github/o1-mini
|
20 |
+
model_name: openai/o1-mini
|
21 |
+
- litellm_params:
|
22 |
+
api_key: os.environ/GITHUB_API_KEY
|
23 |
+
model: github/text-embedding-3-small
|
24 |
+
model_name: openai/text-embedding-3-small
|
25 |
+
- litellm_params:
|
26 |
+
api_key: os.environ/GITHUB_API_KEY
|
27 |
+
model: github/text-embedding-3-large
|
28 |
+
model_name: openai/text-embedding-3-large
|
29 |
+
- litellm_params:
|
30 |
+
api_key: os.environ/GITHUB_API_KEY
|
31 |
+
model: github/Mistral-small
|
32 |
+
model_name: mistral/small
|
33 |
+
- litellm_params:
|
34 |
+
api_key: os.environ/GITHUB_API_KEY
|
35 |
+
model: github/Mistral-large
|
36 |
+
model_name: mistral/large
|
37 |
+
- litellm_params:
|
38 |
+
api_key: os.environ/GITHUB_API_KEY
|
39 |
+
model: github/Mistral-Nemo
|
40 |
+
model_name: mistral/nemo
|
41 |
+
- litellm_params:
|
42 |
+
api_key: os.environ/GITHUB_API_KEY
|
43 |
+
model: github/Cohere-command-r
|
44 |
+
model_name: cohere/command-r
|
45 |
+
- litellm_params:
|
46 |
+
api_key: os.environ/GITHUB_API_KEY
|
47 |
+
model: github/Cohere-command-r-plus
|
48 |
+
model_name: cohere/command-r-plus
|
49 |
+
- litellm_params:
|
50 |
+
api_key: os.environ/GITHUB_API_KEY
|
51 |
+
model: github/Meta-Llama-3-70B-Instruct
|
52 |
+
model_name: meta/llama-3.3-70b
|
53 |
+
- litellm_params:
|
54 |
+
api_key: os.environ/GITHUB_API_KEY
|
55 |
+
model: github/Meta-Llama-3-8B-Instruct
|
56 |
+
model_name: meta/llama-3.3-8b
|
generated/groq.yaml
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- litellm_params:
|
2 |
+
api_key: os.environ/GROQ_API_KEY
|
3 |
+
model: groq/llama-3.3-70b-versatile
|
4 |
+
model_name: meta/llama-3.3-70b
|
5 |
+
- litellm_params:
|
6 |
+
api_key: os.environ/GROQ_API_KEY
|
7 |
+
model: groq/deepseek-r1-distill-llama-70b
|
8 |
+
model_name: deepseek/r1-distill-llama-70b
|
generated/huggingface.yaml
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- litellm_params:
|
2 |
+
model: huggingface/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
3 |
+
model_name: deepseek/r1-distill-qwen-32b
|
4 |
+
- litellm_params:
|
5 |
+
model: huggingface/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
6 |
+
model_name: deepseek/r1-distill-qwen-1.5b
|
7 |
+
- litellm_params:
|
8 |
+
model: huggingface/google/gemma-2-2b-it
|
9 |
+
model_name: google/gemma-2-2b-it
|
10 |
+
- litellm_params:
|
11 |
+
model: huggingface/google/gemma-2-9b-it
|
12 |
+
model_name: google/gemma-2-9b-it
|
13 |
+
- litellm_params:
|
14 |
+
model: huggingface/google/gemma-2-27b-it
|
15 |
+
model_name: google/gemma-2-27b-it
|
generated/openai.yaml
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- litellm_params:
|
2 |
+
api_key: os.environ/OPENAI_API_KEY
|
3 |
+
model: openai/gpt-4o
|
4 |
+
model_name: openai/gpt-4o
|
5 |
+
- litellm_params:
|
6 |
+
api_key: os.environ/OPENAI_API_KEY
|
7 |
+
model: openai/gpt-4o-mini
|
8 |
+
model_name: openai/gpt-4o-mini
|
9 |
+
- litellm_params:
|
10 |
+
api_key: os.environ/OPENAI_API_KEY
|
11 |
+
model: openai/o1-mini
|
12 |
+
model_name: openai/o1-mini
|
13 |
+
- litellm_params:
|
14 |
+
api_key: os.environ/OPENAI_API_KEY
|
15 |
+
model: openai/o1-preview
|
16 |
+
model_name: openai/o1-preview
|
17 |
+
- litellm_params:
|
18 |
+
api_key: os.environ/OPENAI_API_KEY
|
19 |
+
model: openai/gpt-3.5-turbo
|
20 |
+
model_name: openai/gpt-3.5-turbo
|
21 |
+
- litellm_params:
|
22 |
+
api_key: os.environ/OPENAI_API_KEY
|
23 |
+
model: openai/text-embedding-3-small
|
24 |
+
model_name: openai/text-embedding-3-small
|
25 |
+
- litellm_params:
|
26 |
+
api_key: os.environ/OPENAI_API_KEY
|
27 |
+
model: openai/text-embedding-3-large
|
28 |
+
model_name: openai/text-embedding-3-large
|
29 |
+
- litellm_params:
|
30 |
+
api_key: os.environ/OPENAI_API_KEY
|
31 |
+
model: openai/dall-e-3
|
32 |
+
model_name: openai/dall-e-3
|
generated/perplexity.yaml
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- litellm_params:
|
2 |
+
api_key: os.environ/PERPLEXITY_API_KEY
|
3 |
+
model: perplexity/sonar-pro
|
4 |
+
model_name: perplexity/sonar-pro
|
5 |
+
- litellm_params:
|
6 |
+
api_key: os.environ/PERPLEXITY_API_KEY
|
7 |
+
model: perplexity/sonar-reasoning
|
8 |
+
model_name: perplexity/sonar-reasoning
|
9 |
+
- litellm_params:
|
10 |
+
api_key: os.environ/PERPLEXITY_API_KEY
|
11 |
+
model: perplexity/sonar
|
12 |
+
model_name: perplexity/sonar
|
start.sh
CHANGED
@@ -3,42 +3,34 @@ set -euo pipefail
|
|
3 |
|
4 |
pids=()
|
5 |
|
6 |
-
|
7 |
-
|
8 |
-
key_env="$2"
|
9 |
|
10 |
-
if [[
|
11 |
-
>&2 echo "[
|
12 |
-
cat
|
13 |
-
|
14 |
-
litellm_params:
|
15 |
-
model: "${model}"
|
16 |
-
api_key: "os.environ/${key_env}"
|
17 |
-
EOF
|
18 |
fi
|
19 |
-
}
|
20 |
|
21 |
-
generate_litellm_provider() {
|
22 |
-
provider="$1"
|
23 |
key_env="$2"
|
24 |
|
25 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
}
|
27 |
|
28 |
generate_litellm_config() {
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
EOF
|
35 |
generate_litellm_provider "openai" "OPENAI_API_KEY"
|
36 |
generate_litellm_provider "perplexity" "PERPLEXITY_API_KEY"
|
37 |
-
generate_litellm_provider "anthropic" "ANTHROPIC_API_KEY"
|
38 |
-
|
39 |
-
while read -r model; do
|
40 |
-
generate_litellm_model "github/${model}" "GITHUB_API_KEY"
|
41 |
-
done </assets/azure-models.txt
|
42 |
|
43 |
if [[ -n "${LITELLM_MODELS_BASE64:-}" ]]; then
|
44 |
echo "${LITELLM_MODELS_BASE64}" | base64 -d
|
|
|
3 |
|
4 |
pids=()
|
5 |
|
6 |
+
generate_litellm_provider() {
|
7 |
+
provider="$1"
|
|
|
8 |
|
9 |
+
if [[ "$#" -lt 2 ]]; then
|
10 |
+
>&2 echo "[+] Provider ${provider}: No key needed"
|
11 |
+
cat "/assets/${provider}.yaml"
|
12 |
+
return
|
|
|
|
|
|
|
|
|
13 |
fi
|
|
|
14 |
|
|
|
|
|
15 |
key_env="$2"
|
16 |
|
17 |
+
if [[ -n "${!key_env:-}" ]]; then
|
18 |
+
>&2 echo "[+] Provider ${provider}: Found key ${key_env}"
|
19 |
+
cat "/assets/${provider}.yaml"
|
20 |
+
return
|
21 |
+
fi
|
22 |
+
|
23 |
+
>&2 echo "[!] Provider ${provider}: No key found (set ${key_env})"
|
24 |
}
|
25 |
|
26 |
generate_litellm_config() {
|
27 |
+
echo "model_list:"
|
28 |
+
generate_litellm_provider "anthropic" "ANTHROPIC_API_KEY"
|
29 |
+
generate_litellm_provider "github" "GITHUB_API_KEY"
|
30 |
+
generate_litellm_provider "groq" "GROQ_API_KEY"
|
31 |
+
generate_litellm_provider "huggingface"
|
|
|
32 |
generate_litellm_provider "openai" "OPENAI_API_KEY"
|
33 |
generate_litellm_provider "perplexity" "PERPLEXITY_API_KEY"
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
if [[ -n "${LITELLM_MODELS_BASE64:-}" ]]; then
|
36 |
echo "${LITELLM_MODELS_BASE64}" | base64 -d
|