mr06cpp commited on
Commit
453e653
·
1 Parent(s): 1b69db8

Bump, pull in generated configs from personal setup

Browse files
Dockerfile CHANGED
@@ -1,16 +1,16 @@
1
  # Building from source is too slow, so let's just customize the existing image
2
- FROM ghcr.io/open-webui/open-webui:git-7228b39
3
 
4
  WORKDIR /app
5
 
6
  USER 0:0
7
 
8
  # HACK for huggingface.co iframe
9
- RUN sed -i "s|samesite=WEBUI_SESSION_COOKIE_SAME_SITE|samesite='none'|g" backend/open_webui/apps/webui/routers/auths.py
10
 
11
- RUN pip install "litellm[proxy]==1.51.2" && chown -R 1000:0 /app
12
  USER 1000:0
13
 
14
- COPY ./azure-models.txt /assets/azure-models.txt
15
  COPY ./start.sh /start.sh
16
  CMD [ "bash", "/start.sh" ]
 
1
  # Building from source is too slow, so let's just customize the existing image
2
+ FROM ghcr.io/open-webui/open-webui:git-b72150c
3
 
4
  WORKDIR /app
5
 
6
  USER 0:0
7
 
8
  # HACK for huggingface.co iframe
9
+ RUN sed -i "s|samesite=WEBUI_SESSION_COOKIE_SAME_SITE|samesite='none'|g" backend/open_webui/routers/auths.py
10
 
11
+ RUN pip install "litellm[proxy]==1.59.9" && chown -R 1000:0 /app
12
  USER 1000:0
13
 
14
+ COPY ./generated /assets/
15
  COPY ./start.sh /start.sh
16
  CMD [ "bash", "/start.sh" ]
azure-models.txt DELETED
@@ -1,27 +0,0 @@
1
- AI21-Jamba-Instruct
2
- Cohere-command-r
3
- Cohere-command-r-plus
4
- Cohere-embed-v3-english
5
- Cohere-embed-v3-multilingual
6
- Meta-Llama-3-70B-Instruct
7
- Meta-Llama-3-8B-Instruct
8
- Meta-Llama-3.1-405B-Instruct
9
- Meta-Llama-3.1-70B-Instruct
10
- Meta-Llama-3.1-8B-Instruct
11
- Mistral-large
12
- Mistral-large-2407
13
- Mistral-Nemo
14
- Mistral-small
15
- gpt-4o
16
- gpt-4o-mini
17
- text-embedding-3-large
18
- text-embedding-3-small
19
- Phi-3-medium-128k-instruct
20
- Phi-3-medium-4k-instruct
21
- Phi-3-mini-128k-instruct
22
- Phi-3-mini-4k-instruct
23
- Phi-3-small-128k-instruct
24
- Phi-3-small-8k-instruct
25
- Phi-3.5-mini-instruct
26
- o1-preview
27
- o1-mini
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generated/anthropic.yaml ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - litellm_params:
2
+ api_key: os.environ/ANTHROPIC_API_KEY
3
+ model: anthropic/claude-3-5-haiku-latest
4
+ model_name: anthropic/claude-3-5-haiku
5
+ - litellm_params:
6
+ api_key: os.environ/ANTHROPIC_API_KEY
7
+ model: anthropic/claude-3-5-sonnet-latest
8
+ model_name: anthropic/claude-3-5-sonnet
9
+ - litellm_params:
10
+ api_key: os.environ/ANTHROPIC_API_KEY
11
+ model: anthropic/claude-3-5-opus-latest
12
+ model_name: anthropic/claude-3-5-opus
13
+ - litellm_params:
14
+ api_key: os.environ/ANTHROPIC_API_KEY
15
+ model: anthropic/claude-3-haiku-latest
16
+ model_name: anthropic/claude-3-haiku
17
+ - litellm_params:
18
+ api_key: os.environ/ANTHROPIC_API_KEY
19
+ model: anthropic/claude-3-sonnet-latest
20
+ model_name: anthropic/claude-3-sonnet
21
+ - litellm_params:
22
+ api_key: os.environ/ANTHROPIC_API_KEY
23
+ model: anthropic/claude-3-opus-latest
24
+ model_name: anthropic/claude-3-opus
generated/github.yaml ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - litellm_params:
2
+ api_key: os.environ/GITHUB_API_KEY
3
+ model: github/DeepSeek-R1
4
+ model_name: deepseek/r1
5
+ - litellm_params:
6
+ api_key: os.environ/GITHUB_API_KEY
7
+ model: github/gpt-4o
8
+ model_name: openai/gpt-4o
9
+ - litellm_params:
10
+ api_key: os.environ/GITHUB_API_KEY
11
+ model: github/gpt-4o-mini
12
+ model_name: openai/gpt-4o-mini
13
+ - litellm_params:
14
+ api_key: os.environ/GITHUB_API_KEY
15
+ model: github/o1
16
+ model_name: openai/o1
17
+ - litellm_params:
18
+ api_key: os.environ/GITHUB_API_KEY
19
+ model: github/o1-mini
20
+ model_name: openai/o1-mini
21
+ - litellm_params:
22
+ api_key: os.environ/GITHUB_API_KEY
23
+ model: github/text-embedding-3-small
24
+ model_name: openai/text-embedding-3-small
25
+ - litellm_params:
26
+ api_key: os.environ/GITHUB_API_KEY
27
+ model: github/text-embedding-3-large
28
+ model_name: openai/text-embedding-3-large
29
+ - litellm_params:
30
+ api_key: os.environ/GITHUB_API_KEY
31
+ model: github/Mistral-small
32
+ model_name: mistral/small
33
+ - litellm_params:
34
+ api_key: os.environ/GITHUB_API_KEY
35
+ model: github/Mistral-large
36
+ model_name: mistral/large
37
+ - litellm_params:
38
+ api_key: os.environ/GITHUB_API_KEY
39
+ model: github/Mistral-Nemo
40
+ model_name: mistral/nemo
41
+ - litellm_params:
42
+ api_key: os.environ/GITHUB_API_KEY
43
+ model: github/Cohere-command-r
44
+ model_name: cohere/command-r
45
+ - litellm_params:
46
+ api_key: os.environ/GITHUB_API_KEY
47
+ model: github/Cohere-command-r-plus
48
+ model_name: cohere/command-r-plus
49
+ - litellm_params:
50
+ api_key: os.environ/GITHUB_API_KEY
51
+ model: github/Meta-Llama-3-70B-Instruct
52
+ model_name: meta/llama-3.3-70b
53
+ - litellm_params:
54
+ api_key: os.environ/GITHUB_API_KEY
55
+ model: github/Meta-Llama-3-8B-Instruct
56
+ model_name: meta/llama-3.3-8b
generated/groq.yaml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ - litellm_params:
2
+ api_key: os.environ/GROQ_API_KEY
3
+ model: groq/llama-3.3-70b-versatile
4
+ model_name: meta/llama-3.3-70b
5
+ - litellm_params:
6
+ api_key: os.environ/GROQ_API_KEY
7
+ model: groq/deepseek-r1-distill-llama-70b
8
+ model_name: deepseek/r1-distill-llama-70b
generated/huggingface.yaml ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - litellm_params:
2
+ model: huggingface/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
3
+ model_name: deepseek/r1-distill-qwen-32b
4
+ - litellm_params:
5
+ model: huggingface/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
6
+ model_name: deepseek/r1-distill-qwen-1.5b
7
+ - litellm_params:
8
+ model: huggingface/google/gemma-2-2b-it
9
+ model_name: google/gemma-2-2b-it
10
+ - litellm_params:
11
+ model: huggingface/google/gemma-2-9b-it
12
+ model_name: google/gemma-2-9b-it
13
+ - litellm_params:
14
+ model: huggingface/google/gemma-2-27b-it
15
+ model_name: google/gemma-2-27b-it
generated/openai.yaml ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - litellm_params:
2
+ api_key: os.environ/OPENAI_API_KEY
3
+ model: openai/gpt-4o
4
+ model_name: openai/gpt-4o
5
+ - litellm_params:
6
+ api_key: os.environ/OPENAI_API_KEY
7
+ model: openai/gpt-4o-mini
8
+ model_name: openai/gpt-4o-mini
9
+ - litellm_params:
10
+ api_key: os.environ/OPENAI_API_KEY
11
+ model: openai/o1-mini
12
+ model_name: openai/o1-mini
13
+ - litellm_params:
14
+ api_key: os.environ/OPENAI_API_KEY
15
+ model: openai/o1-preview
16
+ model_name: openai/o1-preview
17
+ - litellm_params:
18
+ api_key: os.environ/OPENAI_API_KEY
19
+ model: openai/gpt-3.5-turbo
20
+ model_name: openai/gpt-3.5-turbo
21
+ - litellm_params:
22
+ api_key: os.environ/OPENAI_API_KEY
23
+ model: openai/text-embedding-3-small
24
+ model_name: openai/text-embedding-3-small
25
+ - litellm_params:
26
+ api_key: os.environ/OPENAI_API_KEY
27
+ model: openai/text-embedding-3-large
28
+ model_name: openai/text-embedding-3-large
29
+ - litellm_params:
30
+ api_key: os.environ/OPENAI_API_KEY
31
+ model: openai/dall-e-3
32
+ model_name: openai/dall-e-3
generated/perplexity.yaml ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - litellm_params:
2
+ api_key: os.environ/PERPLEXITY_API_KEY
3
+ model: perplexity/sonar-pro
4
+ model_name: perplexity/sonar-pro
5
+ - litellm_params:
6
+ api_key: os.environ/PERPLEXITY_API_KEY
7
+ model: perplexity/sonar-reasoning
8
+ model_name: perplexity/sonar-reasoning
9
+ - litellm_params:
10
+ api_key: os.environ/PERPLEXITY_API_KEY
11
+ model: perplexity/sonar
12
+ model_name: perplexity/sonar
start.sh CHANGED
@@ -3,42 +3,34 @@ set -euo pipefail
3
 
4
  pids=()
5
 
6
- generate_litellm_model() {
7
- model="$1"
8
- key_env="$2"
9
 
10
- if [[ -n "${!key_env:-}" ]]; then
11
- >&2 echo "[!] Found key ${key_env} for ${provider}"
12
- cat <<EOF
13
- - model_name: "${model}"
14
- litellm_params:
15
- model: "${model}"
16
- api_key: "os.environ/${key_env}"
17
- EOF
18
  fi
19
- }
20
 
21
- generate_litellm_provider() {
22
- provider="$1"
23
  key_env="$2"
24
 
25
- generate_litellm_model "${provider}/*" "${key_env}"
 
 
 
 
 
 
26
  }
27
 
28
  generate_litellm_config() {
29
- cat <<EOF
30
- model_list:
31
- - model_name: huggingface/google/gemma-2-2b-it
32
- litellm_params:
33
- model: huggingface/google/gemma-2-2b-it
34
- EOF
35
  generate_litellm_provider "openai" "OPENAI_API_KEY"
36
  generate_litellm_provider "perplexity" "PERPLEXITY_API_KEY"
37
- generate_litellm_provider "anthropic" "ANTHROPIC_API_KEY"
38
-
39
- while read -r model; do
40
- generate_litellm_model "github/${model}" "GITHUB_API_KEY"
41
- done </assets/azure-models.txt
42
 
43
  if [[ -n "${LITELLM_MODELS_BASE64:-}" ]]; then
44
  echo "${LITELLM_MODELS_BASE64}" | base64 -d
 
3
 
4
  pids=()
5
 
6
+ generate_litellm_provider() {
7
+ provider="$1"
 
8
 
9
+ if [[ "$#" -lt 2 ]]; then
10
+ >&2 echo "[+] Provider ${provider}: No key needed"
11
+ cat "/assets/${provider}.yaml"
12
+ return
 
 
 
 
13
  fi
 
14
 
 
 
15
  key_env="$2"
16
 
17
+ if [[ -n "${!key_env:-}" ]]; then
18
+ >&2 echo "[+] Provider ${provider}: Found key ${key_env}"
19
+ cat "/assets/${provider}.yaml"
20
+ return
21
+ fi
22
+
23
+ >&2 echo "[!] Provider ${provider}: No key found (set ${key_env})"
24
  }
25
 
26
  generate_litellm_config() {
27
+ echo "model_list:"
28
+ generate_litellm_provider "anthropic" "ANTHROPIC_API_KEY"
29
+ generate_litellm_provider "github" "GITHUB_API_KEY"
30
+ generate_litellm_provider "groq" "GROQ_API_KEY"
31
+ generate_litellm_provider "huggingface"
 
32
  generate_litellm_provider "openai" "OPENAI_API_KEY"
33
  generate_litellm_provider "perplexity" "PERPLEXITY_API_KEY"
 
 
 
 
 
34
 
35
  if [[ -n "${LITELLM_MODELS_BASE64:-}" ]]; then
36
  echo "${LITELLM_MODELS_BASE64}" | base64 -d