Spaces:
Sleeping
Sleeping
update
Browse files- config.yaml +14 -0
config.yaml
CHANGED
|
@@ -3,6 +3,20 @@ model_list:
|
|
| 3 |
litellm_params:
|
| 4 |
model: ollama/mistral:7b
|
| 5 |
api_base: https://zhengr-ollama.hf.space
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
|
| 7 |
litellm_settings: # module level litellm settings - https://github.com/BerriAI/litellm/blob/main/litellm/__init__.py
|
| 8 |
drop_params: True
|
|
|
|
| 3 |
litellm_params:
|
| 4 |
model: ollama/mistral:7b
|
| 5 |
api_base: https://zhengr-ollama.hf.space
|
| 6 |
+
# Model-specific parameters
|
| 7 |
+
#model: "huggingface/mistralai/Mistral-7B-Instruct-v0.1"
|
| 8 |
+
#api_base: "<your-api-base>"
|
| 9 |
+
#api_key: "<your-api-key>" # [OPTIONAL] for hf inference endpoints
|
| 10 |
+
#initial_prompt_value: "\n"
|
| 11 |
+
#roles: {"system":{"pre_message":"<|im_start|>system\n", "post_message":"<|im_end|>"}, "assistant":{"pre_message":"<|im_start|>assistant\n","post_message":"<|im_end|>"}, "user":{"pre_message":"<|im_start|>user\n","post_message":"<|im_end|>"}}
|
| 12 |
+
#final_prompt_value: "\n"
|
| 13 |
+
#bos_token: "<s>"
|
| 14 |
+
#eos_token: "</s>"
|
| 15 |
+
#max_tokens: 4096
|
| 16 |
+
- model_name: xinference-llama-3-instruct
|
| 17 |
+
litellm_params:
|
| 18 |
+
model: xinference/llama-3-instruct
|
| 19 |
+
api_base: https://zhengr-xinference.hf.space
|
| 20 |
|
| 21 |
litellm_settings: # module level litellm settings - https://github.com/BerriAI/litellm/blob/main/litellm/__init__.py
|
| 22 |
drop_params: True
|