Spaces:
Sleeping
Sleeping
Deploy DeepSeek LLM chatbotV5
Browse files
app.py
CHANGED
@@ -4,9 +4,12 @@ import torch
|
|
4 |
|
5 |
# Load DeepSeek LLM
|
6 |
model_name = "deepseek-ai/deepseek-llm-7b-chat"
|
|
|
7 |
|
|
|
8 |
|
9 |
-
tokenizer =
|
|
|
10 |
# tokenizer = AutoTokenizer.from_pretrained(model_name)
|
11 |
# model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
|
12 |
|
|
|
4 |
|
5 |
# Load DeepSeek LLM
|
6 |
model_name = "deepseek-ai/deepseek-llm-7b-chat"
|
7 |
+
# llm = ChatOllama(model="deepseek-r1:1.5b", temperature=0)
|
8 |
|
9 |
+
model_name = "deepseek-ai/deepseek-llm-7b-chat"
|
10 |
|
11 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
12 |
+
# tokenizer = LlamaTokenizer.from_pretrained(model_name) # Explicitly use LlamaTokenizer
|
13 |
# tokenizer = AutoTokenizer.from_pretrained(model_name)
|
14 |
# model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
|
15 |
|