Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -24,7 +24,7 @@ TOKEN = os.getenv("HF_TOKEN")
|
|
24 |
def load_embedding_mode():
|
25 |
# embedding_model_dict = {"m3e-base": "/home/xiongwen/m3e-base"}
|
26 |
encode_kwargs = {"normalize_embeddings": False}
|
27 |
-
model_kwargs = {"device": '
|
28 |
return HuggingFaceEmbeddings(model_name="BAAI/bge-m3",
|
29 |
model_kwargs=model_kwargs,
|
30 |
encode_kwargs=encode_kwargs)
|
@@ -81,7 +81,7 @@ retriever = db.as_retriever()
|
|
81 |
def langchain_chat(message, temperature, top_p, max_tokens):
|
82 |
llm = ChatOpenAI(
|
83 |
# model="meta-llama/Meta-Llama-3-8B-Instruct",
|
84 |
-
model="Qwen/
|
85 |
temperature=temperature,
|
86 |
top_p=top_p,
|
87 |
max_tokens=max_tokens)
|
@@ -134,7 +134,7 @@ def respond(
|
|
134 |
|
135 |
for message in client.chat.completions.create(
|
136 |
# model="meta-llama/Meta-Llama-3-8B-Instruct",
|
137 |
-
model="Qwen/
|
138 |
max_tokens=max_tokens,
|
139 |
stream=True,
|
140 |
temperature=temperature,
|
|
|
24 |
def load_embedding_mode():
|
25 |
# embedding_model_dict = {"m3e-base": "/home/xiongwen/m3e-base"}
|
26 |
encode_kwargs = {"normalize_embeddings": False}
|
27 |
+
model_kwargs = {"device": 'cpu'}
|
28 |
return HuggingFaceEmbeddings(model_name="BAAI/bge-m3",
|
29 |
model_kwargs=model_kwargs,
|
30 |
encode_kwargs=encode_kwargs)
|
|
|
81 |
def langchain_chat(message, temperature, top_p, max_tokens):
|
82 |
llm = ChatOpenAI(
|
83 |
# model="meta-llama/Meta-Llama-3-8B-Instruct",
|
84 |
+
model="Qwen/Qwen1.5-4B-Chat",
|
85 |
temperature=temperature,
|
86 |
top_p=top_p,
|
87 |
max_tokens=max_tokens)
|
|
|
134 |
|
135 |
for message in client.chat.completions.create(
|
136 |
# model="meta-llama/Meta-Llama-3-8B-Instruct",
|
137 |
+
model="Qwen/Qwen1.5-4B-Chat",
|
138 |
max_tokens=max_tokens,
|
139 |
stream=True,
|
140 |
temperature=temperature,
|