kenghuoxiong commited on
Commit
c413a53
·
verified ·
1 Parent(s): e40635e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -24,7 +24,7 @@ TOKEN = os.getenv("HF_TOKEN")
24
  def load_embedding_mode():
25
  # embedding_model_dict = {"m3e-base": "/home/xiongwen/m3e-base"}
26
  encode_kwargs = {"normalize_embeddings": False}
27
- model_kwargs = {"device": 'cuda'}
28
  return HuggingFaceEmbeddings(model_name="BAAI/bge-m3",
29
  model_kwargs=model_kwargs,
30
  encode_kwargs=encode_kwargs)
@@ -81,7 +81,7 @@ retriever = db.as_retriever()
81
  def langchain_chat(message, temperature, top_p, max_tokens):
82
  llm = ChatOpenAI(
83
  # model="meta-llama/Meta-Llama-3-8B-Instruct",
84
- model="Qwen/Qwen2-7B-Instruct",
85
  temperature=temperature,
86
  top_p=top_p,
87
  max_tokens=max_tokens)
@@ -134,7 +134,7 @@ def respond(
134
 
135
  for message in client.chat.completions.create(
136
  # model="meta-llama/Meta-Llama-3-8B-Instruct",
137
- model="Qwen/Qwen2-7B-Instruct",
138
  max_tokens=max_tokens,
139
  stream=True,
140
  temperature=temperature,
 
24
  def load_embedding_mode():
25
  # embedding_model_dict = {"m3e-base": "/home/xiongwen/m3e-base"}
26
  encode_kwargs = {"normalize_embeddings": False}
27
+ model_kwargs = {"device": 'cpu'}
28
  return HuggingFaceEmbeddings(model_name="BAAI/bge-m3",
29
  model_kwargs=model_kwargs,
30
  encode_kwargs=encode_kwargs)
 
81
  def langchain_chat(message, temperature, top_p, max_tokens):
82
  llm = ChatOpenAI(
83
  # model="meta-llama/Meta-Llama-3-8B-Instruct",
84
+ model="Qwen/Qwen1.5-4B-Chat",
85
  temperature=temperature,
86
  top_p=top_p,
87
  max_tokens=max_tokens)
 
134
 
135
  for message in client.chat.completions.create(
136
  # model="meta-llama/Meta-Llama-3-8B-Instruct",
137
+ model="Qwen/Qwen1.5-4B-Chat",
138
  max_tokens=max_tokens,
139
  stream=True,
140
  temperature=temperature,