cmulgy commited on
Commit
c580078
·
1 Parent(s): 872dfb8
Files changed (1) hide show
  1. demo.py +5 -5
demo.py CHANGED
@@ -213,7 +213,7 @@ def ensure_client_available():
213
  def model_prompting(
214
  llm_model: str,
215
  prompt: str,
216
- max_token_num: Optional[int] = 2048,
217
  temperature: Optional[float] = 0.2,
218
  top_p: Optional[float] = 0.7,
219
  stream: Optional[bool] = True,
@@ -766,7 +766,7 @@ class GNNLLMSystem:
766
  response = model_prompting(
767
  llm_model=mapped_model_name,
768
  prompt=formatted_query,
769
- max_token_num=4096,
770
  temperature=0.0,
771
  top_p=0.9,
772
  stream=True
@@ -859,7 +859,7 @@ Keep the description concise and informative. Respond with just the task descrip
859
  task_description = model_prompting(
860
  llm_model="meta/llama-3.1-8b-instruct",
861
  prompt=prompt,
862
- max_token_num=2048,
863
  temperature=0.1,
864
  top_p=0.9,
865
  stream=True
@@ -1073,7 +1073,7 @@ def process_query(query):
1073
  response = model_prompting(
1074
  llm_model=api_model,
1075
  prompt=final_prompt,
1076
- max_token_num=4096,
1077
  temperature=0.0,
1078
  top_p=0.9,
1079
  stream=True
@@ -1187,7 +1187,7 @@ def process_thought_template_query(query, template_style, task_description, top_
1187
  llama_response = model_prompting(
1188
  llm_model="meta/llama-3.1-8b-instruct",
1189
  prompt=enhanced_query,
1190
- max_token_num=4096,
1191
  temperature=0.0,
1192
  top_p=0.9,
1193
  stream=True
 
213
  def model_prompting(
214
  llm_model: str,
215
  prompt: str,
216
+ max_token_num: Optional[int] = 1024, # Changed from 2048 to 1024
217
  temperature: Optional[float] = 0.2,
218
  top_p: Optional[float] = 0.7,
219
  stream: Optional[bool] = True,
 
766
  response = model_prompting(
767
  llm_model=mapped_model_name,
768
  prompt=formatted_query,
769
+ max_token_num=1024, # Changed from 4096 to 1024
770
  temperature=0.0,
771
  top_p=0.9,
772
  stream=True
 
859
  task_description = model_prompting(
860
  llm_model="meta/llama-3.1-8b-instruct",
861
  prompt=prompt,
862
+ max_token_num=1024, # Changed from 2048 to 1024
863
  temperature=0.1,
864
  top_p=0.9,
865
  stream=True
 
1073
  response = model_prompting(
1074
  llm_model=api_model,
1075
  prompt=final_prompt,
1076
+ max_token_num=1024, # Changed from 4096 to 1024
1077
  temperature=0.0,
1078
  top_p=0.9,
1079
  stream=True
 
1187
  llama_response = model_prompting(
1188
  llm_model="meta/llama-3.1-8b-instruct",
1189
  prompt=enhanced_query,
1190
+ max_token_num=1024, # Changed from 4096 to 1024
1191
  temperature=0.0,
1192
  top_p=0.9,
1193
  stream=True