NCTCMumbai commited on
Commit
f8871ca
·
verified ·
1 Parent(s): 74a4f1f

Update backend/query_llm.py

Browse files
Files changed (1) hide show
  1. backend/query_llm.py +2 -0
backend/query_llm.py CHANGED
@@ -104,6 +104,7 @@ def generate_hf(prompt: str, history: str, temperature: float = 0.9, max_new_tok
104
  stream=True, details=True, return_full_text=False)
105
  output = ""
106
  for response in stream:
 
107
  output += response.token.text
108
  yield output
109
 
@@ -205,6 +206,7 @@ def generate_gemini(prompt: str, history: str, temperature: float = 0.9, max_new
205
  stream=True)
206
  output = ""
207
  for response in stream:
 
208
  output += response.text
209
  yield output
210
 
 
104
  stream=True, details=True, return_full_text=False)
105
  output = ""
106
  for response in stream:
107
+ print(response.token.text)
108
  output += response.token.text
109
  yield output
110
 
 
206
  stream=True)
207
  output = ""
208
  for response in stream:
209
+ print(response.text)
210
  output += response.text
211
  yield output
212