Namitg02 commited on
Commit
8f3d678
·
verified ·
1 Parent(s): 303482b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -12
app.py CHANGED
@@ -114,7 +114,7 @@ def talk(prompt, history):
114
  # the chat template structure should be based on text generation model format
115
 
116
  # indicates the end of a sequence
117
- stream = model.create_chat_completion(messages = messages, max_tokens=1000, stop=["</s>"], stream=True)
118
  # print(f"{stream}")
119
  print("check 7")
120
  # print(stream['choices'][0]['message']['content'])
@@ -135,17 +135,7 @@ def talk(prompt, history):
135
  # only the tokens comprising the top_p probability mass are considered for responses
136
  # This output is a data structure containing all the information returned by generate(), but that can also be used as tuple or dictionary.
137
 
138
- # start = time.time()
139
- # NUM_TOKENS=0
140
-
141
- # time_generate = time.time() - start
142
- # print('\n')
143
- # print('-'*4+'End Generation'+'-'*4)
144
- # print(f'Num of generated tokens: {NUM_TOKENS}')
145
- # print(f'Time for complete generation: {time_generate}s')
146
- # print(f'Tokens per secound: {NUM_TOKENS/time_generate}')
147
- # print(f'Time per token: {(time_generate/NUM_TOKENS)*1000}ms')
148
-
149
 
150
  TITLE = "AI Copilot for Diabetes Patients"
151
 
 
114
  # the chat template structure should be based on text generation model format
115
 
116
  # indicates the end of a sequence
117
+ stream = model.create_chat_completion(messages = messages, max_tokens=1000, stop=["</s>"], stream=False)
118
  # print(f"{stream}")
119
  print("check 7")
120
  # print(stream['choices'][0]['message']['content'])
 
135
  # only the tokens comprising the top_p probability mass are considered for responses
136
  # This output is a data structure containing all the information returned by generate(), but that can also be used as tuple or dictionary.
137
 
138
+
 
 
 
 
 
 
 
 
 
 
139
 
140
  TITLE = "AI Copilot for Diabetes Patients"
141