han1997 commited on
Commit
0d355d4
·
verified ·
1 Parent(s): 00e38ee

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +17 -17
app.py CHANGED
@@ -76,26 +76,26 @@ def bot_streaming(message, history, temperature, top_k, max_new_tokens):
76
  }
77
 
78
  # Generate from the VLM
79
- buffer = vlm.generate(
80
- **generation_kwargs
81
- )
82
- # thread = Thread(target=vlm.generate, kwargs=generation_kwargs)
83
- # thread.start()
84
-
85
- # buffer = ""
86
- # output_started = False
87
- # for new_text in streamer:
88
- # if not output_started:
89
- # if "<|assistant|>\n" in new_text:
90
- # output_started = True
91
- # continue
92
- # buffer += new_text
93
- # if len(buffer) > 1:
94
- # yield buffer
95
 
96
  prompt_builder.add_turn(role="gpt", message=buffer)
97
 
98
- yield buffer
99
 
100
 
101
  demo = gr.ChatInterface(fn=bot_streaming,
 
76
  }
77
 
78
  # Generate from the VLM
79
+ # buffer = vlm.generate(
80
+ # **generation_kwargs
81
+ # )
82
+ thread = Thread(target=vlm.generate, kwargs=generation_kwargs)
83
+ thread.start()
84
+
85
+ buffer = ""
86
+ output_started = False
87
+ for new_text in streamer:
88
+ if not output_started:
89
+ if "<|assistant|>\n" in new_text:
90
+ output_started = True
91
+ continue
92
+ buffer += new_text
93
+ if len(buffer) > 1:
94
+ yield buffer
95
 
96
  prompt_builder.add_turn(role="gpt", message=buffer)
97
 
98
+ return buffer
99
 
100
 
101
  demo = gr.ChatInterface(fn=bot_streaming,