Update app.py
Browse files
app.py
CHANGED
@@ -353,11 +353,10 @@ def process_llm_response(llm_response):
|
|
353 |
|
354 |
|
355 |
|
356 |
-
|
357 |
|
358 |
@spaces.GPU(duration=60)
|
359 |
def llm_ans(message, history):
|
360 |
-
tok, model = build_model()
|
361 |
terminators = [tok.eos_token_id, 32007, 32011, 32001, 32000]
|
362 |
pipe = pipeline(task="text-generation", model=model, tokenizer=tok, eos_token_id=terminators, do_sample=True, max_new_tokens=CFG.max_new_tokens, temperature=CFG.temperature, top_p=CFG.top_p, repetition_penalty=CFG.repetition_penalty)
|
363 |
llm = HuggingFacePipeline(pipeline=pipe)
|
|
|
353 |
|
354 |
|
355 |
|
356 |
+
tok, model = build_model()
|
357 |
|
358 |
@spaces.GPU(duration=60)
|
359 |
def llm_ans(message, history):
|
|
|
360 |
terminators = [tok.eos_token_id, 32007, 32011, 32001, 32000]
|
361 |
pipe = pipeline(task="text-generation", model=model, tokenizer=tok, eos_token_id=terminators, do_sample=True, max_new_tokens=CFG.max_new_tokens, temperature=CFG.temperature, top_p=CFG.top_p, repetition_penalty=CFG.repetition_penalty)
|
362 |
llm = HuggingFacePipeline(pipeline=pipe)
|