Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -8,8 +8,8 @@ import os
|
|
8 |
|
9 |
#sft_model = "somosnlp/ComeBien_mistral-7b-instruct-v0.2-bnb-4bit"
|
10 |
#base_model_name = "unsloth/Mistral-7B-Instruct-v0.2"
|
11 |
-
|
12 |
-
sft_model = "somosnlp/RecetasDeLaAbuela5k_gemma-2b-bnb-4bit"
|
13 |
base_model_name = "unsloth/gemma-2b-bnb-4bit"
|
14 |
#base_model_name = "unsloth/gemma-2b-it-bnb-4bit"
|
15 |
|
@@ -48,7 +48,7 @@ stopping_criteria = ListOfTokensStoppingCriteria(tokenizer, stop_tokens)
|
|
48 |
# Añade tu criterio de parada a una StoppingCriteriaList
|
49 |
stopping_criteria_list = StoppingCriteriaList([stopping_criteria])
|
50 |
|
51 |
-
def generate_text(prompt, context,
|
52 |
prompt=prompt.replace("\n", "").replace("¿","").replace("?","")
|
53 |
input_text = f'''<bos><start_of_turn>system\n{context}?<end_of_turn><start_of_turn>user\n{prompt}<end_of_turn><start_of_turn>model\n'''
|
54 |
inputs = tokenizer.encode(input_text, return_tensors="pt", add_special_tokens=False).to("cuda:0")
|
@@ -62,9 +62,9 @@ def generate_text(prompt, context, model, tokenizer, max_length=2100):
|
|
62 |
outputs = model.generate(generation_config=generation_config, input_ids=inputs, stopping_criteria=stopping_criteria_list,)
|
63 |
return tokenizer.decode(outputs[0], skip_special_tokens=False) #True
|
64 |
|
65 |
-
def mostrar_respuesta(pregunta, contexto
|
66 |
try:
|
67 |
-
res= generate_text(pregunta, contexto,
|
68 |
return str(res)
|
69 |
except Exception as e:
|
70 |
return str(e)
|
|
|
8 |
|
9 |
#sft_model = "somosnlp/ComeBien_mistral-7b-instruct-v0.2-bnb-4bit"
|
10 |
#base_model_name = "unsloth/Mistral-7B-Instruct-v0.2"
|
11 |
+
sft_model = "somosnlp/ComeBien_gemma-2b-it-bnb-4bit"
|
12 |
+
#sft_model = "somosnlp/RecetasDeLaAbuela5k_gemma-2b-bnb-4bit"
|
13 |
base_model_name = "unsloth/gemma-2b-bnb-4bit"
|
14 |
#base_model_name = "unsloth/gemma-2b-it-bnb-4bit"
|
15 |
|
|
|
48 |
# Añade tu criterio de parada a una StoppingCriteriaList
|
49 |
stopping_criteria_list = StoppingCriteriaList([stopping_criteria])
|
50 |
|
51 |
+
def generate_text(prompt, context, max_length=2100):
|
52 |
prompt=prompt.replace("\n", "").replace("¿","").replace("?","")
|
53 |
input_text = f'''<bos><start_of_turn>system\n{context}?<end_of_turn><start_of_turn>user\n{prompt}<end_of_turn><start_of_turn>model\n'''
|
54 |
inputs = tokenizer.encode(input_text, return_tensors="pt", add_special_tokens=False).to("cuda:0")
|
|
|
62 |
outputs = model.generate(generation_config=generation_config, input_ids=inputs, stopping_criteria=stopping_criteria_list,)
|
63 |
return tokenizer.decode(outputs[0], skip_special_tokens=False) #True
|
64 |
|
65 |
+
def mostrar_respuesta(pregunta, contexto):
|
66 |
try:
|
67 |
+
res= generate_text(pregunta, contexto, max_length=700)
|
68 |
return str(res)
|
69 |
except Exception as e:
|
70 |
return str(e)
|