rovi27 commited on
Commit
cd5bec3
·
verified ·
1 Parent(s): 454596a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -8,8 +8,8 @@ import os
8
 
9
  #sft_model = "somosnlp/ComeBien_mistral-7b-instruct-v0.2-bnb-4bit"
10
  #base_model_name = "unsloth/Mistral-7B-Instruct-v0.2"
11
- #sft_model = "somosnlp/ComeBien_gemma-2b-it-bnb-4bit"
12
- sft_model = "somosnlp/RecetasDeLaAbuela5k_gemma-2b-bnb-4bit"
13
  base_model_name = "unsloth/gemma-2b-bnb-4bit"
14
  #base_model_name = "unsloth/gemma-2b-it-bnb-4bit"
15
 
@@ -48,7 +48,7 @@ stopping_criteria = ListOfTokensStoppingCriteria(tokenizer, stop_tokens)
48
  # Añade tu criterio de parada a una StoppingCriteriaList
49
  stopping_criteria_list = StoppingCriteriaList([stopping_criteria])
50
 
51
- def generate_text(prompt, context, model, tokenizer, max_length=2100):
52
  prompt=prompt.replace("\n", "").replace("¿","").replace("?","")
53
  input_text = f'''<bos><start_of_turn>system\n{context}?<end_of_turn><start_of_turn>user\n{prompt}<end_of_turn><start_of_turn>model\n'''
54
  inputs = tokenizer.encode(input_text, return_tensors="pt", add_special_tokens=False).to("cuda:0")
@@ -62,9 +62,9 @@ def generate_text(prompt, context, model, tokenizer, max_length=2100):
62
  outputs = model.generate(generation_config=generation_config, input_ids=inputs, stopping_criteria=stopping_criteria_list,)
63
  return tokenizer.decode(outputs[0], skip_special_tokens=False) #True
64
 
65
- def mostrar_respuesta(pregunta, contexto, model, tokenizer):
66
  try:
67
- res= generate_text(pregunta, contexto, model, tokenizer, max_length=700)
68
  return str(res)
69
  except Exception as e:
70
  return str(e)
 
8
 
9
  #sft_model = "somosnlp/ComeBien_mistral-7b-instruct-v0.2-bnb-4bit"
10
  #base_model_name = "unsloth/Mistral-7B-Instruct-v0.2"
11
+ sft_model = "somosnlp/ComeBien_gemma-2b-it-bnb-4bit"
12
+ #sft_model = "somosnlp/RecetasDeLaAbuela5k_gemma-2b-bnb-4bit"
13
  base_model_name = "unsloth/gemma-2b-bnb-4bit"
14
  #base_model_name = "unsloth/gemma-2b-it-bnb-4bit"
15
 
 
48
  # Añade tu criterio de parada a una StoppingCriteriaList
49
  stopping_criteria_list = StoppingCriteriaList([stopping_criteria])
50
 
51
+ def generate_text(prompt, context, max_length=2100):
52
  prompt=prompt.replace("\n", "").replace("¿","").replace("?","")
53
  input_text = f'''<bos><start_of_turn>system\n{context}?<end_of_turn><start_of_turn>user\n{prompt}<end_of_turn><start_of_turn>model\n'''
54
  inputs = tokenizer.encode(input_text, return_tensors="pt", add_special_tokens=False).to("cuda:0")
 
62
  outputs = model.generate(generation_config=generation_config, input_ids=inputs, stopping_criteria=stopping_criteria_list,)
63
  return tokenizer.decode(outputs[0], skip_special_tokens=False) #True
64
 
65
+ def mostrar_respuesta(pregunta, contexto):
66
  try:
67
+ res= generate_text(pregunta, contexto, max_length=700)
68
  return str(res)
69
  except Exception as e:
70
  return str(e)