Namitg02 commited on
Commit
4ab17b7
·
verified ·
1 Parent(s): 76d7181

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -1
app.py CHANGED
@@ -63,8 +63,13 @@ tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0")
63
  # temperature=0.4,
64
  # stream = True
65
  # eos_token_id=terminators
66
- )
67
  # send additional parameters to model for generation
 
 
 
 
 
68
 
69
  #model = llama_cpp.Llama(model_path = tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf AutoModelForCausalLM.from_pretrained(llm_model, model_file = "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", model_type="llama", gpu_layers=0, config = generation_config)
70
  model = Llama.from_pretrained(
@@ -74,6 +79,7 @@ model = Llama.from_pretrained(
74
  temperature=0.75,
75
  max_tokens=500,
76
  top_p=0.95,
 
77
  # callback_manager=callback_manager,
78
  # verbose=True, # Verbose is required to pass to the callback manager
79
  )
 
63
  # temperature=0.4,
64
  # stream = True
65
  # eos_token_id=terminators
66
+ #)
67
  # send additional parameters to model for generation
68
+ terminators = [
69
+ tokenizer.eos_token_id, # End-of-Sequence Token that indicates where the model should consider the text sequence to be complete
70
+ tokenizer.convert_tokens_to_ids("<|eot_id|>") # Converts a token strings in a single/ sequence of integer id using the vocabulary
71
+ ]
72
+ # indicates the end of a sequence
73
 
74
  #model = llama_cpp.Llama(model_path = tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf AutoModelForCausalLM.from_pretrained(llm_model, model_file = "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", model_type="llama", gpu_layers=0, config = generation_config)
75
  model = Llama.from_pretrained(
 
79
  temperature=0.75,
80
  max_tokens=500,
81
  top_p=0.95,
82
+ eos_token_id=terminators
83
  # callback_manager=callback_manager,
84
  # verbose=True, # Verbose is required to pass to the callback manager
85
  )