lliu01 commited on
Commit
f6578b1
·
verified ·
1 Parent(s): a8ba0ae

Fix tokenizer.eos_token_id

Browse files
Files changed (1) hide show
  1. app.py +1 -2
app.py CHANGED
@@ -47,7 +47,7 @@ def predict(message, history, system_prompt, temperature, max_tokens):
47
  temperature=temperature,
48
  max_new_tokens=DEFAULT_MAX_NEW_TOKENS,
49
  use_cache=True,
50
- eos_token_id=100278 # <|im_end|>
51
  )
52
  t = Thread(target=model.generate, kwargs=generate_kwargs)
53
  t.start()
@@ -61,7 +61,6 @@ def predict(message, history, system_prompt, temperature, max_tokens):
61
  if __name__ == "__main__":
62
  args = parse_args()
63
  tokenizer = AutoTokenizer.from_pretrained("lliu01/fortios_cli")
64
- tokenizer = AutoTokenizer.from_pretrained("lliu01/fortios_cli")
65
  model = AutoModelForCausalLM.from_pretrained(
66
  "lliu01/fortios_cli",
67
  torch_dtype=torch.bfloat16,
 
47
  temperature=temperature,
48
  max_new_tokens=DEFAULT_MAX_NEW_TOKENS,
49
  use_cache=True,
50
+ eos_token_id=tokenizer.eos_token_id # <|im_end|>
51
  )
52
  t = Thread(target=model.generate, kwargs=generate_kwargs)
53
  t.start()
 
61
  if __name__ == "__main__":
62
  args = parse_args()
63
  tokenizer = AutoTokenizer.from_pretrained("lliu01/fortios_cli")
 
64
  model = AutoModelForCausalLM.from_pretrained(
65
  "lliu01/fortios_cli",
66
  torch_dtype=torch.bfloat16,