KishoreK commited on
Commit
c9cae33
·
1 Parent(s): 128158c
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -81,9 +81,9 @@ If the given question lacks the parameters required by the function, also point
81
  "content": json.dumps(convert_to_xlam_tool(tools))
82
  }]
83
 
84
- model = AutoModelForCausalLM.from_pretrained("KishoreK/ActionGemma-9B", load_in_4bit=True, device_map="auto", use_cache=True,low_cpu_mem_usage=True )
85
  tokenizer = AutoTokenizer.from_pretrained("KishoreK/ActionGemma-9B")
86
- inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True)
87
  outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id)
88
  return tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
89
 
 
81
  "content": json.dumps(convert_to_xlam_tool(tools))
82
  }]
83
 
84
+ model = AutoModelForCausalLM.from_pretrained("KishoreK/ActionGemma-9B", device_map="auto", use_cache=True,low_cpu_mem_usage=True )
85
  tokenizer = AutoTokenizer.from_pretrained("KishoreK/ActionGemma-9B")
86
+ inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
87
  outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id)
88
  return tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
89