Spaces:
Sleeping
Sleeping
req.txt
Browse files
app.py
CHANGED
@@ -81,9 +81,9 @@ If the given question lacks the parameters required by the function, also point
|
|
81 |
"content": json.dumps(convert_to_xlam_tool(tools))
|
82 |
}]
|
83 |
|
84 |
-
model = AutoModelForCausalLM.from_pretrained("KishoreK/ActionGemma-9B",
|
85 |
tokenizer = AutoTokenizer.from_pretrained("KishoreK/ActionGemma-9B")
|
86 |
-
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True)
|
87 |
outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id)
|
88 |
return tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
|
89 |
|
|
|
81 |
"content": json.dumps(convert_to_xlam_tool(tools))
|
82 |
}]
|
83 |
|
84 |
+
model = AutoModelForCausalLM.from_pretrained("KishoreK/ActionGemma-9B", device_map="auto", use_cache=True,low_cpu_mem_usage=True )
|
85 |
tokenizer = AutoTokenizer.from_pretrained("KishoreK/ActionGemma-9B")
|
86 |
+
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
|
87 |
outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id)
|
88 |
return tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
|
89 |
|