Spaces:
Runtime error
Runtime error
File size: 802 Bytes
dc2b56f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 |
def test_model(model, tokenizer, input_text):
"""
Test the fine-tuned model with a given input.
Args:
model: The fine-tuned model to test.
tokenizer: The tokenizer associated with the model.
input_text (str): The input text to generate a response for.
Returns:
str: The generated response from the model.
"""
messages = [
{"role": "user", "content": input_text},
]
inputs = tokenizer.apply_chat_template(
messages,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt"
).to("cuda")
outputs = model.generate(input_ids=inputs, max_new_tokens=64, use_cache=True,
temperature=1.5, min_p=0.1)
return tokenizer.batch_decode(outputs)[0] |