from transformers import pipeline # Initialize the pipeline with your fine-tuned model pipe = pipeline("text-generation", model="crystal99/my-fine-tuned-model") # Define input prompts input_texts = [ "Solve this: 23 + 19?", "Simplify: 48 / 6.", "What is the square root of 64?" ] # Generate outputs outputs = pipe("<|startoftext|> <|user|> what comes after 7? <|bot|>", max_length=150, temperature=0.7, top_k=50, top_p=0.9) print(outputs) # Display results # for i, out in enumerate(outputs): # print(f"Input {i+1}: {input_texts[i]}") # print(f"Output {i+1}: {out['generated_text']}\n")