Spaces:
Sleeping
Sleeping
File size: 851 Bytes
49c0097 959e25e 49c0097 d56e863 959e25e 49c0097 959e25e 309768d 38239df d56e863 38239df d56e863 38239df d56e863 38239df 49c0097 959e25e 309768d d56e863 309768d 49c0097 959e25e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
import gradio as gr
from transformers import pipeline
# Load the model
model_name = "gpt2"
generator = pipeline("text-generation", model=model_name)
# Inference function
def generate_response(prompt):
# Adjust the prompt to guide the model
modified_prompt = f"Answer the question directly: {prompt}"
response = generator(
modified_prompt,
max_length=150, # Maintain this to allow for longer responses
num_return_sequences=1,
temperature=0.7,
top_k=50,
top_p=0.95
)
return response[0]['generated_text'].strip()
# Gradio interface
interface = gr.Interface(
fn=generate_response,
inputs="text",
outputs="text",
title="Conversational LLM",
description="Enter a prompt to generate a relevant and coherent response."
)
# Launch the interface
interface.launch()
|