Spaces:
Runtime error
Runtime error
File size: 1,441 Bytes
91ae465 daecaae 91ae465 17f1f96 91ae465 d7744cf 91ae465 d7744cf 91ae465 f99c184 daecaae 91ae465 d7744cf 91ae465 c3ce568 991b767 91ae465 c3ce568 22770d6 e361677 91ae465 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 |
import gradio as gr
from transformers import pipeline
import torch
import subprocess
import spaces
import os
# Initialize the model pipeline
generator = pipeline('text-generation', model='okeanos/uptimeai-8273')
@spaces.GPU
def generate_text(prompt):
# Generate text using the model
generator.model.cuda()
generator.device = torch.device("cuda")
prompt = f"<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n"
outputs = generator(
prompt,
do_sample=True
return_full_text=False
)
# Extract the generated text and return it
generated_text = outputs[0]['generated_text']
return generated_text
# Create the Gradio interface
iface = gr.Interface(
fn=generate_text,
inputs=[
gr.Textbox(label="Prompt", lines=2, placeholder="Type a prompt..."),
gr.Slider(minimum=0.1, maximum=2.0, step=0.01, value=0.8, label="Temperature"),
gr.Slider(minimum=0.0, maximum=1.0, step=0.01, value=0.95, label="Top p"),
gr.Slider(minimum=0, maximum=100, step=1, value=40, label="Top k"),
gr.Slider(minimum=1.0, maximum=2.0, step=0.01, value=1.10, label="Repetition Penalty"),
gr.Slider(minimum=5, maximum=4096, step=5, value=1024, label="Max Length")
],
outputs=gr.Textbox(label="Generated Text"),
title="ChatHercules-2.5-Mistral-7B",
description="Try out the ChatHercules-2.5-Mistral-7B model for free!"
)
iface.launch() |