Spaces:
Runtime error
Runtime error
import gradio as gr | |
import gradio_client | |
from gradio_client import Client | |
title = """# Welcome To Tonic's Easy YI-6B-200K""" | |
description = """This is a [connector that looks like open ai](https://huggingface.co/spaces/Tonic1/TonicsYI-6B-200k) on the inputs , then ignores all that and connects to my [YI-6B-200K endpoint](https://huggingface.co/01-ai/Yi-6B-200K). This helps using [Yi for AGI here](https://huggingface.co/spaces/Tonic/AGYIntelligence) | |
Join us : 🌟TeamTonic🌟 is always making cool demos! Join our active builder's🛠️community on 👻Discord: [Discord](https://discord.gg/nXx5wbX9) On 🤗Huggingface: [TeamTonic](https://huggingface.co/TeamTonic) & [MultiTransformer](https://huggingface.co/MultiTransformer) On 🌐Github: [Polytonic](https://github.com/tonic-ai) & contribute to 🌟 [PolyGPT](https://github.com/tonic-ai/polygpt-alpha) | |
""" | |
client = Client("https://tonic1-tonicsyi-6b-200k.hf.space/--replicas/cxxbc/") | |
def chat_api_interface(messages, model, frequency_penalty=0, logit_bias=None, max_tokens=None, n=1, presence_penalty=0, response_format="text", seed=None, stop=None, stream=False, temperature=0.9, top_p=1, tools=None, tool_choice="none", user=None): | |
try: | |
result = client.predict( | |
your_message=messages, | |
system_prompt="", | |
max_new_tokens=16000, | |
temperature=temperature, | |
top_p=top_p, | |
top_k=presence_penalty, | |
disable_for_faster_inference=False, | |
api_name="/predict" | |
) | |
return result | |
except Exception as e: | |
return f"An error occurred: {str(e)}" | |
with gr.Blocks() as demo: | |
with gr.Row(): | |
with gr.Column(): | |
messages = gr.Textbox(label="Messages (JSON format)", placeholder="Enter messages as a JSON array", value="[{\"role\": \"user\", \"content\": \"Hello, how can I help you today?\"}, {\"role\": \"assistant\", \"content\": \"I'm looking for information on renewable energy sources.\"}]") | |
model = gr.Textbox(label="Model ID", placeholder="Enter the model ID", value="text-davinci-003") | |
frequency_penalty = gr.Number(label="Frequency Penalty", value=0.5) | |
logit_bias = gr.Textbox(label="Logit Bias (JSON format)", placeholder="Enter logit bias as JSON", value="{\"50256\": -100, \"50257\": 100}") | |
max_tokens = gr.Number(label="Max Tokens", value=150) | |
n = gr.Number(label="N", value=3) | |
presence_penalty = gr.Number(label="Presence Penalty", value=0.6) | |
response_format = gr.Radio(label="Response Format", choices=["text", "json_object"], value="text") | |
seed = gr.Number(label="Seed", value=42) | |
stop = gr.Textbox(label="Stop Sequence(s)", placeholder="Enter stop sequences", value="[\"\\n\", \" end\"]") | |
stream = gr.Checkbox(label="Stream", value=False) | |
temperature = gr.Slider(label="Temperature", minimum=0, maximum=2, value=0.7) | |
top_p = gr.Slider(label="Top P", minimum=0, maximum=1, value=0.9) | |
tools = gr.Textbox(label="Tools (JSON format)", placeholder="Enter tools as JSON", value="{\"spellcheck\": true, \"grammar_check\": false}") | |
tool_choice = gr.Textbox(label="Tool Choice", value="spellcheck", placeholder="Enter tool choice") | |
user = gr.Textbox(label="User", placeholder="Enter user identifier", value="user123") | |
with gr.Column(): | |
submit_btn = gr.Button("Submit") | |
output = gr.Textbox(label="API Response") | |
submit_btn.click( | |
chat_api_interface, | |
inputs=[messages, model, frequency_penalty, logit_bias, max_tokens, n, presence_penalty, response_format, seed, stop, stream, temperature, top_p, tools, tool_choice, user], | |
outputs=output | |
) | |
demo.launch() |