Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -21,6 +21,7 @@ client = InferenceClient("meta-llama/Llama-3.1-8B-Instruct", token=os.getenv('de
|
|
21 |
def respond(
|
22 |
message,
|
23 |
history: list[tuple[str, str]],
|
|
|
24 |
system_message,
|
25 |
max_tokens,
|
26 |
temperature,
|
@@ -59,8 +60,7 @@ For information on how to customize the ChatInterface, peruse the gradio docs: h
|
|
59 |
|
60 |
demo = gr.ChatInterface(
|
61 |
inputs = respond,
|
62 |
-
|
63 |
-
description = "Ask me anything!",
|
64 |
additional_inputs=[
|
65 |
gr.Dropdown(["Qwen1.5", "Llama3-8b-Instruct", "Llama3.1-8b-Instruct"], label="Select Model"),
|
66 |
gr.Textbox(value="You are a friendly and helpful Chatbot, be concise and straight to the point, avoid excessive reasoning.", label="System message"),
|
|
|
21 |
def respond(
|
22 |
message,
|
23 |
history: list[tuple[str, str]],
|
24 |
+
model,
|
25 |
system_message,
|
26 |
max_tokens,
|
27 |
temperature,
|
|
|
60 |
|
61 |
demo = gr.ChatInterface(
|
62 |
inputs = respond,
|
63 |
+
|
|
|
64 |
additional_inputs=[
|
65 |
gr.Dropdown(["Qwen1.5", "Llama3-8b-Instruct", "Llama3.1-8b-Instruct"], label="Select Model"),
|
66 |
gr.Textbox(value="You are a friendly and helpful Chatbot, be concise and straight to the point, avoid excessive reasoning.", label="System message"),
|