Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -25,7 +25,7 @@ if not torch.cuda.is_available():
|
|
25 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
|
26 |
|
27 |
if torch.cuda.is_available():
|
28 |
-
modelA_id = "meta-llama/Llama-2-7b-
|
29 |
bnb_config = BitsAndBytesConfig(
|
30 |
load_in_4bit=True,
|
31 |
bnb_4bit_use_double_quant=False,
|
@@ -106,7 +106,7 @@ logging.basicConfig(level=logging.DEBUG)
|
|
106 |
# Gradio Interface Setup
|
107 |
chat_interface = gr.ChatInterface(
|
108 |
fn=generate,
|
109 |
-
additional_inputs=[gr.Dropdown(
|
110 |
fill_height=True,
|
111 |
stop_btn=None,
|
112 |
examples=[
|
|
|
25 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
|
26 |
|
27 |
if torch.cuda.is_available():
|
28 |
+
modelA_id = "meta-llama/Llama-2-7b-hf"
|
29 |
bnb_config = BitsAndBytesConfig(
|
30 |
load_in_4bit=True,
|
31 |
bnb_4bit_use_double_quant=False,
|
|
|
106 |
# Gradio Interface Setup
|
107 |
chat_interface = gr.ChatInterface(
|
108 |
fn=generate,
|
109 |
+
additional_inputs=[gr.Dropdown(["A", "B"],label="Model", info="Will add more animals later!")],
|
110 |
fill_height=True,
|
111 |
stop_btn=None,
|
112 |
examples=[
|