Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -42,7 +42,9 @@ if torch.cuda.is_available():
|
|
42 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
43 |
tokenizer.use_default_system_prompt = False
|
44 |
|
45 |
-
|
|
|
|
|
46 |
@spaces.GPU
|
47 |
def generate(
|
48 |
message: str,
|
@@ -56,7 +58,7 @@ def generate(
|
|
56 |
conversation = []
|
57 |
for user, assistant in chat_history:
|
58 |
conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
|
59 |
-
conversation.append({"role": "user", "content": message})
|
60 |
|
61 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt")
|
62 |
if input_ids.shape[1] > MAX_INPUT_TOKEN_LENGTH:
|
@@ -99,7 +101,6 @@ chat_interface = gr.ChatInterface(
|
|
99 |
|
100 |
with gr.Blocks(css="style.css") as demo:
|
101 |
gr.Markdown(DESCRIPTION)
|
102 |
-
gr.DuplicateButton(value="Duplicate Space for private use", elem_id="duplicate-button")
|
103 |
chat_interface.render()
|
104 |
gr.Markdown(LICENSE)
|
105 |
|
|
|
42 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
43 |
tokenizer.use_default_system_prompt = False
|
44 |
|
45 |
+
def make_prompt(entry):
|
46 |
+
return f"### Human: YOUR INSTRUCTION HERE: {entry} ### Assistant:"
|
47 |
+
|
48 |
@spaces.GPU
|
49 |
def generate(
|
50 |
message: str,
|
|
|
58 |
conversation = []
|
59 |
for user, assistant in chat_history:
|
60 |
conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
|
61 |
+
conversation.append({"role": "user", "content": make_prompt(message)})
|
62 |
|
63 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt")
|
64 |
if input_ids.shape[1] > MAX_INPUT_TOKEN_LENGTH:
|
|
|
101 |
|
102 |
with gr.Blocks(css="style.css") as demo:
|
103 |
gr.Markdown(DESCRIPTION)
|
|
|
104 |
chat_interface.render()
|
105 |
gr.Markdown(LICENSE)
|
106 |
|