Update app.py
Browse files
app.py
CHANGED
@@ -102,11 +102,11 @@ def respond(
|
|
102 |
|
103 |
messages= json_obj
|
104 |
|
105 |
-
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt").to('cuda') # .to(accelerator.device)
|
106 |
-
input_ids2 = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True, return_tensors="pt") #.to('cuda')
|
107 |
-
print(f"Converted input_ids dtype: {input_ids.dtype}")
|
108 |
-
input_str= str(input_ids2)
|
109 |
-
print('input str = ', input_str)
|
110 |
|
111 |
|
112 |
# gen_tokens = model.generate(
|
|
|
102 |
|
103 |
messages= json_obj
|
104 |
|
105 |
+
# input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt").to('cuda') # .to(accelerator.device)
|
106 |
+
# input_ids2 = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True, return_tensors="pt") #.to('cuda')
|
107 |
+
# print(f"Converted input_ids dtype: {input_ids.dtype}")
|
108 |
+
# input_str= str(input_ids2)
|
109 |
+
# print('input str = ', input_str)
|
110 |
|
111 |
|
112 |
# gen_tokens = model.generate(
|