- app.py +1 -1
- app_test.py +1 -1
app.py
CHANGED
@@ -342,7 +342,7 @@ def evaluate_v1(inputs, model, quantizer, tokenizer, width, height, device, do_s
|
|
342 |
input_intension = '{"wholecaption":"' + json_example["wholecaption"] + '","layout":[{"layer":'
|
343 |
inputs = tokenizer(
|
344 |
input_intension, return_tensors="pt"
|
345 |
-
).to(
|
346 |
|
347 |
stopping_criteria = StoppingCriteriaList()
|
348 |
stopping_criteria.append(StopAtSpecificTokenCriteria(token_id_list=[128000]))
|
|
|
342 |
input_intension = '{"wholecaption":"' + json_example["wholecaption"] + '","layout":[{"layer":'
|
343 |
inputs = tokenizer(
|
344 |
input_intension, return_tensors="pt"
|
345 |
+
).to(model.lm.device)
|
346 |
|
347 |
stopping_criteria = StoppingCriteriaList()
|
348 |
stopping_criteria.append(StopAtSpecificTokenCriteria(token_id_list=[128000]))
|
app_test.py
CHANGED
@@ -342,7 +342,7 @@ def evaluate_v1(inputs, model, quantizer, tokenizer, width, height, device, do_s
|
|
342 |
input_intension = '{"wholecaption":"' + json_example["wholecaption"] + '","layout":[{"layer":'
|
343 |
inputs = tokenizer(
|
344 |
input_intension, return_tensors="pt"
|
345 |
-
).to(
|
346 |
|
347 |
stopping_criteria = StoppingCriteriaList()
|
348 |
stopping_criteria.append(StopAtSpecificTokenCriteria(token_id_list=[128000]))
|
|
|
342 |
input_intension = '{"wholecaption":"' + json_example["wholecaption"] + '","layout":[{"layer":'
|
343 |
inputs = tokenizer(
|
344 |
input_intension, return_tensors="pt"
|
345 |
+
).to(model.lm.device)
|
346 |
|
347 |
stopping_criteria = StoppingCriteriaList()
|
348 |
stopping_criteria.append(StopAtSpecificTokenCriteria(token_id_list=[128000]))
|