Spaces:
Build error
Build error
no padding
Browse files- llm_toolkit/llm_utils.py +2 -2
- scripts/eval-epochs.sh +1 -1
- scripts/eval-mgtv-qwen2.5_4bit.sh +2 -1
- scripts/eval-shots.sh +1 -1
llm_toolkit/llm_utils.py
CHANGED
@@ -146,7 +146,7 @@ def eval_model(
|
|
146 |
inputs = tokenizer(
|
147 |
batch_prompts,
|
148 |
return_tensors="pt",
|
149 |
-
padding=
|
150 |
).to(device)
|
151 |
|
152 |
outputs = model.generate(
|
@@ -155,7 +155,7 @@ def eval_model(
|
|
155 |
repetition_penalty=repetition_penalty,
|
156 |
use_cache=False,
|
157 |
)
|
158 |
-
outputs = outputs[:, inputs["input_ids"].shape[1]
|
159 |
decoded_output = tokenizer.batch_decode(
|
160 |
outputs, skip_special_tokens=True
|
161 |
) # Skip special tokens for clean output
|
|
|
146 |
inputs = tokenizer(
|
147 |
batch_prompts,
|
148 |
return_tensors="pt",
|
149 |
+
padding=batch_size > 1, # Ensure all inputs in the batch have the same length
|
150 |
).to(device)
|
151 |
|
152 |
outputs = model.generate(
|
|
|
155 |
repetition_penalty=repetition_penalty,
|
156 |
use_cache=False,
|
157 |
)
|
158 |
+
outputs = outputs[:, inputs["input_ids"].shape[1]:]
|
159 |
decoded_output = tokenizer.batch_decode(
|
160 |
outputs, skip_special_tokens=True
|
161 |
) # Skip special tokens for clean output
|
scripts/eval-epochs.sh
CHANGED
@@ -7,7 +7,7 @@ pwd
|
|
7 |
|
8 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
9 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
10 |
-
export USING_LLAMA_FACTORY=true
|
11 |
export USING_P1_PROMPT_TEMPLATE=false
|
12 |
|
13 |
export ORG_NAME=$1
|
|
|
7 |
|
8 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
9 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
10 |
+
# export USING_LLAMA_FACTORY=true
|
11 |
export USING_P1_PROMPT_TEMPLATE=false
|
12 |
|
13 |
export ORG_NAME=$1
|
scripts/eval-mgtv-qwen2.5_4bit.sh
CHANGED
@@ -13,17 +13,18 @@ cat /etc/os-release
|
|
13 |
lscpu
|
14 |
grep MemTotal /proc/meminfo
|
15 |
|
|
|
16 |
export LOAD_IN_4BIT=true
|
17 |
$BASEDIR/scripts/eval-epochs.sh Qwen Qwen2.5-72B-Instruct
|
18 |
|
19 |
export START_NUM_SHOTS=5
|
20 |
$BASEDIR/scripts/eval-shots_4bit.sh Qwen Qwen2.5-72B-Instruct
|
21 |
|
22 |
-
|
23 |
export START_NUM_SHOTS=40
|
24 |
$BASEDIR/scripts/eval-shots_4bit.sh shenzhi-wang Llama3.1-70B-Chinese-Chat
|
25 |
|
26 |
export LOAD_IN_4BIT=false
|
|
|
27 |
$BASEDIR/scripts/eval-shots.sh internlm internlm2_5-7b-chat
|
28 |
|
29 |
export START_NUM_SHOTS=50
|
|
|
13 |
lscpu
|
14 |
grep MemTotal /proc/meminfo
|
15 |
|
16 |
+
export USING_LLAMA_FACTORY=false
|
17 |
export LOAD_IN_4BIT=true
|
18 |
$BASEDIR/scripts/eval-epochs.sh Qwen Qwen2.5-72B-Instruct
|
19 |
|
20 |
export START_NUM_SHOTS=5
|
21 |
$BASEDIR/scripts/eval-shots_4bit.sh Qwen Qwen2.5-72B-Instruct
|
22 |
|
|
|
23 |
export START_NUM_SHOTS=40
|
24 |
$BASEDIR/scripts/eval-shots_4bit.sh shenzhi-wang Llama3.1-70B-Chinese-Chat
|
25 |
|
26 |
export LOAD_IN_4BIT=false
|
27 |
+
export USING_LLAMA_FACTORY=true
|
28 |
$BASEDIR/scripts/eval-shots.sh internlm internlm2_5-7b-chat
|
29 |
|
30 |
export START_NUM_SHOTS=50
|
scripts/eval-shots.sh
CHANGED
@@ -7,7 +7,7 @@ pwd
|
|
7 |
|
8 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
9 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
10 |
-
export USING_LLAMA_FACTORY=true
|
11 |
export USING_P1_PROMPT_TEMPLATE=false
|
12 |
export LOAD_IN_4BIT=false
|
13 |
|
|
|
7 |
|
8 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
9 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
10 |
+
# export USING_LLAMA_FACTORY=true
|
11 |
export USING_P1_PROMPT_TEMPLATE=false
|
12 |
export LOAD_IN_4BIT=false
|
13 |
|