Update README.md
Browse files
README.md
CHANGED
@@ -47,22 +47,22 @@ cd auto-round/examples/language-modeling
|
|
47 |
python3 eval_042/evluation.py --model_name "Intel/Qwen2-7B-int4-inc" --eval_bs 16 --tasks lambada_openai,hellaswag,piqa,winogrande,truthfulqa_mc1,openbookqa,boolq,arc_easy,arc_challenge,mmlu,gsm8k,cmmlu,ceval-valid
|
48 |
```
|
49 |
|
50 |
-
| Metric | BF16 | INT4 |
|
51 |
-
| -------------- | ------ | ------ |
|
52 |
-
| Avg | 0.6655 | 0.6624 |
|
53 |
-
| mmlu | 0.6706 | 0.6650 |
|
54 |
-
| cmmlu | 0.8310 | 0.8151 |
|
55 |
-
| ceval-valid | 0.8232 | 0.8053 |
|
56 |
-
| lambada_openai | 0.7182 | 0.7254 |
|
57 |
-
| hellaswag | 0.5823 | 0.5758 |
|
58 |
-
| winogrande | 0.7222 | 0.7174 |
|
59 |
-
| piqa | 0.7911 | 0.7905 |
|
60 |
-
| truthfulqa_mc1 | 0.3647 | 0.3599 |
|
61 |
-
| openbookqa | 0.3520 | 0.3420 |
|
62 |
-
| boolq | 0.8183 | 0.8168 |
|
63 |
-
| arc_easy | 0.7660 | 0.7643 |
|
64 |
-
| arc_challenge | 0.4505 | 0.4625 |
|
65 |
-
| gsm8k 5 shots | 0.7619 | 0.7710 |
|
66 |
|
67 |
|
68 |
|
|
|
47 |
python3 eval_042/evluation.py --model_name "Intel/Qwen2-7B-int4-inc" --eval_bs 16 --tasks lambada_openai,hellaswag,piqa,winogrande,truthfulqa_mc1,openbookqa,boolq,arc_easy,arc_challenge,mmlu,gsm8k,cmmlu,ceval-valid
|
48 |
```
|
49 |
|
50 |
+
| Metric | BF16 | INT4 |
|
51 |
+
| -------------- | ------ | ------ |
|
52 |
+
| Avg | 0.6655 | 0.6624 |
|
53 |
+
| mmlu | 0.6706 | 0.6650 |
|
54 |
+
| cmmlu | 0.8310 | 0.8151 |
|
55 |
+
| ceval-valid | 0.8232 | 0.8053 |
|
56 |
+
| lambada_openai | 0.7182 | 0.7254 |
|
57 |
+
| hellaswag | 0.5823 | 0.5758 |
|
58 |
+
| winogrande | 0.7222 | 0.7174 |
|
59 |
+
| piqa | 0.7911 | 0.7905 |
|
60 |
+
| truthfulqa_mc1 | 0.3647 | 0.3599 |
|
61 |
+
| openbookqa | 0.3520 | 0.3420 |
|
62 |
+
| boolq | 0.8183 | 0.8168 |
|
63 |
+
| arc_easy | 0.7660 | 0.7643 |
|
64 |
+
| arc_challenge | 0.4505 | 0.4625 |
|
65 |
+
| gsm8k 5 shots | 0.7619 | 0.7710 |
|
66 |
|
67 |
|
68 |
|