Update README.md
Browse files
README.md
CHANGED
@@ -31,4 +31,102 @@ widget:
|
|
31 |
pipeline_tag: text-generation
|
32 |
---
|
33 |
|
34 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
pipeline_tag: text-generation
|
32 |
---
|
33 |
|
34 |
+
.
|
35 |
+
|
36 |
+
## Introduction
|
37 |
+
|
38 |
+
**SUS-CHhat** is powered by SUSTech x IDEA-CCNL, based on `01-ai/Yi-34B`
|
39 |
+
|
40 |
+
## News
|
41 |
+
|
42 |
+
<details open>
|
43 |
+
<summary>π― <b>2023/11/23</b>: The chat models are open to public.</summary>
|
44 |
+
|
45 |
+
This release contains two chat models based on previous released base models, two 8-bits models quantized by GPTQ, two 4-bits models quantized by AWQ.
|
46 |
+
|
47 |
+
- `Yi-34B-Chat`
|
48 |
+
- `Yi-34B-Chat-4bits`
|
49 |
+
- `Yi-34B-Chat-8bits`
|
50 |
+
- `Yi-6B-Chat`
|
51 |
+
- `Yi-6B-Chat-4bits`
|
52 |
+
- `Yi-6B-Chat-8bits`
|
53 |
+
|
54 |
+
You can try some of them interactively at:
|
55 |
+
|
56 |
+
- [HuggingFace](https://huggingface.co/spaces/01-ai/Yi-34B-Chat)
|
57 |
+
- [Replicate](https://replicate.com/01-ai)
|
58 |
+
</details>
|
59 |
+
|
60 |
+
<details open>
|
61 |
+
<summary>π <b>2023/11/23</b>: The Yi Series Models Community License Agreement is updated to v2.1.</summary>
|
62 |
+
</details>
|
63 |
+
|
64 |
+
<details>
|
65 |
+
<summary>π₯ <b>2023/11/08</b>: Invited test of Yi-34B chat model.</summary>
|
66 |
+
|
67 |
+
Application form:
|
68 |
+
|
69 |
+
- [English](https://cn.mikecrm.com/l91ODJf)
|
70 |
+
- [Chinese](https://cn.mikecrm.com/gnEZjiQ)
|
71 |
+
|
72 |
+
</details>
|
73 |
+
|
74 |
+
<details>
|
75 |
+
<summary>π― <b>2023/11/05</b>: The base model of <code>Yi-6B-200K</code> and <code>Yi-34B-200K</code>.</summary>
|
76 |
+
|
77 |
+
This release contains two base models with the same parameter sizes of previous
|
78 |
+
release, except that the context window is extended to 200K.
|
79 |
+
|
80 |
+
</details>
|
81 |
+
|
82 |
+
<details>
|
83 |
+
<summary>π― <b>2023/11/02</b>: The base model of <code>Yi-6B</code> and <code>Yi-34B</code>.</summary>
|
84 |
+
|
85 |
+
The first public release contains two bilingual (English/Chinese) base models
|
86 |
+
with the parameter sizes of 6B and 34B. Both of them are trained with 4K
|
87 |
+
sequence length and can be extended to 32K during inference time.
|
88 |
+
|
89 |
+
</details>
|
90 |
+
|
91 |
+
## Model Performance
|
92 |
+
|
93 |
+
### Base Model Performance
|
94 |
+
|
95 |
+
| Model | MMLU | CMMLU | C-Eval | GAOKAO | BBH | Common-sense Reasoning | Reading Comprehension | Math & Code |
|
96 |
+
| :------------ | :------: | :------: | :------: | :------: | :------: | :--------------------: | :-------------------: | :---------: |
|
97 |
+
| | 5-shot | 5-shot | 5-shot | 0-shot | 3-shot@1 | - | - | - |
|
98 |
+
| LLaMA2-34B | 62.6 | - | - | - | 44.1 | 69.9 | 68.0 | 26.0 |
|
99 |
+
| LLaMA2-70B | 68.9 | 53.3 | - | 49.8 | 51.2 | 71.9 | 69.4 | 36.8 |
|
100 |
+
| Baichuan2-13B | 59.2 | 62.0 | 58.1 | 54.3 | 48.8 | 64.3 | 62.4 | 23.0 |
|
101 |
+
| Qwen-14B | 66.3 | 71.0 | 72.1 | 62.5 | 53.4 | 73.3 | 72.5 | **39.8** |
|
102 |
+
| Skywork-13B | 62.1 | 61.8 | 60.6 | 68.1 | 41.7 | 72.4 | 61.4 | 24.9 |
|
103 |
+
| InternLM-20B | 62.1 | 59.0 | 58.8 | 45.5 | 52.5 | 78.3 | - | 30.4 |
|
104 |
+
| Aquila-34B | 67.8 | 71.4 | 63.1 | - | - | - | - | - |
|
105 |
+
| Falcon-180B | 70.4 | 58.0 | 57.8 | 59.0 | 54.0 | 77.3 | 68.8 | 34.0 |
|
106 |
+
| Yi-6B | 63.2 | 75.5 | 72.0 | 72.2 | 42.8 | 72.3 | 68.7 | 19.8 |
|
107 |
+
| Yi-6B-200K | 64.0 | 75.3 | 73.5 | 73.9 | 42.0 | 72.0 | 69.1 | 19.0 |
|
108 |
+
| **Yi-34B** | **76.3** | **83.7** | 81.4 | 82.8 | **54.3** | **80.1** | 76.4 | 37.1 |
|
109 |
+
| Yi-34B-200K | 76.1 | 83.6 | **81.9** | **83.4** | 52.7 | 79.7 | **76.6** | 36.3 |
|
110 |
+
|
111 |
+
While benchmarking open-source models, we have observed a disparity between the
|
112 |
+
results generated by our pipeline and those reported in public sources (e.g.
|
113 |
+
OpenCompass). Upon conducting a more in-depth investigation of this difference,
|
114 |
+
we have discovered that various models may employ different prompts,
|
115 |
+
post-processing strategies, and sampling techniques, potentially resulting in
|
116 |
+
significant variations in the outcomes. Our prompt and post-processing strategy
|
117 |
+
remains consistent with the original benchmark, and greedy decoding is employed
|
118 |
+
during evaluation without any post-processing for the generated content. For
|
119 |
+
scores that were not reported by the original authors (including scores reported
|
120 |
+
with different settings), we try to get results with our pipeline.
|
121 |
+
|
122 |
+
To evaluate the model's capability extensively, we adopted the methodology
|
123 |
+
outlined in Llama2. Specifically, we included PIQA, SIQA, HellaSwag, WinoGrande,
|
124 |
+
ARC, OBQA, and CSQA to assess common sense reasoning. SquAD, QuAC, and BoolQ
|
125 |
+
were incorporated to evaluate reading comprehension. CSQA was exclusively tested
|
126 |
+
using a 7-shot setup, while all other tests were conducted with a 0-shot
|
127 |
+
configuration. Additionally, we introduced GSM8K (8-shot@1), MATH (4-shot@1),
|
128 |
+
HumanEval (0-shot@1), and MBPP (3-shot@1) under the category "Math & Code". Due
|
129 |
+
to technical constraints, we did not test Falcon-180 on QuAC and OBQA; the score
|
130 |
+
is derived by averaging the scores on the remaining tasks. Since the scores for
|
131 |
+
these two tasks are generally lower than the average, we believe that
|
132 |
+
Falcon-180B's performance was not underestimated.
|