qanthony-z
commited on
Commit
•
7274961
1
Parent(s):
1cb4251
update bar charts
Browse files
README.md
CHANGED
@@ -54,11 +54,12 @@ print((tokenizer.decode(outputs[0])))
|
|
54 |
Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat performance for a model of its size and matches strong models significantly larger. For instance, Zamba2-1.2B-Instruct outperforms Gemma2-2B-Instruct, a very strong model over 2x its size.
|
55 |
|
56 |
<center>
|
57 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/
|
58 |
</center>
|
59 |
|
60 |
-
|
61 |
-
|
|
|
62 |
| **Zamba2-1.2B-Instruct** | 1.2B | **59.53** | **41.45** |
|
63 |
| Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
|
64 |
| H2O-Danube-1.8B-Chat | 1.6B | 49.78 | 27.95 |
|
@@ -69,7 +70,7 @@ Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat
|
|
69 |
Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
|
70 |
|
71 |
<center>
|
72 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/
|
73 |
</center>
|
74 |
|
75 |
|
|
|
54 |
Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat performance for a model of its size and matches strong models significantly larger. For instance, Zamba2-1.2B-Instruct outperforms Gemma2-2B-Instruct, a very strong model over 2x its size.
|
55 |
|
56 |
<center>
|
57 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/ceOUHVeJPhBgwTDCsR9Y6.png" width="900"/>
|
58 |
</center>
|
59 |
|
60 |
+
|
61 |
+
| Model | Size | Aggregate MT-Bench | IFEval |
|
62 |
+
|:-------------:|:----:|:-------------:|:----:|
|
63 |
| **Zamba2-1.2B-Instruct** | 1.2B | **59.53** | **41.45** |
|
64 |
| Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
|
65 |
| H2O-Danube-1.8B-Chat | 1.6B | 49.78 | 27.95 |
|
|
|
70 |
Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
|
71 |
|
72 |
<center>
|
73 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/tQ-j1krA634EfTU1Lp3E7.png" width="700" alt="Zamba performance">
|
74 |
</center>
|
75 |
|
76 |
|