qanthony-z commited on
Commit
7274961
1 Parent(s): 1cb4251

update bar charts

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -54,11 +54,12 @@ print((tokenizer.decode(outputs[0])))
54
  Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat performance for a model of its size and matches strong models significantly larger. For instance, Zamba2-1.2B-Instruct outperforms Gemma2-2B-Instruct, a very strong model over 2x its size.
55
 
56
  <center>
57
- <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/UdlseqQFDxYvEdLV5xurw.png" width="900"/>
58
  </center>
59
 
60
- | Model | Size | MT-Bench | IFEval |
61
- |-------------|----|----|----|
 
62
  | **Zamba2-1.2B-Instruct** | 1.2B | **59.53** | **41.45** |
63
  | Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
64
  | H2O-Danube-1.8B-Chat | 1.6B | 49.78 | 27.95 |
@@ -69,7 +70,7 @@ Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat
69
  Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
70
 
71
  <center>
72
- <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/Q82BVdIppSyqPBHYEAjAl.png" width="700" alt="Zamba performance">
73
  </center>
74
 
75
 
 
54
  Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat performance for a model of its size and matches strong models significantly larger. For instance, Zamba2-1.2B-Instruct outperforms Gemma2-2B-Instruct, a very strong model over 2x its size.
55
 
56
  <center>
57
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/ceOUHVeJPhBgwTDCsR9Y6.png" width="900"/>
58
  </center>
59
 
60
+
61
+ | Model | Size | Aggregate MT-Bench | IFEval |
62
+ |:-------------:|:----:|:-------------:|:----:|
63
  | **Zamba2-1.2B-Instruct** | 1.2B | **59.53** | **41.45** |
64
  | Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
65
  | H2O-Danube-1.8B-Chat | 1.6B | 49.78 | 27.95 |
 
70
  Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
71
 
72
  <center>
73
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/tQ-j1krA634EfTU1Lp3E7.png" width="700" alt="Zamba performance">
74
  </center>
75
 
76