MultivexAI commited on
Commit
06fd6a6
·
verified ·
1 Parent(s): 63a192d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -25
README.md CHANGED
@@ -10,31 +10,28 @@ tags:
10
  - sft
11
  ---
12
 
13
- * **Model size: 3.21B parameters**
14
-
15
- # **V2**
16
-
17
- **This version is a further fine-tuned version of Llama-3.2-3B-Instruct with a slightly larger dataset and more epochs.**
18
-
19
- **The V2 version performs better in multiple benchmarks than the V1 version, and even outmatches both 1211 and V1 in mathematics: https://huggingface.co/MultivexAI/Gladiator-Mini-Exp-1221-3B-Instruct**
20
-
21
- **Examples:**
22
- ----------------
23
- **MATH**
24
- **1211: 13.44 %**
25
- **V1: 13.07 %**
26
- **V2: 13.75 %**
27
- ----------------
28
- ----------------
29
- **IFEval**
30
- **V1: 60.79 %**
31
- **V2: 62.15 %**
32
- ----------------
33
- ----------------
34
- **BBH**
35
- **V1: 20.40 %**
36
- **V2: 20.65 %**
37
- ----------------
38
  # Gladiator-Mini-exp-1221-Instruct
39
 
40
  **Gladiator-Mini-exp-1221** is a 3-billion parameter language model focused on **complex reasoning**. Built upon the foundation of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct), this experimental model is designed to explore what's achievable with smaller models in analytical thinking. It's all about pushing boundaries and learning what's possible in resource-efficient AI. We believe small models represent the future of open source language models, making AI more accessible and adaptable for a wider range of users and applications.
 
10
  - sft
11
  ---
12
 
13
+ ## Gladiator-Mini-Exp-1221-3B-Instruct - V2: Enhanced Performance
14
+
15
+ **This is V2, an improved iteration of our Gladiator-Mini-Exp-1221-3B-Instruct model, fine-tuned from Llama-3.2-3B-Instruct with a slightly expanded dataset and increased training epochs.**
16
+
17
+ **Major Improvements in V2:**
18
+
19
+ * **Superior Performance:** V2 demonstrates enhanced performance across multiple benchmarks compared to V1.
20
+ * **Mathematics Boost:** Notably, V2 surpasses both the 1211 model and V1 in mathematical reasoning, as evidenced by the MATH benchmark results below.
21
+ * **Model Size:** 3.21 Billion parameters
22
+
23
+ **Benchmark Highlights:**
24
+
25
+ | Benchmark | 1211 | V1 | V2 |
26
+ | :-------- | :-------- | :-------- | :-------- |
27
+ | **MATH** | **13.44%** | **13.07%** | **13.75%** |
28
+ | IFEval | | 60.79% | 62.15% |
29
+ | BBH | | 20.40% | 20.65% |
30
+
31
+ [Link to V1](https://huggingface.co/MultivexAI/Gladiator-Mini-Exp-1221-3B-Instruct)
32
+
33
+ **In summary, V2 offers a noticeable performance upgrade over V1, particularly in mathematical tasks. Explore the model and experience the improvements!**
34
+
 
 
 
35
  # Gladiator-Mini-exp-1221-Instruct
36
 
37
  **Gladiator-Mini-exp-1221** is a 3-billion parameter language model focused on **complex reasoning**. Built upon the foundation of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct), this experimental model is designed to explore what's achievable with smaller models in analytical thinking. It's all about pushing boundaries and learning what's possible in resource-efficient AI. We believe small models represent the future of open source language models, making AI more accessible and adaptable for a wider range of users and applications.