Update README.md
Browse files
README.md
CHANGED
@@ -10,31 +10,28 @@ tags:
|
|
10 |
- sft
|
11 |
---
|
12 |
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
**
|
18 |
-
|
19 |
-
**
|
20 |
-
|
21 |
-
**
|
22 |
-
|
23 |
-
**
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
**V1: 20.40 %**
|
36 |
-
**V2: 20.65 %**
|
37 |
-
----------------
|
38 |
# Gladiator-Mini-exp-1221-Instruct
|
39 |
|
40 |
**Gladiator-Mini-exp-1221** is a 3-billion parameter language model focused on **complex reasoning**. Built upon the foundation of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct), this experimental model is designed to explore what's achievable with smaller models in analytical thinking. It's all about pushing boundaries and learning what's possible in resource-efficient AI. We believe small models represent the future of open source language models, making AI more accessible and adaptable for a wider range of users and applications.
|
|
|
10 |
- sft
|
11 |
---
|
12 |
|
13 |
+
## Gladiator-Mini-Exp-1221-3B-Instruct - V2: Enhanced Performance
|
14 |
+
|
15 |
+
**This is V2, an improved iteration of our Gladiator-Mini-Exp-1221-3B-Instruct model, fine-tuned from Llama-3.2-3B-Instruct with a slightly expanded dataset and increased training epochs.**
|
16 |
+
|
17 |
+
**Major Improvements in V2:**
|
18 |
+
|
19 |
+
* **Superior Performance:** V2 demonstrates enhanced performance across multiple benchmarks compared to V1.
|
20 |
+
* **Mathematics Boost:** Notably, V2 surpasses both the 1211 model and V1 in mathematical reasoning, as evidenced by the MATH benchmark results below.
|
21 |
+
* **Model Size:** 3.21 Billion parameters
|
22 |
+
|
23 |
+
**Benchmark Highlights:**
|
24 |
+
|
25 |
+
| Benchmark | 1211 | V1 | V2 |
|
26 |
+
| :-------- | :-------- | :-------- | :-------- |
|
27 |
+
| **MATH** | **13.44%** | **13.07%** | **13.75%** |
|
28 |
+
| IFEval | | 60.79% | 62.15% |
|
29 |
+
| BBH | | 20.40% | 20.65% |
|
30 |
+
|
31 |
+
[Link to V1](https://huggingface.co/MultivexAI/Gladiator-Mini-Exp-1221-3B-Instruct)
|
32 |
+
|
33 |
+
**In summary, V2 offers a noticeable performance upgrade over V1, particularly in mathematical tasks. Explore the model and experience the improvements!**
|
34 |
+
|
|
|
|
|
|
|
35 |
# Gladiator-Mini-exp-1221-Instruct
|
36 |
|
37 |
**Gladiator-Mini-exp-1221** is a 3-billion parameter language model focused on **complex reasoning**. Built upon the foundation of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct), this experimental model is designed to explore what's achievable with smaller models in analytical thinking. It's all about pushing boundaries and learning what's possible in resource-efficient AI. We believe small models represent the future of open source language models, making AI more accessible and adaptable for a wider range of users and applications.
|