Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- sw
|
5 |
+
- bn
|
6 |
+
- te
|
7 |
+
- th
|
8 |
+
- ja
|
9 |
+
- zh
|
10 |
+
- ru
|
11 |
+
- es
|
12 |
+
- fr
|
13 |
+
- de
|
14 |
+
- en
|
15 |
+
tags:
|
16 |
+
- text-generation
|
17 |
+
---
|
18 |
+
# mCoT: Multilingual Instruction Tuning for Reasoning Consistency in Language Models
|
19 |
+
|
20 |
+
Paper: []()
|
21 |
+
|
22 |
+
Code: []()
|
23 |
+
|
24 |
+
Dataset: []()
|
25 |
+
|
26 |
+
### Introduction
|
27 |
+
We introduce mCoT-MATH, a 7B parameter model for multilingual math reasoning, which achieves impressive consistency across languages. mCoT is based on [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) and trained on mCoT-MATH, the first large-scale multilingual math CoT reasoning dataset containing around 6.3 million samples for 11 diverse languages.
|
28 |
+
|
29 |
+
### 🤗 Dataset: [mCoT-MATH](https://huggingface.co/datasets/laihuiyuan/mCoT-MATH)
|
30 |
+
|
31 |
+
Based on [MetaMathQA](https://github.com/meta-math/MetaMath) and [MathInstruct](https://github.com/TIGER-AI-Lab/MAmmoTH)
|
32 |
+
, we compile [mCoT-MATH](https://huggingface.co/datasets/laihuiyuan/mCoT-MATH) using machine translation.
|
33 |
+
|
34 |
+
| Language | SW | BN | TE | TH | JA | ZH | RU | ES | FR | DE | DE |Overall |
|
35 |
+
|:----------|:------|:------|:------|:------|:------|:------|:------|:------|:------|:------|:------|--------|
|
36 |
+
| mCoT-MATH | ~580K | ~580K | ~580K | ~580K | ~580K | ~580K | ~580K | ~580K | ~580K | ~580K | ~580K | ~6.3M |
|
37 |
+
|
38 |
+
### Results on [MGSM](https://arxiv.org/abs/2210.03057v1)
|
39 |
+
| Language | SW | BN | TE | TH | JA | ZH | RU | ES | FR | DE | EN |
|
40 |
+
|:-----------------------|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|
|
41 |
+
| GPT-3 few-shot | 11.2 | 6.4 | 0.4 | 0.8 | 26.0 | 40.0 | 28.4 | 40.4 | 37.6 | 36.0 | 53.6 |
|
42 |
+
| GPT-3.5-En 2-shot | 40.0 | 7.6 | - | 15.6 | 46.8 | 52.8 | 50.4 | 61.2 | 59.2 | 62.0 | 67.2 |
|
43 |
+
| GPT4-En 2-shot | 64.4 | 17.6 | - | 40.4 | 71.6 | 70.0 | 64.0 | 71.2 | 72.0 | 73.6 | 80.0 |
|
44 |
+
| PaLM-540B few-shot | 35.2 | 46.0 | 45.6 | 52.8 | 40.0 | 46.8 | 48.4 | 56.8 | 46.4 | 49.2 | 62.4 |
|
45 |
+
| WizardMath-7B | 3.4 | 2.0 | - | 4.0 | 24.0 | 22.4 | 30.8 | 34.8 | 30.4 | 30.4 | 47.6 |
|
46 |
+
| MathOctopus-7B | 38.4 | 33.2 | - | 36.4 | 35.6 | 45.2 | 48.4 | 45.2 | 38.0 | 43.6 | 54.8 |
|
47 |
+
| MathOctopus-Mistral-7B | 51.6 | 44.0 | - | 48.8 | 48.0 | 51.6 | 49.6 | 53.2 | 47.2 | 50.0 | 58.4 |
|
48 |
+
| xCoT-7B | 48.4 | 40.4 | 42.8 | 49.2 | 50.0 | 50.0 | 50.0 | 48.8 | 49.6 | 47.2 | 48.4 |
|
49 |
+
| WizardMath-13B | 5.6 | 6.4 | - | 5.6 | 22.0 | 28.0 | 34.4 | 45.6 | 42.0 | 40.4 | 52.8 |
|
50 |
+
| MathOctopus-13B | 46.0 | 42.0 | - | 46.0 | 39.6 | 51.2 | 47.6 | 53.2 | 49.6 | 49.2 | 51.6 |
|
51 |
+
| xCoT-13B | 51.6 | 50.0 | 47.2 | 50.0 | 49.6 | 54.0 | 56.8 | 54.8 | 46.4 | 52.4 | 54.4 |
|
52 |
+
| mCoT-7B | 67.2 | 65.6 | 62.4 | 67.6 | 65.2 | 64.8 | 66.8 | 68.4 | 63.8 | 61.2 | 71.6 |
|
53 |
+
|
54 |
+
|
55 |
+
### Results on [MSVAMP](https://arxiv.org/abs/2310.20246)
|
56 |
+
| Language | SW | BN | TH | JA | ZH | RU | ES | FR | DE | EN | AVG |
|
57 |
+
|:-----------------------|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|:-----|
|
58 |
+
| GPT-3.5-En zero-shot | 63.2 | 3.1 | 24.4 | 63.3 | 72.4 | 62.3 | 69.5 | 71.9 | 66.7 | 76.1 | 57.3 |
|
59 |
+
| GPT-3.5-En 2-shot | 68.4 | 14.4 | 46.0 | 74.0 | 78.4 | 70.9 | 74.6 | 78.2 | 73.9 | 81.2 | 66.0 |
|
60 |
+
| GPT4-En 2-shot | 75.7 | 31.2 | 68.1 | 74.8 | 78.9 | 77.9 | 81.5 | 83.9 | 78.1 | 80.1 | 73.0 |
|
61 |
+
| PaLM-540B few-shot | 35.2 | 46.0 | 45.6 | 52.8 | 40.0 | 46.8 | 48.4 | 56.8 | 46.4 | 49.2 | 62.4 |
|
62 |
+
| WizardMath-7B | 10.3 | 16.1 | 6.3 | 26.7 | 26.8 | 33.7 | 42.9 | 39.9 | 39.6 | 45.1 | 27.0 |
|
63 |
+
| MathOctopus-7B | 42.3 | 32.8 | 40.5 | 43.2 | 43.2 | 42.1 | 44.5 | 45.3 | 43.1 | 46.8 | 42.4 |
|
64 |
+
| MathOctopus-Mistral-7B | 41.2 | 36.7 | 40.2 | 41.5 | 43.1 | 44.0 | 47.0 | 49.0 | 46.4 | 49.7 | 43.9 |
|
65 |
+
| WizardMath-13B | 12.5 | 13.7 | 16.3 | 29.5 | 37.0 | 43.8 | 50.4 | 49.4 | 48.7 | 56.3 | 35.8 |
|
66 |
+
| MathOctopus-13B | 43.4 | 34.2 | 39.5 | 43.1 | 46.4 | 48.2 | 48.2 | 49.9 | 47.7 | 44.6 | 44.5 |
|
67 |
+
| mCoT-7B | 55.0 | 53.7 | 56.4 | 58.8 | 58.2 | 58.1 | 58.9 | 58.8 | 61.1 | 58.3 | 57.7 |
|
68 |
+
|
69 |
+
### Prompt Template
|
70 |
+
```bash
|
71 |
+
# Language
|
72 |
+
bn = "আসুন ধাপে ধাপে চিন্তা করি।"
|
73 |
+
de = "Denken wir Schritt für Schritt."
|
74 |
+
en = "Let's think step by step."
|
75 |
+
es = "Pensemos paso a paso."
|
76 |
+
fr = "Réfléchissons étape par étape."
|
77 |
+
ja = "段階的に考えてみましょう。"
|
78 |
+
ru = "Давайте думать поэтапно."
|
79 |
+
sw = "Hebu fikiria hatua kwa hatua."
|
80 |
+
te = "అంచెలంచెలుగా ఆలోచిద్దాం."
|
81 |
+
th = "ลองคิดทีละขั้นตอน"
|
82 |
+
zh = "让我们一步步思考。"
|
83 |
+
|
84 |
+
# Math Question
|
85 |
+
math = "A robe takes 2 bolts of blue fiber and half that much white fiber. How many bolts in total does it take?"
|
86 |
+
|
87 |
+
Prompt = "Question: \n[Math Question] \nAnswer: \n[Language]\n[CoT Reasoning]"
|
88 |
+
```
|
89 |
+
|
90 |
+
### Citation
|
91 |
+
If you use any content from this repository, please cite our paper:
|
92 |
+
```
|
93 |
+
@inproceedings{lai-etal-2024-mcot,
|
94 |
+
title = "mCoT: Multilingual Instruction Tuning for Reasoning Consistency
|
95 |
+
in Language Models",
|
96 |
+
author = "Lai, Huiyuan and Nissim, Malvina",
|
97 |
+
booktitle = "Proceedings of the 62nd Annual Meeting of the Association
|
98 |
+
for Computational Linguistics,
|
99 |
+
month = aug,
|
100 |
+
address = "Bangkok, Thailand",
|
101 |
+
year = "2024",
|
102 |
+
publisher = "Association for Computational Linguistics"
|
103 |
+
}
|
104 |
+
```
|