huangzixian
commited on
Commit
·
81fff73
1
Parent(s):
4381330
update readme
Browse files
README.md
CHANGED
@@ -1,10 +1,8 @@
|
|
1 |
### Model Sources
|
2 |
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
Repository: https://github.com/CONE-MT/
|
8 |
|
9 |
### Model Description
|
10 |
|
@@ -17,7 +15,7 @@ We collected extensive training sets in 102 languages for continued pre-training
|
|
17 |
LLaMAX supports translation between more than 100 languages, surpassing the performance of similarly scaled LLMs.
|
18 |
|
19 |
```angular2html
|
20 |
-
def
|
21 |
instruction = f'Translate the following sentences from {src_language} to {trg_language}.'
|
22 |
prompt = (
|
23 |
'Below is an instruction that describes a task, paired with an input that provides further context. '
|
@@ -47,19 +45,19 @@ tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokeniza
|
|
47 |
|
48 |
### 🔥 Excellent Translation Performance
|
49 |
|
50 |
-
LLaMAX3-Alpaca achieves an average spBLEU score improvement of over **5 points** compared to the LLaMA3-8B-Alpaca model on the Flores-101 dataset.
|
51 |
|
52 |
|
53 |
| System | Size | en-X (COMET) | en-X (BLEU) | zh-X (COMET)| zh-X (BLEU) | de-X (COMET) | de-X (BLEU) | ne-X (COMET) | ne-X (BLEU) |ar-X (COMET) | ar-X (BLEU) | az-X (COMET) | az-X (BLEU) | ceb-X (COMET) | ceb-X (BLEU)|
|
54 |
|--------------------|------|--------------------|-------------| ----| ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |
|
55 |
-
| LLaMA3-Alpaca | 8B |67.97|17.23|64.65|10.14|64.67|13.62|62.95|7.96|63.45|11.27|60.61|6.98|55.26|8.52|
|
56 |
-
| LLaMAX3-Alpaca | 8B |75.52|22.77|73.16|14.43|73.47|18.95|75.13|15.32|72.29|16.42|72.06|12.41|68.88|15.85|
|
57 |
|
58 |
|
59 |
| System | Size | X-en (COMET) | X-en (BLEU) | X-zh (COMET)| X-zh (BLEU) | X-de (COMET) | X-de (BLEU) | X-ne (COMET) | X-ne (BLEU) |X-ar (COMET) | X-ar (BLEU) | X-az (COMET) | X-az (BLEU) | X-ceb (COMET) | X-ceb (BLEU) |
|
60 |
|--------------------|------|----------------|-------------| ----| ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |--------------|
|
61 |
-
| LLaMA3-Alpaca | 8B |77.43|26.55|73.56|13.17|71.59|16.82|46.56|3.83|66.49|10.20|58.30|4.81|52.68|4.18|
|
62 |
-
| LLaMAX3-Alpaca | 8B |81.28|31.85|78.34|16.46|76.23|20.64|65.83|14.16|75.84|15.45|70.61|9.32|63.35|12.66|
|
63 |
|
64 |
|
65 |
s
|
@@ -72,7 +70,7 @@ We implement multiple versions of the LLaMAX model, the model links are as follo
|
|
72 |
| Model | LLaMAX | LLaMAX-Alpaca |
|
73 |
|---------|----------------------------------------------------------|-----------------------------------------------------------------|
|
74 |
| Llama-2 | [Link](https://huggingface.co/LLaMAX/LLaMAX2-7B) | [Link](https://huggingface.co/LLaMAX/LLaMAX2-7B-Alpaca) |
|
75 |
-
| Llama-3 | [Link](https://huggingface.co/LLaMAX/LLaMAX3-8B) | [Link](https://huggingface.co/LLaMAX/LLaMAX3-8B-Alpaca) |
|
76 |
|
77 |
### Citation
|
78 |
If our model helps your work, please cite this paper:
|
|
|
1 |
### Model Sources
|
2 |
|
3 |
+
- **Math Reasoning**: https://huggingface.co/LLaMAX/LLaMAX2-7B-MetaMath
|
4 |
+
- **Commonsense Reasoning**: https://huggingface.co/LLaMAX/LLaMAX2-7B-X-CSQA
|
5 |
+
- **Natural Language Inference**: https://huggingface.co/LLaMAX/LLaMAX2-7B-XNLI
|
|
|
|
|
6 |
|
7 |
### Model Description
|
8 |
|
|
|
15 |
LLaMAX supports translation between more than 100 languages, surpassing the performance of similarly scaled LLMs.
|
16 |
|
17 |
```angular2html
|
18 |
+
def Prompt_template(query, src_language, trg_language):
|
19 |
instruction = f'Translate the following sentences from {src_language} to {trg_language}.'
|
20 |
prompt = (
|
21 |
'Below is an instruction that describes a task, paired with an input that provides further context. '
|
|
|
45 |
|
46 |
### 🔥 Excellent Translation Performance
|
47 |
|
48 |
+
LLaMAX3-8B-Alpaca achieves an average spBLEU score improvement of over **5 points** compared to the LLaMA3-8B-Alpaca model on the Flores-101 dataset.
|
49 |
|
50 |
|
51 |
| System | Size | en-X (COMET) | en-X (BLEU) | zh-X (COMET)| zh-X (BLEU) | de-X (COMET) | de-X (BLEU) | ne-X (COMET) | ne-X (BLEU) |ar-X (COMET) | ar-X (BLEU) | az-X (COMET) | az-X (BLEU) | ceb-X (COMET) | ceb-X (BLEU)|
|
52 |
|--------------------|------|--------------------|-------------| ----| ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |
|
53 |
+
| LLaMA3-8B-Alpaca | 8B |67.97|17.23|64.65|10.14|64.67|13.62|62.95|7.96|63.45|11.27|60.61|6.98|55.26|8.52|
|
54 |
+
| LLaMAX3-8B-Alpaca | 8B |75.52|22.77|73.16|14.43|73.47|18.95|75.13|15.32|72.29|16.42|72.06|12.41|68.88|15.85|
|
55 |
|
56 |
|
57 |
| System | Size | X-en (COMET) | X-en (BLEU) | X-zh (COMET)| X-zh (BLEU) | X-de (COMET) | X-de (BLEU) | X-ne (COMET) | X-ne (BLEU) |X-ar (COMET) | X-ar (BLEU) | X-az (COMET) | X-az (BLEU) | X-ceb (COMET) | X-ceb (BLEU) |
|
58 |
|--------------------|------|----------------|-------------| ----| ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |--------------|
|
59 |
+
| LLaMA3-8B-Alpaca | 8B |77.43|26.55|73.56|13.17|71.59|16.82|46.56|3.83|66.49|10.20|58.30|4.81|52.68|4.18|
|
60 |
+
| LLaMAX3-8B-Alpaca | 8B |81.28|31.85|78.34|16.46|76.23|20.64|65.83|14.16|75.84|15.45|70.61|9.32|63.35|12.66|
|
61 |
|
62 |
|
63 |
s
|
|
|
70 |
| Model | LLaMAX | LLaMAX-Alpaca |
|
71 |
|---------|----------------------------------------------------------|-----------------------------------------------------------------|
|
72 |
| Llama-2 | [Link](https://huggingface.co/LLaMAX/LLaMAX2-7B) | [Link](https://huggingface.co/LLaMAX/LLaMAX2-7B-Alpaca) |
|
73 |
+
| Llama-3 | [Link](https://huggingface.co/LLaMAX/LLaMAX3-8B-8B) | [Link](https://huggingface.co/LLaMAX/LLaMAX3-8B-8B-Alpaca) |
|
74 |
|
75 |
### Citation
|
76 |
If our model helps your work, please cite this paper:
|