add NTREX evaluation to model card
Browse files
README.md
CHANGED
@@ -118,7 +118,7 @@ The model was trained for a total of 36.000 updates. Weights were saved every 10
|
|
118 |
|
119 |
### Variable and metrics
|
120 |
|
121 |
-
We use the BLEU score for evaluation on the
|
122 |
|
123 |
### Evaluation results
|
124 |
|
@@ -128,7 +128,8 @@ Below are the evaluation results on the machine translation from Catalan to Ital
|
|
128 |
|----------------------|------------|------------------|---------------|
|
129 |
| Flores 101 dev | 24,3 | **28,5** | 26,1 |
|
130 |
| Flores 101 devtest |24,7 | **29,1** | 26,3 |
|
131 |
-
|
|
|
|
132 |
|
133 |
## Additional information
|
134 |
|
|
|
118 |
|
119 |
### Variable and metrics
|
120 |
|
121 |
+
We use the BLEU score for evaluation on the [Flores-101](https://github.com/facebookresearch/flores), and [NTREX](https://github.com/MicrosoftTranslator/NTREX) evaluation datasets.
|
122 |
|
123 |
### Evaluation results
|
124 |
|
|
|
128 |
|----------------------|------------|------------------|---------------|
|
129 |
| Flores 101 dev | 24,3 | **28,5** | 26,1 |
|
130 |
| Flores 101 devtest |24,7 | **29,1** | 26,3 |
|
131 |
+
| NTREX | 27,2 | **31,6** | 28,3 |
|
132 |
+
| Average | 25,4 | **29,7** | 26,9 |
|
133 |
|
134 |
## Additional information
|
135 |
|