Update README.md
Browse files
README.md
CHANGED
@@ -25,7 +25,7 @@ Modern: Bulgarian (bg), Croatian (hr), Czech (cs), Danish (da), Dutch (nl), Engl
|
|
25 |
Medieval: Middle High German (mhd), Latin (la), Middle Low German (gml), Old French (fro), Old Church Slavonic (chu), Early New High German (fnhd), Ancient and Medieval Greek (grc)
|
26 |
|
27 |
## Training and evaluation data
|
28 |
-
The model was fine-tuned using the Monasterium and Wikipedia datasets, which consist of text sequences in
|
29 |
|
30 |
## Training procedure
|
31 |
Fine-tuning was done via the Trainer API with WeightedLossTrainer.
|
|
|
25 |
Medieval: Middle High German (mhd), Latin (la), Middle Low German (gml), Old French (fro), Old Church Slavonic (chu), Early New High German (fnhd), Ancient and Medieval Greek (grc)
|
26 |
|
27 |
## Training and evaluation data
|
28 |
+
The model was fine-tuned using the Monasterium and Wikipedia datasets, which consist of text sequences in 41 languages. The training set contains 80k samples, while the validation and test sets contain 16k. The average accuracy on the test set is 99.59% (this matches the average macro/weighted F1-score, the test set being perfectly balanced).
|
29 |
|
30 |
## Training procedure
|
31 |
Fine-tuning was done via the Trainer API with WeightedLossTrainer.
|