RVN commited on
Commit
c6d978e
1 Parent(s): f4a568a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -35,7 +35,7 @@ After de-duplicating the data, we were left with a total of 54.5 GB of Bulgarian
35
 
36
  # Benchmark performance
37
 
38
- We tested the performance of BERTovski on benchmarks of XPOS, UPOS and NER. For Bulgarian, we used the data from the [Universal Dependencies](https://universaldependencies.org/) project. For Macedonian, we used the data sets created in the [babushka-bench](https://github.com/clarinsi/babushka-bench/) project. We also tested on a human translated version of the COPA data set (for details see our [Github repo](https://github.com/RikVN/COPA)). We compare performance to the strong multi-lingual models XLMR-base and XLMR-large. For details regarding the fine-tuning procedure you can checkout our [Github](https://github.com/macocu/LanguageModels).
39
 
40
  Scores are averages of three runs, except for COPA, for which we use 10 runs. We use the same hyperparameter settings for all models.
41
 
 
35
 
36
  # Benchmark performance
37
 
38
+ We tested the performance of BERTovski on benchmarks of XPOS, UPOS and NER. For Bulgarian, we used the data from the [Universal Dependencies](https://universaldependencies.org/) project. For Macedonian, we used the data sets created in the [babushka-bench](https://github.com/clarinsi/babushka-bench/) project. We also tested on a Google translated version of the COPA data set (for details see our [Github repo](https://github.com/RikVN/COPA)). We compare performance to the strong multi-lingual models XLMR-base and XLMR-large. For details regarding the fine-tuning procedure you can checkout our [Github](https://github.com/macocu/LanguageModels).
39
 
40
  Scores are averages of three runs, except for COPA, for which we use 10 runs. We use the same hyperparameter settings for all models.
41