mlm-spanish-roberta-base
This model has a RoBERTa base architecture and was trained from scratch with 3.6 GB of raw text over 10 epochs. 4 Tesla V-100 GPUs were used for the training.
To test the quality of the resulting model we evaluate it over the GLUES benchmark for Spanish NLU. The results are the following:
Task | Score (metric) |
---|---|
XNLI | 71.99 (accuracy) |
Paraphrasing | 74.85 (accuracy) |
NER | 85.34 (F1) |
POS | 97.49 (accuracy) |
Dependency Parsing | 85.14/81.08 (UAS/LAS) |
Document Classification | 93.00 (accuracy) |
- Downloads last month
- 134
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.