File size: 511 Bytes
de906df 1789c36 394fff2 37d4558 394fff2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 |
---
license: mit
---
## Model description
This is a Turkish RoBERTa base model pretrained on Turkish Wikipedia, Turkish OSCAR, and some news websites.
The final training corpus has a size of 38 GB and 329.720.508 sentences.
# Usage
Load transformers library with:
```
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("burakaytan/roberta-base-turkish-uncased")
model = AutoModelForMaskedLM.from_pretrained("burakaytan/roberta-base-turkish-uncased")
```
|