File size: 511 Bytes
de906df
 
 
1789c36
 
 
 
 
394fff2
 
 
 
 
37d4558
 
394fff2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
---
license: mit
---
## Model description
This is a Turkish RoBERTa base model pretrained on Turkish Wikipedia, Turkish OSCAR, and some news websites.

The final training corpus has a size of 38 GB and 329.720.508 sentences.

# Usage
Load transformers library with:
```
from transformers import AutoTokenizer, AutoModelForMaskedLM
  
tokenizer = AutoTokenizer.from_pretrained("burakaytan/roberta-base-turkish-uncased")
model = AutoModelForMaskedLM.from_pretrained("burakaytan/roberta-base-turkish-uncased")
```