update readme
Browse files
README.md
CHANGED
@@ -1,21 +1,21 @@
|
|
1 |
Hugging Face's logo
|
2 |
---
|
3 |
-
language:
|
4 |
datasets:
|
5 |
|
6 |
---
|
7 |
-
# xlm-roberta-base-finetuned-
|
8 |
## Model description
|
9 |
-
**xlm-roberta-base-finetuned-
|
10 |
|
11 |
-
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on
|
12 |
## Intended uses & limitations
|
13 |
#### How to use
|
14 |
You can use this model with Transformers *pipeline* for masked token prediction.
|
15 |
```python
|
16 |
>>> from transformers import pipeline
|
17 |
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-naija')
|
18 |
-
>>>
|
19 |
|
20 |
|
21 |
|
@@ -23,15 +23,15 @@ You can use this model with Transformers *pipeline* for masked token prediction.
|
|
23 |
#### Limitations and bias
|
24 |
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
|
25 |
## Training data
|
26 |
-
This model was fine-tuned on JW300 + [
|
27 |
|
28 |
## Training procedure
|
29 |
This model was trained on a single NVIDIA V100 GPU
|
30 |
|
31 |
## Eval results on Test set (F-score, average over 5 runs)
|
32 |
-
Dataset| XLM-R F1 |
|
33 |
-|-|-
|
34 |
-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) |
|
35 |
|
36 |
### BibTeX entry and citation info
|
37 |
By David Adelani
|
|
|
1 |
Hugging Face's logo
|
2 |
---
|
3 |
+
language: pcm
|
4 |
datasets:
|
5 |
|
6 |
---
|
7 |
+
# xlm-roberta-base-finetuned-naija
|
8 |
## Model description
|
9 |
+
**xlm-roberta-base-finetuned-naija** is a **Nigerian Pidgin RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Nigerian Pidgin language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
|
10 |
|
11 |
+
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Nigerian Pidgin corpus.
|
12 |
## Intended uses & limitations
|
13 |
#### How to use
|
14 |
You can use this model with Transformers *pipeline* for masked token prediction.
|
15 |
```python
|
16 |
>>> from transformers import pipeline
|
17 |
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-naija')
|
18 |
+
>>> unmasker("Another attack on ambulance happen for Koforidua in March <mask> year where robbers kill Ambulance driver")
|
19 |
|
20 |
|
21 |
|
|
|
23 |
#### Limitations and bias
|
24 |
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
|
25 |
## Training data
|
26 |
+
This model was fine-tuned on JW300 + [BBC Pidgin](https://www.bbc.com/pidgin)
|
27 |
|
28 |
## Training procedure
|
29 |
This model was trained on a single NVIDIA V100 GPU
|
30 |
|
31 |
## Eval results on Test set (F-score, average over 5 runs)
|
32 |
+
Dataset| XLM-R F1 | pcm_roberta F1
|
33 |
-|-|-
|
34 |
+
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 87.26 | 90.00
|
35 |
|
36 |
### BibTeX entry and citation info
|
37 |
By David Adelani
|