Overview
- Model Name: FILL-MAsk-RoBERTa-base
- Task: Masked Language Modeling (FILL-MAsk)
- Dataset: WikiText2
Model Description
FILL-MAsk-RoBERTa-base is a distilled version of the RoBERTa-base model, designed for the Masked Language Modeling task. This model follows a similar training procedure as DistilBERT, resulting in a smaller model with 6 layers, 768 dimensions, and 12 attention heads. It contains a total of 82 million parameters, making it more lightweight compared to the original RoBERTa-base, which has 125 million parameters. On average, DistilRoBERTa is approximately twice as fast as RoBERTa-base.
Usage
FILL-MAsk-RoBERTa-base can be used for both direct and downstream tasks. It is suitable for masked language modeling tasks, where tokens are masked, and the model must predict the masked tokens. It is also intended to be fine-tuned on downstream tasks such as sequence classification, token classification, or question answering. Users can explore the Hugging Face Model Hub to find fine-tuned versions of this model for specific tasks of interest.
Limitations
Bias: Significant research has explored bias and fairness issues with language models. Predictions generated by this model may contain biases, including harmful stereotypes related to protected classes, identity characteristics, and sensitive social and occupational groups.
Fairness: It's essential to be aware of fairness considerations when using this model and to ensure that its predictions do not contribute to unfair or harmful outcomes.
Ethical Use: Users are encouraged to use this model ethically and responsibly, taking into account the potential for bias and ensuring that it does not generate harmful or offensive content.
Usage :
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("fill-mask", model="ayoubkirouane/FILL-MAsk-RoBERTa-base")
result = pipe("The capital of Algeria is <mask>.")
print(result)
- Downloads last month
- 16
Model tree for ayoubkirouane/FILL-MAsk-RoBERTa-base
Base model
distilbert/distilroberta-base