--- widget: - text: Em uma bela manhã de - text: Em uma cidade tão grande como - text: Maria e Joana são license: mit datasets: - mc4 language: - pt metrics: - perplexity library_name: transformers pipeline_tag: text-generation --- # OPT-125M finetuned Portuguese Fine-tuning the [OPT-125M](https://huggingface.co/facebook/opt-125m) model on a reduced corpus of MC4-Portuguese with approximately 300M tokens. In this training a sequence length of 512 tokens was used, batch of 32 for 2 epochs. With an A100 with 40GB of RAM, the training took around 3 hours **Perplexity:** 9.4 ## Sample Use ```python from transformers import pipeline generator = pipeline('text-generation', model='Mirelle/opt-125M-pt-br-finetuned', max_length=100, do_sample=True) generator("Em uma bela manhã de") ```