--- language: en inference: false tags: - text-generation - opt license: other commercial: false --- ## Intro This is a OPT-125m model trained with HF dataset on a single 3090 GPU. ### How to use You can use this model directly with a pipeline for text generation. ```python >>> from transformers import pipeline >>> generator = pipeline('text-generation', model="facebook/opt-125m") >>> generator("Hello, I'm am conscious and") [{'generated_text': 'Hello, I am conscious and aware of the fact that I am a woman. I am aware of'}] ``` By default, generation is deterministic. In order to use the top-k sampling, please set `do_sample` to `True`. ```python >>> from transformers import pipeline, set_seed >>> set_seed(32) >>> generator = pipeline('text-generation', model="facebook/opt-125m", do_sample=True) >>> generator("Hello, I'm am conscious and") [{'generated_text': 'Hello, I am conscious and active member of the Khaosan Group, a private, self'}] ``` ## Training data This model uses AHRLHF for RL https://huggingface.co/datasets/Anthropic/hh-rlhf