--- base_model: HuggingFaceTB/SmolLM2-135M library_name: transformers model_name: smollm-smtalk-v1 tags: - generated_from_trainer - trl - sft licence: license license: mit datasets: - HuggingFaceTB/smol-smoltalk pipeline_tag: text-generation --- # Model Card for smollm-smtalk-v1 This model is a fine-tuned version of [HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="sfarrukh/smollm-smtalk-v1", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [Visualize in Weights & Biases](https://wandb.ai/farrukhmehmood-nts-the-islamia-university-of-bahawalpur/huggingface/runs/dbkkuti9) This model was trained with SFT. ### Framework versions - TRL: 0.13.0 - Transformers: 4.48.1 - Pytorch: 2.5.1+cu121 - Datasets: 3.2.0 - Tokenizers: 0.21.0