wdli's picture
Update README.md
7cbe8a6 verified
---
base_model: unsloth/llama-3-8b-Instruct-bnb-4bit
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
---
# Uploaded model
- **Developed by:** wdli
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-Instruct-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
The model is trained on reddit_depression_dataset, The epoch = 2.
The training is in dialog format, but the user's input is ignored.
For example
```python
def formatting_prompts_func(examples):
texts_dataset = examples['text']
formatted_prompts = []
for text in texts_dataset:
dialog = [
{"role": "system", "content": "You are a patient undergoing depression."},
# {"role": "user", "content": ""},
{"role": "assistant", "content": text}
]
formatted_prompt = tokenizer.apply_chat_template(dialog, tokenize=False, add_generation_prompt=False)
formatted_prompts.append(formatted_prompt)
return {"text": formatted_prompts}
```