|
--- |
|
license: "apache-2.0" |
|
--- |
|
|
|
*This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.* |
|
|
|
**Benchmarks** |
|
|
|
Average 59.62 |
|
|
|
ARC 59.39 |
|
|
|
HellaSwag 82.48 |
|
|
|
MMLU 62.61 |
|
|
|
TruthfulQA 40.38 |
|
|
|
Winogrande 78.37 |
|
|
|
GSM8K 34.5 |
|
|
|
**Training Details** |
|
|
|
Duration: ~9 hours on one Kaggle T4 with Unsloth |
|
|
|
Model: https://huggingface.co/unsloth/mistral-7b-v0.2-bnb-4bit |
|
|
|
Dataset: https://huggingface.co/datasets/argilla/dpo-mix-7k |
|
|
|
Rank: 8 |
|
|
|
Alpha: 16 |
|
|
|
Learning rate: 5e-5 |
|
|
|
Beta: 0.1 |
|
|
|
Batch size: 8 |
|
|
|
Epochs: 1 |
|
|
|
Learning rate scheduler: Linear |
|
|
|
Prompt Format: ChatML |
|
``` |
|
<|im_start|>system |
|
You are a helpful assistant.<|im_end|> |
|
<|im_start|>user |
|
Why is the sky blue?<|im_end|> |
|
<|im_start|>assistant |
|
``` |
|
|
|
|
|
**WanDB Reports** |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/VaskOeT7IrbLpDzwC8jI-.png) |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/O_Uf0NrJC3qIMv9J0cEx9.png) |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/HipEtK-oxyT--IM5nHYQn.png) |
|
|
|
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) |