Llama3.1-Allades-8B
Allades finetunes abliterated Llama 3.1 with 5 datasets to improve creative writing, reasoning, and roleplay.
Datasets
- jondurbin/gutenberg-dpo-v0.1
- nbeerbower/gutenberg2-dpo
- jondurbin/truthy-dpo-v0.1
- kyujinpy/orca_math_dpo
- antiven0m/physical-reasoning-dpo
Training
ORPO tuned for 1 epoch with 2x RTX 3090 (sponsored by Schneewolf Labs).
Data was prepared with Llama 3.1 Instruct.
- Downloads last month
- 36
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for nbeerbower/Llama3.1-Allades-8B
Base model
meta-llama/Llama-3.1-8B
Finetuned
meta-llama/Llama-3.1-8B-Instruct