--- license: apache-2.0 datasets: - abacusai/MetaMathFewshot --- DPO finetune of our [MetaMath SFT Model](https://huggingface.co/abacusai/MM-Orc-Vic-bagel-34b-c1000) on the [Truthy DPO dataset](https://huggingface.co/datasets/jondurbin/truthy-dpo-v0.1) ### Evaluation Results | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | | --- | --- | --- | --- | --- | --- | --- | | 75.54 | 69.20 | 84.34 | 76.46 | 67.58 | 82.87 | 72.78 |