license: apache-2.0 | |
datasets: | |
- abacusai/MetaMathFewshot | |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64c14f6b02e1f8f67c73bd05/pf4d6FA7DriRtVq5HCkxd.png) | |
DPO finetune of our [MetaMath SFT Model](https://huggingface.co/abacusai/MM-Orc-Vic-bagel-34b-c1000) on the [Truthy DPO dataset](https://huggingface.co/datasets/jondurbin/truthy-dpo-v0.1) | |
### Evaluation Results | |
| Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | | |
| --- | --- | --- | --- | --- | --- | --- | | |
| 75.54 | 69.20 | 84.34 | 76.46 | 67.58 | 82.87 | 72.78 | |