DeepKarkhanis's picture
Update README.md
eee432c verified
|
raw
history blame
575 Bytes
---
license: apache-2.0
datasets:
- abacusai/MetaMathFewshot
---
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64c14f6b02e1f8f67c73bd05/pf4d6FA7DriRtVq5HCkxd.png)
DPO finetune of our [MetaMath SFT Model](https://huggingface.co/abacusai/MM-Orc-Vic-bagel-34b-c1000) on the [Truthy DPO dataset](https://huggingface.co/datasets/jondurbin/truthy-dpo-v0.1)
### Evaluation Results
| Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
| --- | --- | --- | --- | --- | --- | --- |
| 75.54 | 69.20 | 84.34 | 76.46 | 67.58 | 82.87 | 72.78 |