|
--- |
|
license: other |
|
datasets: |
|
- euclaise/MiniCoT |
|
- euclaise/SciCoT |
|
- euclaise/symtune_mini |
|
- euclaise/gsm8k_self_correct |
|
- euclaise/mathoverflow-accepted |
|
- euirim/goodwiki |
|
--- |
|
|
|
A pre-finetuning finetuned version of Mistral 7B 0.1, focused on CoT reasoning tasks. |
|
|
|
Probably decent at reasoning, but also probably not great as a chat assistant- it's designed to be finetuned further to give it a friendlier style. As such, it is intentionally somewhat undertrained. |
|
|
|
Current benchmarks aren't great for instruct models, so I've temporarily omitted them. I'm working on a benchmark suite for instruct models though, and will update this with scores when that is released. |
|
|
|
Uses ChatML prompt formatting. |
|
|
|
I reserve no rights to the model. To the extent possible under law, I release it as public domain. However, the datasets used have various licenses that may impact how the model may be used in your jurisdiction. |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_euclaise__Ferret-7B) |
|
|
|
| Metric | Value | |
|
|-----------------------|---------------------------| |
|
| Avg. | 47.81 | |
|
| ARC (25-shot) | 62.2 | |
|
| HellaSwag (10-shot) | 81.75 | |
|
| MMLU (5-shot) | 60.82 | |
|
| TruthfulQA (0-shot) | 40.94 | |
|
| Winogrande (5-shot) | 77.35 | |
|
| GSM8K (5-shot) | 5.76 | |
|
| DROP (3-shot) | 5.87 | |
|
|
|
I'm not sure what's going on with GSM8K. Since GSK8K (train split) data was included in the Ferret dataset, I suspect that either it is over-correcting itself or the eval is broken. |
|
|