image/png

This is the fourth in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of Mistral-Nemo-Base-2407.

Prompting

Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:

"""<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""

Credits

This model has been a team effort, and the credits goes to all members of Anthracite.

Training

The training was done for 2 epochs. We used 8x NVIDIA H100 Tensor Core GPUs for the full-parameter fine-tuning of the model.

Built with Axolotl

Safety

...

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 18.68
IFEval (0-Shot) 37.62
BBH (3-Shot) 28.79
MATH Lvl 5 (4-Shot) 4.76
GPQA (0-shot) 5.48
MuSR (0-shot) 11.37
MMLU-PRO (5-shot) 24.08
Downloads last month
881
Safetensors
Model size
12.2B params
Tensor type
BF16
Β·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for anthracite-org/magnum-v2-12b

Finetuned
(45)
this model
Finetunes
2 models
Merges
20 models
Quantizations
17 models

Spaces using anthracite-org/magnum-v2-12b 6

Collection including anthracite-org/magnum-v2-12b

Evaluation results