image/png

This is the fourth in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of Mistral-Nemo-Base-2407.

Prompting

Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:

"""<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""

Credits

This model has been a team effort, and the credits goes to all members of Anthracite.

Training

The training was done for 2 epochs. We used 8x NVIDIA H100 Tensor Core GPUs for the full-parameter fine-tuning of the model.

Built with Axolotl

Safety

...

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 18.68
IFEval (0-Shot) 37.62
BBH (3-Shot) 28.79
MATH Lvl 5 (4-Shot) 4.76
GPQA (0-shot) 5.48
MuSR (0-shot) 11.37
MMLU-PRO (5-shot) 24.08
Downloads last month
2,715
Safetensors
Model size
12.2B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for anthracite-org/magnum-v2-12b

Finetuned
(56)
this model
Finetunes
2 models
Merges
28 models
Quantizations
14 models

Spaces using anthracite-org/magnum-v2-12b 6

Collection including anthracite-org/magnum-v2-12b

Evaluation results