IMPORTANT

In case you got the following error:

exception: data did not match any variant of untagged enum modelwrapper at line 1251003 column 3

Please upgrade your transformer package, that is, use the following code:

pip install --upgrade "transformers>=4.45"

Uploaded model

  • Developed by: NotASI
  • License: apache-2.0
  • Finetuned from model : unsloth/Llama-3.2-3B-Instruct-bnb-4bit

Details

This model was trained on mlabonne/FineTome-100k for 2 epochs with rslora + qlora, and achieve the final training loss: 0.596400.

This model follows the same chat template as the base model one.

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 16.60
IFEval (0-Shot) 54.74
BBH (3-Shot) 19.52
MATH Lvl 5 (4-Shot) 5.29
GPQA (0-shot) 0.11
MuSR (0-shot) 3.96
MMLU-PRO (5-shot) 15.96
Downloads last month
11
Safetensors
Model size
3.21B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for NotASI/FineTome-Llama3.2-3B-1002

Finetuned
(198)
this model
Merges
1 model
Quantizations
2 models

Dataset used to train NotASI/FineTome-Llama3.2-3B-1002

Collection including NotASI/FineTome-Llama3.2-3B-1002

Evaluation results