IMPORTANT

In case you got the following error:

exception: data did not match any variant of untagged enum modelwrapper at line 1251003 column 3

Please upgrade your transformer package, that is, use the following code:

pip install --upgrade "transformers>=4.45"

Uploaded model

  • Developed by: NotASI
  • License: apache-2.0
  • Finetuned from model : unsloth/Llama-3.2-3B-Instruct-bnb-4bit

Details

This model was trained on mlabonne/FineTome-100k for 2 epochs with rslora + qlora, and achieve the final training loss: 0.596400.

This model follows the same chat template as the base model one.

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 16.60
IFEval (0-Shot) 54.74
BBH (3-Shot) 19.52
MATH Lvl 5 (4-Shot) 5.29
GPQA (0-shot) 0.11
MuSR (0-shot) 3.96
MMLU-PRO (5-shot) 15.96
Downloads last month
39
Safetensors
Model size
3.21B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for NotASI/FineTome-Llama3.2-3B-1002

Finetuned
(146)
this model
Merges
1 model
Quantizations
2 models

Dataset used to train NotASI/FineTome-Llama3.2-3B-1002

Collection including NotASI/FineTome-Llama3.2-3B-1002

Evaluation results