lesso04's picture
End of training
6630dab verified
metadata
library_name: peft
license: apache-2.0
base_model: unsloth/SmolLM2-1.7B
tags:
  - axolotl
  - generated_from_trainer
model-index:
  - name: c631b668-a7fe-4682-949b-1d81685337f2
    results: []

Built with Axolotl

c631b668-a7fe-4682-949b-1d81685337f2

This model is a fine-tuned version of unsloth/SmolLM2-1.7B on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2226

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.000204
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 8
  • optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 50
  • training_steps: 500

Training results

Training Loss Epoch Step Validation Loss
No log 0.0002 1 1.5368
1.3959 0.0116 50 1.3478
1.3112 0.0231 100 1.3007
1.2917 0.0347 150 1.2785
1.2684 0.0463 200 1.2622
1.257 0.0579 250 1.2484
1.2571 0.0694 300 1.2384
1.2338 0.0810 350 1.2304
1.2192 0.0926 400 1.2252
1.1918 0.1041 450 1.2230
1.2379 0.1157 500 1.2226

Framework versions

  • PEFT 0.13.2
  • Transformers 4.46.0
  • Pytorch 2.5.0+cu124
  • Datasets 3.0.1
  • Tokenizers 0.20.1