llama-7b_oasst1_l0.0002_32-16-16

This model is a fine-tuned version of huggyllama/llama-7b on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4964

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0002
  • train_batch_size: 1
  • eval_batch_size: 1
  • seed: 0
  • gradient_accumulation_steps: 16
  • total_train_batch_size: 16
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: constant
  • lr_scheduler_warmup_ratio: 0.03
  • training_steps: 1875

Training results

Training Loss Epoch Step Validation Loss
1.5015 0.0018 1 1.6846
1.5051 0.3392 187 1.3161
1.1342 0.6783 374 1.3054
1.355 1.0175 561 1.3059
1.1253 1.3566 748 1.3328
1.1364 1.6958 935 1.3135
0.7877 2.0349 1122 1.3669
0.9603 2.3741 1309 1.4090
0.6496 2.7132 1496 1.4736
0.5245 3.0524 1683 1.5293
0.5728 3.3915 1870 1.4740

Framework versions

  • PEFT 0.12.1.dev0
  • Transformers 4.45.0.dev0
  • Pytorch 2.3.0+cu121
  • Datasets 2.19.0
  • Tokenizers 0.19.1
Downloads last month
0
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for alexander-hm/llama-7b_oasst1_l0.0002_32-16-16

Adapter
(122)
this model