fpadovani's picture
Model save
dd8c5a6 verified
|
raw
history blame
2.01 kB
metadata
library_name: transformers
license: mit
base_model: gpt2
tags:
  - generated_from_trainer
model-index:
  - name: random_sentence_french
    results: []

random_sentence_french

This model is a fine-tuned version of gpt2 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.1513

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: reduce_lr_on_plateau
  • lr_scheduler_warmup_steps: 500
  • num_epochs: 1

Training results

Training Loss Epoch Step Validation Loss
0.7697 0.0036 2000 0.1915
0.1884 0.0072 4000 0.1777
0.178 0.0109 6000 0.1698
0.1714 0.0145 8000 0.1655
0.1669 0.0181 10000 0.1620
0.1647 0.0217 12000 0.1595
0.1608 0.0253 14000 0.1578
0.1601 0.0289 16000 0.1565
0.1607 0.0326 18000 0.1552
0.1565 0.0362 20000 0.1544
0.1555 0.0398 22000 0.1533
0.1549 0.0434 24000 0.1526
0.1544 0.0470 26000 0.1519
0.1529 0.0506 28000 0.1513

Framework versions

  • Transformers 4.45.2
  • Pytorch 2.4.1
  • Datasets 3.0.1
  • Tokenizers 0.20.1