Edit model card

t5-base-finetuned-depression

This model is a fine-tuned version of google-t5/t5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2260
  • Rouge1: 89.7655
  • Rouge2: 24.4136
  • Rougel: 89.7655
  • Rougelsum: 89.7655
  • Gen Len: 2.2719
  • Precision: 0.8856
  • Recall: 0.8807
  • F1: 0.8817
  • Accuracy: 0.8977

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 20
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len Precision Recall F1 Accuracy
No log 1.0 469 0.3428 69.6162 9.7015 69.5096 69.6162 2.1087 0.8545 0.4409 0.4375 0.6962
0.7863 2.0 938 0.2674 79.5309 19.0832 79.5309 79.5309 2.2058 0.8192 0.5744 0.6052 0.7953
0.3128 3.0 1407 0.2317 84.0085 21.322 84.0085 84.0085 2.2239 0.9053 0.6654 0.721 0.8401
0.2367 4.0 1876 0.1736 86.887 22.3881 86.887 86.887 2.242 0.6608 0.586 0.6155 0.8689
0.1844 5.0 2345 0.1802 88.5928 22.7079 88.5928 88.5928 2.2388 0.9113 0.8252 0.8597 0.8859
0.135 6.0 2814 0.2000 88.4861 22.2814 88.4861 88.4861 2.2345 0.9045 0.8405 0.8655 0.8849
0.1247 7.0 3283 0.2048 89.5522 23.5608 89.4989 89.5522 2.2495 0.9108 0.8526 0.8769 0.8955
0.1071 8.0 3752 0.2361 89.1258 23.7207 89.1258 89.1258 2.2591 0.6783 0.6467 0.6603 0.8913
0.0832 9.0 4221 0.2486 89.8721 24.5203 89.8721 89.8721 2.2889 0.6695 0.6532 0.6603 0.8987
0.0652 10.0 4690 0.3051 89.339 23.1343 89.339 89.339 2.2473 0.9065 0.8642 0.8811 0.8934
0.0674 11.0 5159 0.3269 89.7655 23.9872 89.7655 89.7655 2.2623 0.8973 0.8711 0.8819 0.8977
0.0575 12.0 5628 0.3241 89.4456 23.8806 89.4456 89.4456 2.2633 0.8903 0.8652 0.8756 0.8945
0.0422 13.0 6097 0.3088 90.0853 24.5203 90.0853 90.0853 2.2729 0.6754 0.6595 0.6664 0.9009
0.0395 14.0 6566 0.2781 90.0853 25.3731 90.0853 90.0853 2.2889 0.6801 0.6575 0.6681 0.9009
0.0341 15.0 7035 0.2658 90.1919 24.5203 90.1919 90.1919 2.2719 0.9043 0.8836 0.8926 0.9019
0.0336 16.0 7504 0.2433 90.0853 24.8401 90.0853 90.0853 2.2772 0.9048 0.8769 0.8896 0.9009
0.0336 17.0 7973 0.2363 89.8721 24.6269 89.8721 89.8721 2.274 0.6717 0.6563 0.6631 0.8987
0.0274 18.0 8442 0.2297 90.4051 25.2132 90.4051 90.4051 2.2814 0.904 0.8882 0.8953 0.9041
0.0298 19.0 8911 0.2275 89.7655 24.4136 89.7655 89.7655 2.2719 0.8886 0.8807 0.8832 0.8977
0.0261 20.0 9380 0.2260 89.7655 24.4136 89.7655 89.7655 2.2719 0.8856 0.8807 0.8817 0.8977

Framework versions

  • Transformers 4.41.2
  • Pytorch 2.3.0+cu121
  • Datasets 2.20.0
  • Tokenizers 0.19.1
Downloads last month
12
Safetensors
Model size
223M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for welsachy/t5-base-finetuned-depression

Base model

google-t5/t5-base
Finetuned
(380)
this model