Edit model card

my_fine_tuning_summary_t5_large_model_hf

This model is a fine-tuned version of google-t5/t5-large on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: nan
  • Rouge1: 0.1329
  • Rouge2: 0.0498
  • Rougel: 0.1105
  • Rougelsum: 0.1106
  • Gen Len: 19.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 1
  • eval_batch_size: 1
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 20
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
0.0 1.0 989 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 2.0 1978 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 3.0 2967 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 4.0 3956 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 5.0 4945 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 6.0 5934 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 7.0 6923 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 8.0 7912 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 9.0 8901 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 10.0 9890 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 11.0 10879 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 12.0 11868 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 13.0 12857 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 14.0 13846 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 15.0 14835 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 16.0 15824 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 17.0 16813 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 18.0 17802 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 19.0 18791 nan 0.1329 0.0498 0.1105 0.1106 19.0
0.0 20.0 19780 nan 0.1329 0.0498 0.1105 0.1106 19.0

Framework versions

  • Transformers 4.37.2
  • Pytorch 2.0.1+cu117
  • Datasets 2.14.5
  • Tokenizers 0.15.2
Downloads last month
65
Safetensors
Model size
738M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Kudod/my_fine_tuning_summary_t5_large_model_hf

Base model

google-t5/t5-large
Finetuned
(68)
this model