--- license: apache-2.0 base_model: google/flan-t5-base tags: - generated_from_trainer metrics: - rouge model-index: - name: only_vary_model1 results: [] --- # only_vary_model1 This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2400 - Rouge1: 62.207 - Rouge2: 48.2441 - Rougel: 57.479 - Rougelsum: 62.039 - Gen Len: 16.3689 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 200 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 11.6904 | 0.7 | 200 | 0.5014 | 40.0947 | 28.9027 | 39.0539 | 40.0066 | 6.986 | | 0.4064 | 1.4 | 400 | 0.2939 | 58.7275 | 43.2853 | 55.0201 | 58.5308 | 16.264 | | 0.3125 | 2.1 | 600 | 0.2749 | 60.4914 | 45.7417 | 56.4753 | 60.3543 | 16.1503 | | 0.2836 | 2.8 | 800 | 0.2576 | 60.3004 | 45.6356 | 55.9723 | 60.1254 | 16.2692 | | 0.2634 | 3.5 | 1000 | 0.2519 | 61.3543 | 47.3391 | 57.2049 | 61.1618 | 16.3724 | | 0.2608 | 4.2 | 1200 | 0.2522 | 61.538 | 46.8595 | 56.8033 | 61.3561 | 16.3409 | | 0.2387 | 4.9 | 1400 | 0.2439 | 60.7253 | 46.2206 | 56.342 | 60.57 | 16.3357 | | 0.2253 | 5.59 | 1600 | 0.2479 | 60.9499 | 46.7452 | 56.8289 | 60.8009 | 16.4178 | | 0.2299 | 6.29 | 1800 | 0.2416 | 61.6976 | 47.212 | 57.1834 | 61.558 | 16.3077 | | 0.2174 | 6.99 | 2000 | 0.2418 | 61.829 | 47.2977 | 57.3766 | 61.718 | 16.3112 | | 0.2089 | 7.69 | 2200 | 0.2389 | 62.2459 | 47.6275 | 57.1022 | 62.1048 | 16.299 | | 0.2127 | 8.39 | 2400 | 0.2402 | 61.8535 | 47.7161 | 57.2504 | 61.7496 | 16.4196 | | 0.2097 | 9.09 | 2600 | 0.2402 | 62.242 | 47.6438 | 57.0938 | 62.0981 | 16.299 | | 0.2036 | 9.79 | 2800 | 0.2400 | 62.207 | 48.2441 | 57.479 | 62.039 | 16.3689 | ### Framework versions - Transformers 4.34.0 - Pytorch 2.0.1+cu117 - Datasets 2.14.4 - Tokenizers 0.14.0