outputs

This model is a fine-tuned version of mrm8488/t5-base-finetuned-common_gen on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 3.2833
  • Rouge1: 79.0721
  • Rouge2: 59.355
  • Rougel: 70.9787
  • Rougelsum: 70.9177
  • Gen Len: 16.3819

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 241 2.3934 75.6585 53.2221 65.7693 65.7796 16.4214
No log 2.0 482 2.4223 76.75 55.466 67.3377 67.3425 16.4381
0.5668 3.0 723 2.5147 77.5314 56.5654 68.4518 68.4263 16.4235
0.5668 4.0 964 2.5164 78.3267 57.4609 69.1759 69.1835 16.4381
0.4546 5.0 1205 2.6058 78.6699 58.2725 69.8265 69.8802 16.358
0.4546 6.0 1446 2.6295 78.8943 58.182 69.5933 69.619 16.4464
0.4014 7.0 1687 2.6592 79.0497 58.7135 70.2046 70.2047 16.4152
0.4014 8.0 1928 2.6610 79.1756 59.143 70.5631 70.5854 16.4235
0.3652 9.0 2169 2.6636 79.2156 59.3209 70.5418 70.5757 16.4308
0.3652 10.0 2410 2.6647 78.9922 59.0252 70.49 70.5262 16.3569
0.341 11.0 2651 2.7105 79.0372 58.7891 70.2666 70.3333 16.4131
0.341 12.0 2892 2.7829 79.2323 59.4941 70.6877 70.7283 16.3684
0.3199 13.0 3133 2.7344 79.0326 58.8869 70.2648 70.2983 16.4329
0.3199 14.0 3374 2.7568 79.172 59.1695 70.5229 70.5505 16.435
0.301 15.0 3615 2.7773 78.9152 58.8776 70.368 70.3922 16.3923
0.301 16.0 3856 2.8178 79.1939 59.1455 70.6938 70.7153 16.4693
0.2883 17.0 4097 2.8105 79.1307 59.114 70.6573 70.67 16.4287
0.2883 18.0 4338 2.8360 78.8425 58.6959 70.4532 70.3955 16.358
0.2779 19.0 4579 2.8606 79.3713 59.3462 70.6531 70.648 16.3632
0.2779 20.0 4820 2.8513 79.0083 58.817 70.5779 70.5775 16.435
0.2664 21.0 5061 2.8882 79.1277 59.1771 70.8215 70.7963 16.4194
0.2664 22.0 5302 2.8852 79.0191 59.0226 70.6604 70.6499 16.3715
0.258 23.0 5543 2.9253 79.3219 59.1689 70.8438 70.8462 16.41
0.258 24.0 5784 2.9020 79.1941 59.044 70.6849 70.6449 16.3996
0.2494 25.0 6025 2.9579 79.5266 59.8762 71.5284 71.5085 16.4422
0.2494 26.0 6266 2.9660 79.1653 59.3299 70.7888 70.7755 16.4475
0.2415 27.0 6507 2.9682 79.3267 59.7158 71.1062 71.0864 16.411
0.2415 28.0 6748 3.0086 79.0646 59.1926 70.8172 70.78 16.4422
0.2415 29.0 6989 3.0190 78.9127 58.8431 70.4618 70.4597 16.4308
0.2361 30.0 7230 2.9958 79.1999 59.5494 70.9411 70.9 16.3684
0.2361 31.0 7471 2.9824 79.0219 59.0778 70.6465 70.6445 16.3809
0.2295 32.0 7712 3.0079 79.1633 59.4146 71.0352 70.9144 16.4017
0.2295 33.0 7953 2.9894 79.106 59.0905 70.644 70.622 16.4069
0.2247 34.0 8194 3.0256 78.8789 59.0956 70.8588 70.796 16.3954
0.2247 35.0 8435 3.0451 78.7977 58.889 70.6925 70.6408 16.3975
0.2196 36.0 8676 3.0475 78.8942 58.8944 70.4803 70.4858 16.3871
0.2196 37.0 8917 3.0289 78.7668 58.7055 70.3606 70.3058 16.411
0.2157 38.0 9158 3.0748 78.833 58.6404 70.271 70.2338 16.4266
0.2157 39.0 9399 3.0665 79.2216 59.3481 70.9647 70.9076 16.4058
0.2104 40.0 9640 3.0773 78.9717 58.8784 70.6392 70.6359 16.3704
0.2104 41.0 9881 3.0823 78.901 58.9511 70.737 70.7051 16.3788
0.2061 42.0 10122 3.0637 78.8689 58.8098 70.754 70.7783 16.36
0.2061 43.0 10363 3.0965 79.4409 59.5757 71.2304 71.2553 16.4422
0.2027 44.0 10604 3.1178 78.8964 58.9388 70.5416 70.5367 16.4048
0.2027 45.0 10845 3.1078 79.2306 59.5613 71.1202 71.0741 16.4089
0.1998 46.0 11086 3.1432 79.2393 59.5587 71.1138 71.0502 16.4485
0.1998 47.0 11327 3.1493 79.3553 59.5023 71.0459 71.0508 16.4256
0.1966 48.0 11568 3.1670 79.0395 59.1524 70.9374 70.921 16.4006
0.1966 49.0 11809 3.1580 79.2359 59.2038 71.0061 70.9725 16.4568
0.1941 50.0 12050 3.1458 79.1914 59.4751 71.2448 71.2131 16.437
0.1941 51.0 12291 3.1379 79.2891 59.6776 71.3416 71.2787 16.4287
0.1903 52.0 12532 3.1618 79.4208 59.7335 71.4722 71.4196 16.4173
0.1903 53.0 12773 3.1712 79.2976 59.736 71.1151 71.0491 16.4454
0.188 54.0 13014 3.1795 79.2538 59.2768 71.0137 70.9602 16.4183
0.188 55.0 13255 3.1836 79.4504 59.3201 71.0908 71.0614 16.4277
0.188 56.0 13496 3.1800 79.067 59.1309 70.7069 70.6896 16.3996
0.1858 57.0 13737 3.1660 79.2369 59.5092 71.0077 70.9537 16.3871
0.1858 58.0 13978 3.1886 79.0404 58.8652 70.6484 70.6162 16.3798
0.1849 59.0 14219 3.1413 79.1468 59.1964 70.8362 70.8076 16.4121
0.1849 60.0 14460 3.1778 79.0189 58.9894 70.597 70.5447 16.3788
0.1824 61.0 14701 3.1674 78.8303 58.7354 70.4517 70.4009 16.3871
0.1824 62.0 14942 3.1756 78.8914 58.7551 70.3439 70.3031 16.3965
0.1793 63.0 15183 3.1943 78.9093 58.8599 70.415 70.3998 16.4152
0.1793 64.0 15424 3.1970 78.8353 58.8096 70.4231 70.3813 16.4194
0.1768 65.0 15665 3.2301 79.0958 59.2143 70.837 70.7973 16.4162
0.1768 66.0 15906 3.2415 79.1112 59.1872 70.7534 70.7145 16.4173
0.177 67.0 16147 3.2332 79.0807 59.3005 70.6878 70.6351 16.3757
0.177 68.0 16388 3.2415 79.0587 59.1872 70.6691 70.6295 16.3819
0.1744 69.0 16629 3.2454 79.3829 59.5478 70.997 70.9835 16.4037
0.1744 70.0 16870 3.2239 79.1289 59.1661 70.7455 70.684 16.4412
0.1715 71.0 17111 3.2056 79.2074 59.1811 70.7366 70.682 16.4339
0.1715 72.0 17352 3.2101 79.1756 59.1883 70.8208 70.821 16.4079
0.1722 73.0 17593 3.2086 79.4652 59.7631 71.3223 71.2917 16.4214
0.1722 74.0 17834 3.2269 79.0154 59.0411 70.6754 70.6184 16.3621
0.17 75.0 18075 3.2389 79.1212 59.3143 70.7629 70.6935 16.3444
0.17 76.0 18316 3.2387 79.1148 59.4001 70.897 70.8525 16.3663
0.1682 77.0 18557 3.2505 79.0955 59.2033 70.8455 70.8303 16.3632
0.1682 78.0 18798 3.2670 79.1781 59.1849 70.8704 70.8286 16.3611
0.1686 79.0 19039 3.2825 79.0246 59.1434 70.7866 70.7353 16.3559
0.1686 80.0 19280 3.2816 79.0431 58.9944 70.6819 70.6257 16.3486
0.1664 81.0 19521 3.2567 79.0287 59.0805 70.7245 70.6889 16.3392
0.1664 82.0 19762 3.2750 78.9783 59.1123 70.6961 70.6369 16.3465
0.1659 83.0 20003 3.2757 78.9774 59.2248 70.7978 70.7407 16.3455
0.1659 84.0 20244 3.2800 79.0205 59.2371 70.8016 70.7532 16.3704
0.1659 85.0 20485 3.2723 78.9507 59.1004 70.6419 70.5825 16.3663
0.165 86.0 20726 3.2723 78.9698 59.1362 70.7738 70.7018 16.3652
0.165 87.0 20967 3.2740 79.0156 59.2383 70.7799 70.7075 16.384
0.1637 88.0 21208 3.2628 79.0806 59.3468 70.8158 70.753 16.3725
0.1637 89.0 21449 3.2605 78.9712 59.223 70.7653 70.7021 16.3621
0.1642 90.0 21690 3.2738 79.0793 59.3309 70.8753 70.8282 16.385
0.1642 91.0 21931 3.2674 78.9877 59.1468 70.7031 70.6598 16.3694
0.1636 92.0 22172 3.2706 79.006 59.1366 70.7069 70.6638 16.3757
0.1636 93.0 22413 3.2783 79.0874 59.2523 70.7881 70.7397 16.3746
0.1629 94.0 22654 3.2801 79.1126 59.2826 70.8352 70.7865 16.3715
0.1629 95.0 22895 3.2797 79.0586 59.3089 70.8126 70.7697 16.3757
0.1625 96.0 23136 3.2826 79.1019 59.3294 70.8399 70.803 16.3777
0.1625 97.0 23377 3.2844 79.0739 59.3167 70.9062 70.846 16.3767
0.1622 98.0 23618 3.2826 79.0739 59.3167 70.9062 70.846 16.3767
0.1622 99.0 23859 3.2830 79.0721 59.355 70.9787 70.9177 16.3819
0.161 100.0 24100 3.2833 79.0721 59.355 70.9787 70.9177 16.3819

Framework versions

  • Transformers 4.20.1
  • Pytorch 1.11.0
  • Datasets 2.1.0
  • Tokenizers 0.12.1
Downloads last month
11
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.