tanvirsrbd1 commited on
Commit
171ab04
·
1 Parent(s): bbde18e

flan_vary_merged_filter1

Browse files
Files changed (5) hide show
  1. README.md +38 -38
  2. config.json +1 -1
  3. generation_config.json +1 -1
  4. pytorch_model.bin +1 -1
  5. training_args.bin +1 -1
README.md CHANGED
@@ -17,12 +17,12 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.1246
21
- - Rouge1: 73.6996
22
- - Rouge2: 66.3297
23
- - Rougel: 72.9503
24
- - Rougelsum: 73.5497
25
- - Gen Len: 15.9842
26
 
27
  ## Model description
28
 
@@ -54,39 +54,39 @@ The following hyperparameters were used during training:
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
57
- | 11.6985 | 0.35 | 200 | 0.4780 | 42.7521 | 33.9716 | 42.5027 | 42.7074 | 7.7939 |
58
- | 0.4363 | 0.7 | 400 | 0.2630 | 72.2102 | 63.1149 | 71.3335 | 72.1077 | 15.9351 |
59
- | 0.2885 | 1.05 | 600 | 0.2311 | 72.6405 | 63.9543 | 71.7757 | 72.4666 | 16.0579 |
60
- | 0.2576 | 1.4 | 800 | 0.2061 | 73.051 | 64.6174 | 72.2036 | 72.8541 | 15.8904 |
61
- | 0.233 | 1.75 | 1000 | 0.1957 | 72.8728 | 64.4924 | 71.9725 | 72.6908 | 15.9825 |
62
- | 0.2147 | 2.11 | 1200 | 0.1820 | 73.1831 | 65.1141 | 72.2787 | 72.9905 | 16.0246 |
63
- | 0.2123 | 2.46 | 1400 | 0.1775 | 72.951 | 64.9622 | 72.1818 | 72.7514 | 16.0439 |
64
- | 0.1801 | 2.81 | 1600 | 0.1662 | 73.2681 | 65.3002 | 72.4169 | 73.0978 | 15.9307 |
65
- | 0.1895 | 3.16 | 1800 | 0.1615 | 73.4215 | 65.4955 | 72.6108 | 73.2215 | 15.9868 |
66
- | 0.1631 | 3.51 | 2000 | 0.1597 | 73.4484 | 65.7314 | 72.5242 | 73.1706 | 16.043 |
67
- | 0.1728 | 3.86 | 2200 | 0.1513 | 73.7441 | 66.2993 | 72.957 | 73.5977 | 16.0123 |
68
- | 0.1585 | 4.21 | 2400 | 0.1506 | 73.564 | 65.8746 | 72.7345 | 73.4216 | 15.9439 |
69
- | 0.1501 | 4.56 | 2600 | 0.1462 | 73.4627 | 65.6971 | 72.6544 | 73.2861 | 15.9605 |
70
- | 0.1524 | 4.91 | 2800 | 0.1428 | 73.6264 | 66.0797 | 72.7434 | 73.4739 | 15.9596 |
71
- | 0.1452 | 5.26 | 3000 | 0.1385 | 73.4734 | 65.9915 | 72.6215 | 73.278 | 16.0675 |
72
- | 0.1387 | 5.61 | 3200 | 0.1375 | 73.7164 | 66.0897 | 72.8753 | 73.5286 | 15.95 |
73
- | 0.1429 | 5.96 | 3400 | 0.1351 | 73.7749 | 66.0932 | 72.9629 | 73.5989 | 15.9316 |
74
- | 0.138 | 6.32 | 3600 | 0.1342 | 73.7193 | 66.1623 | 72.8531 | 73.5521 | 15.9667 |
75
- | 0.1317 | 6.67 | 3800 | 0.1321 | 73.7627 | 66.212 | 72.9562 | 73.5713 | 15.9886 |
76
- | 0.1311 | 7.02 | 4000 | 0.1298 | 73.5849 | 66.0413 | 72.638 | 73.3629 | 16.0114 |
77
- | 0.1254 | 7.37 | 4200 | 0.1302 | 73.7472 | 66.224 | 72.9199 | 73.4865 | 16.0114 |
78
- | 0.1237 | 7.72 | 4400 | 0.1264 | 73.5486 | 65.8615 | 72.578 | 73.3444 | 15.993 |
79
- | 0.1273 | 8.07 | 4600 | 0.1261 | 73.9193 | 66.4538 | 73.103 | 73.7133 | 15.9491 |
80
- | 0.1201 | 8.42 | 4800 | 0.1269 | 73.4446 | 66.0467 | 72.6864 | 73.2561 | 16.0439 |
81
- | 0.1192 | 8.77 | 5000 | 0.1243 | 73.8268 | 66.4231 | 72.9562 | 73.6422 | 16.0026 |
82
- | 0.1217 | 9.12 | 5200 | 0.1253 | 73.9342 | 66.6018 | 73.0531 | 73.7114 | 16.0035 |
83
- | 0.1212 | 9.47 | 5400 | 0.1241 | 73.8482 | 66.6244 | 73.0402 | 73.6438 | 15.9789 |
84
- | 0.1159 | 9.82 | 5600 | 0.1246 | 73.6996 | 66.3297 | 72.9503 | 73.5497 | 15.9842 |
85
 
86
 
87
  ### Framework versions
88
 
89
- - Transformers 4.33.2
90
- - Pytorch 2.0.1
91
- - Datasets 2.14.5
92
- - Tokenizers 0.13.3
 
17
 
18
  This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.1597
21
+ - Rouge1: 66.8856
22
+ - Rouge2: 55.6869
23
+ - Rougel: 63.8241
24
+ - Rougelsum: 66.7005
25
+ - Gen Len: 16.3392
26
 
27
  ## Model description
28
 
 
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
57
+ | 11.8095 | 0.35 | 200 | 0.5275 | 38.2792 | 29.3331 | 37.9276 | 38.1283 | 8.0624 |
58
+ | 0.4481 | 0.7 | 400 | 0.3046 | 64.4437 | 52.3632 | 62.0225 | 64.2515 | 16.4262 |
59
+ | 0.3616 | 1.05 | 600 | 0.2656 | 64.9871 | 53.1185 | 62.4919 | 64.739 | 16.4279 |
60
+ | 0.2944 | 1.41 | 800 | 0.2412 | 65.2117 | 53.5512 | 62.6779 | 64.9318 | 16.4464 |
61
+ | 0.264 | 1.76 | 1000 | 0.2295 | 65.5748 | 54.0948 | 62.9803 | 65.3339 | 16.3866 |
62
+ | 0.2571 | 2.11 | 1200 | 0.2223 | 65.7216 | 53.793 | 62.9877 | 65.491 | 16.1898 |
63
+ | 0.2364 | 2.46 | 1400 | 0.2164 | 65.5444 | 53.9296 | 62.9975 | 65.3055 | 16.3172 |
64
+ | 0.2293 | 2.81 | 1600 | 0.2029 | 65.7977 | 54.3067 | 63.1851 | 65.5544 | 16.1766 |
65
+ | 0.2129 | 3.16 | 1800 | 0.2006 | 65.8342 | 53.9105 | 63.163 | 65.6175 | 16.1757 |
66
+ | 0.2184 | 3.51 | 2000 | 0.1931 | 65.1608 | 53.7707 | 62.6719 | 64.9743 | 16.1547 |
67
+ | 0.1952 | 3.87 | 2200 | 0.1873 | 66.3361 | 54.8382 | 63.2054 | 66.0954 | 16.3155 |
68
+ | 0.1992 | 4.22 | 2400 | 0.1847 | 66.316 | 55.0379 | 63.5154 | 66.0694 | 16.3594 |
69
+ | 0.1873 | 4.57 | 2600 | 0.1811 | 66.4999 | 55.263 | 63.8319 | 66.2513 | 16.3146 |
70
+ | 0.1839 | 4.92 | 2800 | 0.1783 | 66.0055 | 54.3406 | 62.9554 | 65.7387 | 16.3304 |
71
+ | 0.1748 | 5.27 | 3000 | 0.1777 | 66.1592 | 54.8048 | 63.407 | 66.0067 | 16.3348 |
72
+ | 0.1844 | 5.62 | 3200 | 0.1736 | 66.7642 | 55.3404 | 63.7069 | 66.5324 | 16.2996 |
73
+ | 0.1745 | 5.98 | 3400 | 0.1698 | 66.3946 | 55.1716 | 63.5596 | 66.1663 | 16.3216 |
74
+ | 0.1739 | 6.33 | 3600 | 0.1678 | 66.4472 | 55.1785 | 63.602 | 66.2704 | 16.3049 |
75
+ | 0.1633 | 6.68 | 3800 | 0.1680 | 66.6666 | 55.4584 | 63.8058 | 66.4708 | 16.3445 |
76
+ | 0.1659 | 7.03 | 4000 | 0.1682 | 66.6592 | 55.3712 | 63.5841 | 66.4587 | 16.2953 |
77
+ | 0.1557 | 7.38 | 4200 | 0.1634 | 66.876 | 55.423 | 63.8431 | 66.5569 | 16.2434 |
78
+ | 0.158 | 7.73 | 4400 | 0.1622 | 66.6165 | 55.2948 | 63.5996 | 66.4314 | 16.3849 |
79
+ | 0.1647 | 8.08 | 4600 | 0.1622 | 66.7592 | 55.5552 | 63.7194 | 66.5229 | 16.2794 |
80
+ | 0.1579 | 8.44 | 4800 | 0.1614 | 66.7889 | 55.5768 | 63.8266 | 66.5511 | 16.3181 |
81
+ | 0.1526 | 8.79 | 5000 | 0.1610 | 66.7516 | 55.5383 | 63.6509 | 66.5754 | 16.261 |
82
+ | 0.1506 | 9.14 | 5200 | 0.1608 | 66.9266 | 55.6277 | 63.7712 | 66.6668 | 16.3445 |
83
+ | 0.1502 | 9.49 | 5400 | 0.1604 | 66.9759 | 55.6586 | 63.8856 | 66.7849 | 16.3251 |
84
+ | 0.158 | 9.84 | 5600 | 0.1597 | 66.8856 | 55.6869 | 63.8241 | 66.7005 | 16.3392 |
85
 
86
 
87
  ### Framework versions
88
 
89
+ - Transformers 4.34.0
90
+ - Pytorch 2.0.1+cu117
91
+ - Datasets 2.14.4
92
+ - Tokenizers 0.14.0
config.json CHANGED
@@ -56,7 +56,7 @@
56
  },
57
  "tie_word_embeddings": false,
58
  "torch_dtype": "float32",
59
- "transformers_version": "4.33.2",
60
  "use_cache": true,
61
  "vocab_size": 32128
62
  }
 
56
  },
57
  "tie_word_embeddings": false,
58
  "torch_dtype": "float32",
59
+ "transformers_version": "4.34.0",
60
  "use_cache": true,
61
  "vocab_size": 32128
62
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
- "transformers_version": "4.33.2"
6
  }
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
+ "transformers_version": "4.34.0"
6
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0118864daddde40b1f1a748ba44428858440f4cab951547cd8cdc262643a0bf
3
  size 990408885
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:934cc52079d888276ebb39430954c85584a58a0e79a6f2f523675b4db2292bcd
3
  size 990408885
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee4e2a5bab7ae385f59877940ecf9955bdd2df87578368a448d446b296738de8
3
  size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cebf456f4678a1e40d857db43c211a1d2114c0ebbeb554c11fcbbfafd25e9ef
3
  size 4155