Commit
·
171ab04
1
Parent(s):
bbde18e
flan_vary_merged_filter1
Browse files- README.md +38 -38
- config.json +1 -1
- generation_config.json +1 -1
- pytorch_model.bin +1 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -17,12 +17,12 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
-
- Loss: 0.
|
21 |
-
- Rouge1:
|
22 |
-
- Rouge2:
|
23 |
-
- Rougel:
|
24 |
-
- Rougelsum:
|
25 |
-
- Gen Len:
|
26 |
|
27 |
## Model description
|
28 |
|
@@ -54,39 +54,39 @@ The following hyperparameters were used during training:
|
|
54 |
|
55 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|
56 |
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
|
57 |
-
| 11.
|
58 |
-
| 0.
|
59 |
-
| 0.
|
60 |
-
| 0.
|
61 |
-
| 0.
|
62 |
-
| 0.
|
63 |
-
| 0.
|
64 |
-
| 0.
|
65 |
-
| 0.
|
66 |
-
| 0.
|
67 |
-
| 0.
|
68 |
-
| 0.
|
69 |
-
| 0.
|
70 |
-
| 0.
|
71 |
-
| 0.
|
72 |
-
| 0.
|
73 |
-
| 0.
|
74 |
-
| 0.
|
75 |
-
| 0.
|
76 |
-
| 0.
|
77 |
-
| 0.
|
78 |
-
| 0.
|
79 |
-
| 0.
|
80 |
-
| 0.
|
81 |
-
| 0.
|
82 |
-
| 0.
|
83 |
-
| 0.
|
84 |
-
| 0.
|
85 |
|
86 |
|
87 |
### Framework versions
|
88 |
|
89 |
-
- Transformers 4.
|
90 |
-
- Pytorch 2.0.1
|
91 |
-
- Datasets 2.14.
|
92 |
-
- Tokenizers 0.
|
|
|
17 |
|
18 |
This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 0.1597
|
21 |
+
- Rouge1: 66.8856
|
22 |
+
- Rouge2: 55.6869
|
23 |
+
- Rougel: 63.8241
|
24 |
+
- Rougelsum: 66.7005
|
25 |
+
- Gen Len: 16.3392
|
26 |
|
27 |
## Model description
|
28 |
|
|
|
54 |
|
55 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|
56 |
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
|
57 |
+
| 11.8095 | 0.35 | 200 | 0.5275 | 38.2792 | 29.3331 | 37.9276 | 38.1283 | 8.0624 |
|
58 |
+
| 0.4481 | 0.7 | 400 | 0.3046 | 64.4437 | 52.3632 | 62.0225 | 64.2515 | 16.4262 |
|
59 |
+
| 0.3616 | 1.05 | 600 | 0.2656 | 64.9871 | 53.1185 | 62.4919 | 64.739 | 16.4279 |
|
60 |
+
| 0.2944 | 1.41 | 800 | 0.2412 | 65.2117 | 53.5512 | 62.6779 | 64.9318 | 16.4464 |
|
61 |
+
| 0.264 | 1.76 | 1000 | 0.2295 | 65.5748 | 54.0948 | 62.9803 | 65.3339 | 16.3866 |
|
62 |
+
| 0.2571 | 2.11 | 1200 | 0.2223 | 65.7216 | 53.793 | 62.9877 | 65.491 | 16.1898 |
|
63 |
+
| 0.2364 | 2.46 | 1400 | 0.2164 | 65.5444 | 53.9296 | 62.9975 | 65.3055 | 16.3172 |
|
64 |
+
| 0.2293 | 2.81 | 1600 | 0.2029 | 65.7977 | 54.3067 | 63.1851 | 65.5544 | 16.1766 |
|
65 |
+
| 0.2129 | 3.16 | 1800 | 0.2006 | 65.8342 | 53.9105 | 63.163 | 65.6175 | 16.1757 |
|
66 |
+
| 0.2184 | 3.51 | 2000 | 0.1931 | 65.1608 | 53.7707 | 62.6719 | 64.9743 | 16.1547 |
|
67 |
+
| 0.1952 | 3.87 | 2200 | 0.1873 | 66.3361 | 54.8382 | 63.2054 | 66.0954 | 16.3155 |
|
68 |
+
| 0.1992 | 4.22 | 2400 | 0.1847 | 66.316 | 55.0379 | 63.5154 | 66.0694 | 16.3594 |
|
69 |
+
| 0.1873 | 4.57 | 2600 | 0.1811 | 66.4999 | 55.263 | 63.8319 | 66.2513 | 16.3146 |
|
70 |
+
| 0.1839 | 4.92 | 2800 | 0.1783 | 66.0055 | 54.3406 | 62.9554 | 65.7387 | 16.3304 |
|
71 |
+
| 0.1748 | 5.27 | 3000 | 0.1777 | 66.1592 | 54.8048 | 63.407 | 66.0067 | 16.3348 |
|
72 |
+
| 0.1844 | 5.62 | 3200 | 0.1736 | 66.7642 | 55.3404 | 63.7069 | 66.5324 | 16.2996 |
|
73 |
+
| 0.1745 | 5.98 | 3400 | 0.1698 | 66.3946 | 55.1716 | 63.5596 | 66.1663 | 16.3216 |
|
74 |
+
| 0.1739 | 6.33 | 3600 | 0.1678 | 66.4472 | 55.1785 | 63.602 | 66.2704 | 16.3049 |
|
75 |
+
| 0.1633 | 6.68 | 3800 | 0.1680 | 66.6666 | 55.4584 | 63.8058 | 66.4708 | 16.3445 |
|
76 |
+
| 0.1659 | 7.03 | 4000 | 0.1682 | 66.6592 | 55.3712 | 63.5841 | 66.4587 | 16.2953 |
|
77 |
+
| 0.1557 | 7.38 | 4200 | 0.1634 | 66.876 | 55.423 | 63.8431 | 66.5569 | 16.2434 |
|
78 |
+
| 0.158 | 7.73 | 4400 | 0.1622 | 66.6165 | 55.2948 | 63.5996 | 66.4314 | 16.3849 |
|
79 |
+
| 0.1647 | 8.08 | 4600 | 0.1622 | 66.7592 | 55.5552 | 63.7194 | 66.5229 | 16.2794 |
|
80 |
+
| 0.1579 | 8.44 | 4800 | 0.1614 | 66.7889 | 55.5768 | 63.8266 | 66.5511 | 16.3181 |
|
81 |
+
| 0.1526 | 8.79 | 5000 | 0.1610 | 66.7516 | 55.5383 | 63.6509 | 66.5754 | 16.261 |
|
82 |
+
| 0.1506 | 9.14 | 5200 | 0.1608 | 66.9266 | 55.6277 | 63.7712 | 66.6668 | 16.3445 |
|
83 |
+
| 0.1502 | 9.49 | 5400 | 0.1604 | 66.9759 | 55.6586 | 63.8856 | 66.7849 | 16.3251 |
|
84 |
+
| 0.158 | 9.84 | 5600 | 0.1597 | 66.8856 | 55.6869 | 63.8241 | 66.7005 | 16.3392 |
|
85 |
|
86 |
|
87 |
### Framework versions
|
88 |
|
89 |
+
- Transformers 4.34.0
|
90 |
+
- Pytorch 2.0.1+cu117
|
91 |
+
- Datasets 2.14.4
|
92 |
+
- Tokenizers 0.14.0
|
config.json
CHANGED
@@ -56,7 +56,7 @@
|
|
56 |
},
|
57 |
"tie_word_embeddings": false,
|
58 |
"torch_dtype": "float32",
|
59 |
-
"transformers_version": "4.
|
60 |
"use_cache": true,
|
61 |
"vocab_size": 32128
|
62 |
}
|
|
|
56 |
},
|
57 |
"tie_word_embeddings": false,
|
58 |
"torch_dtype": "float32",
|
59 |
+
"transformers_version": "4.34.0",
|
60 |
"use_cache": true,
|
61 |
"vocab_size": 32128
|
62 |
}
|
generation_config.json
CHANGED
@@ -2,5 +2,5 @@
|
|
2 |
"decoder_start_token_id": 0,
|
3 |
"eos_token_id": 1,
|
4 |
"pad_token_id": 0,
|
5 |
-
"transformers_version": "4.
|
6 |
}
|
|
|
2 |
"decoder_start_token_id": 0,
|
3 |
"eos_token_id": 1,
|
4 |
"pad_token_id": 0,
|
5 |
+
"transformers_version": "4.34.0"
|
6 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 990408885
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:934cc52079d888276ebb39430954c85584a58a0e79a6f2f523675b4db2292bcd
|
3 |
size 990408885
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4155
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6cebf456f4678a1e40d857db43c211a1d2114c0ebbeb554c11fcbbfafd25e9ef
|
3 |
size 4155
|