abdiharyadi
commited on
Training in progress, epoch 7
Browse files- last-checkpoint/model.safetensors +1 -1
- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/special_tokens_map.json +6 -42
- last-checkpoint/trainer_state.json +5 -1173
- last-checkpoint/training_args.bin +2 -2
- model.safetensors +1 -1
- run-2024-10-28T03:01:14+00:00.log +0 -0
- val_outputs/val_generated_predictions_36735830.txt +0 -0
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:044e91a7c52f50e4043dc236fc046775db89767dd283738cb05dae269d0fe1f9
|
3 |
size 1575259780
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3150397656
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06fd3b48586725ea6ba928db3bb6432129af3d8438eb526dfaaa3ca8c57e58e5
|
3 |
size 3150397656
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8337dd673478657a1e3e59ab5c0126da6f87ecc51591bad61e39324efa7f5883
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10535ed970c1f8b1967fdb1bcf70b29e64c063da0c7c6d212af5b4ef07621922
|
3 |
size 1064
|
last-checkpoint/special_tokens_map.json
CHANGED
@@ -53,27 +53,9 @@
|
|
53 |
"gl_ES",
|
54 |
"sl_SI"
|
55 |
],
|
56 |
-
"bos_token":
|
57 |
-
|
58 |
-
|
59 |
-
"normalized": false,
|
60 |
-
"rstrip": false,
|
61 |
-
"single_word": false
|
62 |
-
},
|
63 |
-
"cls_token": {
|
64 |
-
"content": "<s>",
|
65 |
-
"lstrip": false,
|
66 |
-
"normalized": false,
|
67 |
-
"rstrip": false,
|
68 |
-
"single_word": false
|
69 |
-
},
|
70 |
-
"eos_token": {
|
71 |
-
"content": "</s>",
|
72 |
-
"lstrip": false,
|
73 |
-
"normalized": false,
|
74 |
-
"rstrip": false,
|
75 |
-
"single_word": false
|
76 |
-
},
|
77 |
"mask_token": {
|
78 |
"content": "<mask>",
|
79 |
"lstrip": true,
|
@@ -81,25 +63,7 @@
|
|
81 |
"rstrip": false,
|
82 |
"single_word": false
|
83 |
},
|
84 |
-
"pad_token":
|
85 |
-
|
86 |
-
|
87 |
-
"normalized": false,
|
88 |
-
"rstrip": false,
|
89 |
-
"single_word": false
|
90 |
-
},
|
91 |
-
"sep_token": {
|
92 |
-
"content": "</s>",
|
93 |
-
"lstrip": false,
|
94 |
-
"normalized": false,
|
95 |
-
"rstrip": false,
|
96 |
-
"single_word": false
|
97 |
-
},
|
98 |
-
"unk_token": {
|
99 |
-
"content": "<unk>",
|
100 |
-
"lstrip": false,
|
101 |
-
"normalized": false,
|
102 |
-
"rstrip": false,
|
103 |
-
"single_word": false
|
104 |
-
}
|
105 |
}
|
|
|
53 |
"gl_ES",
|
54 |
"sl_SI"
|
55 |
],
|
56 |
+
"bos_token": "<s>",
|
57 |
+
"cls_token": "<s>",
|
58 |
+
"eos_token": "</s>",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
59 |
"mask_token": {
|
60 |
"content": "<mask>",
|
61 |
"lstrip": true,
|
|
|
63 |
"rstrip": false,
|
64 |
"single_word": false
|
65 |
},
|
66 |
+
"pad_token": "<pad>",
|
67 |
+
"sep_token": "</s>",
|
68 |
+
"unk_token": "<unk>"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
69 |
}
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"best_metric":
|
3 |
-
"best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -5865,1174 +5865,6 @@
|
|
5865 |
"eval_samples_per_second": 1.019,
|
5866 |
"eval_steps_per_second": 0.51,
|
5867 |
"step": 19347
|
5868 |
-
},
|
5869 |
-
{
|
5870 |
-
"epoch": 5.0038764698281435,
|
5871 |
-
"learning_rate": 1.6894852845844678e-07,
|
5872 |
-
"loss": 2.6522,
|
5873 |
-
"step": 19360
|
5874 |
-
},
|
5875 |
-
{
|
5876 |
-
"epoch": 5.009045096265667,
|
5877 |
-
"learning_rate": 1.6891611564890445e-07,
|
5878 |
-
"loss": 2.6008,
|
5879 |
-
"step": 19380
|
5880 |
-
},
|
5881 |
-
{
|
5882 |
-
"epoch": 5.014213722703191,
|
5883 |
-
"learning_rate": 1.688837028393621e-07,
|
5884 |
-
"loss": 2.6294,
|
5885 |
-
"step": 19400
|
5886 |
-
},
|
5887 |
-
{
|
5888 |
-
"epoch": 5.019382349140716,
|
5889 |
-
"learning_rate": 1.6885129002981977e-07,
|
5890 |
-
"loss": 2.6519,
|
5891 |
-
"step": 19420
|
5892 |
-
},
|
5893 |
-
{
|
5894 |
-
"epoch": 5.02455097557824,
|
5895 |
-
"learning_rate": 1.6881887722027743e-07,
|
5896 |
-
"loss": 2.6537,
|
5897 |
-
"step": 19440
|
5898 |
-
},
|
5899 |
-
{
|
5900 |
-
"epoch": 5.029719602015764,
|
5901 |
-
"learning_rate": 1.6878646441073512e-07,
|
5902 |
-
"loss": 2.6256,
|
5903 |
-
"step": 19460
|
5904 |
-
},
|
5905 |
-
{
|
5906 |
-
"epoch": 5.034888228453289,
|
5907 |
-
"learning_rate": 1.6875405160119278e-07,
|
5908 |
-
"loss": 2.6721,
|
5909 |
-
"step": 19480
|
5910 |
-
},
|
5911 |
-
{
|
5912 |
-
"epoch": 5.040056854890813,
|
5913 |
-
"learning_rate": 1.6872163879165044e-07,
|
5914 |
-
"loss": 2.6894,
|
5915 |
-
"step": 19500
|
5916 |
-
},
|
5917 |
-
{
|
5918 |
-
"epoch": 5.045225481328337,
|
5919 |
-
"learning_rate": 1.6868922598210813e-07,
|
5920 |
-
"loss": 2.6702,
|
5921 |
-
"step": 19520
|
5922 |
-
},
|
5923 |
-
{
|
5924 |
-
"epoch": 5.050394107765861,
|
5925 |
-
"learning_rate": 1.686568131725658e-07,
|
5926 |
-
"loss": 2.7041,
|
5927 |
-
"step": 19540
|
5928 |
-
},
|
5929 |
-
{
|
5930 |
-
"epoch": 5.055562734203385,
|
5931 |
-
"learning_rate": 1.6862440036302346e-07,
|
5932 |
-
"loss": 2.7243,
|
5933 |
-
"step": 19560
|
5934 |
-
},
|
5935 |
-
{
|
5936 |
-
"epoch": 5.06073136064091,
|
5937 |
-
"learning_rate": 1.6859198755348114e-07,
|
5938 |
-
"loss": 2.7082,
|
5939 |
-
"step": 19580
|
5940 |
-
},
|
5941 |
-
{
|
5942 |
-
"epoch": 5.065899987078434,
|
5943 |
-
"learning_rate": 1.685595747439388e-07,
|
5944 |
-
"loss": 2.6755,
|
5945 |
-
"step": 19600
|
5946 |
-
},
|
5947 |
-
{
|
5948 |
-
"epoch": 5.071068613515958,
|
5949 |
-
"learning_rate": 1.6852716193439647e-07,
|
5950 |
-
"loss": 2.6075,
|
5951 |
-
"step": 19620
|
5952 |
-
},
|
5953 |
-
{
|
5954 |
-
"epoch": 5.076237239953483,
|
5955 |
-
"learning_rate": 1.6849474912485413e-07,
|
5956 |
-
"loss": 2.6402,
|
5957 |
-
"step": 19640
|
5958 |
-
},
|
5959 |
-
{
|
5960 |
-
"epoch": 5.081405866391006,
|
5961 |
-
"learning_rate": 1.684623363153118e-07,
|
5962 |
-
"loss": 2.6928,
|
5963 |
-
"step": 19660
|
5964 |
-
},
|
5965 |
-
{
|
5966 |
-
"epoch": 5.086574492828531,
|
5967 |
-
"learning_rate": 1.6842992350576948e-07,
|
5968 |
-
"loss": 2.6689,
|
5969 |
-
"step": 19680
|
5970 |
-
},
|
5971 |
-
{
|
5972 |
-
"epoch": 5.091743119266055,
|
5973 |
-
"learning_rate": 1.6839751069622714e-07,
|
5974 |
-
"loss": 2.6848,
|
5975 |
-
"step": 19700
|
5976 |
-
},
|
5977 |
-
{
|
5978 |
-
"epoch": 5.096911745703579,
|
5979 |
-
"learning_rate": 1.683650978866848e-07,
|
5980 |
-
"loss": 2.6315,
|
5981 |
-
"step": 19720
|
5982 |
-
},
|
5983 |
-
{
|
5984 |
-
"epoch": 5.102080372141104,
|
5985 |
-
"learning_rate": 1.683326850771425e-07,
|
5986 |
-
"loss": 2.6936,
|
5987 |
-
"step": 19740
|
5988 |
-
},
|
5989 |
-
{
|
5990 |
-
"epoch": 5.107248998578628,
|
5991 |
-
"learning_rate": 1.6830027226760015e-07,
|
5992 |
-
"loss": 2.6354,
|
5993 |
-
"step": 19760
|
5994 |
-
},
|
5995 |
-
{
|
5996 |
-
"epoch": 5.112417625016152,
|
5997 |
-
"learning_rate": 1.6826785945805782e-07,
|
5998 |
-
"loss": 2.6376,
|
5999 |
-
"step": 19780
|
6000 |
-
},
|
6001 |
-
{
|
6002 |
-
"epoch": 5.117586251453676,
|
6003 |
-
"learning_rate": 1.6823544664851548e-07,
|
6004 |
-
"loss": 2.7595,
|
6005 |
-
"step": 19800
|
6006 |
-
},
|
6007 |
-
{
|
6008 |
-
"epoch": 5.1227548778912,
|
6009 |
-
"learning_rate": 1.6820303383897317e-07,
|
6010 |
-
"loss": 2.5688,
|
6011 |
-
"step": 19820
|
6012 |
-
},
|
6013 |
-
{
|
6014 |
-
"epoch": 5.127923504328725,
|
6015 |
-
"learning_rate": 1.6817062102943083e-07,
|
6016 |
-
"loss": 2.6504,
|
6017 |
-
"step": 19840
|
6018 |
-
},
|
6019 |
-
{
|
6020 |
-
"epoch": 5.133092130766249,
|
6021 |
-
"learning_rate": 1.681382082198885e-07,
|
6022 |
-
"loss": 2.701,
|
6023 |
-
"step": 19860
|
6024 |
-
},
|
6025 |
-
{
|
6026 |
-
"epoch": 5.138260757203773,
|
6027 |
-
"learning_rate": 1.6810579541034615e-07,
|
6028 |
-
"loss": 2.6691,
|
6029 |
-
"step": 19880
|
6030 |
-
},
|
6031 |
-
{
|
6032 |
-
"epoch": 5.143429383641298,
|
6033 |
-
"learning_rate": 1.6807338260080384e-07,
|
6034 |
-
"loss": 2.6396,
|
6035 |
-
"step": 19900
|
6036 |
-
},
|
6037 |
-
{
|
6038 |
-
"epoch": 5.148598010078821,
|
6039 |
-
"learning_rate": 1.680409697912615e-07,
|
6040 |
-
"loss": 2.6752,
|
6041 |
-
"step": 19920
|
6042 |
-
},
|
6043 |
-
{
|
6044 |
-
"epoch": 5.1537666365163455,
|
6045 |
-
"learning_rate": 1.6800855698171916e-07,
|
6046 |
-
"loss": 2.567,
|
6047 |
-
"step": 19940
|
6048 |
-
},
|
6049 |
-
{
|
6050 |
-
"epoch": 5.15893526295387,
|
6051 |
-
"learning_rate": 1.6797614417217685e-07,
|
6052 |
-
"loss": 2.6691,
|
6053 |
-
"step": 19960
|
6054 |
-
},
|
6055 |
-
{
|
6056 |
-
"epoch": 5.164103889391394,
|
6057 |
-
"learning_rate": 1.6794373136263451e-07,
|
6058 |
-
"loss": 2.6172,
|
6059 |
-
"step": 19980
|
6060 |
-
},
|
6061 |
-
{
|
6062 |
-
"epoch": 5.169272515828919,
|
6063 |
-
"learning_rate": 1.6791131855309218e-07,
|
6064 |
-
"loss": 2.6856,
|
6065 |
-
"step": 20000
|
6066 |
-
},
|
6067 |
-
{
|
6068 |
-
"epoch": 5.174441142266443,
|
6069 |
-
"learning_rate": 1.6787890574354984e-07,
|
6070 |
-
"loss": 2.6874,
|
6071 |
-
"step": 20020
|
6072 |
-
},
|
6073 |
-
{
|
6074 |
-
"epoch": 5.179609768703967,
|
6075 |
-
"learning_rate": 1.678464929340075e-07,
|
6076 |
-
"loss": 2.6737,
|
6077 |
-
"step": 20040
|
6078 |
-
},
|
6079 |
-
{
|
6080 |
-
"epoch": 5.184778395141491,
|
6081 |
-
"learning_rate": 1.678140801244652e-07,
|
6082 |
-
"loss": 2.6503,
|
6083 |
-
"step": 20060
|
6084 |
-
},
|
6085 |
-
{
|
6086 |
-
"epoch": 5.189947021579015,
|
6087 |
-
"learning_rate": 1.6778166731492285e-07,
|
6088 |
-
"loss": 2.6155,
|
6089 |
-
"step": 20080
|
6090 |
-
},
|
6091 |
-
{
|
6092 |
-
"epoch": 5.1951156480165395,
|
6093 |
-
"learning_rate": 1.677492545053805e-07,
|
6094 |
-
"loss": 2.7035,
|
6095 |
-
"step": 20100
|
6096 |
-
},
|
6097 |
-
{
|
6098 |
-
"epoch": 5.200284274454064,
|
6099 |
-
"learning_rate": 1.677168416958382e-07,
|
6100 |
-
"loss": 2.6192,
|
6101 |
-
"step": 20120
|
6102 |
-
},
|
6103 |
-
{
|
6104 |
-
"epoch": 5.205452900891588,
|
6105 |
-
"learning_rate": 1.6768442888629586e-07,
|
6106 |
-
"loss": 2.5974,
|
6107 |
-
"step": 20140
|
6108 |
-
},
|
6109 |
-
{
|
6110 |
-
"epoch": 5.2106215273291125,
|
6111 |
-
"learning_rate": 1.6765201607675352e-07,
|
6112 |
-
"loss": 2.694,
|
6113 |
-
"step": 20160
|
6114 |
-
},
|
6115 |
-
{
|
6116 |
-
"epoch": 5.215790153766637,
|
6117 |
-
"learning_rate": 1.676196032672112e-07,
|
6118 |
-
"loss": 2.6391,
|
6119 |
-
"step": 20180
|
6120 |
-
},
|
6121 |
-
{
|
6122 |
-
"epoch": 5.22095878020416,
|
6123 |
-
"learning_rate": 1.6758719045766887e-07,
|
6124 |
-
"loss": 2.6428,
|
6125 |
-
"step": 20200
|
6126 |
-
},
|
6127 |
-
{
|
6128 |
-
"epoch": 5.226127406641685,
|
6129 |
-
"learning_rate": 1.6755477764812654e-07,
|
6130 |
-
"loss": 2.6796,
|
6131 |
-
"step": 20220
|
6132 |
-
},
|
6133 |
-
{
|
6134 |
-
"epoch": 5.231296033079209,
|
6135 |
-
"learning_rate": 1.675223648385842e-07,
|
6136 |
-
"loss": 2.6819,
|
6137 |
-
"step": 20240
|
6138 |
-
},
|
6139 |
-
{
|
6140 |
-
"epoch": 5.2364646595167335,
|
6141 |
-
"learning_rate": 1.6748995202904186e-07,
|
6142 |
-
"loss": 2.6714,
|
6143 |
-
"step": 20260
|
6144 |
-
},
|
6145 |
-
{
|
6146 |
-
"epoch": 5.241633285954258,
|
6147 |
-
"learning_rate": 1.6745753921949955e-07,
|
6148 |
-
"loss": 2.6392,
|
6149 |
-
"step": 20280
|
6150 |
-
},
|
6151 |
-
{
|
6152 |
-
"epoch": 5.246801912391782,
|
6153 |
-
"learning_rate": 1.674251264099572e-07,
|
6154 |
-
"loss": 2.7057,
|
6155 |
-
"step": 20300
|
6156 |
-
},
|
6157 |
-
{
|
6158 |
-
"epoch": 5.2519705388293065,
|
6159 |
-
"learning_rate": 1.6739271360041487e-07,
|
6160 |
-
"loss": 2.6676,
|
6161 |
-
"step": 20320
|
6162 |
-
},
|
6163 |
-
{
|
6164 |
-
"epoch": 5.25713916526683,
|
6165 |
-
"learning_rate": 1.6736030079087256e-07,
|
6166 |
-
"loss": 2.7037,
|
6167 |
-
"step": 20340
|
6168 |
-
},
|
6169 |
-
{
|
6170 |
-
"epoch": 5.262307791704354,
|
6171 |
-
"learning_rate": 1.6732788798133022e-07,
|
6172 |
-
"loss": 2.7595,
|
6173 |
-
"step": 20360
|
6174 |
-
},
|
6175 |
-
{
|
6176 |
-
"epoch": 5.267476418141879,
|
6177 |
-
"learning_rate": 1.6729547517178788e-07,
|
6178 |
-
"loss": 2.6665,
|
6179 |
-
"step": 20380
|
6180 |
-
},
|
6181 |
-
{
|
6182 |
-
"epoch": 5.272645044579403,
|
6183 |
-
"learning_rate": 1.6726306236224554e-07,
|
6184 |
-
"loss": 2.6281,
|
6185 |
-
"step": 20400
|
6186 |
-
},
|
6187 |
-
{
|
6188 |
-
"epoch": 5.277813671016927,
|
6189 |
-
"learning_rate": 1.6723064955270323e-07,
|
6190 |
-
"loss": 2.6047,
|
6191 |
-
"step": 20420
|
6192 |
-
},
|
6193 |
-
{
|
6194 |
-
"epoch": 5.282982297454452,
|
6195 |
-
"learning_rate": 1.671982367431609e-07,
|
6196 |
-
"loss": 2.6466,
|
6197 |
-
"step": 20440
|
6198 |
-
},
|
6199 |
-
{
|
6200 |
-
"epoch": 5.288150923891976,
|
6201 |
-
"learning_rate": 1.6716582393361856e-07,
|
6202 |
-
"loss": 2.6026,
|
6203 |
-
"step": 20460
|
6204 |
-
},
|
6205 |
-
{
|
6206 |
-
"epoch": 5.2933195503295,
|
6207 |
-
"learning_rate": 1.6713341112407622e-07,
|
6208 |
-
"loss": 2.6714,
|
6209 |
-
"step": 20480
|
6210 |
-
},
|
6211 |
-
{
|
6212 |
-
"epoch": 5.298488176767024,
|
6213 |
-
"learning_rate": 1.671009983145339e-07,
|
6214 |
-
"loss": 2.6351,
|
6215 |
-
"step": 20500
|
6216 |
-
},
|
6217 |
-
{
|
6218 |
-
"epoch": 5.303656803204548,
|
6219 |
-
"learning_rate": 1.6706858550499157e-07,
|
6220 |
-
"loss": 2.6293,
|
6221 |
-
"step": 20520
|
6222 |
-
},
|
6223 |
-
{
|
6224 |
-
"epoch": 5.308825429642073,
|
6225 |
-
"learning_rate": 1.6703617269544923e-07,
|
6226 |
-
"loss": 2.6368,
|
6227 |
-
"step": 20540
|
6228 |
-
},
|
6229 |
-
{
|
6230 |
-
"epoch": 5.313994056079597,
|
6231 |
-
"learning_rate": 1.6700375988590692e-07,
|
6232 |
-
"loss": 2.6963,
|
6233 |
-
"step": 20560
|
6234 |
-
},
|
6235 |
-
{
|
6236 |
-
"epoch": 5.319162682517121,
|
6237 |
-
"learning_rate": 1.6697134707636458e-07,
|
6238 |
-
"loss": 2.6401,
|
6239 |
-
"step": 20580
|
6240 |
-
},
|
6241 |
-
{
|
6242 |
-
"epoch": 5.324331308954645,
|
6243 |
-
"learning_rate": 1.6693893426682224e-07,
|
6244 |
-
"loss": 2.669,
|
6245 |
-
"step": 20600
|
6246 |
-
},
|
6247 |
-
{
|
6248 |
-
"epoch": 5.329499935392169,
|
6249 |
-
"learning_rate": 1.669065214572799e-07,
|
6250 |
-
"loss": 2.6384,
|
6251 |
-
"step": 20620
|
6252 |
-
},
|
6253 |
-
{
|
6254 |
-
"epoch": 5.334668561829694,
|
6255 |
-
"learning_rate": 1.6687410864773757e-07,
|
6256 |
-
"loss": 2.7073,
|
6257 |
-
"step": 20640
|
6258 |
-
},
|
6259 |
-
{
|
6260 |
-
"epoch": 5.339837188267218,
|
6261 |
-
"learning_rate": 1.6684169583819526e-07,
|
6262 |
-
"loss": 2.6507,
|
6263 |
-
"step": 20660
|
6264 |
-
},
|
6265 |
-
{
|
6266 |
-
"epoch": 5.345005814704742,
|
6267 |
-
"learning_rate": 1.6680928302865292e-07,
|
6268 |
-
"loss": 2.6378,
|
6269 |
-
"step": 20680
|
6270 |
-
},
|
6271 |
-
{
|
6272 |
-
"epoch": 5.350174441142267,
|
6273 |
-
"learning_rate": 1.6677687021911058e-07,
|
6274 |
-
"loss": 2.6756,
|
6275 |
-
"step": 20700
|
6276 |
-
},
|
6277 |
-
{
|
6278 |
-
"epoch": 5.355343067579791,
|
6279 |
-
"learning_rate": 1.6674445740956827e-07,
|
6280 |
-
"loss": 2.6667,
|
6281 |
-
"step": 20720
|
6282 |
-
},
|
6283 |
-
{
|
6284 |
-
"epoch": 5.3605116940173145,
|
6285 |
-
"learning_rate": 1.6671204460002593e-07,
|
6286 |
-
"loss": 2.6714,
|
6287 |
-
"step": 20740
|
6288 |
-
},
|
6289 |
-
{
|
6290 |
-
"epoch": 5.365680320454839,
|
6291 |
-
"learning_rate": 1.666796317904836e-07,
|
6292 |
-
"loss": 2.6117,
|
6293 |
-
"step": 20760
|
6294 |
-
},
|
6295 |
-
{
|
6296 |
-
"epoch": 5.370848946892363,
|
6297 |
-
"learning_rate": 1.6664721898094128e-07,
|
6298 |
-
"loss": 2.6513,
|
6299 |
-
"step": 20780
|
6300 |
-
},
|
6301 |
-
{
|
6302 |
-
"epoch": 5.376017573329888,
|
6303 |
-
"learning_rate": 1.6661480617139894e-07,
|
6304 |
-
"loss": 2.6997,
|
6305 |
-
"step": 20800
|
6306 |
-
},
|
6307 |
-
{
|
6308 |
-
"epoch": 5.381186199767412,
|
6309 |
-
"learning_rate": 1.665823933618566e-07,
|
6310 |
-
"loss": 2.6395,
|
6311 |
-
"step": 20820
|
6312 |
-
},
|
6313 |
-
{
|
6314 |
-
"epoch": 5.386354826204936,
|
6315 |
-
"learning_rate": 1.6654998055231426e-07,
|
6316 |
-
"loss": 2.6615,
|
6317 |
-
"step": 20840
|
6318 |
-
},
|
6319 |
-
{
|
6320 |
-
"epoch": 5.39152345264246,
|
6321 |
-
"learning_rate": 1.6651756774277193e-07,
|
6322 |
-
"loss": 2.6915,
|
6323 |
-
"step": 20860
|
6324 |
-
},
|
6325 |
-
{
|
6326 |
-
"epoch": 5.396692079079984,
|
6327 |
-
"learning_rate": 1.6648515493322961e-07,
|
6328 |
-
"loss": 2.6248,
|
6329 |
-
"step": 20880
|
6330 |
-
},
|
6331 |
-
{
|
6332 |
-
"epoch": 5.4018607055175085,
|
6333 |
-
"learning_rate": 1.6645274212368728e-07,
|
6334 |
-
"loss": 2.7061,
|
6335 |
-
"step": 20900
|
6336 |
-
},
|
6337 |
-
{
|
6338 |
-
"epoch": 5.407029331955033,
|
6339 |
-
"learning_rate": 1.6642032931414494e-07,
|
6340 |
-
"loss": 2.642,
|
6341 |
-
"step": 20920
|
6342 |
-
},
|
6343 |
-
{
|
6344 |
-
"epoch": 5.412197958392557,
|
6345 |
-
"learning_rate": 1.6638791650460263e-07,
|
6346 |
-
"loss": 2.6377,
|
6347 |
-
"step": 20940
|
6348 |
-
},
|
6349 |
-
{
|
6350 |
-
"epoch": 5.417366584830082,
|
6351 |
-
"learning_rate": 1.663555036950603e-07,
|
6352 |
-
"loss": 2.6452,
|
6353 |
-
"step": 20960
|
6354 |
-
},
|
6355 |
-
{
|
6356 |
-
"epoch": 5.422535211267606,
|
6357 |
-
"learning_rate": 1.6632309088551795e-07,
|
6358 |
-
"loss": 2.7206,
|
6359 |
-
"step": 20980
|
6360 |
-
},
|
6361 |
-
{
|
6362 |
-
"epoch": 5.427703837705129,
|
6363 |
-
"learning_rate": 1.662906780759756e-07,
|
6364 |
-
"loss": 2.6524,
|
6365 |
-
"step": 21000
|
6366 |
-
},
|
6367 |
-
{
|
6368 |
-
"epoch": 5.432872464142654,
|
6369 |
-
"learning_rate": 1.662582652664333e-07,
|
6370 |
-
"loss": 2.6398,
|
6371 |
-
"step": 21020
|
6372 |
-
},
|
6373 |
-
{
|
6374 |
-
"epoch": 5.438041090580178,
|
6375 |
-
"learning_rate": 1.6622585245689096e-07,
|
6376 |
-
"loss": 2.6486,
|
6377 |
-
"step": 21040
|
6378 |
-
},
|
6379 |
-
{
|
6380 |
-
"epoch": 5.4432097170177025,
|
6381 |
-
"learning_rate": 1.6619343964734862e-07,
|
6382 |
-
"loss": 2.6466,
|
6383 |
-
"step": 21060
|
6384 |
-
},
|
6385 |
-
{
|
6386 |
-
"epoch": 5.448378343455227,
|
6387 |
-
"learning_rate": 1.6616102683780629e-07,
|
6388 |
-
"loss": 2.6818,
|
6389 |
-
"step": 21080
|
6390 |
-
},
|
6391 |
-
{
|
6392 |
-
"epoch": 5.453546969892751,
|
6393 |
-
"learning_rate": 1.6612861402826397e-07,
|
6394 |
-
"loss": 2.6826,
|
6395 |
-
"step": 21100
|
6396 |
-
},
|
6397 |
-
{
|
6398 |
-
"epoch": 5.458715596330276,
|
6399 |
-
"learning_rate": 1.6609620121872164e-07,
|
6400 |
-
"loss": 2.665,
|
6401 |
-
"step": 21120
|
6402 |
-
},
|
6403 |
-
{
|
6404 |
-
"epoch": 5.463884222767799,
|
6405 |
-
"learning_rate": 1.660637884091793e-07,
|
6406 |
-
"loss": 2.6018,
|
6407 |
-
"step": 21140
|
6408 |
-
},
|
6409 |
-
{
|
6410 |
-
"epoch": 5.469052849205323,
|
6411 |
-
"learning_rate": 1.66031375599637e-07,
|
6412 |
-
"loss": 2.6867,
|
6413 |
-
"step": 21160
|
6414 |
-
},
|
6415 |
-
{
|
6416 |
-
"epoch": 5.474221475642848,
|
6417 |
-
"learning_rate": 1.6599896279009465e-07,
|
6418 |
-
"loss": 2.6972,
|
6419 |
-
"step": 21180
|
6420 |
-
},
|
6421 |
-
{
|
6422 |
-
"epoch": 5.479390102080372,
|
6423 |
-
"learning_rate": 1.659665499805523e-07,
|
6424 |
-
"loss": 2.6775,
|
6425 |
-
"step": 21200
|
6426 |
-
},
|
6427 |
-
{
|
6428 |
-
"epoch": 5.4845587285178965,
|
6429 |
-
"learning_rate": 1.6593413717100997e-07,
|
6430 |
-
"loss": 2.7239,
|
6431 |
-
"step": 21220
|
6432 |
-
},
|
6433 |
-
{
|
6434 |
-
"epoch": 5.489727354955421,
|
6435 |
-
"learning_rate": 1.6590172436146763e-07,
|
6436 |
-
"loss": 2.6349,
|
6437 |
-
"step": 21240
|
6438 |
-
},
|
6439 |
-
{
|
6440 |
-
"epoch": 5.494895981392945,
|
6441 |
-
"learning_rate": 1.6586931155192532e-07,
|
6442 |
-
"loss": 2.6607,
|
6443 |
-
"step": 21260
|
6444 |
-
},
|
6445 |
-
{
|
6446 |
-
"epoch": 5.500064607830469,
|
6447 |
-
"learning_rate": 1.6583689874238298e-07,
|
6448 |
-
"loss": 2.6806,
|
6449 |
-
"step": 21280
|
6450 |
-
},
|
6451 |
-
{
|
6452 |
-
"epoch": 5.505233234267993,
|
6453 |
-
"learning_rate": 1.6580448593284065e-07,
|
6454 |
-
"loss": 2.7212,
|
6455 |
-
"step": 21300
|
6456 |
-
},
|
6457 |
-
{
|
6458 |
-
"epoch": 5.510401860705517,
|
6459 |
-
"learning_rate": 1.6577207312329833e-07,
|
6460 |
-
"loss": 2.6997,
|
6461 |
-
"step": 21320
|
6462 |
-
},
|
6463 |
-
{
|
6464 |
-
"epoch": 5.515570487143042,
|
6465 |
-
"learning_rate": 1.65739660313756e-07,
|
6466 |
-
"loss": 2.6277,
|
6467 |
-
"step": 21340
|
6468 |
-
},
|
6469 |
-
{
|
6470 |
-
"epoch": 5.520739113580566,
|
6471 |
-
"learning_rate": 1.6570724750421366e-07,
|
6472 |
-
"loss": 2.6953,
|
6473 |
-
"step": 21360
|
6474 |
-
},
|
6475 |
-
{
|
6476 |
-
"epoch": 5.5259077400180905,
|
6477 |
-
"learning_rate": 1.6567483469467135e-07,
|
6478 |
-
"loss": 2.7586,
|
6479 |
-
"step": 21380
|
6480 |
-
},
|
6481 |
-
{
|
6482 |
-
"epoch": 5.531076366455615,
|
6483 |
-
"learning_rate": 1.65642421885129e-07,
|
6484 |
-
"loss": 2.5887,
|
6485 |
-
"step": 21400
|
6486 |
-
},
|
6487 |
-
{
|
6488 |
-
"epoch": 5.536244992893138,
|
6489 |
-
"learning_rate": 1.6561000907558667e-07,
|
6490 |
-
"loss": 2.6364,
|
6491 |
-
"step": 21420
|
6492 |
-
},
|
6493 |
-
{
|
6494 |
-
"epoch": 5.541413619330663,
|
6495 |
-
"learning_rate": 1.6557759626604433e-07,
|
6496 |
-
"loss": 2.5951,
|
6497 |
-
"step": 21440
|
6498 |
-
},
|
6499 |
-
{
|
6500 |
-
"epoch": 5.546582245768187,
|
6501 |
-
"learning_rate": 1.65545183456502e-07,
|
6502 |
-
"loss": 2.6595,
|
6503 |
-
"step": 21460
|
6504 |
-
},
|
6505 |
-
{
|
6506 |
-
"epoch": 5.551750872205711,
|
6507 |
-
"learning_rate": 1.6551277064695968e-07,
|
6508 |
-
"loss": 2.602,
|
6509 |
-
"step": 21480
|
6510 |
-
},
|
6511 |
-
{
|
6512 |
-
"epoch": 5.556919498643236,
|
6513 |
-
"learning_rate": 1.6548035783741734e-07,
|
6514 |
-
"loss": 2.5996,
|
6515 |
-
"step": 21500
|
6516 |
-
},
|
6517 |
-
{
|
6518 |
-
"epoch": 5.56208812508076,
|
6519 |
-
"learning_rate": 1.65447945027875e-07,
|
6520 |
-
"loss": 2.6529,
|
6521 |
-
"step": 21520
|
6522 |
-
},
|
6523 |
-
{
|
6524 |
-
"epoch": 5.5672567515182845,
|
6525 |
-
"learning_rate": 1.654155322183327e-07,
|
6526 |
-
"loss": 2.6917,
|
6527 |
-
"step": 21540
|
6528 |
-
},
|
6529 |
-
{
|
6530 |
-
"epoch": 5.572425377955808,
|
6531 |
-
"learning_rate": 1.6538311940879036e-07,
|
6532 |
-
"loss": 2.6509,
|
6533 |
-
"step": 21560
|
6534 |
-
},
|
6535 |
-
{
|
6536 |
-
"epoch": 5.577594004393332,
|
6537 |
-
"learning_rate": 1.6535070659924802e-07,
|
6538 |
-
"loss": 2.6524,
|
6539 |
-
"step": 21580
|
6540 |
-
},
|
6541 |
-
{
|
6542 |
-
"epoch": 5.582762630830857,
|
6543 |
-
"learning_rate": 1.6531829378970568e-07,
|
6544 |
-
"loss": 2.6894,
|
6545 |
-
"step": 21600
|
6546 |
-
},
|
6547 |
-
{
|
6548 |
-
"epoch": 5.587931257268381,
|
6549 |
-
"learning_rate": 1.6528588098016334e-07,
|
6550 |
-
"loss": 2.6451,
|
6551 |
-
"step": 21620
|
6552 |
-
},
|
6553 |
-
{
|
6554 |
-
"epoch": 5.593099883705905,
|
6555 |
-
"learning_rate": 1.6525346817062103e-07,
|
6556 |
-
"loss": 2.6636,
|
6557 |
-
"step": 21640
|
6558 |
-
},
|
6559 |
-
{
|
6560 |
-
"epoch": 5.59826851014343,
|
6561 |
-
"learning_rate": 1.652210553610787e-07,
|
6562 |
-
"loss": 2.6696,
|
6563 |
-
"step": 21660
|
6564 |
-
},
|
6565 |
-
{
|
6566 |
-
"epoch": 5.603437136580954,
|
6567 |
-
"learning_rate": 1.6518864255153635e-07,
|
6568 |
-
"loss": 2.6577,
|
6569 |
-
"step": 21680
|
6570 |
-
},
|
6571 |
-
{
|
6572 |
-
"epoch": 5.608605763018478,
|
6573 |
-
"learning_rate": 1.6515622974199404e-07,
|
6574 |
-
"loss": 2.5968,
|
6575 |
-
"step": 21700
|
6576 |
-
},
|
6577 |
-
{
|
6578 |
-
"epoch": 5.613774389456002,
|
6579 |
-
"learning_rate": 1.651238169324517e-07,
|
6580 |
-
"loss": 2.6492,
|
6581 |
-
"step": 21720
|
6582 |
-
},
|
6583 |
-
{
|
6584 |
-
"epoch": 5.618943015893526,
|
6585 |
-
"learning_rate": 1.6509140412290937e-07,
|
6586 |
-
"loss": 2.6357,
|
6587 |
-
"step": 21740
|
6588 |
-
},
|
6589 |
-
{
|
6590 |
-
"epoch": 5.624111642331051,
|
6591 |
-
"learning_rate": 1.6505899131336705e-07,
|
6592 |
-
"loss": 2.6401,
|
6593 |
-
"step": 21760
|
6594 |
-
},
|
6595 |
-
{
|
6596 |
-
"epoch": 5.629280268768575,
|
6597 |
-
"learning_rate": 1.650265785038247e-07,
|
6598 |
-
"loss": 2.5984,
|
6599 |
-
"step": 21780
|
6600 |
-
},
|
6601 |
-
{
|
6602 |
-
"epoch": 5.6344488952060985,
|
6603 |
-
"learning_rate": 1.6499416569428238e-07,
|
6604 |
-
"loss": 2.6678,
|
6605 |
-
"step": 21800
|
6606 |
-
},
|
6607 |
-
{
|
6608 |
-
"epoch": 5.639617521643623,
|
6609 |
-
"learning_rate": 1.6496175288474004e-07,
|
6610 |
-
"loss": 2.6161,
|
6611 |
-
"step": 21820
|
6612 |
-
},
|
6613 |
-
{
|
6614 |
-
"epoch": 5.644786148081147,
|
6615 |
-
"learning_rate": 1.649293400751977e-07,
|
6616 |
-
"loss": 2.6262,
|
6617 |
-
"step": 21840
|
6618 |
-
},
|
6619 |
-
{
|
6620 |
-
"epoch": 5.649954774518672,
|
6621 |
-
"learning_rate": 1.648969272656554e-07,
|
6622 |
-
"loss": 2.6514,
|
6623 |
-
"step": 21860
|
6624 |
-
},
|
6625 |
-
{
|
6626 |
-
"epoch": 5.655123400956196,
|
6627 |
-
"learning_rate": 1.6486451445611305e-07,
|
6628 |
-
"loss": 2.6629,
|
6629 |
-
"step": 21880
|
6630 |
-
},
|
6631 |
-
{
|
6632 |
-
"epoch": 5.66029202739372,
|
6633 |
-
"learning_rate": 1.6483210164657071e-07,
|
6634 |
-
"loss": 2.6764,
|
6635 |
-
"step": 21900
|
6636 |
-
},
|
6637 |
-
{
|
6638 |
-
"epoch": 5.665460653831245,
|
6639 |
-
"learning_rate": 1.647996888370284e-07,
|
6640 |
-
"loss": 2.6414,
|
6641 |
-
"step": 21920
|
6642 |
-
},
|
6643 |
-
{
|
6644 |
-
"epoch": 5.670629280268768,
|
6645 |
-
"learning_rate": 1.6476727602748604e-07,
|
6646 |
-
"loss": 2.5379,
|
6647 |
-
"step": 21940
|
6648 |
-
},
|
6649 |
-
{
|
6650 |
-
"epoch": 5.6757979067062925,
|
6651 |
-
"learning_rate": 1.6473486321794373e-07,
|
6652 |
-
"loss": 2.6744,
|
6653 |
-
"step": 21960
|
6654 |
-
},
|
6655 |
-
{
|
6656 |
-
"epoch": 5.680966533143817,
|
6657 |
-
"learning_rate": 1.647024504084014e-07,
|
6658 |
-
"loss": 2.7254,
|
6659 |
-
"step": 21980
|
6660 |
-
},
|
6661 |
-
{
|
6662 |
-
"epoch": 5.686135159581341,
|
6663 |
-
"learning_rate": 1.6467003759885905e-07,
|
6664 |
-
"loss": 2.6408,
|
6665 |
-
"step": 22000
|
6666 |
-
},
|
6667 |
-
{
|
6668 |
-
"epoch": 5.6913037860188656,
|
6669 |
-
"learning_rate": 1.6463762478931674e-07,
|
6670 |
-
"loss": 2.6751,
|
6671 |
-
"step": 22020
|
6672 |
-
},
|
6673 |
-
{
|
6674 |
-
"epoch": 5.69647241245639,
|
6675 |
-
"learning_rate": 1.646052119797744e-07,
|
6676 |
-
"loss": 2.6391,
|
6677 |
-
"step": 22040
|
6678 |
-
},
|
6679 |
-
{
|
6680 |
-
"epoch": 5.701641038893914,
|
6681 |
-
"learning_rate": 1.6457279917023206e-07,
|
6682 |
-
"loss": 2.625,
|
6683 |
-
"step": 22060
|
6684 |
-
},
|
6685 |
-
{
|
6686 |
-
"epoch": 5.706809665331438,
|
6687 |
-
"learning_rate": 1.6454038636068975e-07,
|
6688 |
-
"loss": 2.607,
|
6689 |
-
"step": 22080
|
6690 |
-
},
|
6691 |
-
{
|
6692 |
-
"epoch": 5.711978291768962,
|
6693 |
-
"learning_rate": 1.6450797355114739e-07,
|
6694 |
-
"loss": 2.6629,
|
6695 |
-
"step": 22100
|
6696 |
-
},
|
6697 |
-
{
|
6698 |
-
"epoch": 5.7171469182064865,
|
6699 |
-
"learning_rate": 1.6447556074160507e-07,
|
6700 |
-
"loss": 2.6358,
|
6701 |
-
"step": 22120
|
6702 |
-
},
|
6703 |
-
{
|
6704 |
-
"epoch": 5.722315544644011,
|
6705 |
-
"learning_rate": 1.6444314793206274e-07,
|
6706 |
-
"loss": 2.6962,
|
6707 |
-
"step": 22140
|
6708 |
-
},
|
6709 |
-
{
|
6710 |
-
"epoch": 5.727484171081535,
|
6711 |
-
"learning_rate": 1.644107351225204e-07,
|
6712 |
-
"loss": 2.6403,
|
6713 |
-
"step": 22160
|
6714 |
-
},
|
6715 |
-
{
|
6716 |
-
"epoch": 5.7326527975190595,
|
6717 |
-
"learning_rate": 1.6437832231297809e-07,
|
6718 |
-
"loss": 2.641,
|
6719 |
-
"step": 22180
|
6720 |
-
},
|
6721 |
-
{
|
6722 |
-
"epoch": 5.737821423956584,
|
6723 |
-
"learning_rate": 1.6434590950343575e-07,
|
6724 |
-
"loss": 2.6213,
|
6725 |
-
"step": 22200
|
6726 |
-
},
|
6727 |
-
{
|
6728 |
-
"epoch": 5.742990050394107,
|
6729 |
-
"learning_rate": 1.643134966938934e-07,
|
6730 |
-
"loss": 2.6508,
|
6731 |
-
"step": 22220
|
6732 |
-
},
|
6733 |
-
{
|
6734 |
-
"epoch": 5.748158676831632,
|
6735 |
-
"learning_rate": 1.642810838843511e-07,
|
6736 |
-
"loss": 2.6528,
|
6737 |
-
"step": 22240
|
6738 |
-
},
|
6739 |
-
{
|
6740 |
-
"epoch": 5.753327303269156,
|
6741 |
-
"learning_rate": 1.6424867107480873e-07,
|
6742 |
-
"loss": 2.5921,
|
6743 |
-
"step": 22260
|
6744 |
-
},
|
6745 |
-
{
|
6746 |
-
"epoch": 5.7584959297066804,
|
6747 |
-
"learning_rate": 1.6421625826526642e-07,
|
6748 |
-
"loss": 2.6379,
|
6749 |
-
"step": 22280
|
6750 |
-
},
|
6751 |
-
{
|
6752 |
-
"epoch": 5.763664556144205,
|
6753 |
-
"learning_rate": 1.641838454557241e-07,
|
6754 |
-
"loss": 2.6838,
|
6755 |
-
"step": 22300
|
6756 |
-
},
|
6757 |
-
{
|
6758 |
-
"epoch": 5.768833182581729,
|
6759 |
-
"learning_rate": 1.6415143264618175e-07,
|
6760 |
-
"loss": 2.6511,
|
6761 |
-
"step": 22320
|
6762 |
-
},
|
6763 |
-
{
|
6764 |
-
"epoch": 5.7740018090192535,
|
6765 |
-
"learning_rate": 1.6411901983663943e-07,
|
6766 |
-
"loss": 2.6547,
|
6767 |
-
"step": 22340
|
6768 |
-
},
|
6769 |
-
{
|
6770 |
-
"epoch": 5.779170435456777,
|
6771 |
-
"learning_rate": 1.640866070270971e-07,
|
6772 |
-
"loss": 2.5962,
|
6773 |
-
"step": 22360
|
6774 |
-
},
|
6775 |
-
{
|
6776 |
-
"epoch": 5.784339061894301,
|
6777 |
-
"learning_rate": 1.6405419421755476e-07,
|
6778 |
-
"loss": 2.6074,
|
6779 |
-
"step": 22380
|
6780 |
-
},
|
6781 |
-
{
|
6782 |
-
"epoch": 5.789507688331826,
|
6783 |
-
"learning_rate": 1.6402178140801245e-07,
|
6784 |
-
"loss": 2.5927,
|
6785 |
-
"step": 22400
|
6786 |
-
},
|
6787 |
-
{
|
6788 |
-
"epoch": 5.79467631476935,
|
6789 |
-
"learning_rate": 1.6398936859847008e-07,
|
6790 |
-
"loss": 2.6697,
|
6791 |
-
"step": 22420
|
6792 |
-
},
|
6793 |
-
{
|
6794 |
-
"epoch": 5.799844941206874,
|
6795 |
-
"learning_rate": 1.6395695578892777e-07,
|
6796 |
-
"loss": 2.6267,
|
6797 |
-
"step": 22440
|
6798 |
-
},
|
6799 |
-
{
|
6800 |
-
"epoch": 5.805013567644399,
|
6801 |
-
"learning_rate": 1.6392454297938546e-07,
|
6802 |
-
"loss": 2.6833,
|
6803 |
-
"step": 22460
|
6804 |
-
},
|
6805 |
-
{
|
6806 |
-
"epoch": 5.810182194081923,
|
6807 |
-
"learning_rate": 1.638921301698431e-07,
|
6808 |
-
"loss": 2.6647,
|
6809 |
-
"step": 22480
|
6810 |
-
},
|
6811 |
-
{
|
6812 |
-
"epoch": 5.815350820519447,
|
6813 |
-
"learning_rate": 1.6385971736030078e-07,
|
6814 |
-
"loss": 2.6619,
|
6815 |
-
"step": 22500
|
6816 |
-
},
|
6817 |
-
{
|
6818 |
-
"epoch": 5.820519446956971,
|
6819 |
-
"learning_rate": 1.6382730455075847e-07,
|
6820 |
-
"loss": 2.6341,
|
6821 |
-
"step": 22520
|
6822 |
-
},
|
6823 |
-
{
|
6824 |
-
"epoch": 5.825688073394495,
|
6825 |
-
"learning_rate": 1.637948917412161e-07,
|
6826 |
-
"loss": 2.604,
|
6827 |
-
"step": 22540
|
6828 |
-
},
|
6829 |
-
{
|
6830 |
-
"epoch": 5.83085669983202,
|
6831 |
-
"learning_rate": 1.637624789316738e-07,
|
6832 |
-
"loss": 2.6344,
|
6833 |
-
"step": 22560
|
6834 |
-
},
|
6835 |
-
{
|
6836 |
-
"epoch": 5.836025326269544,
|
6837 |
-
"learning_rate": 1.6373006612213146e-07,
|
6838 |
-
"loss": 2.6453,
|
6839 |
-
"step": 22580
|
6840 |
-
},
|
6841 |
-
{
|
6842 |
-
"epoch": 5.841193952707068,
|
6843 |
-
"learning_rate": 1.6369765331258912e-07,
|
6844 |
-
"loss": 2.6433,
|
6845 |
-
"step": 22600
|
6846 |
-
},
|
6847 |
-
{
|
6848 |
-
"epoch": 5.846362579144593,
|
6849 |
-
"learning_rate": 1.636652405030468e-07,
|
6850 |
-
"loss": 2.5673,
|
6851 |
-
"step": 22620
|
6852 |
-
},
|
6853 |
-
{
|
6854 |
-
"epoch": 5.851531205582116,
|
6855 |
-
"learning_rate": 1.6363282769350444e-07,
|
6856 |
-
"loss": 2.6455,
|
6857 |
-
"step": 22640
|
6858 |
-
},
|
6859 |
-
{
|
6860 |
-
"epoch": 5.856699832019641,
|
6861 |
-
"learning_rate": 1.6360041488396213e-07,
|
6862 |
-
"loss": 2.6617,
|
6863 |
-
"step": 22660
|
6864 |
-
},
|
6865 |
-
{
|
6866 |
-
"epoch": 5.861868458457165,
|
6867 |
-
"learning_rate": 1.6356800207441982e-07,
|
6868 |
-
"loss": 2.6395,
|
6869 |
-
"step": 22680
|
6870 |
-
},
|
6871 |
-
{
|
6872 |
-
"epoch": 5.867037084894689,
|
6873 |
-
"learning_rate": 1.6353558926487745e-07,
|
6874 |
-
"loss": 2.6358,
|
6875 |
-
"step": 22700
|
6876 |
-
},
|
6877 |
-
{
|
6878 |
-
"epoch": 5.872205711332214,
|
6879 |
-
"learning_rate": 1.6350317645533514e-07,
|
6880 |
-
"loss": 2.618,
|
6881 |
-
"step": 22720
|
6882 |
-
},
|
6883 |
-
{
|
6884 |
-
"epoch": 5.877374337769738,
|
6885 |
-
"learning_rate": 1.634707636457928e-07,
|
6886 |
-
"loss": 2.6219,
|
6887 |
-
"step": 22740
|
6888 |
-
},
|
6889 |
-
{
|
6890 |
-
"epoch": 5.882542964207262,
|
6891 |
-
"learning_rate": 1.6343835083625047e-07,
|
6892 |
-
"loss": 2.6028,
|
6893 |
-
"step": 22760
|
6894 |
-
},
|
6895 |
-
{
|
6896 |
-
"epoch": 5.887711590644786,
|
6897 |
-
"learning_rate": 1.6340593802670815e-07,
|
6898 |
-
"loss": 2.599,
|
6899 |
-
"step": 22780
|
6900 |
-
},
|
6901 |
-
{
|
6902 |
-
"epoch": 5.89288021708231,
|
6903 |
-
"learning_rate": 1.633735252171658e-07,
|
6904 |
-
"loss": 2.6062,
|
6905 |
-
"step": 22800
|
6906 |
-
},
|
6907 |
-
{
|
6908 |
-
"epoch": 5.898048843519835,
|
6909 |
-
"learning_rate": 1.6334111240762348e-07,
|
6910 |
-
"loss": 2.5854,
|
6911 |
-
"step": 22820
|
6912 |
-
},
|
6913 |
-
{
|
6914 |
-
"epoch": 5.903217469957359,
|
6915 |
-
"learning_rate": 1.6330869959808117e-07,
|
6916 |
-
"loss": 2.5416,
|
6917 |
-
"step": 22840
|
6918 |
-
},
|
6919 |
-
{
|
6920 |
-
"epoch": 5.908386096394883,
|
6921 |
-
"learning_rate": 1.632762867885388e-07,
|
6922 |
-
"loss": 2.6731,
|
6923 |
-
"step": 22860
|
6924 |
-
},
|
6925 |
-
{
|
6926 |
-
"epoch": 5.913554722832407,
|
6927 |
-
"learning_rate": 1.632438739789965e-07,
|
6928 |
-
"loss": 2.6271,
|
6929 |
-
"step": 22880
|
6930 |
-
},
|
6931 |
-
{
|
6932 |
-
"epoch": 5.918723349269931,
|
6933 |
-
"learning_rate": 1.6321146116945418e-07,
|
6934 |
-
"loss": 2.5896,
|
6935 |
-
"step": 22900
|
6936 |
-
},
|
6937 |
-
{
|
6938 |
-
"epoch": 5.9238919757074555,
|
6939 |
-
"learning_rate": 1.6317904835991181e-07,
|
6940 |
-
"loss": 2.6794,
|
6941 |
-
"step": 22920
|
6942 |
-
},
|
6943 |
-
{
|
6944 |
-
"epoch": 5.92906060214498,
|
6945 |
-
"learning_rate": 1.631466355503695e-07,
|
6946 |
-
"loss": 2.6051,
|
6947 |
-
"step": 22940
|
6948 |
-
},
|
6949 |
-
{
|
6950 |
-
"epoch": 5.934229228582504,
|
6951 |
-
"learning_rate": 1.6311422274082716e-07,
|
6952 |
-
"loss": 2.6901,
|
6953 |
-
"step": 22960
|
6954 |
-
},
|
6955 |
-
{
|
6956 |
-
"epoch": 5.939397855020029,
|
6957 |
-
"learning_rate": 1.6308180993128483e-07,
|
6958 |
-
"loss": 2.6288,
|
6959 |
-
"step": 22980
|
6960 |
-
},
|
6961 |
-
{
|
6962 |
-
"epoch": 5.944566481457553,
|
6963 |
-
"learning_rate": 1.6304939712174251e-07,
|
6964 |
-
"loss": 2.6806,
|
6965 |
-
"step": 23000
|
6966 |
-
},
|
6967 |
-
{
|
6968 |
-
"epoch": 5.949735107895076,
|
6969 |
-
"learning_rate": 1.6301698431220015e-07,
|
6970 |
-
"loss": 2.6003,
|
6971 |
-
"step": 23020
|
6972 |
-
},
|
6973 |
-
{
|
6974 |
-
"epoch": 5.954903734332601,
|
6975 |
-
"learning_rate": 1.6298457150265784e-07,
|
6976 |
-
"loss": 2.6595,
|
6977 |
-
"step": 23040
|
6978 |
-
},
|
6979 |
-
{
|
6980 |
-
"epoch": 5.960072360770125,
|
6981 |
-
"learning_rate": 1.6295215869311553e-07,
|
6982 |
-
"loss": 2.6563,
|
6983 |
-
"step": 23060
|
6984 |
-
},
|
6985 |
-
{
|
6986 |
-
"epoch": 5.9652409872076495,
|
6987 |
-
"learning_rate": 1.6291974588357316e-07,
|
6988 |
-
"loss": 2.6499,
|
6989 |
-
"step": 23080
|
6990 |
-
},
|
6991 |
-
{
|
6992 |
-
"epoch": 5.970409613645174,
|
6993 |
-
"learning_rate": 1.6288733307403085e-07,
|
6994 |
-
"loss": 2.6836,
|
6995 |
-
"step": 23100
|
6996 |
-
},
|
6997 |
-
{
|
6998 |
-
"epoch": 5.975578240082698,
|
6999 |
-
"learning_rate": 1.6285492026448854e-07,
|
7000 |
-
"loss": 2.6193,
|
7001 |
-
"step": 23120
|
7002 |
-
},
|
7003 |
-
{
|
7004 |
-
"epoch": 5.980746866520223,
|
7005 |
-
"learning_rate": 1.6282250745494617e-07,
|
7006 |
-
"loss": 2.6293,
|
7007 |
-
"step": 23140
|
7008 |
-
},
|
7009 |
-
{
|
7010 |
-
"epoch": 5.985915492957746,
|
7011 |
-
"learning_rate": 1.6279009464540386e-07,
|
7012 |
-
"loss": 2.7275,
|
7013 |
-
"step": 23160
|
7014 |
-
},
|
7015 |
-
{
|
7016 |
-
"epoch": 5.99108411939527,
|
7017 |
-
"learning_rate": 1.6275768183586152e-07,
|
7018 |
-
"loss": 2.6021,
|
7019 |
-
"step": 23180
|
7020 |
-
},
|
7021 |
-
{
|
7022 |
-
"epoch": 5.996252745832795,
|
7023 |
-
"learning_rate": 1.6272526902631919e-07,
|
7024 |
-
"loss": 2.6331,
|
7025 |
-
"step": 23200
|
7026 |
-
},
|
7027 |
-
{
|
7028 |
-
"epoch": 5.999870784339062,
|
7029 |
-
"eval_bleu": 1.6991,
|
7030 |
-
"eval_gen_len": 66.9245,
|
7031 |
-
"eval_loss": 2.596095085144043,
|
7032 |
-
"eval_runtime": 1347.5897,
|
7033 |
-
"eval_samples_per_second": 1.278,
|
7034 |
-
"eval_steps_per_second": 0.639,
|
7035 |
-
"step": 23214
|
7036 |
}
|
7037 |
],
|
7038 |
"logging_steps": 20,
|
@@ -7052,7 +5884,7 @@
|
|
7052 |
"attributes": {}
|
7053 |
}
|
7054 |
},
|
7055 |
-
"total_flos": 1.
|
7056 |
"train_batch_size": 2,
|
7057 |
"trial_name": null,
|
7058 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.5081,
|
3 |
+
"best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-19347",
|
4 |
+
"epoch": 4.999870784339062,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 19347,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
5865 |
"eval_samples_per_second": 1.019,
|
5866 |
"eval_steps_per_second": 0.51,
|
5867 |
"step": 19347
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5868 |
}
|
5869 |
],
|
5870 |
"logging_steps": 20,
|
|
|
5884 |
"attributes": {}
|
5885 |
}
|
5886 |
},
|
5887 |
+
"total_flos": 1.2701974646813491e+17,
|
5888 |
"train_batch_size": 2,
|
5889 |
"trial_name": null,
|
5890 |
"trial_params": null
|
last-checkpoint/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:581ac533a892239697e45b2f3ff1f250e5eee0be2c398f8f8ef23a485abea95b
|
3 |
+
size 5688
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:00790e6eb3aaf7be06f42ed5ebbbf19f51da90eabde4485d44e5afbbaa258042
|
3 |
size 1575259780
|
run-2024-10-28T03:01:14+00:00.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
val_outputs/val_generated_predictions_36735830.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|