abdiharyadi commited on
Commit
e629175
·
verified ·
1 Parent(s): eba0947

Training in progress, epoch 7

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f4f18bd8d1702c4a179d3d22bfeea1352e0e066d35c6cff7c5d20e32493e5e2
3
  size 1575259780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:044e91a7c52f50e4043dc236fc046775db89767dd283738cb05dae269d0fe1f9
3
  size 1575259780
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c57667d5cf758779a2f65c2c0c0b2fa8a4adb3689446f494b25f34a1442bc21e
3
  size 3150397656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06fd3b48586725ea6ba928db3bb6432129af3d8438eb526dfaaa3ca8c57e58e5
3
  size 3150397656
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1826d529dae546578bde8e0e091cc664820bb9c9ab8d88a16c573f7faf51765
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8337dd673478657a1e3e59ab5c0126da6f87ecc51591bad61e39324efa7f5883
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd2954b3420c30ff450402b71de271b92c41ea14f65dc1ae2ca0b42b01f02c60
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10535ed970c1f8b1967fdb1bcf70b29e64c063da0c7c6d212af5b4ef07621922
3
  size 1064
last-checkpoint/special_tokens_map.json CHANGED
@@ -53,27 +53,9 @@
53
  "gl_ES",
54
  "sl_SI"
55
  ],
56
- "bos_token": {
57
- "content": "<s>",
58
- "lstrip": false,
59
- "normalized": false,
60
- "rstrip": false,
61
- "single_word": false
62
- },
63
- "cls_token": {
64
- "content": "<s>",
65
- "lstrip": false,
66
- "normalized": false,
67
- "rstrip": false,
68
- "single_word": false
69
- },
70
- "eos_token": {
71
- "content": "</s>",
72
- "lstrip": false,
73
- "normalized": false,
74
- "rstrip": false,
75
- "single_word": false
76
- },
77
  "mask_token": {
78
  "content": "<mask>",
79
  "lstrip": true,
@@ -81,25 +63,7 @@
81
  "rstrip": false,
82
  "single_word": false
83
  },
84
- "pad_token": {
85
- "content": "<pad>",
86
- "lstrip": false,
87
- "normalized": false,
88
- "rstrip": false,
89
- "single_word": false
90
- },
91
- "sep_token": {
92
- "content": "</s>",
93
- "lstrip": false,
94
- "normalized": false,
95
- "rstrip": false,
96
- "single_word": false
97
- },
98
- "unk_token": {
99
- "content": "<unk>",
100
- "lstrip": false,
101
- "normalized": false,
102
- "rstrip": false,
103
- "single_word": false
104
- }
105
  }
 
53
  "gl_ES",
54
  "sl_SI"
55
  ],
56
+ "bos_token": "<s>",
57
+ "cls_token": "<s>",
58
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
59
  "mask_token": {
60
  "content": "<mask>",
61
  "lstrip": true,
 
63
  "rstrip": false,
64
  "single_word": false
65
  },
66
+ "pad_token": "<pad>",
67
+ "sep_token": "</s>",
68
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
  }
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 1.6991,
3
- "best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-23214",
4
- "epoch": 5.999870784339062,
5
  "eval_steps": 500,
6
- "global_step": 23214,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -5865,1174 +5865,6 @@
5865
  "eval_samples_per_second": 1.019,
5866
  "eval_steps_per_second": 0.51,
5867
  "step": 19347
5868
- },
5869
- {
5870
- "epoch": 5.0038764698281435,
5871
- "learning_rate": 1.6894852845844678e-07,
5872
- "loss": 2.6522,
5873
- "step": 19360
5874
- },
5875
- {
5876
- "epoch": 5.009045096265667,
5877
- "learning_rate": 1.6891611564890445e-07,
5878
- "loss": 2.6008,
5879
- "step": 19380
5880
- },
5881
- {
5882
- "epoch": 5.014213722703191,
5883
- "learning_rate": 1.688837028393621e-07,
5884
- "loss": 2.6294,
5885
- "step": 19400
5886
- },
5887
- {
5888
- "epoch": 5.019382349140716,
5889
- "learning_rate": 1.6885129002981977e-07,
5890
- "loss": 2.6519,
5891
- "step": 19420
5892
- },
5893
- {
5894
- "epoch": 5.02455097557824,
5895
- "learning_rate": 1.6881887722027743e-07,
5896
- "loss": 2.6537,
5897
- "step": 19440
5898
- },
5899
- {
5900
- "epoch": 5.029719602015764,
5901
- "learning_rate": 1.6878646441073512e-07,
5902
- "loss": 2.6256,
5903
- "step": 19460
5904
- },
5905
- {
5906
- "epoch": 5.034888228453289,
5907
- "learning_rate": 1.6875405160119278e-07,
5908
- "loss": 2.6721,
5909
- "step": 19480
5910
- },
5911
- {
5912
- "epoch": 5.040056854890813,
5913
- "learning_rate": 1.6872163879165044e-07,
5914
- "loss": 2.6894,
5915
- "step": 19500
5916
- },
5917
- {
5918
- "epoch": 5.045225481328337,
5919
- "learning_rate": 1.6868922598210813e-07,
5920
- "loss": 2.6702,
5921
- "step": 19520
5922
- },
5923
- {
5924
- "epoch": 5.050394107765861,
5925
- "learning_rate": 1.686568131725658e-07,
5926
- "loss": 2.7041,
5927
- "step": 19540
5928
- },
5929
- {
5930
- "epoch": 5.055562734203385,
5931
- "learning_rate": 1.6862440036302346e-07,
5932
- "loss": 2.7243,
5933
- "step": 19560
5934
- },
5935
- {
5936
- "epoch": 5.06073136064091,
5937
- "learning_rate": 1.6859198755348114e-07,
5938
- "loss": 2.7082,
5939
- "step": 19580
5940
- },
5941
- {
5942
- "epoch": 5.065899987078434,
5943
- "learning_rate": 1.685595747439388e-07,
5944
- "loss": 2.6755,
5945
- "step": 19600
5946
- },
5947
- {
5948
- "epoch": 5.071068613515958,
5949
- "learning_rate": 1.6852716193439647e-07,
5950
- "loss": 2.6075,
5951
- "step": 19620
5952
- },
5953
- {
5954
- "epoch": 5.076237239953483,
5955
- "learning_rate": 1.6849474912485413e-07,
5956
- "loss": 2.6402,
5957
- "step": 19640
5958
- },
5959
- {
5960
- "epoch": 5.081405866391006,
5961
- "learning_rate": 1.684623363153118e-07,
5962
- "loss": 2.6928,
5963
- "step": 19660
5964
- },
5965
- {
5966
- "epoch": 5.086574492828531,
5967
- "learning_rate": 1.6842992350576948e-07,
5968
- "loss": 2.6689,
5969
- "step": 19680
5970
- },
5971
- {
5972
- "epoch": 5.091743119266055,
5973
- "learning_rate": 1.6839751069622714e-07,
5974
- "loss": 2.6848,
5975
- "step": 19700
5976
- },
5977
- {
5978
- "epoch": 5.096911745703579,
5979
- "learning_rate": 1.683650978866848e-07,
5980
- "loss": 2.6315,
5981
- "step": 19720
5982
- },
5983
- {
5984
- "epoch": 5.102080372141104,
5985
- "learning_rate": 1.683326850771425e-07,
5986
- "loss": 2.6936,
5987
- "step": 19740
5988
- },
5989
- {
5990
- "epoch": 5.107248998578628,
5991
- "learning_rate": 1.6830027226760015e-07,
5992
- "loss": 2.6354,
5993
- "step": 19760
5994
- },
5995
- {
5996
- "epoch": 5.112417625016152,
5997
- "learning_rate": 1.6826785945805782e-07,
5998
- "loss": 2.6376,
5999
- "step": 19780
6000
- },
6001
- {
6002
- "epoch": 5.117586251453676,
6003
- "learning_rate": 1.6823544664851548e-07,
6004
- "loss": 2.7595,
6005
- "step": 19800
6006
- },
6007
- {
6008
- "epoch": 5.1227548778912,
6009
- "learning_rate": 1.6820303383897317e-07,
6010
- "loss": 2.5688,
6011
- "step": 19820
6012
- },
6013
- {
6014
- "epoch": 5.127923504328725,
6015
- "learning_rate": 1.6817062102943083e-07,
6016
- "loss": 2.6504,
6017
- "step": 19840
6018
- },
6019
- {
6020
- "epoch": 5.133092130766249,
6021
- "learning_rate": 1.681382082198885e-07,
6022
- "loss": 2.701,
6023
- "step": 19860
6024
- },
6025
- {
6026
- "epoch": 5.138260757203773,
6027
- "learning_rate": 1.6810579541034615e-07,
6028
- "loss": 2.6691,
6029
- "step": 19880
6030
- },
6031
- {
6032
- "epoch": 5.143429383641298,
6033
- "learning_rate": 1.6807338260080384e-07,
6034
- "loss": 2.6396,
6035
- "step": 19900
6036
- },
6037
- {
6038
- "epoch": 5.148598010078821,
6039
- "learning_rate": 1.680409697912615e-07,
6040
- "loss": 2.6752,
6041
- "step": 19920
6042
- },
6043
- {
6044
- "epoch": 5.1537666365163455,
6045
- "learning_rate": 1.6800855698171916e-07,
6046
- "loss": 2.567,
6047
- "step": 19940
6048
- },
6049
- {
6050
- "epoch": 5.15893526295387,
6051
- "learning_rate": 1.6797614417217685e-07,
6052
- "loss": 2.6691,
6053
- "step": 19960
6054
- },
6055
- {
6056
- "epoch": 5.164103889391394,
6057
- "learning_rate": 1.6794373136263451e-07,
6058
- "loss": 2.6172,
6059
- "step": 19980
6060
- },
6061
- {
6062
- "epoch": 5.169272515828919,
6063
- "learning_rate": 1.6791131855309218e-07,
6064
- "loss": 2.6856,
6065
- "step": 20000
6066
- },
6067
- {
6068
- "epoch": 5.174441142266443,
6069
- "learning_rate": 1.6787890574354984e-07,
6070
- "loss": 2.6874,
6071
- "step": 20020
6072
- },
6073
- {
6074
- "epoch": 5.179609768703967,
6075
- "learning_rate": 1.678464929340075e-07,
6076
- "loss": 2.6737,
6077
- "step": 20040
6078
- },
6079
- {
6080
- "epoch": 5.184778395141491,
6081
- "learning_rate": 1.678140801244652e-07,
6082
- "loss": 2.6503,
6083
- "step": 20060
6084
- },
6085
- {
6086
- "epoch": 5.189947021579015,
6087
- "learning_rate": 1.6778166731492285e-07,
6088
- "loss": 2.6155,
6089
- "step": 20080
6090
- },
6091
- {
6092
- "epoch": 5.1951156480165395,
6093
- "learning_rate": 1.677492545053805e-07,
6094
- "loss": 2.7035,
6095
- "step": 20100
6096
- },
6097
- {
6098
- "epoch": 5.200284274454064,
6099
- "learning_rate": 1.677168416958382e-07,
6100
- "loss": 2.6192,
6101
- "step": 20120
6102
- },
6103
- {
6104
- "epoch": 5.205452900891588,
6105
- "learning_rate": 1.6768442888629586e-07,
6106
- "loss": 2.5974,
6107
- "step": 20140
6108
- },
6109
- {
6110
- "epoch": 5.2106215273291125,
6111
- "learning_rate": 1.6765201607675352e-07,
6112
- "loss": 2.694,
6113
- "step": 20160
6114
- },
6115
- {
6116
- "epoch": 5.215790153766637,
6117
- "learning_rate": 1.676196032672112e-07,
6118
- "loss": 2.6391,
6119
- "step": 20180
6120
- },
6121
- {
6122
- "epoch": 5.22095878020416,
6123
- "learning_rate": 1.6758719045766887e-07,
6124
- "loss": 2.6428,
6125
- "step": 20200
6126
- },
6127
- {
6128
- "epoch": 5.226127406641685,
6129
- "learning_rate": 1.6755477764812654e-07,
6130
- "loss": 2.6796,
6131
- "step": 20220
6132
- },
6133
- {
6134
- "epoch": 5.231296033079209,
6135
- "learning_rate": 1.675223648385842e-07,
6136
- "loss": 2.6819,
6137
- "step": 20240
6138
- },
6139
- {
6140
- "epoch": 5.2364646595167335,
6141
- "learning_rate": 1.6748995202904186e-07,
6142
- "loss": 2.6714,
6143
- "step": 20260
6144
- },
6145
- {
6146
- "epoch": 5.241633285954258,
6147
- "learning_rate": 1.6745753921949955e-07,
6148
- "loss": 2.6392,
6149
- "step": 20280
6150
- },
6151
- {
6152
- "epoch": 5.246801912391782,
6153
- "learning_rate": 1.674251264099572e-07,
6154
- "loss": 2.7057,
6155
- "step": 20300
6156
- },
6157
- {
6158
- "epoch": 5.2519705388293065,
6159
- "learning_rate": 1.6739271360041487e-07,
6160
- "loss": 2.6676,
6161
- "step": 20320
6162
- },
6163
- {
6164
- "epoch": 5.25713916526683,
6165
- "learning_rate": 1.6736030079087256e-07,
6166
- "loss": 2.7037,
6167
- "step": 20340
6168
- },
6169
- {
6170
- "epoch": 5.262307791704354,
6171
- "learning_rate": 1.6732788798133022e-07,
6172
- "loss": 2.7595,
6173
- "step": 20360
6174
- },
6175
- {
6176
- "epoch": 5.267476418141879,
6177
- "learning_rate": 1.6729547517178788e-07,
6178
- "loss": 2.6665,
6179
- "step": 20380
6180
- },
6181
- {
6182
- "epoch": 5.272645044579403,
6183
- "learning_rate": 1.6726306236224554e-07,
6184
- "loss": 2.6281,
6185
- "step": 20400
6186
- },
6187
- {
6188
- "epoch": 5.277813671016927,
6189
- "learning_rate": 1.6723064955270323e-07,
6190
- "loss": 2.6047,
6191
- "step": 20420
6192
- },
6193
- {
6194
- "epoch": 5.282982297454452,
6195
- "learning_rate": 1.671982367431609e-07,
6196
- "loss": 2.6466,
6197
- "step": 20440
6198
- },
6199
- {
6200
- "epoch": 5.288150923891976,
6201
- "learning_rate": 1.6716582393361856e-07,
6202
- "loss": 2.6026,
6203
- "step": 20460
6204
- },
6205
- {
6206
- "epoch": 5.2933195503295,
6207
- "learning_rate": 1.6713341112407622e-07,
6208
- "loss": 2.6714,
6209
- "step": 20480
6210
- },
6211
- {
6212
- "epoch": 5.298488176767024,
6213
- "learning_rate": 1.671009983145339e-07,
6214
- "loss": 2.6351,
6215
- "step": 20500
6216
- },
6217
- {
6218
- "epoch": 5.303656803204548,
6219
- "learning_rate": 1.6706858550499157e-07,
6220
- "loss": 2.6293,
6221
- "step": 20520
6222
- },
6223
- {
6224
- "epoch": 5.308825429642073,
6225
- "learning_rate": 1.6703617269544923e-07,
6226
- "loss": 2.6368,
6227
- "step": 20540
6228
- },
6229
- {
6230
- "epoch": 5.313994056079597,
6231
- "learning_rate": 1.6700375988590692e-07,
6232
- "loss": 2.6963,
6233
- "step": 20560
6234
- },
6235
- {
6236
- "epoch": 5.319162682517121,
6237
- "learning_rate": 1.6697134707636458e-07,
6238
- "loss": 2.6401,
6239
- "step": 20580
6240
- },
6241
- {
6242
- "epoch": 5.324331308954645,
6243
- "learning_rate": 1.6693893426682224e-07,
6244
- "loss": 2.669,
6245
- "step": 20600
6246
- },
6247
- {
6248
- "epoch": 5.329499935392169,
6249
- "learning_rate": 1.669065214572799e-07,
6250
- "loss": 2.6384,
6251
- "step": 20620
6252
- },
6253
- {
6254
- "epoch": 5.334668561829694,
6255
- "learning_rate": 1.6687410864773757e-07,
6256
- "loss": 2.7073,
6257
- "step": 20640
6258
- },
6259
- {
6260
- "epoch": 5.339837188267218,
6261
- "learning_rate": 1.6684169583819526e-07,
6262
- "loss": 2.6507,
6263
- "step": 20660
6264
- },
6265
- {
6266
- "epoch": 5.345005814704742,
6267
- "learning_rate": 1.6680928302865292e-07,
6268
- "loss": 2.6378,
6269
- "step": 20680
6270
- },
6271
- {
6272
- "epoch": 5.350174441142267,
6273
- "learning_rate": 1.6677687021911058e-07,
6274
- "loss": 2.6756,
6275
- "step": 20700
6276
- },
6277
- {
6278
- "epoch": 5.355343067579791,
6279
- "learning_rate": 1.6674445740956827e-07,
6280
- "loss": 2.6667,
6281
- "step": 20720
6282
- },
6283
- {
6284
- "epoch": 5.3605116940173145,
6285
- "learning_rate": 1.6671204460002593e-07,
6286
- "loss": 2.6714,
6287
- "step": 20740
6288
- },
6289
- {
6290
- "epoch": 5.365680320454839,
6291
- "learning_rate": 1.666796317904836e-07,
6292
- "loss": 2.6117,
6293
- "step": 20760
6294
- },
6295
- {
6296
- "epoch": 5.370848946892363,
6297
- "learning_rate": 1.6664721898094128e-07,
6298
- "loss": 2.6513,
6299
- "step": 20780
6300
- },
6301
- {
6302
- "epoch": 5.376017573329888,
6303
- "learning_rate": 1.6661480617139894e-07,
6304
- "loss": 2.6997,
6305
- "step": 20800
6306
- },
6307
- {
6308
- "epoch": 5.381186199767412,
6309
- "learning_rate": 1.665823933618566e-07,
6310
- "loss": 2.6395,
6311
- "step": 20820
6312
- },
6313
- {
6314
- "epoch": 5.386354826204936,
6315
- "learning_rate": 1.6654998055231426e-07,
6316
- "loss": 2.6615,
6317
- "step": 20840
6318
- },
6319
- {
6320
- "epoch": 5.39152345264246,
6321
- "learning_rate": 1.6651756774277193e-07,
6322
- "loss": 2.6915,
6323
- "step": 20860
6324
- },
6325
- {
6326
- "epoch": 5.396692079079984,
6327
- "learning_rate": 1.6648515493322961e-07,
6328
- "loss": 2.6248,
6329
- "step": 20880
6330
- },
6331
- {
6332
- "epoch": 5.4018607055175085,
6333
- "learning_rate": 1.6645274212368728e-07,
6334
- "loss": 2.7061,
6335
- "step": 20900
6336
- },
6337
- {
6338
- "epoch": 5.407029331955033,
6339
- "learning_rate": 1.6642032931414494e-07,
6340
- "loss": 2.642,
6341
- "step": 20920
6342
- },
6343
- {
6344
- "epoch": 5.412197958392557,
6345
- "learning_rate": 1.6638791650460263e-07,
6346
- "loss": 2.6377,
6347
- "step": 20940
6348
- },
6349
- {
6350
- "epoch": 5.417366584830082,
6351
- "learning_rate": 1.663555036950603e-07,
6352
- "loss": 2.6452,
6353
- "step": 20960
6354
- },
6355
- {
6356
- "epoch": 5.422535211267606,
6357
- "learning_rate": 1.6632309088551795e-07,
6358
- "loss": 2.7206,
6359
- "step": 20980
6360
- },
6361
- {
6362
- "epoch": 5.427703837705129,
6363
- "learning_rate": 1.662906780759756e-07,
6364
- "loss": 2.6524,
6365
- "step": 21000
6366
- },
6367
- {
6368
- "epoch": 5.432872464142654,
6369
- "learning_rate": 1.662582652664333e-07,
6370
- "loss": 2.6398,
6371
- "step": 21020
6372
- },
6373
- {
6374
- "epoch": 5.438041090580178,
6375
- "learning_rate": 1.6622585245689096e-07,
6376
- "loss": 2.6486,
6377
- "step": 21040
6378
- },
6379
- {
6380
- "epoch": 5.4432097170177025,
6381
- "learning_rate": 1.6619343964734862e-07,
6382
- "loss": 2.6466,
6383
- "step": 21060
6384
- },
6385
- {
6386
- "epoch": 5.448378343455227,
6387
- "learning_rate": 1.6616102683780629e-07,
6388
- "loss": 2.6818,
6389
- "step": 21080
6390
- },
6391
- {
6392
- "epoch": 5.453546969892751,
6393
- "learning_rate": 1.6612861402826397e-07,
6394
- "loss": 2.6826,
6395
- "step": 21100
6396
- },
6397
- {
6398
- "epoch": 5.458715596330276,
6399
- "learning_rate": 1.6609620121872164e-07,
6400
- "loss": 2.665,
6401
- "step": 21120
6402
- },
6403
- {
6404
- "epoch": 5.463884222767799,
6405
- "learning_rate": 1.660637884091793e-07,
6406
- "loss": 2.6018,
6407
- "step": 21140
6408
- },
6409
- {
6410
- "epoch": 5.469052849205323,
6411
- "learning_rate": 1.66031375599637e-07,
6412
- "loss": 2.6867,
6413
- "step": 21160
6414
- },
6415
- {
6416
- "epoch": 5.474221475642848,
6417
- "learning_rate": 1.6599896279009465e-07,
6418
- "loss": 2.6972,
6419
- "step": 21180
6420
- },
6421
- {
6422
- "epoch": 5.479390102080372,
6423
- "learning_rate": 1.659665499805523e-07,
6424
- "loss": 2.6775,
6425
- "step": 21200
6426
- },
6427
- {
6428
- "epoch": 5.4845587285178965,
6429
- "learning_rate": 1.6593413717100997e-07,
6430
- "loss": 2.7239,
6431
- "step": 21220
6432
- },
6433
- {
6434
- "epoch": 5.489727354955421,
6435
- "learning_rate": 1.6590172436146763e-07,
6436
- "loss": 2.6349,
6437
- "step": 21240
6438
- },
6439
- {
6440
- "epoch": 5.494895981392945,
6441
- "learning_rate": 1.6586931155192532e-07,
6442
- "loss": 2.6607,
6443
- "step": 21260
6444
- },
6445
- {
6446
- "epoch": 5.500064607830469,
6447
- "learning_rate": 1.6583689874238298e-07,
6448
- "loss": 2.6806,
6449
- "step": 21280
6450
- },
6451
- {
6452
- "epoch": 5.505233234267993,
6453
- "learning_rate": 1.6580448593284065e-07,
6454
- "loss": 2.7212,
6455
- "step": 21300
6456
- },
6457
- {
6458
- "epoch": 5.510401860705517,
6459
- "learning_rate": 1.6577207312329833e-07,
6460
- "loss": 2.6997,
6461
- "step": 21320
6462
- },
6463
- {
6464
- "epoch": 5.515570487143042,
6465
- "learning_rate": 1.65739660313756e-07,
6466
- "loss": 2.6277,
6467
- "step": 21340
6468
- },
6469
- {
6470
- "epoch": 5.520739113580566,
6471
- "learning_rate": 1.6570724750421366e-07,
6472
- "loss": 2.6953,
6473
- "step": 21360
6474
- },
6475
- {
6476
- "epoch": 5.5259077400180905,
6477
- "learning_rate": 1.6567483469467135e-07,
6478
- "loss": 2.7586,
6479
- "step": 21380
6480
- },
6481
- {
6482
- "epoch": 5.531076366455615,
6483
- "learning_rate": 1.65642421885129e-07,
6484
- "loss": 2.5887,
6485
- "step": 21400
6486
- },
6487
- {
6488
- "epoch": 5.536244992893138,
6489
- "learning_rate": 1.6561000907558667e-07,
6490
- "loss": 2.6364,
6491
- "step": 21420
6492
- },
6493
- {
6494
- "epoch": 5.541413619330663,
6495
- "learning_rate": 1.6557759626604433e-07,
6496
- "loss": 2.5951,
6497
- "step": 21440
6498
- },
6499
- {
6500
- "epoch": 5.546582245768187,
6501
- "learning_rate": 1.65545183456502e-07,
6502
- "loss": 2.6595,
6503
- "step": 21460
6504
- },
6505
- {
6506
- "epoch": 5.551750872205711,
6507
- "learning_rate": 1.6551277064695968e-07,
6508
- "loss": 2.602,
6509
- "step": 21480
6510
- },
6511
- {
6512
- "epoch": 5.556919498643236,
6513
- "learning_rate": 1.6548035783741734e-07,
6514
- "loss": 2.5996,
6515
- "step": 21500
6516
- },
6517
- {
6518
- "epoch": 5.56208812508076,
6519
- "learning_rate": 1.65447945027875e-07,
6520
- "loss": 2.6529,
6521
- "step": 21520
6522
- },
6523
- {
6524
- "epoch": 5.5672567515182845,
6525
- "learning_rate": 1.654155322183327e-07,
6526
- "loss": 2.6917,
6527
- "step": 21540
6528
- },
6529
- {
6530
- "epoch": 5.572425377955808,
6531
- "learning_rate": 1.6538311940879036e-07,
6532
- "loss": 2.6509,
6533
- "step": 21560
6534
- },
6535
- {
6536
- "epoch": 5.577594004393332,
6537
- "learning_rate": 1.6535070659924802e-07,
6538
- "loss": 2.6524,
6539
- "step": 21580
6540
- },
6541
- {
6542
- "epoch": 5.582762630830857,
6543
- "learning_rate": 1.6531829378970568e-07,
6544
- "loss": 2.6894,
6545
- "step": 21600
6546
- },
6547
- {
6548
- "epoch": 5.587931257268381,
6549
- "learning_rate": 1.6528588098016334e-07,
6550
- "loss": 2.6451,
6551
- "step": 21620
6552
- },
6553
- {
6554
- "epoch": 5.593099883705905,
6555
- "learning_rate": 1.6525346817062103e-07,
6556
- "loss": 2.6636,
6557
- "step": 21640
6558
- },
6559
- {
6560
- "epoch": 5.59826851014343,
6561
- "learning_rate": 1.652210553610787e-07,
6562
- "loss": 2.6696,
6563
- "step": 21660
6564
- },
6565
- {
6566
- "epoch": 5.603437136580954,
6567
- "learning_rate": 1.6518864255153635e-07,
6568
- "loss": 2.6577,
6569
- "step": 21680
6570
- },
6571
- {
6572
- "epoch": 5.608605763018478,
6573
- "learning_rate": 1.6515622974199404e-07,
6574
- "loss": 2.5968,
6575
- "step": 21700
6576
- },
6577
- {
6578
- "epoch": 5.613774389456002,
6579
- "learning_rate": 1.651238169324517e-07,
6580
- "loss": 2.6492,
6581
- "step": 21720
6582
- },
6583
- {
6584
- "epoch": 5.618943015893526,
6585
- "learning_rate": 1.6509140412290937e-07,
6586
- "loss": 2.6357,
6587
- "step": 21740
6588
- },
6589
- {
6590
- "epoch": 5.624111642331051,
6591
- "learning_rate": 1.6505899131336705e-07,
6592
- "loss": 2.6401,
6593
- "step": 21760
6594
- },
6595
- {
6596
- "epoch": 5.629280268768575,
6597
- "learning_rate": 1.650265785038247e-07,
6598
- "loss": 2.5984,
6599
- "step": 21780
6600
- },
6601
- {
6602
- "epoch": 5.6344488952060985,
6603
- "learning_rate": 1.6499416569428238e-07,
6604
- "loss": 2.6678,
6605
- "step": 21800
6606
- },
6607
- {
6608
- "epoch": 5.639617521643623,
6609
- "learning_rate": 1.6496175288474004e-07,
6610
- "loss": 2.6161,
6611
- "step": 21820
6612
- },
6613
- {
6614
- "epoch": 5.644786148081147,
6615
- "learning_rate": 1.649293400751977e-07,
6616
- "loss": 2.6262,
6617
- "step": 21840
6618
- },
6619
- {
6620
- "epoch": 5.649954774518672,
6621
- "learning_rate": 1.648969272656554e-07,
6622
- "loss": 2.6514,
6623
- "step": 21860
6624
- },
6625
- {
6626
- "epoch": 5.655123400956196,
6627
- "learning_rate": 1.6486451445611305e-07,
6628
- "loss": 2.6629,
6629
- "step": 21880
6630
- },
6631
- {
6632
- "epoch": 5.66029202739372,
6633
- "learning_rate": 1.6483210164657071e-07,
6634
- "loss": 2.6764,
6635
- "step": 21900
6636
- },
6637
- {
6638
- "epoch": 5.665460653831245,
6639
- "learning_rate": 1.647996888370284e-07,
6640
- "loss": 2.6414,
6641
- "step": 21920
6642
- },
6643
- {
6644
- "epoch": 5.670629280268768,
6645
- "learning_rate": 1.6476727602748604e-07,
6646
- "loss": 2.5379,
6647
- "step": 21940
6648
- },
6649
- {
6650
- "epoch": 5.6757979067062925,
6651
- "learning_rate": 1.6473486321794373e-07,
6652
- "loss": 2.6744,
6653
- "step": 21960
6654
- },
6655
- {
6656
- "epoch": 5.680966533143817,
6657
- "learning_rate": 1.647024504084014e-07,
6658
- "loss": 2.7254,
6659
- "step": 21980
6660
- },
6661
- {
6662
- "epoch": 5.686135159581341,
6663
- "learning_rate": 1.6467003759885905e-07,
6664
- "loss": 2.6408,
6665
- "step": 22000
6666
- },
6667
- {
6668
- "epoch": 5.6913037860188656,
6669
- "learning_rate": 1.6463762478931674e-07,
6670
- "loss": 2.6751,
6671
- "step": 22020
6672
- },
6673
- {
6674
- "epoch": 5.69647241245639,
6675
- "learning_rate": 1.646052119797744e-07,
6676
- "loss": 2.6391,
6677
- "step": 22040
6678
- },
6679
- {
6680
- "epoch": 5.701641038893914,
6681
- "learning_rate": 1.6457279917023206e-07,
6682
- "loss": 2.625,
6683
- "step": 22060
6684
- },
6685
- {
6686
- "epoch": 5.706809665331438,
6687
- "learning_rate": 1.6454038636068975e-07,
6688
- "loss": 2.607,
6689
- "step": 22080
6690
- },
6691
- {
6692
- "epoch": 5.711978291768962,
6693
- "learning_rate": 1.6450797355114739e-07,
6694
- "loss": 2.6629,
6695
- "step": 22100
6696
- },
6697
- {
6698
- "epoch": 5.7171469182064865,
6699
- "learning_rate": 1.6447556074160507e-07,
6700
- "loss": 2.6358,
6701
- "step": 22120
6702
- },
6703
- {
6704
- "epoch": 5.722315544644011,
6705
- "learning_rate": 1.6444314793206274e-07,
6706
- "loss": 2.6962,
6707
- "step": 22140
6708
- },
6709
- {
6710
- "epoch": 5.727484171081535,
6711
- "learning_rate": 1.644107351225204e-07,
6712
- "loss": 2.6403,
6713
- "step": 22160
6714
- },
6715
- {
6716
- "epoch": 5.7326527975190595,
6717
- "learning_rate": 1.6437832231297809e-07,
6718
- "loss": 2.641,
6719
- "step": 22180
6720
- },
6721
- {
6722
- "epoch": 5.737821423956584,
6723
- "learning_rate": 1.6434590950343575e-07,
6724
- "loss": 2.6213,
6725
- "step": 22200
6726
- },
6727
- {
6728
- "epoch": 5.742990050394107,
6729
- "learning_rate": 1.643134966938934e-07,
6730
- "loss": 2.6508,
6731
- "step": 22220
6732
- },
6733
- {
6734
- "epoch": 5.748158676831632,
6735
- "learning_rate": 1.642810838843511e-07,
6736
- "loss": 2.6528,
6737
- "step": 22240
6738
- },
6739
- {
6740
- "epoch": 5.753327303269156,
6741
- "learning_rate": 1.6424867107480873e-07,
6742
- "loss": 2.5921,
6743
- "step": 22260
6744
- },
6745
- {
6746
- "epoch": 5.7584959297066804,
6747
- "learning_rate": 1.6421625826526642e-07,
6748
- "loss": 2.6379,
6749
- "step": 22280
6750
- },
6751
- {
6752
- "epoch": 5.763664556144205,
6753
- "learning_rate": 1.641838454557241e-07,
6754
- "loss": 2.6838,
6755
- "step": 22300
6756
- },
6757
- {
6758
- "epoch": 5.768833182581729,
6759
- "learning_rate": 1.6415143264618175e-07,
6760
- "loss": 2.6511,
6761
- "step": 22320
6762
- },
6763
- {
6764
- "epoch": 5.7740018090192535,
6765
- "learning_rate": 1.6411901983663943e-07,
6766
- "loss": 2.6547,
6767
- "step": 22340
6768
- },
6769
- {
6770
- "epoch": 5.779170435456777,
6771
- "learning_rate": 1.640866070270971e-07,
6772
- "loss": 2.5962,
6773
- "step": 22360
6774
- },
6775
- {
6776
- "epoch": 5.784339061894301,
6777
- "learning_rate": 1.6405419421755476e-07,
6778
- "loss": 2.6074,
6779
- "step": 22380
6780
- },
6781
- {
6782
- "epoch": 5.789507688331826,
6783
- "learning_rate": 1.6402178140801245e-07,
6784
- "loss": 2.5927,
6785
- "step": 22400
6786
- },
6787
- {
6788
- "epoch": 5.79467631476935,
6789
- "learning_rate": 1.6398936859847008e-07,
6790
- "loss": 2.6697,
6791
- "step": 22420
6792
- },
6793
- {
6794
- "epoch": 5.799844941206874,
6795
- "learning_rate": 1.6395695578892777e-07,
6796
- "loss": 2.6267,
6797
- "step": 22440
6798
- },
6799
- {
6800
- "epoch": 5.805013567644399,
6801
- "learning_rate": 1.6392454297938546e-07,
6802
- "loss": 2.6833,
6803
- "step": 22460
6804
- },
6805
- {
6806
- "epoch": 5.810182194081923,
6807
- "learning_rate": 1.638921301698431e-07,
6808
- "loss": 2.6647,
6809
- "step": 22480
6810
- },
6811
- {
6812
- "epoch": 5.815350820519447,
6813
- "learning_rate": 1.6385971736030078e-07,
6814
- "loss": 2.6619,
6815
- "step": 22500
6816
- },
6817
- {
6818
- "epoch": 5.820519446956971,
6819
- "learning_rate": 1.6382730455075847e-07,
6820
- "loss": 2.6341,
6821
- "step": 22520
6822
- },
6823
- {
6824
- "epoch": 5.825688073394495,
6825
- "learning_rate": 1.637948917412161e-07,
6826
- "loss": 2.604,
6827
- "step": 22540
6828
- },
6829
- {
6830
- "epoch": 5.83085669983202,
6831
- "learning_rate": 1.637624789316738e-07,
6832
- "loss": 2.6344,
6833
- "step": 22560
6834
- },
6835
- {
6836
- "epoch": 5.836025326269544,
6837
- "learning_rate": 1.6373006612213146e-07,
6838
- "loss": 2.6453,
6839
- "step": 22580
6840
- },
6841
- {
6842
- "epoch": 5.841193952707068,
6843
- "learning_rate": 1.6369765331258912e-07,
6844
- "loss": 2.6433,
6845
- "step": 22600
6846
- },
6847
- {
6848
- "epoch": 5.846362579144593,
6849
- "learning_rate": 1.636652405030468e-07,
6850
- "loss": 2.5673,
6851
- "step": 22620
6852
- },
6853
- {
6854
- "epoch": 5.851531205582116,
6855
- "learning_rate": 1.6363282769350444e-07,
6856
- "loss": 2.6455,
6857
- "step": 22640
6858
- },
6859
- {
6860
- "epoch": 5.856699832019641,
6861
- "learning_rate": 1.6360041488396213e-07,
6862
- "loss": 2.6617,
6863
- "step": 22660
6864
- },
6865
- {
6866
- "epoch": 5.861868458457165,
6867
- "learning_rate": 1.6356800207441982e-07,
6868
- "loss": 2.6395,
6869
- "step": 22680
6870
- },
6871
- {
6872
- "epoch": 5.867037084894689,
6873
- "learning_rate": 1.6353558926487745e-07,
6874
- "loss": 2.6358,
6875
- "step": 22700
6876
- },
6877
- {
6878
- "epoch": 5.872205711332214,
6879
- "learning_rate": 1.6350317645533514e-07,
6880
- "loss": 2.618,
6881
- "step": 22720
6882
- },
6883
- {
6884
- "epoch": 5.877374337769738,
6885
- "learning_rate": 1.634707636457928e-07,
6886
- "loss": 2.6219,
6887
- "step": 22740
6888
- },
6889
- {
6890
- "epoch": 5.882542964207262,
6891
- "learning_rate": 1.6343835083625047e-07,
6892
- "loss": 2.6028,
6893
- "step": 22760
6894
- },
6895
- {
6896
- "epoch": 5.887711590644786,
6897
- "learning_rate": 1.6340593802670815e-07,
6898
- "loss": 2.599,
6899
- "step": 22780
6900
- },
6901
- {
6902
- "epoch": 5.89288021708231,
6903
- "learning_rate": 1.633735252171658e-07,
6904
- "loss": 2.6062,
6905
- "step": 22800
6906
- },
6907
- {
6908
- "epoch": 5.898048843519835,
6909
- "learning_rate": 1.6334111240762348e-07,
6910
- "loss": 2.5854,
6911
- "step": 22820
6912
- },
6913
- {
6914
- "epoch": 5.903217469957359,
6915
- "learning_rate": 1.6330869959808117e-07,
6916
- "loss": 2.5416,
6917
- "step": 22840
6918
- },
6919
- {
6920
- "epoch": 5.908386096394883,
6921
- "learning_rate": 1.632762867885388e-07,
6922
- "loss": 2.6731,
6923
- "step": 22860
6924
- },
6925
- {
6926
- "epoch": 5.913554722832407,
6927
- "learning_rate": 1.632438739789965e-07,
6928
- "loss": 2.6271,
6929
- "step": 22880
6930
- },
6931
- {
6932
- "epoch": 5.918723349269931,
6933
- "learning_rate": 1.6321146116945418e-07,
6934
- "loss": 2.5896,
6935
- "step": 22900
6936
- },
6937
- {
6938
- "epoch": 5.9238919757074555,
6939
- "learning_rate": 1.6317904835991181e-07,
6940
- "loss": 2.6794,
6941
- "step": 22920
6942
- },
6943
- {
6944
- "epoch": 5.92906060214498,
6945
- "learning_rate": 1.631466355503695e-07,
6946
- "loss": 2.6051,
6947
- "step": 22940
6948
- },
6949
- {
6950
- "epoch": 5.934229228582504,
6951
- "learning_rate": 1.6311422274082716e-07,
6952
- "loss": 2.6901,
6953
- "step": 22960
6954
- },
6955
- {
6956
- "epoch": 5.939397855020029,
6957
- "learning_rate": 1.6308180993128483e-07,
6958
- "loss": 2.6288,
6959
- "step": 22980
6960
- },
6961
- {
6962
- "epoch": 5.944566481457553,
6963
- "learning_rate": 1.6304939712174251e-07,
6964
- "loss": 2.6806,
6965
- "step": 23000
6966
- },
6967
- {
6968
- "epoch": 5.949735107895076,
6969
- "learning_rate": 1.6301698431220015e-07,
6970
- "loss": 2.6003,
6971
- "step": 23020
6972
- },
6973
- {
6974
- "epoch": 5.954903734332601,
6975
- "learning_rate": 1.6298457150265784e-07,
6976
- "loss": 2.6595,
6977
- "step": 23040
6978
- },
6979
- {
6980
- "epoch": 5.960072360770125,
6981
- "learning_rate": 1.6295215869311553e-07,
6982
- "loss": 2.6563,
6983
- "step": 23060
6984
- },
6985
- {
6986
- "epoch": 5.9652409872076495,
6987
- "learning_rate": 1.6291974588357316e-07,
6988
- "loss": 2.6499,
6989
- "step": 23080
6990
- },
6991
- {
6992
- "epoch": 5.970409613645174,
6993
- "learning_rate": 1.6288733307403085e-07,
6994
- "loss": 2.6836,
6995
- "step": 23100
6996
- },
6997
- {
6998
- "epoch": 5.975578240082698,
6999
- "learning_rate": 1.6285492026448854e-07,
7000
- "loss": 2.6193,
7001
- "step": 23120
7002
- },
7003
- {
7004
- "epoch": 5.980746866520223,
7005
- "learning_rate": 1.6282250745494617e-07,
7006
- "loss": 2.6293,
7007
- "step": 23140
7008
- },
7009
- {
7010
- "epoch": 5.985915492957746,
7011
- "learning_rate": 1.6279009464540386e-07,
7012
- "loss": 2.7275,
7013
- "step": 23160
7014
- },
7015
- {
7016
- "epoch": 5.99108411939527,
7017
- "learning_rate": 1.6275768183586152e-07,
7018
- "loss": 2.6021,
7019
- "step": 23180
7020
- },
7021
- {
7022
- "epoch": 5.996252745832795,
7023
- "learning_rate": 1.6272526902631919e-07,
7024
- "loss": 2.6331,
7025
- "step": 23200
7026
- },
7027
- {
7028
- "epoch": 5.999870784339062,
7029
- "eval_bleu": 1.6991,
7030
- "eval_gen_len": 66.9245,
7031
- "eval_loss": 2.596095085144043,
7032
- "eval_runtime": 1347.5897,
7033
- "eval_samples_per_second": 1.278,
7034
- "eval_steps_per_second": 0.639,
7035
- "step": 23214
7036
  }
7037
  ],
7038
  "logging_steps": 20,
@@ -7052,7 +5884,7 @@
7052
  "attributes": {}
7053
  }
7054
  },
7055
- "total_flos": 1.5241017744329933e+17,
7056
  "train_batch_size": 2,
7057
  "trial_name": null,
7058
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.5081,
3
+ "best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-19347",
4
+ "epoch": 4.999870784339062,
5
  "eval_steps": 500,
6
+ "global_step": 19347,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
5865
  "eval_samples_per_second": 1.019,
5866
  "eval_steps_per_second": 0.51,
5867
  "step": 19347
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5868
  }
5869
  ],
5870
  "logging_steps": 20,
 
5884
  "attributes": {}
5885
  }
5886
  },
5887
+ "total_flos": 1.2701974646813491e+17,
5888
  "train_batch_size": 2,
5889
  "trial_name": null,
5890
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d11a3b4036b5ce40442d47a4051217581bc97fdd9805ef76304488e63693998
3
- size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:581ac533a892239697e45b2f3ff1f250e5eee0be2c398f8f8ef23a485abea95b
3
+ size 5688
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f4f18bd8d1702c4a179d3d22bfeea1352e0e066d35c6cff7c5d20e32493e5e2
3
  size 1575259780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00790e6eb3aaf7be06f42ed5ebbbf19f51da90eabde4485d44e5afbbaa258042
3
  size 1575259780
run-2024-10-28T03:01:14+00:00.log CHANGED
The diff for this file is too large to render. See raw diff
 
val_outputs/val_generated_predictions_36735830.txt ADDED
The diff for this file is too large to render. See raw diff