Check commited on
Commit
71a1e60
Β·
1 Parent(s): 23775cb

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630048138.6581674/events.out.tfevents.1630048138.52f5c7e305a3.886.81 +3 -0
  11. model-bin/finetune/base/log/1630048619.9533648/events.out.tfevents.1630048619.52f5c7e305a3.886.83 +3 -0
  12. model-bin/finetune/base/log/1630049102.1531558/events.out.tfevents.1630049102.52f5c7e305a3.886.85 +3 -0
  13. model-bin/finetune/base/log/1630049572.6758537/events.out.tfevents.1630049572.52f5c7e305a3.886.87 +3 -0
  14. model-bin/finetune/base/log/1630050046.2035077/events.out.tfevents.1630050046.52f5c7e305a3.886.89 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630048138.52f5c7e305a3.886.80 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630048619.52f5c7e305a3.886.82 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630049102.52f5c7e305a3.886.84 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630049572.52f5c7e305a3.886.86 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630050046.52f5c7e305a3.886.88 +3 -0
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1b8197993eb37d9ee36acbb12a1f9605ec9f468a36950f0b19cfe128de2b5f0
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1573691ea7d0fc61236cc2759103b23e73e155f43b48959ab586bdb034de93db
3
  size 722165393
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8232bc1150a76590269afa478f75b4236d9220bb172b83e409418781d8efeebd
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ad107c141d3f7809507fa2921dcccdf0788b1defb6ed5e37e0ce2eaa7d50eec
3
  size 377909911
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a14f61f7dddaa53dbcb08f44b28b663a43e59c31dc4d069de7af3154f39d5215
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0940cfeead080e2f3ccd94ee15dab324a13e5461273c03d52367d56baa29b599
3
  size 14503
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f60fd12782841de6f909cf814170025073e913f168ae702ff0cdd93272700668
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd0c1baf2f5398bf54715f4be73eadcf661402972348e07ee31ecfc7ef64bd75
3
  size 559
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e9a559e52c3b4b08e9232cdf54f10d4fa2d1e564b84998252aef70d42eaf820
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcf1b5511c9d2f60343a826d533e62e2cd2de37fa4f207684f8625a938a917e1
3
  size 623
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 898.0,
5
- "global_step": 111379,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -241071,11 +241071,806 @@
241071
  "eval_steps_per_second": 0.642,
241072
  "eval_wer": 0.18447376051029285,
241073
  "step": 111379
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
241074
  }
241075
  ],
241076
  "max_steps": 620000,
241077
  "num_train_epochs": 5000,
241078
- "total_flos": 3.134467323546747e+20,
241079
  "trial_name": null,
241080
  "trial_params": null
241081
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 902.995983935743,
5
+ "global_step": 112002,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
241071
  "eval_steps_per_second": 0.642,
241072
  "eval_wer": 0.18447376051029285,
241073
  "step": 111379
241074
+ },
241075
+ {
241076
+ "epoch": 898.01,
241077
+ "learning_rate": 8.217512116316641e-06,
241078
+ "loss": 0.3168,
241079
+ "step": 111380
241080
+ },
241081
+ {
241082
+ "epoch": 898.05,
241083
+ "learning_rate": 8.217431340872375e-06,
241084
+ "loss": 0.2956,
241085
+ "step": 111385
241086
+ },
241087
+ {
241088
+ "epoch": 898.09,
241089
+ "learning_rate": 8.217350565428111e-06,
241090
+ "loss": 0.2882,
241091
+ "step": 111390
241092
+ },
241093
+ {
241094
+ "epoch": 898.13,
241095
+ "learning_rate": 8.217269789983845e-06,
241096
+ "loss": 0.2941,
241097
+ "step": 111395
241098
+ },
241099
+ {
241100
+ "epoch": 898.17,
241101
+ "learning_rate": 8.217189014539581e-06,
241102
+ "loss": 0.5359,
241103
+ "step": 111400
241104
+ },
241105
+ {
241106
+ "epoch": 898.21,
241107
+ "learning_rate": 8.217108239095315e-06,
241108
+ "loss": 1.293,
241109
+ "step": 111405
241110
+ },
241111
+ {
241112
+ "epoch": 898.25,
241113
+ "learning_rate": 8.217027463651051e-06,
241114
+ "loss": 0.2586,
241115
+ "step": 111410
241116
+ },
241117
+ {
241118
+ "epoch": 898.29,
241119
+ "learning_rate": 8.216946688206785e-06,
241120
+ "loss": 0.2983,
241121
+ "step": 111415
241122
+ },
241123
+ {
241124
+ "epoch": 898.33,
241125
+ "learning_rate": 8.216865912762521e-06,
241126
+ "loss": 0.3252,
241127
+ "step": 111420
241128
+ },
241129
+ {
241130
+ "epoch": 898.37,
241131
+ "learning_rate": 8.216785137318255e-06,
241132
+ "loss": 0.4947,
241133
+ "step": 111425
241134
+ },
241135
+ {
241136
+ "epoch": 898.41,
241137
+ "learning_rate": 8.216704361873991e-06,
241138
+ "loss": 1.025,
241139
+ "step": 111430
241140
+ },
241141
+ {
241142
+ "epoch": 898.45,
241143
+ "learning_rate": 8.216623586429727e-06,
241144
+ "loss": 0.3136,
241145
+ "step": 111435
241146
+ },
241147
+ {
241148
+ "epoch": 898.49,
241149
+ "learning_rate": 8.216542810985461e-06,
241150
+ "loss": 0.2458,
241151
+ "step": 111440
241152
+ },
241153
+ {
241154
+ "epoch": 898.53,
241155
+ "learning_rate": 8.216462035541197e-06,
241156
+ "loss": 0.3212,
241157
+ "step": 111445
241158
+ },
241159
+ {
241160
+ "epoch": 898.57,
241161
+ "learning_rate": 8.216381260096931e-06,
241162
+ "loss": 0.5968,
241163
+ "step": 111450
241164
+ },
241165
+ {
241166
+ "epoch": 898.61,
241167
+ "learning_rate": 8.216300484652667e-06,
241168
+ "loss": 1.074,
241169
+ "step": 111455
241170
+ },
241171
+ {
241172
+ "epoch": 898.65,
241173
+ "learning_rate": 8.216219709208401e-06,
241174
+ "loss": 0.2987,
241175
+ "step": 111460
241176
+ },
241177
+ {
241178
+ "epoch": 898.69,
241179
+ "learning_rate": 8.216138933764137e-06,
241180
+ "loss": 0.3227,
241181
+ "step": 111465
241182
+ },
241183
+ {
241184
+ "epoch": 898.73,
241185
+ "learning_rate": 8.216058158319871e-06,
241186
+ "loss": 0.5044,
241187
+ "step": 111470
241188
+ },
241189
+ {
241190
+ "epoch": 898.77,
241191
+ "learning_rate": 8.215977382875607e-06,
241192
+ "loss": 0.5049,
241193
+ "step": 111475
241194
+ },
241195
+ {
241196
+ "epoch": 898.81,
241197
+ "learning_rate": 8.215896607431341e-06,
241198
+ "loss": 1.1103,
241199
+ "step": 111480
241200
+ },
241201
+ {
241202
+ "epoch": 898.85,
241203
+ "learning_rate": 8.215815831987077e-06,
241204
+ "loss": 0.292,
241205
+ "step": 111485
241206
+ },
241207
+ {
241208
+ "epoch": 898.89,
241209
+ "learning_rate": 8.215735056542811e-06,
241210
+ "loss": 0.2973,
241211
+ "step": 111490
241212
+ },
241213
+ {
241214
+ "epoch": 898.93,
241215
+ "learning_rate": 8.215654281098547e-06,
241216
+ "loss": 0.3051,
241217
+ "step": 111495
241218
+ },
241219
+ {
241220
+ "epoch": 898.97,
241221
+ "learning_rate": 8.215573505654283e-06,
241222
+ "loss": 0.5463,
241223
+ "step": 111500
241224
+ },
241225
+ {
241226
+ "epoch": 899.0,
241227
+ "eval_loss": 0.45212462544441223,
241228
+ "eval_runtime": 41.8965,
241229
+ "eval_samples_per_second": 19.954,
241230
+ "eval_steps_per_second": 0.644,
241231
+ "eval_wer": 0.18223921193854298,
241232
+ "step": 111503
241233
+ },
241234
+ {
241235
+ "epoch": 892.02,
241236
+ "learning_rate": 8.215492730210017e-06,
241237
+ "loss": 0.5205,
241238
+ "step": 111505
241239
+ },
241240
+ {
241241
+ "epoch": 892.06,
241242
+ "learning_rate": 8.215411954765753e-06,
241243
+ "loss": 0.2519,
241244
+ "step": 111510
241245
+ },
241246
+ {
241247
+ "epoch": 892.1,
241248
+ "learning_rate": 8.215331179321487e-06,
241249
+ "loss": 0.2812,
241250
+ "step": 111515
241251
+ },
241252
+ {
241253
+ "epoch": 892.14,
241254
+ "learning_rate": 8.215250403877223e-06,
241255
+ "loss": 0.3625,
241256
+ "step": 111520
241257
+ },
241258
+ {
241259
+ "epoch": 892.18,
241260
+ "learning_rate": 8.215169628432957e-06,
241261
+ "loss": 0.6445,
241262
+ "step": 111525
241263
+ },
241264
+ {
241265
+ "epoch": 892.22,
241266
+ "learning_rate": 8.215088852988693e-06,
241267
+ "loss": 0.9554,
241268
+ "step": 111530
241269
+ },
241270
+ {
241271
+ "epoch": 892.25,
241272
+ "learning_rate": 8.215008077544427e-06,
241273
+ "loss": 0.2955,
241274
+ "step": 111535
241275
+ },
241276
+ {
241277
+ "epoch": 892.29,
241278
+ "learning_rate": 8.214927302100163e-06,
241279
+ "loss": 0.2583,
241280
+ "step": 111540
241281
+ },
241282
+ {
241283
+ "epoch": 892.33,
241284
+ "learning_rate": 8.214846526655897e-06,
241285
+ "loss": 0.3004,
241286
+ "step": 111545
241287
+ },
241288
+ {
241289
+ "epoch": 892.37,
241290
+ "learning_rate": 8.214765751211633e-06,
241291
+ "loss": 0.6362,
241292
+ "step": 111550
241293
+ },
241294
+ {
241295
+ "epoch": 892.41,
241296
+ "learning_rate": 8.214684975767367e-06,
241297
+ "loss": 1.0824,
241298
+ "step": 111555
241299
+ },
241300
+ {
241301
+ "epoch": 892.45,
241302
+ "learning_rate": 8.214604200323102e-06,
241303
+ "loss": 0.2687,
241304
+ "step": 111560
241305
+ },
241306
+ {
241307
+ "epoch": 892.49,
241308
+ "learning_rate": 8.214523424878837e-06,
241309
+ "loss": 0.3912,
241310
+ "step": 111565
241311
+ },
241312
+ {
241313
+ "epoch": 892.53,
241314
+ "learning_rate": 8.214442649434572e-06,
241315
+ "loss": 0.4064,
241316
+ "step": 111570
241317
+ },
241318
+ {
241319
+ "epoch": 892.57,
241320
+ "learning_rate": 8.214361873990308e-06,
241321
+ "loss": 0.6121,
241322
+ "step": 111575
241323
+ },
241324
+ {
241325
+ "epoch": 892.61,
241326
+ "learning_rate": 8.214281098546042e-06,
241327
+ "loss": 1.0386,
241328
+ "step": 111580
241329
+ },
241330
+ {
241331
+ "epoch": 892.65,
241332
+ "learning_rate": 8.214200323101778e-06,
241333
+ "loss": 0.2815,
241334
+ "step": 111585
241335
+ },
241336
+ {
241337
+ "epoch": 892.69,
241338
+ "learning_rate": 8.214119547657512e-06,
241339
+ "loss": 0.3376,
241340
+ "step": 111590
241341
+ },
241342
+ {
241343
+ "epoch": 892.73,
241344
+ "learning_rate": 8.214038772213248e-06,
241345
+ "loss": 0.3432,
241346
+ "step": 111595
241347
+ },
241348
+ {
241349
+ "epoch": 892.77,
241350
+ "learning_rate": 8.213957996768982e-06,
241351
+ "loss": 0.6253,
241352
+ "step": 111600
241353
+ },
241354
+ {
241355
+ "epoch": 892.81,
241356
+ "learning_rate": 8.213877221324718e-06,
241357
+ "loss": 0.8157,
241358
+ "step": 111605
241359
+ },
241360
+ {
241361
+ "epoch": 892.85,
241362
+ "learning_rate": 8.213796445880452e-06,
241363
+ "loss": 0.282,
241364
+ "step": 111610
241365
+ },
241366
+ {
241367
+ "epoch": 892.89,
241368
+ "learning_rate": 8.213715670436188e-06,
241369
+ "loss": 0.2759,
241370
+ "step": 111615
241371
+ },
241372
+ {
241373
+ "epoch": 892.93,
241374
+ "learning_rate": 8.213634894991922e-06,
241375
+ "loss": 0.3045,
241376
+ "step": 111620
241377
+ },
241378
+ {
241379
+ "epoch": 892.97,
241380
+ "learning_rate": 8.213554119547658e-06,
241381
+ "loss": 0.5921,
241382
+ "step": 111625
241383
+ },
241384
+ {
241385
+ "epoch": 893.0,
241386
+ "eval_loss": 0.4683041274547577,
241387
+ "eval_runtime": 42.5438,
241388
+ "eval_samples_per_second": 19.65,
241389
+ "eval_steps_per_second": 0.635,
241390
+ "eval_wer": 0.18959383338274532,
241391
+ "step": 111628
241392
+ },
241393
+ {
241394
+ "epoch": 893.02,
241395
+ "learning_rate": 8.213473344103392e-06,
241396
+ "loss": 0.3051,
241397
+ "step": 111630
241398
+ },
241399
+ {
241400
+ "epoch": 893.06,
241401
+ "learning_rate": 8.213392568659128e-06,
241402
+ "loss": 0.2988,
241403
+ "step": 111635
241404
+ },
241405
+ {
241406
+ "epoch": 893.1,
241407
+ "learning_rate": 8.213311793214864e-06,
241408
+ "loss": 0.3022,
241409
+ "step": 111640
241410
+ },
241411
+ {
241412
+ "epoch": 893.14,
241413
+ "learning_rate": 8.213231017770598e-06,
241414
+ "loss": 0.3361,
241415
+ "step": 111645
241416
+ },
241417
+ {
241418
+ "epoch": 893.18,
241419
+ "learning_rate": 8.213150242326334e-06,
241420
+ "loss": 0.6711,
241421
+ "step": 111650
241422
+ },
241423
+ {
241424
+ "epoch": 893.22,
241425
+ "learning_rate": 8.213069466882068e-06,
241426
+ "loss": 1.1263,
241427
+ "step": 111655
241428
+ },
241429
+ {
241430
+ "epoch": 893.26,
241431
+ "learning_rate": 8.212988691437804e-06,
241432
+ "loss": 0.3122,
241433
+ "step": 111660
241434
+ },
241435
+ {
241436
+ "epoch": 893.3,
241437
+ "learning_rate": 8.212907915993538e-06,
241438
+ "loss": 0.2739,
241439
+ "step": 111665
241440
+ },
241441
+ {
241442
+ "epoch": 893.34,
241443
+ "learning_rate": 8.212827140549274e-06,
241444
+ "loss": 0.3829,
241445
+ "step": 111670
241446
+ },
241447
+ {
241448
+ "epoch": 893.38,
241449
+ "learning_rate": 8.212746365105008e-06,
241450
+ "loss": 0.6704,
241451
+ "step": 111675
241452
+ },
241453
+ {
241454
+ "epoch": 893.42,
241455
+ "learning_rate": 8.212665589660744e-06,
241456
+ "loss": 1.0724,
241457
+ "step": 111680
241458
+ },
241459
+ {
241460
+ "epoch": 893.46,
241461
+ "learning_rate": 8.212584814216478e-06,
241462
+ "loss": 0.2658,
241463
+ "step": 111685
241464
+ },
241465
+ {
241466
+ "epoch": 893.5,
241467
+ "learning_rate": 8.212504038772214e-06,
241468
+ "loss": 0.3109,
241469
+ "step": 111690
241470
+ },
241471
+ {
241472
+ "epoch": 893.54,
241473
+ "learning_rate": 8.212423263327948e-06,
241474
+ "loss": 0.3202,
241475
+ "step": 111695
241476
+ },
241477
+ {
241478
+ "epoch": 893.58,
241479
+ "learning_rate": 8.212342487883684e-06,
241480
+ "loss": 0.6632,
241481
+ "step": 111700
241482
+ },
241483
+ {
241484
+ "epoch": 893.62,
241485
+ "learning_rate": 8.21226171243942e-06,
241486
+ "loss": 1.0617,
241487
+ "step": 111705
241488
+ },
241489
+ {
241490
+ "epoch": 893.66,
241491
+ "learning_rate": 8.212180936995154e-06,
241492
+ "loss": 0.2693,
241493
+ "step": 111710
241494
+ },
241495
+ {
241496
+ "epoch": 893.7,
241497
+ "learning_rate": 8.21210016155089e-06,
241498
+ "loss": 0.32,
241499
+ "step": 111715
241500
+ },
241501
+ {
241502
+ "epoch": 893.74,
241503
+ "learning_rate": 8.212019386106624e-06,
241504
+ "loss": 0.4235,
241505
+ "step": 111720
241506
+ },
241507
+ {
241508
+ "epoch": 893.78,
241509
+ "learning_rate": 8.21193861066236e-06,
241510
+ "loss": 0.6357,
241511
+ "step": 111725
241512
+ },
241513
+ {
241514
+ "epoch": 893.82,
241515
+ "learning_rate": 8.211857835218094e-06,
241516
+ "loss": 0.9449,
241517
+ "step": 111730
241518
+ },
241519
+ {
241520
+ "epoch": 893.86,
241521
+ "learning_rate": 8.21177705977383e-06,
241522
+ "loss": 0.2906,
241523
+ "step": 111735
241524
+ },
241525
+ {
241526
+ "epoch": 893.9,
241527
+ "learning_rate": 8.211696284329564e-06,
241528
+ "loss": 0.298,
241529
+ "step": 111740
241530
+ },
241531
+ {
241532
+ "epoch": 893.94,
241533
+ "learning_rate": 8.2116155088853e-06,
241534
+ "loss": 0.3205,
241535
+ "step": 111745
241536
+ },
241537
+ {
241538
+ "epoch": 893.98,
241539
+ "learning_rate": 8.211534733441034e-06,
241540
+ "loss": 0.602,
241541
+ "step": 111750
241542
+ },
241543
+ {
241544
+ "epoch": 894.0,
241545
+ "eval_loss": 0.3670633137226105,
241546
+ "eval_runtime": 42.7136,
241547
+ "eval_samples_per_second": 19.572,
241548
+ "eval_steps_per_second": 0.632,
241549
+ "eval_wer": 0.18261776982707215,
241550
+ "step": 111753
241551
+ },
241552
+ {
241553
+ "epoch": 894.02,
241554
+ "learning_rate": 8.21145395799677e-06,
241555
+ "loss": 0.3963,
241556
+ "step": 111755
241557
+ },
241558
+ {
241559
+ "epoch": 894.06,
241560
+ "learning_rate": 8.211373182552504e-06,
241561
+ "loss": 0.3368,
241562
+ "step": 111760
241563
+ },
241564
+ {
241565
+ "epoch": 894.1,
241566
+ "learning_rate": 8.21129240710824e-06,
241567
+ "loss": 0.2995,
241568
+ "step": 111765
241569
+ },
241570
+ {
241571
+ "epoch": 894.14,
241572
+ "learning_rate": 8.211211631663974e-06,
241573
+ "loss": 0.3304,
241574
+ "step": 111770
241575
+ },
241576
+ {
241577
+ "epoch": 894.18,
241578
+ "learning_rate": 8.21113085621971e-06,
241579
+ "loss": 0.5743,
241580
+ "step": 111775
241581
+ },
241582
+ {
241583
+ "epoch": 894.22,
241584
+ "learning_rate": 8.211050080775445e-06,
241585
+ "loss": 0.9688,
241586
+ "step": 111780
241587
+ },
241588
+ {
241589
+ "epoch": 894.26,
241590
+ "learning_rate": 8.21096930533118e-06,
241591
+ "loss": 0.2755,
241592
+ "step": 111785
241593
+ },
241594
+ {
241595
+ "epoch": 894.3,
241596
+ "learning_rate": 8.210888529886915e-06,
241597
+ "loss": 0.3825,
241598
+ "step": 111790
241599
+ },
241600
+ {
241601
+ "epoch": 894.34,
241602
+ "learning_rate": 8.21080775444265e-06,
241603
+ "loss": 0.3727,
241604
+ "step": 111795
241605
+ },
241606
+ {
241607
+ "epoch": 894.38,
241608
+ "learning_rate": 8.210726978998385e-06,
241609
+ "loss": 0.6055,
241610
+ "step": 111800
241611
+ },
241612
+ {
241613
+ "epoch": 894.42,
241614
+ "learning_rate": 8.21064620355412e-06,
241615
+ "loss": 0.949,
241616
+ "step": 111805
241617
+ },
241618
+ {
241619
+ "epoch": 894.46,
241620
+ "learning_rate": 8.210565428109855e-06,
241621
+ "loss": 0.3053,
241622
+ "step": 111810
241623
+ },
241624
+ {
241625
+ "epoch": 894.5,
241626
+ "learning_rate": 8.21048465266559e-06,
241627
+ "loss": 0.4324,
241628
+ "step": 111815
241629
+ },
241630
+ {
241631
+ "epoch": 894.54,
241632
+ "learning_rate": 8.210403877221325e-06,
241633
+ "loss": 0.4127,
241634
+ "step": 111820
241635
+ },
241636
+ {
241637
+ "epoch": 894.58,
241638
+ "learning_rate": 8.21032310177706e-06,
241639
+ "loss": 0.5997,
241640
+ "step": 111825
241641
+ },
241642
+ {
241643
+ "epoch": 894.62,
241644
+ "learning_rate": 8.210242326332795e-06,
241645
+ "loss": 1.0255,
241646
+ "step": 111830
241647
+ },
241648
+ {
241649
+ "epoch": 894.66,
241650
+ "learning_rate": 8.21016155088853e-06,
241651
+ "loss": 0.2515,
241652
+ "step": 111835
241653
+ },
241654
+ {
241655
+ "epoch": 894.7,
241656
+ "learning_rate": 8.210080775444265e-06,
241657
+ "loss": 0.3091,
241658
+ "step": 111840
241659
+ },
241660
+ {
241661
+ "epoch": 894.74,
241662
+ "learning_rate": 8.210000000000001e-06,
241663
+ "loss": 0.3108,
241664
+ "step": 111845
241665
+ },
241666
+ {
241667
+ "epoch": 894.78,
241668
+ "learning_rate": 8.209919224555735e-06,
241669
+ "loss": 0.6145,
241670
+ "step": 111850
241671
+ },
241672
+ {
241673
+ "epoch": 894.82,
241674
+ "learning_rate": 8.209838449111471e-06,
241675
+ "loss": 1.1353,
241676
+ "step": 111855
241677
+ },
241678
+ {
241679
+ "epoch": 894.86,
241680
+ "learning_rate": 8.209757673667205e-06,
241681
+ "loss": 0.3686,
241682
+ "step": 111860
241683
+ },
241684
+ {
241685
+ "epoch": 894.9,
241686
+ "learning_rate": 8.209676898222941e-06,
241687
+ "loss": 0.281,
241688
+ "step": 111865
241689
+ },
241690
+ {
241691
+ "epoch": 894.94,
241692
+ "learning_rate": 8.209596122778675e-06,
241693
+ "loss": 0.4096,
241694
+ "step": 111870
241695
+ },
241696
+ {
241697
+ "epoch": 894.98,
241698
+ "learning_rate": 8.209515347334411e-06,
241699
+ "loss": 0.6653,
241700
+ "step": 111875
241701
+ },
241702
+ {
241703
+ "epoch": 895.0,
241704
+ "eval_loss": 0.3755977153778076,
241705
+ "eval_runtime": 42.4892,
241706
+ "eval_samples_per_second": 19.676,
241707
+ "eval_steps_per_second": 0.635,
241708
+ "eval_wer": 0.18443218123729305,
241709
+ "step": 111878
241710
+ },
241711
+ {
241712
+ "epoch": 902.02,
241713
+ "learning_rate": 8.209434571890145e-06,
241714
+ "loss": 0.3006,
241715
+ "step": 111880
241716
+ },
241717
+ {
241718
+ "epoch": 902.06,
241719
+ "learning_rate": 8.209353796445881e-06,
241720
+ "loss": 0.2412,
241721
+ "step": 111885
241722
+ },
241723
+ {
241724
+ "epoch": 902.1,
241725
+ "learning_rate": 8.209273021001615e-06,
241726
+ "loss": 0.3399,
241727
+ "step": 111890
241728
+ },
241729
+ {
241730
+ "epoch": 902.14,
241731
+ "learning_rate": 8.209192245557351e-06,
241732
+ "loss": 0.3268,
241733
+ "step": 111895
241734
+ },
241735
+ {
241736
+ "epoch": 902.18,
241737
+ "learning_rate": 8.209111470113085e-06,
241738
+ "loss": 0.6156,
241739
+ "step": 111900
241740
+ },
241741
+ {
241742
+ "epoch": 902.22,
241743
+ "learning_rate": 8.209030694668821e-06,
241744
+ "loss": 1.1313,
241745
+ "step": 111905
241746
+ },
241747
+ {
241748
+ "epoch": 902.26,
241749
+ "learning_rate": 8.208949919224557e-06,
241750
+ "loss": 0.3566,
241751
+ "step": 111910
241752
+ },
241753
+ {
241754
+ "epoch": 902.3,
241755
+ "learning_rate": 8.20886914378029e-06,
241756
+ "loss": 0.3318,
241757
+ "step": 111915
241758
+ },
241759
+ {
241760
+ "epoch": 902.34,
241761
+ "learning_rate": 8.208788368336027e-06,
241762
+ "loss": 0.3042,
241763
+ "step": 111920
241764
+ },
241765
+ {
241766
+ "epoch": 902.38,
241767
+ "learning_rate": 8.20870759289176e-06,
241768
+ "loss": 0.5518,
241769
+ "step": 111925
241770
+ },
241771
+ {
241772
+ "epoch": 902.42,
241773
+ "learning_rate": 8.208626817447497e-06,
241774
+ "loss": 0.9917,
241775
+ "step": 111930
241776
+ },
241777
+ {
241778
+ "epoch": 902.46,
241779
+ "learning_rate": 8.20854604200323e-06,
241780
+ "loss": 0.3356,
241781
+ "step": 111935
241782
+ },
241783
+ {
241784
+ "epoch": 902.5,
241785
+ "learning_rate": 8.208465266558967e-06,
241786
+ "loss": 0.286,
241787
+ "step": 111940
241788
+ },
241789
+ {
241790
+ "epoch": 902.54,
241791
+ "learning_rate": 8.2083844911147e-06,
241792
+ "loss": 0.3935,
241793
+ "step": 111945
241794
+ },
241795
+ {
241796
+ "epoch": 902.58,
241797
+ "learning_rate": 8.208303715670437e-06,
241798
+ "loss": 0.6371,
241799
+ "step": 111950
241800
+ },
241801
+ {
241802
+ "epoch": 902.62,
241803
+ "learning_rate": 8.20822294022617e-06,
241804
+ "loss": 0.9316,
241805
+ "step": 111955
241806
+ },
241807
+ {
241808
+ "epoch": 902.66,
241809
+ "learning_rate": 8.208142164781907e-06,
241810
+ "loss": 0.3458,
241811
+ "step": 111960
241812
+ },
241813
+ {
241814
+ "epoch": 902.7,
241815
+ "learning_rate": 8.20806138933764e-06,
241816
+ "loss": 0.3009,
241817
+ "step": 111965
241818
+ },
241819
+ {
241820
+ "epoch": 902.74,
241821
+ "learning_rate": 8.207980613893377e-06,
241822
+ "loss": 0.3419,
241823
+ "step": 111970
241824
+ },
241825
+ {
241826
+ "epoch": 902.78,
241827
+ "learning_rate": 8.207899838449112e-06,
241828
+ "loss": 0.5863,
241829
+ "step": 111975
241830
+ },
241831
+ {
241832
+ "epoch": 902.82,
241833
+ "learning_rate": 8.207819063004847e-06,
241834
+ "loss": 1.0017,
241835
+ "step": 111980
241836
+ },
241837
+ {
241838
+ "epoch": 902.86,
241839
+ "learning_rate": 8.207738287560582e-06,
241840
+ "loss": 0.2961,
241841
+ "step": 111985
241842
+ },
241843
+ {
241844
+ "epoch": 902.9,
241845
+ "learning_rate": 8.207657512116317e-06,
241846
+ "loss": 0.2509,
241847
+ "step": 111990
241848
+ },
241849
+ {
241850
+ "epoch": 902.94,
241851
+ "learning_rate": 8.207576736672052e-06,
241852
+ "loss": 0.33,
241853
+ "step": 111995
241854
+ },
241855
+ {
241856
+ "epoch": 902.98,
241857
+ "learning_rate": 8.207495961227786e-06,
241858
+ "loss": 0.6344,
241859
+ "step": 112000
241860
+ },
241861
+ {
241862
+ "epoch": 903.0,
241863
+ "eval_loss": 0.4465380311012268,
241864
+ "eval_runtime": 43.2217,
241865
+ "eval_samples_per_second": 19.342,
241866
+ "eval_steps_per_second": 0.625,
241867
+ "eval_wer": 0.1886928860613071,
241868
+ "step": 112002
241869
  }
241870
  ],
241871
  "max_steps": 620000,
241872
  "num_train_epochs": 5000,
241873
+ "total_flos": 3.1520264569981986e+20,
241874
  "trial_name": null,
241875
  "trial_params": null
241876
  }
model-bin/finetune/base/{checkpoint-111379 β†’ checkpoint-112002}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630048138.6581674/events.out.tfevents.1630048138.52f5c7e305a3.886.81 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fd565bfc4df649dc0f4e5aadcf1eead11a12fb058bf7d9f5390ea0410bd88c0
3
+ size 4194
model-bin/finetune/base/log/1630048619.9533648/events.out.tfevents.1630048619.52f5c7e305a3.886.83 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd30dccc3acb9b02a74544425cb00b04f97859f9d1abf3d1eb655dec856c326c
3
+ size 4194
model-bin/finetune/base/log/1630049102.1531558/events.out.tfevents.1630049102.52f5c7e305a3.886.85 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab07217fdffdd042e7a95682d29c5c31f2248936ef812bffa701cd9f704a18f3
3
+ size 4194
model-bin/finetune/base/log/1630049572.6758537/events.out.tfevents.1630049572.52f5c7e305a3.886.87 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d3d43cf4345ea0f605d5516e05ca68007bcf57f75ba4819cc09f69f91395822
3
+ size 4194
model-bin/finetune/base/log/1630050046.2035077/events.out.tfevents.1630050046.52f5c7e305a3.886.89 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:709b11cfce27e86178a99bf7524ea96c43320f71e12f0b0b627dfbda8169b128
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630048138.52f5c7e305a3.886.80 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c9452f94cea99f7d845a7fce5fd6275f9cb7a4fc00b5141501e09163cefe720
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630048619.52f5c7e305a3.886.82 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc002d98a39a9740f6a94d265791543e732e073653e36f3504b9b7813d90f821
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630049102.52f5c7e305a3.886.84 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bac776b58ce1ed85f4af1b4fa33016f75c4b5c31fe37a4fa301132e607e7928
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630049572.52f5c7e305a3.886.86 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:298c644a488ca1e5101e65db1c62c62a2a587a66960116c40947d0643f9cc208
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630050046.52f5c7e305a3.886.88 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afc1ed3a9e82f43336874596bd326ec16c14ee316e9cf41c1c14df3165618864
3
+ size 8622