Check commited on
Commit
f458085
Β·
1 Parent(s): 51bc235

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630248561.4202452/events.out.tfevents.1630248561.cc93b136ebf5.1086.199 +3 -0
  11. model-bin/finetune/base/log/1630248992.3105023/events.out.tfevents.1630248992.cc93b136ebf5.1086.201 +3 -0
  12. model-bin/finetune/base/log/1630249423.7833743/events.out.tfevents.1630249423.cc93b136ebf5.1086.203 +3 -0
  13. model-bin/finetune/base/log/1630249857.8968918/events.out.tfevents.1630249857.cc93b136ebf5.1086.205 +3 -0
  14. model-bin/finetune/base/log/1630250287.4046617/events.out.tfevents.1630250287.cc93b136ebf5.1086.207 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630248561.cc93b136ebf5.1086.198 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630248992.cc93b136ebf5.1086.200 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630249423.cc93b136ebf5.1086.202 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630249857.cc93b136ebf5.1086.204 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630250287.cc93b136ebf5.1086.206 +3 -0
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e9eede5251c3f12c63b46d66a179f1dbd2b6d9f12ab83956ee5715976561602
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88106383b07c091dad16598681c368eb1f7d5dd6754f75dcdede18a251a76193
3
  size 722165393
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6bd337b8f0e2f3b46fc1c4ecb5f3bafbfc576e1e6bde2071ce0f670336e2879
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e90c5a61e1081391cac20c2749dbcf1cb5ad34446b8a2074676f8de81385b325
3
  size 377909911
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a1d6790720c3684b5fab1b80fa056d808968462ad941960785ea5dec6ae7387
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e5c8178a08780b7e94358c9559057df556e07dd698f9fa5dc769d640159e51e
3
  size 14503
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:884cb1f5ec43d0bdfcfad8769f8573a767f703fd8562f4c240659e0fdd8266d6
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fe336fce1f24fb6ba716a3ec77454f1b6f380bbf0069997841f401a2fb77c2a
3
  size 559
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be40c7eaf300a7ee9457562db278d81160ffb7d793ceebd4250c4e629248949f
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1c8fd19316f91f2cc4123c75c86425f3af495dca75999776dc0193fa6dd50cb
3
  size 623
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1220.995983935743,
5
- "global_step": 151448,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -292053,11 +292053,800 @@
292053
  "eval_steps_per_second": 0.639,
292054
  "eval_wer": 0.17625899280575538,
292055
  "step": 151448
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
292056
  }
292057
  ],
292058
  "max_steps": 620000,
292059
  "num_train_epochs": 5000,
292060
- "total_flos": 4.2618930913164073e+20,
292061
  "trial_name": null,
292062
  "trial_params": null
292063
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1225.995983935743,
5
+ "global_step": 152068,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
292053
  "eval_steps_per_second": 0.639,
292054
  "eval_wer": 0.17625899280575538,
292055
  "step": 151448
292056
+ },
292057
+ {
292058
+ "epoch": 1221.02,
292059
+ "learning_rate": 7.570420032310179e-06,
292060
+ "loss": 0.3857,
292061
+ "step": 151450
292062
+ },
292063
+ {
292064
+ "epoch": 1221.06,
292065
+ "learning_rate": 7.570339256865913e-06,
292066
+ "loss": 0.27,
292067
+ "step": 151455
292068
+ },
292069
+ {
292070
+ "epoch": 1221.1,
292071
+ "learning_rate": 7.5702584814216486e-06,
292072
+ "loss": 0.2653,
292073
+ "step": 151460
292074
+ },
292075
+ {
292076
+ "epoch": 1221.14,
292077
+ "learning_rate": 7.570177705977384e-06,
292078
+ "loss": 0.34,
292079
+ "step": 151465
292080
+ },
292081
+ {
292082
+ "epoch": 1221.18,
292083
+ "learning_rate": 7.5700969305331186e-06,
292084
+ "loss": 0.5222,
292085
+ "step": 151470
292086
+ },
292087
+ {
292088
+ "epoch": 1221.22,
292089
+ "learning_rate": 7.570016155088854e-06,
292090
+ "loss": 0.8796,
292091
+ "step": 151475
292092
+ },
292093
+ {
292094
+ "epoch": 1221.26,
292095
+ "learning_rate": 7.5699353796445885e-06,
292096
+ "loss": 0.2732,
292097
+ "step": 151480
292098
+ },
292099
+ {
292100
+ "epoch": 1221.3,
292101
+ "learning_rate": 7.569854604200324e-06,
292102
+ "loss": 0.2637,
292103
+ "step": 151485
292104
+ },
292105
+ {
292106
+ "epoch": 1221.34,
292107
+ "learning_rate": 7.5697738287560585e-06,
292108
+ "loss": 0.3104,
292109
+ "step": 151490
292110
+ },
292111
+ {
292112
+ "epoch": 1221.38,
292113
+ "learning_rate": 7.569693053311794e-06,
292114
+ "loss": 0.5735,
292115
+ "step": 151495
292116
+ },
292117
+ {
292118
+ "epoch": 1221.42,
292119
+ "learning_rate": 7.5696122778675285e-06,
292120
+ "loss": 0.9314,
292121
+ "step": 151500
292122
+ },
292123
+ {
292124
+ "epoch": 1221.46,
292125
+ "learning_rate": 7.569531502423264e-06,
292126
+ "loss": 0.2415,
292127
+ "step": 151505
292128
+ },
292129
+ {
292130
+ "epoch": 1221.5,
292131
+ "learning_rate": 7.5694507269789985e-06,
292132
+ "loss": 0.2433,
292133
+ "step": 151510
292134
+ },
292135
+ {
292136
+ "epoch": 1221.54,
292137
+ "learning_rate": 7.569369951534734e-06,
292138
+ "loss": 0.3259,
292139
+ "step": 151515
292140
+ },
292141
+ {
292142
+ "epoch": 1221.58,
292143
+ "learning_rate": 7.569289176090469e-06,
292144
+ "loss": 0.5686,
292145
+ "step": 151520
292146
+ },
292147
+ {
292148
+ "epoch": 1221.62,
292149
+ "learning_rate": 7.569208400646204e-06,
292150
+ "loss": 0.9219,
292151
+ "step": 151525
292152
+ },
292153
+ {
292154
+ "epoch": 1221.66,
292155
+ "learning_rate": 7.569127625201939e-06,
292156
+ "loss": 0.2692,
292157
+ "step": 151530
292158
+ },
292159
+ {
292160
+ "epoch": 1221.7,
292161
+ "learning_rate": 7.569046849757674e-06,
292162
+ "loss": 0.29,
292163
+ "step": 151535
292164
+ },
292165
+ {
292166
+ "epoch": 1221.74,
292167
+ "learning_rate": 7.56896607431341e-06,
292168
+ "loss": 0.2842,
292169
+ "step": 151540
292170
+ },
292171
+ {
292172
+ "epoch": 1221.78,
292173
+ "learning_rate": 7.568885298869144e-06,
292174
+ "loss": 0.5224,
292175
+ "step": 151545
292176
+ },
292177
+ {
292178
+ "epoch": 1221.82,
292179
+ "learning_rate": 7.56880452342488e-06,
292180
+ "loss": 0.9927,
292181
+ "step": 151550
292182
+ },
292183
+ {
292184
+ "epoch": 1221.86,
292185
+ "learning_rate": 7.568723747980614e-06,
292186
+ "loss": 0.2828,
292187
+ "step": 151555
292188
+ },
292189
+ {
292190
+ "epoch": 1221.9,
292191
+ "learning_rate": 7.56864297253635e-06,
292192
+ "loss": 0.2777,
292193
+ "step": 151560
292194
+ },
292195
+ {
292196
+ "epoch": 1221.94,
292197
+ "learning_rate": 7.568562197092084e-06,
292198
+ "loss": 0.3088,
292199
+ "step": 151565
292200
+ },
292201
+ {
292202
+ "epoch": 1221.98,
292203
+ "learning_rate": 7.56848142164782e-06,
292204
+ "loss": 0.6365,
292205
+ "step": 151570
292206
+ },
292207
+ {
292208
+ "epoch": 1222.0,
292209
+ "eval_loss": 0.5525439977645874,
292210
+ "eval_runtime": 39.5867,
292211
+ "eval_samples_per_second": 21.118,
292212
+ "eval_steps_per_second": 0.682,
292213
+ "eval_wer": 0.1825968316531494,
292214
+ "step": 151572
292215
+ },
292216
+ {
292217
+ "epoch": 1222.02,
292218
+ "learning_rate": 7.568400646203554e-06,
292219
+ "loss": 0.2871,
292220
+ "step": 151575
292221
+ },
292222
+ {
292223
+ "epoch": 1222.06,
292224
+ "learning_rate": 7.56831987075929e-06,
292225
+ "loss": 0.2949,
292226
+ "step": 151580
292227
+ },
292228
+ {
292229
+ "epoch": 1222.1,
292230
+ "learning_rate": 7.568239095315025e-06,
292231
+ "loss": 0.2781,
292232
+ "step": 151585
292233
+ },
292234
+ {
292235
+ "epoch": 1222.14,
292236
+ "learning_rate": 7.56815831987076e-06,
292237
+ "loss": 0.4072,
292238
+ "step": 151590
292239
+ },
292240
+ {
292241
+ "epoch": 1222.18,
292242
+ "learning_rate": 7.568077544426495e-06,
292243
+ "loss": 0.69,
292244
+ "step": 151595
292245
+ },
292246
+ {
292247
+ "epoch": 1222.22,
292248
+ "learning_rate": 7.56799676898223e-06,
292249
+ "loss": 0.7913,
292250
+ "step": 151600
292251
+ },
292252
+ {
292253
+ "epoch": 1222.27,
292254
+ "learning_rate": 7.567915993537966e-06,
292255
+ "loss": 0.3416,
292256
+ "step": 151605
292257
+ },
292258
+ {
292259
+ "epoch": 1222.31,
292260
+ "learning_rate": 7.5678352180937e-06,
292261
+ "loss": 0.2677,
292262
+ "step": 151610
292263
+ },
292264
+ {
292265
+ "epoch": 1222.35,
292266
+ "learning_rate": 7.567754442649436e-06,
292267
+ "loss": 0.3702,
292268
+ "step": 151615
292269
+ },
292270
+ {
292271
+ "epoch": 1222.39,
292272
+ "learning_rate": 7.56767366720517e-06,
292273
+ "loss": 0.7914,
292274
+ "step": 151620
292275
+ },
292276
+ {
292277
+ "epoch": 1222.43,
292278
+ "learning_rate": 7.567592891760906e-06,
292279
+ "loss": 0.7734,
292280
+ "step": 151625
292281
+ },
292282
+ {
292283
+ "epoch": 1222.47,
292284
+ "learning_rate": 7.56751211631664e-06,
292285
+ "loss": 0.4037,
292286
+ "step": 151630
292287
+ },
292288
+ {
292289
+ "epoch": 1222.51,
292290
+ "learning_rate": 7.567431340872376e-06,
292291
+ "loss": 0.2796,
292292
+ "step": 151635
292293
+ },
292294
+ {
292295
+ "epoch": 1222.55,
292296
+ "learning_rate": 7.567350565428111e-06,
292297
+ "loss": 0.4157,
292298
+ "step": 151640
292299
+ },
292300
+ {
292301
+ "epoch": 1222.59,
292302
+ "learning_rate": 7.567269789983846e-06,
292303
+ "loss": 0.7692,
292304
+ "step": 151645
292305
+ },
292306
+ {
292307
+ "epoch": 1222.63,
292308
+ "learning_rate": 7.567189014539581e-06,
292309
+ "loss": 0.8732,
292310
+ "step": 151650
292311
+ },
292312
+ {
292313
+ "epoch": 1222.67,
292314
+ "learning_rate": 7.567108239095316e-06,
292315
+ "loss": 0.256,
292316
+ "step": 151655
292317
+ },
292318
+ {
292319
+ "epoch": 1222.71,
292320
+ "learning_rate": 7.567027463651051e-06,
292321
+ "loss": 0.2665,
292322
+ "step": 151660
292323
+ },
292324
+ {
292325
+ "epoch": 1222.75,
292326
+ "learning_rate": 7.566946688206786e-06,
292327
+ "loss": 0.4612,
292328
+ "step": 151665
292329
+ },
292330
+ {
292331
+ "epoch": 1222.79,
292332
+ "learning_rate": 7.5668659127625214e-06,
292333
+ "loss": 0.7535,
292334
+ "step": 151670
292335
+ },
292336
+ {
292337
+ "epoch": 1222.83,
292338
+ "learning_rate": 7.566785137318256e-06,
292339
+ "loss": 0.9049,
292340
+ "step": 151675
292341
+ },
292342
+ {
292343
+ "epoch": 1222.87,
292344
+ "learning_rate": 7.566704361873991e-06,
292345
+ "loss": 0.3755,
292346
+ "step": 151680
292347
+ },
292348
+ {
292349
+ "epoch": 1222.91,
292350
+ "learning_rate": 7.5666235864297256e-06,
292351
+ "loss": 0.2588,
292352
+ "step": 151685
292353
+ },
292354
+ {
292355
+ "epoch": 1222.95,
292356
+ "learning_rate": 7.566542810985461e-06,
292357
+ "loss": 0.365,
292358
+ "step": 151690
292359
+ },
292360
+ {
292361
+ "epoch": 1222.99,
292362
+ "learning_rate": 7.566462035541196e-06,
292363
+ "loss": 0.7382,
292364
+ "step": 151695
292365
+ },
292366
+ {
292367
+ "epoch": 1223.0,
292368
+ "eval_loss": 0.3656209409236908,
292369
+ "eval_runtime": 41.975,
292370
+ "eval_samples_per_second": 19.917,
292371
+ "eval_steps_per_second": 0.643,
292372
+ "eval_wer": 0.16889752446747267,
292373
+ "step": 151696
292374
+ },
292375
+ {
292376
+ "epoch": 1223.03,
292377
+ "learning_rate": 7.566381260096931e-06,
292378
+ "loss": 0.3057,
292379
+ "step": 151700
292380
+ },
292381
+ {
292382
+ "epoch": 1223.07,
292383
+ "learning_rate": 7.566300484652666e-06,
292384
+ "loss": 0.3804,
292385
+ "step": 151705
292386
+ },
292387
+ {
292388
+ "epoch": 1223.11,
292389
+ "learning_rate": 7.566219709208401e-06,
292390
+ "loss": 0.3112,
292391
+ "step": 151710
292392
+ },
292393
+ {
292394
+ "epoch": 1223.15,
292395
+ "learning_rate": 7.566138933764136e-06,
292396
+ "loss": 0.3271,
292397
+ "step": 151715
292398
+ },
292399
+ {
292400
+ "epoch": 1223.19,
292401
+ "learning_rate": 7.566058158319871e-06,
292402
+ "loss": 0.8103,
292403
+ "step": 151720
292404
+ },
292405
+ {
292406
+ "epoch": 1223.23,
292407
+ "learning_rate": 7.565977382875606e-06,
292408
+ "loss": 0.6019,
292409
+ "step": 151725
292410
+ },
292411
+ {
292412
+ "epoch": 1223.27,
292413
+ "learning_rate": 7.565896607431341e-06,
292414
+ "loss": 0.2601,
292415
+ "step": 151730
292416
+ },
292417
+ {
292418
+ "epoch": 1223.31,
292419
+ "learning_rate": 7.565815831987076e-06,
292420
+ "loss": 0.2529,
292421
+ "step": 151735
292422
+ },
292423
+ {
292424
+ "epoch": 1223.35,
292425
+ "learning_rate": 7.565735056542811e-06,
292426
+ "loss": 0.3854,
292427
+ "step": 151740
292428
+ },
292429
+ {
292430
+ "epoch": 1223.4,
292431
+ "learning_rate": 7.565654281098547e-06,
292432
+ "loss": 0.8185,
292433
+ "step": 151745
292434
+ },
292435
+ {
292436
+ "epoch": 1223.44,
292437
+ "learning_rate": 7.565573505654281e-06,
292438
+ "loss": 0.5565,
292439
+ "step": 151750
292440
+ },
292441
+ {
292442
+ "epoch": 1223.48,
292443
+ "learning_rate": 7.565492730210017e-06,
292444
+ "loss": 0.2581,
292445
+ "step": 151755
292446
+ },
292447
+ {
292448
+ "epoch": 1223.52,
292449
+ "learning_rate": 7.565411954765752e-06,
292450
+ "loss": 0.2536,
292451
+ "step": 151760
292452
+ },
292453
+ {
292454
+ "epoch": 1223.56,
292455
+ "learning_rate": 7.565331179321487e-06,
292456
+ "loss": 0.3955,
292457
+ "step": 151765
292458
+ },
292459
+ {
292460
+ "epoch": 1223.6,
292461
+ "learning_rate": 7.565250403877222e-06,
292462
+ "loss": 0.8402,
292463
+ "step": 151770
292464
+ },
292465
+ {
292466
+ "epoch": 1223.64,
292467
+ "learning_rate": 7.565169628432957e-06,
292468
+ "loss": 0.6756,
292469
+ "step": 151775
292470
+ },
292471
+ {
292472
+ "epoch": 1223.68,
292473
+ "learning_rate": 7.565088852988692e-06,
292474
+ "loss": 0.2822,
292475
+ "step": 151780
292476
+ },
292477
+ {
292478
+ "epoch": 1223.72,
292479
+ "learning_rate": 7.565008077544427e-06,
292480
+ "loss": 0.3014,
292481
+ "step": 151785
292482
+ },
292483
+ {
292484
+ "epoch": 1223.76,
292485
+ "learning_rate": 7.564927302100162e-06,
292486
+ "loss": 0.4023,
292487
+ "step": 151790
292488
+ },
292489
+ {
292490
+ "epoch": 1223.8,
292491
+ "learning_rate": 7.564846526655897e-06,
292492
+ "loss": 0.8358,
292493
+ "step": 151795
292494
+ },
292495
+ {
292496
+ "epoch": 1223.84,
292497
+ "learning_rate": 7.564765751211632e-06,
292498
+ "loss": 0.7451,
292499
+ "step": 151800
292500
+ },
292501
+ {
292502
+ "epoch": 1223.88,
292503
+ "learning_rate": 7.564684975767367e-06,
292504
+ "loss": 0.2822,
292505
+ "step": 151805
292506
+ },
292507
+ {
292508
+ "epoch": 1223.92,
292509
+ "learning_rate": 7.564604200323103e-06,
292510
+ "loss": 0.2669,
292511
+ "step": 151810
292512
+ },
292513
+ {
292514
+ "epoch": 1223.96,
292515
+ "learning_rate": 7.564523424878838e-06,
292516
+ "loss": 0.4182,
292517
+ "step": 151815
292518
+ },
292519
+ {
292520
+ "epoch": 1224.0,
292521
+ "learning_rate": 7.564442649434573e-06,
292522
+ "loss": 1.1147,
292523
+ "step": 151820
292524
+ },
292525
+ {
292526
+ "epoch": 1224.0,
292527
+ "eval_loss": 0.34463566541671753,
292528
+ "eval_runtime": 39.9445,
292529
+ "eval_samples_per_second": 20.929,
292530
+ "eval_steps_per_second": 0.676,
292531
+ "eval_wer": 0.1835759611795446,
292532
+ "step": 151820
292533
+ },
292534
+ {
292535
+ "epoch": 1224.04,
292536
+ "learning_rate": 7.564361873990308e-06,
292537
+ "loss": 0.277,
292538
+ "step": 151825
292539
+ },
292540
+ {
292541
+ "epoch": 1224.08,
292542
+ "learning_rate": 7.564281098546043e-06,
292543
+ "loss": 0.2902,
292544
+ "step": 151830
292545
+ },
292546
+ {
292547
+ "epoch": 1224.12,
292548
+ "learning_rate": 7.564200323101778e-06,
292549
+ "loss": 0.2973,
292550
+ "step": 151835
292551
+ },
292552
+ {
292553
+ "epoch": 1224.16,
292554
+ "learning_rate": 7.564119547657513e-06,
292555
+ "loss": 0.4297,
292556
+ "step": 151840
292557
+ },
292558
+ {
292559
+ "epoch": 1224.2,
292560
+ "learning_rate": 7.564038772213248e-06,
292561
+ "loss": 1.2619,
292562
+ "step": 151845
292563
+ },
292564
+ {
292565
+ "epoch": 1224.24,
292566
+ "learning_rate": 7.563957996768983e-06,
292567
+ "loss": 0.3364,
292568
+ "step": 151850
292569
+ },
292570
+ {
292571
+ "epoch": 1224.28,
292572
+ "learning_rate": 7.563877221324718e-06,
292573
+ "loss": 0.2814,
292574
+ "step": 151855
292575
+ },
292576
+ {
292577
+ "epoch": 1224.32,
292578
+ "learning_rate": 7.563796445880453e-06,
292579
+ "loss": 0.3734,
292580
+ "step": 151860
292581
+ },
292582
+ {
292583
+ "epoch": 1224.36,
292584
+ "learning_rate": 7.563715670436188e-06,
292585
+ "loss": 0.4837,
292586
+ "step": 151865
292587
+ },
292588
+ {
292589
+ "epoch": 1224.4,
292590
+ "learning_rate": 7.5636348949919235e-06,
292591
+ "loss": 1.1229,
292592
+ "step": 151870
292593
+ },
292594
+ {
292595
+ "epoch": 1224.44,
292596
+ "learning_rate": 7.5635541195476585e-06,
292597
+ "loss": 0.3491,
292598
+ "step": 151875
292599
+ },
292600
+ {
292601
+ "epoch": 1224.48,
292602
+ "learning_rate": 7.5634733441033935e-06,
292603
+ "loss": 0.2809,
292604
+ "step": 151880
292605
+ },
292606
+ {
292607
+ "epoch": 1224.52,
292608
+ "learning_rate": 7.5633925686591284e-06,
292609
+ "loss": 0.2803,
292610
+ "step": 151885
292611
+ },
292612
+ {
292613
+ "epoch": 1224.56,
292614
+ "learning_rate": 7.5633117932148634e-06,
292615
+ "loss": 0.4462,
292616
+ "step": 151890
292617
+ },
292618
+ {
292619
+ "epoch": 1224.6,
292620
+ "learning_rate": 7.563231017770598e-06,
292621
+ "loss": 1.3136,
292622
+ "step": 151895
292623
+ },
292624
+ {
292625
+ "epoch": 1224.65,
292626
+ "learning_rate": 7.563150242326333e-06,
292627
+ "loss": 0.2907,
292628
+ "step": 151900
292629
+ },
292630
+ {
292631
+ "epoch": 1224.69,
292632
+ "learning_rate": 7.563069466882068e-06,
292633
+ "loss": 0.2524,
292634
+ "step": 151905
292635
+ },
292636
+ {
292637
+ "epoch": 1224.73,
292638
+ "learning_rate": 7.562988691437803e-06,
292639
+ "loss": 0.3055,
292640
+ "step": 151910
292641
+ },
292642
+ {
292643
+ "epoch": 1224.77,
292644
+ "learning_rate": 7.562907915993538e-06,
292645
+ "loss": 0.4168,
292646
+ "step": 151915
292647
+ },
292648
+ {
292649
+ "epoch": 1224.81,
292650
+ "learning_rate": 7.562827140549273e-06,
292651
+ "loss": 0.9734,
292652
+ "step": 151920
292653
+ },
292654
+ {
292655
+ "epoch": 1224.85,
292656
+ "learning_rate": 7.562746365105008e-06,
292657
+ "loss": 0.2457,
292658
+ "step": 151925
292659
+ },
292660
+ {
292661
+ "epoch": 1224.89,
292662
+ "learning_rate": 7.562665589660743e-06,
292663
+ "loss": 0.2464,
292664
+ "step": 151930
292665
+ },
292666
+ {
292667
+ "epoch": 1224.93,
292668
+ "learning_rate": 7.562584814216479e-06,
292669
+ "loss": 0.2614,
292670
+ "step": 151935
292671
+ },
292672
+ {
292673
+ "epoch": 1224.97,
292674
+ "learning_rate": 7.562504038772213e-06,
292675
+ "loss": 0.4991,
292676
+ "step": 151940
292677
+ },
292678
+ {
292679
+ "epoch": 1225.0,
292680
+ "eval_loss": 0.344937801361084,
292681
+ "eval_runtime": 40.5017,
292682
+ "eval_samples_per_second": 20.666,
292683
+ "eval_steps_per_second": 0.667,
292684
+ "eval_wer": 0.1767148683722655,
292685
+ "step": 151944
292686
+ },
292687
+ {
292688
+ "epoch": 1225.01,
292689
+ "learning_rate": 7.562423263327949e-06,
292690
+ "loss": 0.2553,
292691
+ "step": 151945
292692
+ },
292693
+ {
292694
+ "epoch": 1225.05,
292695
+ "learning_rate": 7.562342487883684e-06,
292696
+ "loss": 0.2594,
292697
+ "step": 151950
292698
+ },
292699
+ {
292700
+ "epoch": 1225.09,
292701
+ "learning_rate": 7.562261712439419e-06,
292702
+ "loss": 0.2837,
292703
+ "step": 151955
292704
+ },
292705
+ {
292706
+ "epoch": 1225.13,
292707
+ "learning_rate": 7.562180936995154e-06,
292708
+ "loss": 0.2949,
292709
+ "step": 151960
292710
+ },
292711
+ {
292712
+ "epoch": 1225.17,
292713
+ "learning_rate": 7.562100161550889e-06,
292714
+ "loss": 0.4855,
292715
+ "step": 151965
292716
+ },
292717
+ {
292718
+ "epoch": 1225.21,
292719
+ "learning_rate": 7.562019386106624e-06,
292720
+ "loss": 1.1717,
292721
+ "step": 151970
292722
+ },
292723
+ {
292724
+ "epoch": 1225.25,
292725
+ "learning_rate": 7.561938610662359e-06,
292726
+ "loss": 0.3164,
292727
+ "step": 151975
292728
+ },
292729
+ {
292730
+ "epoch": 1225.29,
292731
+ "learning_rate": 7.561857835218094e-06,
292732
+ "loss": 0.2502,
292733
+ "step": 151980
292734
+ },
292735
+ {
292736
+ "epoch": 1225.33,
292737
+ "learning_rate": 7.561777059773829e-06,
292738
+ "loss": 0.2958,
292739
+ "step": 151985
292740
+ },
292741
+ {
292742
+ "epoch": 1225.37,
292743
+ "learning_rate": 7.561696284329565e-06,
292744
+ "loss": 0.4835,
292745
+ "step": 151990
292746
+ },
292747
+ {
292748
+ "epoch": 1225.41,
292749
+ "learning_rate": 7.561615508885299e-06,
292750
+ "loss": 1.2292,
292751
+ "step": 151995
292752
+ },
292753
+ {
292754
+ "epoch": 1225.45,
292755
+ "learning_rate": 7.561534733441035e-06,
292756
+ "loss": 0.3822,
292757
+ "step": 152000
292758
+ },
292759
+ {
292760
+ "epoch": 1225.49,
292761
+ "learning_rate": 7.561453957996769e-06,
292762
+ "loss": 0.326,
292763
+ "step": 152005
292764
+ },
292765
+ {
292766
+ "epoch": 1225.53,
292767
+ "learning_rate": 7.561373182552505e-06,
292768
+ "loss": 0.3352,
292769
+ "step": 152010
292770
+ },
292771
+ {
292772
+ "epoch": 1225.57,
292773
+ "learning_rate": 7.56129240710824e-06,
292774
+ "loss": 0.5148,
292775
+ "step": 152015
292776
+ },
292777
+ {
292778
+ "epoch": 1225.61,
292779
+ "learning_rate": 7.561211631663975e-06,
292780
+ "loss": 1.0644,
292781
+ "step": 152020
292782
+ },
292783
+ {
292784
+ "epoch": 1225.65,
292785
+ "learning_rate": 7.56113085621971e-06,
292786
+ "loss": 0.2705,
292787
+ "step": 152025
292788
+ },
292789
+ {
292790
+ "epoch": 1225.69,
292791
+ "learning_rate": 7.561050080775445e-06,
292792
+ "loss": 0.2654,
292793
+ "step": 152030
292794
+ },
292795
+ {
292796
+ "epoch": 1225.73,
292797
+ "learning_rate": 7.56096930533118e-06,
292798
+ "loss": 0.2655,
292799
+ "step": 152035
292800
+ },
292801
+ {
292802
+ "epoch": 1225.77,
292803
+ "learning_rate": 7.560888529886915e-06,
292804
+ "loss": 0.5798,
292805
+ "step": 152040
292806
+ },
292807
+ {
292808
+ "epoch": 1225.81,
292809
+ "learning_rate": 7.560807754442651e-06,
292810
+ "loss": 1.0464,
292811
+ "step": 152045
292812
+ },
292813
+ {
292814
+ "epoch": 1225.85,
292815
+ "learning_rate": 7.560726978998385e-06,
292816
+ "loss": 0.264,
292817
+ "step": 152050
292818
+ },
292819
+ {
292820
+ "epoch": 1225.89,
292821
+ "learning_rate": 7.5606462035541206e-06,
292822
+ "loss": 0.2871,
292823
+ "step": 152055
292824
+ },
292825
+ {
292826
+ "epoch": 1225.93,
292827
+ "learning_rate": 7.560565428109855e-06,
292828
+ "loss": 0.3312,
292829
+ "step": 152060
292830
+ },
292831
+ {
292832
+ "epoch": 1225.97,
292833
+ "learning_rate": 7.5604846526655905e-06,
292834
+ "loss": 0.5482,
292835
+ "step": 152065
292836
+ },
292837
+ {
292838
+ "epoch": 1226.0,
292839
+ "eval_loss": 0.3725050091743469,
292840
+ "eval_runtime": 40.5085,
292841
+ "eval_samples_per_second": 20.662,
292842
+ "eval_steps_per_second": 0.667,
292843
+ "eval_wer": 0.1764970733903647,
292844
+ "step": 152068
292845
  }
292846
  ],
292847
  "max_steps": 620000,
292848
  "num_train_epochs": 5000,
292849
+ "total_flos": 4.2793273628046655e+20,
292850
  "trial_name": null,
292851
  "trial_params": null
292852
  }
model-bin/finetune/base/{checkpoint-151448 β†’ checkpoint-152068}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630248561.4202452/events.out.tfevents.1630248561.cc93b136ebf5.1086.199 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e69c53824a99a78f62d60a3191904b4929d958e37a96aed121cea82ef6f44c6
3
+ size 4194
model-bin/finetune/base/log/1630248992.3105023/events.out.tfevents.1630248992.cc93b136ebf5.1086.201 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18c11de576210df8c1bcd2e338686d0380f1a420b31373ffd06eb3aed8c1f453
3
+ size 4194
model-bin/finetune/base/log/1630249423.7833743/events.out.tfevents.1630249423.cc93b136ebf5.1086.203 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:568ece09adb0211ba3f15921d31806d68262b8cfa9857606486c11d4bae9edd6
3
+ size 4194
model-bin/finetune/base/log/1630249857.8968918/events.out.tfevents.1630249857.cc93b136ebf5.1086.205 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca6c23b7959ed4eff92db23ba2ba96ded88c6e6bd51c3726d065275e78ced855
3
+ size 4194
model-bin/finetune/base/log/1630250287.4046617/events.out.tfevents.1630250287.cc93b136ebf5.1086.207 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d4f64e2892b1172fc9ed28d3b5df11b5c1e87ce978ca5299337293b8153e478
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630248561.cc93b136ebf5.1086.198 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4839731a799003ae2c63f39a2b9d8d658d064e4e4d52528b686ea21a72782e95
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630248992.cc93b136ebf5.1086.200 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0ef6e85b448f1e9daa36e511987fdd78a3da2fe85a0475a40bc89341de9b3cb
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630249423.cc93b136ebf5.1086.202 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ad462f18f3ef5e7674be715c076f3a6912a2bdaa53ab1bde8f0f57357adc4f4
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630249857.cc93b136ebf5.1086.204 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8beff9185670015d1b0453cddc98932f4cf7524e092ff8056b49425d192625e8
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630250287.cc93b136ebf5.1086.206 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c75d446f14703e0a0c02a6515f68440f0868da9b0a01d30d83ccc76744caace
3
+ size 8622