Check commited on
Commit
ac53841
Β·
1 Parent(s): e5fedeb

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630109431.8143556/events.out.tfevents.1630109431.86bb0ddabf9b.1042.11 +3 -0
  11. model-bin/finetune/base/log/1630109864.1808882/events.out.tfevents.1630109864.86bb0ddabf9b.1042.13 +3 -0
  12. model-bin/finetune/base/log/1630110291.1942098/events.out.tfevents.1630110291.86bb0ddabf9b.1042.15 +3 -0
  13. model-bin/finetune/base/log/1630110708.1539922/events.out.tfevents.1630110708.86bb0ddabf9b.1042.17 +3 -0
  14. model-bin/finetune/base/log/1630111124.0089736/events.out.tfevents.1630111124.86bb0ddabf9b.1042.19 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630109431.86bb0ddabf9b.1042.10 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630109864.86bb0ddabf9b.1042.12 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630110291.86bb0ddabf9b.1042.14 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630110708.86bb0ddabf9b.1042.16 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630111124.86bb0ddabf9b.1042.18 +3 -0
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dc554dec301ccc8dca1e759108d10f280284ba56da831c7924e4f12d8241e81
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3bc213f8b1a88e0b2512f257ed2494755fd2795f55af3712d785a4772256d44
3
  size 722165393
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5144d589a8f3be429110fd3438d71c50cbe2b5d96f43ebff77a52854f4e4c466
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:845b93ea6994b83196fa59050206c61d3f4fc9ef2aa5ba6e72e95c173579ac5a
3
  size 377909911
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06429aa76c5c31c9f61e7d1a7c32603b81cc59ed4d9203338065436cd62cc39b
3
  size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c7fcf8524438e496a57b73e453b24b2519c6f58941a6fb8da174143dac50f8d
3
  size 14567
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d93d43bbeecc2e2dd56d18ef4fdffea967e794dca54efdfd4a801ca408b71336
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e28e0530f2551ac6de66de5fb95c14533593794a980009f90dec1f828fe4986
3
  size 559
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffe03c0c156b224ba155a03dcb28f915736509e98b9e680245bcc2241ec2d0d4
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6eac4ec54a649aef54843448fdae717a79bc1578521fca8e21301c3cddb0022
3
  size 623
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 923.995983935743,
5
- "global_step": 114613,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -245187,11 +245187,800 @@
245187
  "eval_steps_per_second": 0.677,
245188
  "eval_wer": 0.18582472364713531,
245189
  "step": 114613
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
245190
  }
245191
  ],
245192
  "max_steps": 620000,
245193
  "num_train_epochs": 5000,
245194
- "total_flos": 3.225517960789425e+20,
245195
  "trial_name": null,
245196
  "trial_params": null
245197
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 928.995983935743,
5
+ "global_step": 115234,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
245187
  "eval_steps_per_second": 0.677,
245188
  "eval_wer": 0.18582472364713531,
245189
  "step": 114613
245190
+ },
245191
+ {
245192
+ "epoch": 924.02,
245193
+ "learning_rate": 8.179983974358975e-06,
245194
+ "loss": 0.3841,
245195
+ "step": 114615
245196
+ },
245197
+ {
245198
+ "epoch": 924.06,
245199
+ "learning_rate": 8.179903846153847e-06,
245200
+ "loss": 0.3366,
245201
+ "step": 114620
245202
+ },
245203
+ {
245204
+ "epoch": 924.1,
245205
+ "learning_rate": 8.179823717948718e-06,
245206
+ "loss": 0.2559,
245207
+ "step": 114625
245208
+ },
245209
+ {
245210
+ "epoch": 924.14,
245211
+ "learning_rate": 8.179743589743591e-06,
245212
+ "loss": 0.3411,
245213
+ "step": 114630
245214
+ },
245215
+ {
245216
+ "epoch": 924.18,
245217
+ "learning_rate": 8.179663461538462e-06,
245218
+ "loss": 0.6367,
245219
+ "step": 114635
245220
+ },
245221
+ {
245222
+ "epoch": 924.22,
245223
+ "learning_rate": 8.179583333333334e-06,
245224
+ "loss": 1.0442,
245225
+ "step": 114640
245226
+ },
245227
+ {
245228
+ "epoch": 924.26,
245229
+ "learning_rate": 8.179503205128205e-06,
245230
+ "loss": 0.3043,
245231
+ "step": 114645
245232
+ },
245233
+ {
245234
+ "epoch": 924.3,
245235
+ "learning_rate": 8.179423076923078e-06,
245236
+ "loss": 0.2539,
245237
+ "step": 114650
245238
+ },
245239
+ {
245240
+ "epoch": 924.34,
245241
+ "learning_rate": 8.17934294871795e-06,
245242
+ "loss": 0.3802,
245243
+ "step": 114655
245244
+ },
245245
+ {
245246
+ "epoch": 924.38,
245247
+ "learning_rate": 8.179262820512821e-06,
245248
+ "loss": 0.6318,
245249
+ "step": 114660
245250
+ },
245251
+ {
245252
+ "epoch": 924.42,
245253
+ "learning_rate": 8.179182692307694e-06,
245254
+ "loss": 0.9742,
245255
+ "step": 114665
245256
+ },
245257
+ {
245258
+ "epoch": 924.46,
245259
+ "learning_rate": 8.179102564102565e-06,
245260
+ "loss": 0.2686,
245261
+ "step": 114670
245262
+ },
245263
+ {
245264
+ "epoch": 924.5,
245265
+ "learning_rate": 8.179022435897437e-06,
245266
+ "loss": 0.3525,
245267
+ "step": 114675
245268
+ },
245269
+ {
245270
+ "epoch": 924.54,
245271
+ "learning_rate": 8.178942307692308e-06,
245272
+ "loss": 0.3878,
245273
+ "step": 114680
245274
+ },
245275
+ {
245276
+ "epoch": 924.58,
245277
+ "learning_rate": 8.178862179487181e-06,
245278
+ "loss": 0.6801,
245279
+ "step": 114685
245280
+ },
245281
+ {
245282
+ "epoch": 924.62,
245283
+ "learning_rate": 8.178798076923078e-06,
245284
+ "loss": 0.976,
245285
+ "step": 114690
245286
+ },
245287
+ {
245288
+ "epoch": 924.66,
245289
+ "learning_rate": 8.17871794871795e-06,
245290
+ "loss": 0.2602,
245291
+ "step": 114695
245292
+ },
245293
+ {
245294
+ "epoch": 924.7,
245295
+ "learning_rate": 8.178637820512821e-06,
245296
+ "loss": 0.2754,
245297
+ "step": 114700
245298
+ },
245299
+ {
245300
+ "epoch": 924.74,
245301
+ "learning_rate": 8.178557692307693e-06,
245302
+ "loss": 0.3819,
245303
+ "step": 114705
245304
+ },
245305
+ {
245306
+ "epoch": 924.78,
245307
+ "learning_rate": 8.178477564102566e-06,
245308
+ "loss": 0.5863,
245309
+ "step": 114710
245310
+ },
245311
+ {
245312
+ "epoch": 924.82,
245313
+ "learning_rate": 8.178397435897435e-06,
245314
+ "loss": 1.1332,
245315
+ "step": 114715
245316
+ },
245317
+ {
245318
+ "epoch": 924.86,
245319
+ "learning_rate": 8.178317307692308e-06,
245320
+ "loss": 0.2884,
245321
+ "step": 114720
245322
+ },
245323
+ {
245324
+ "epoch": 924.9,
245325
+ "learning_rate": 8.178237179487181e-06,
245326
+ "loss": 0.2688,
245327
+ "step": 114725
245328
+ },
245329
+ {
245330
+ "epoch": 924.94,
245331
+ "learning_rate": 8.178157051282051e-06,
245332
+ "loss": 0.4118,
245333
+ "step": 114730
245334
+ },
245335
+ {
245336
+ "epoch": 924.98,
245337
+ "learning_rate": 8.178076923076924e-06,
245338
+ "loss": 0.608,
245339
+ "step": 114735
245340
+ },
245341
+ {
245342
+ "epoch": 925.0,
245343
+ "eval_loss": 0.3686845004558563,
245344
+ "eval_runtime": 40.4476,
245345
+ "eval_samples_per_second": 20.842,
245346
+ "eval_steps_per_second": 0.668,
245347
+ "eval_wer": 0.18588031473327077,
245348
+ "step": 114737
245349
+ },
245350
+ {
245351
+ "epoch": 925.02,
245352
+ "learning_rate": 8.177996794871795e-06,
245353
+ "loss": 0.4001,
245354
+ "step": 114740
245355
+ },
245356
+ {
245357
+ "epoch": 925.06,
245358
+ "learning_rate": 8.177916666666667e-06,
245359
+ "loss": 0.2602,
245360
+ "step": 114745
245361
+ },
245362
+ {
245363
+ "epoch": 925.1,
245364
+ "learning_rate": 8.177836538461538e-06,
245365
+ "loss": 0.3255,
245366
+ "step": 114750
245367
+ },
245368
+ {
245369
+ "epoch": 925.14,
245370
+ "learning_rate": 8.177756410256411e-06,
245371
+ "loss": 0.3547,
245372
+ "step": 114755
245373
+ },
245374
+ {
245375
+ "epoch": 925.18,
245376
+ "learning_rate": 8.177676282051283e-06,
245377
+ "loss": 0.7224,
245378
+ "step": 114760
245379
+ },
245380
+ {
245381
+ "epoch": 925.22,
245382
+ "learning_rate": 8.177596153846154e-06,
245383
+ "loss": 0.8328,
245384
+ "step": 114765
245385
+ },
245386
+ {
245387
+ "epoch": 925.27,
245388
+ "learning_rate": 8.177516025641025e-06,
245389
+ "loss": 0.2706,
245390
+ "step": 114770
245391
+ },
245392
+ {
245393
+ "epoch": 925.31,
245394
+ "learning_rate": 8.177435897435898e-06,
245395
+ "loss": 0.2727,
245396
+ "step": 114775
245397
+ },
245398
+ {
245399
+ "epoch": 925.35,
245400
+ "learning_rate": 8.17735576923077e-06,
245401
+ "loss": 0.3431,
245402
+ "step": 114780
245403
+ },
245404
+ {
245405
+ "epoch": 925.39,
245406
+ "learning_rate": 8.177275641025641e-06,
245407
+ "loss": 0.7878,
245408
+ "step": 114785
245409
+ },
245410
+ {
245411
+ "epoch": 925.43,
245412
+ "learning_rate": 8.177195512820514e-06,
245413
+ "loss": 1.0417,
245414
+ "step": 114790
245415
+ },
245416
+ {
245417
+ "epoch": 925.47,
245418
+ "learning_rate": 8.177115384615385e-06,
245419
+ "loss": 0.3444,
245420
+ "step": 114795
245421
+ },
245422
+ {
245423
+ "epoch": 925.51,
245424
+ "learning_rate": 8.177035256410257e-06,
245425
+ "loss": 0.2683,
245426
+ "step": 114800
245427
+ },
245428
+ {
245429
+ "epoch": 925.55,
245430
+ "learning_rate": 8.176955128205128e-06,
245431
+ "loss": 0.3586,
245432
+ "step": 114805
245433
+ },
245434
+ {
245435
+ "epoch": 925.59,
245436
+ "learning_rate": 8.176875000000001e-06,
245437
+ "loss": 0.7673,
245438
+ "step": 114810
245439
+ },
245440
+ {
245441
+ "epoch": 925.63,
245442
+ "learning_rate": 8.176794871794873e-06,
245443
+ "loss": 0.8954,
245444
+ "step": 114815
245445
+ },
245446
+ {
245447
+ "epoch": 925.67,
245448
+ "learning_rate": 8.176714743589744e-06,
245449
+ "loss": 0.3164,
245450
+ "step": 114820
245451
+ },
245452
+ {
245453
+ "epoch": 925.71,
245454
+ "learning_rate": 8.176634615384617e-06,
245455
+ "loss": 0.3331,
245456
+ "step": 114825
245457
+ },
245458
+ {
245459
+ "epoch": 925.75,
245460
+ "learning_rate": 8.176554487179488e-06,
245461
+ "loss": 0.3298,
245462
+ "step": 114830
245463
+ },
245464
+ {
245465
+ "epoch": 925.79,
245466
+ "learning_rate": 8.17647435897436e-06,
245467
+ "loss": 0.6546,
245468
+ "step": 114835
245469
+ },
245470
+ {
245471
+ "epoch": 925.83,
245472
+ "learning_rate": 8.176394230769231e-06,
245473
+ "loss": 0.8006,
245474
+ "step": 114840
245475
+ },
245476
+ {
245477
+ "epoch": 925.87,
245478
+ "learning_rate": 8.176314102564104e-06,
245479
+ "loss": 0.3202,
245480
+ "step": 114845
245481
+ },
245482
+ {
245483
+ "epoch": 925.91,
245484
+ "learning_rate": 8.176233974358975e-06,
245485
+ "loss": 0.3295,
245486
+ "step": 114850
245487
+ },
245488
+ {
245489
+ "epoch": 925.95,
245490
+ "learning_rate": 8.176153846153847e-06,
245491
+ "loss": 0.3493,
245492
+ "step": 114855
245493
+ },
245494
+ {
245495
+ "epoch": 925.99,
245496
+ "learning_rate": 8.176073717948718e-06,
245497
+ "loss": 0.7909,
245498
+ "step": 114860
245499
+ },
245500
+ {
245501
+ "epoch": 926.0,
245502
+ "eval_loss": 0.38164782524108887,
245503
+ "eval_runtime": 38.7323,
245504
+ "eval_samples_per_second": 21.765,
245505
+ "eval_steps_per_second": 0.697,
245506
+ "eval_wer": 0.180527833222444,
245507
+ "step": 114861
245508
+ },
245509
+ {
245510
+ "epoch": 926.03,
245511
+ "learning_rate": 8.175993589743591e-06,
245512
+ "loss": 0.3478,
245513
+ "step": 114865
245514
+ },
245515
+ {
245516
+ "epoch": 926.07,
245517
+ "learning_rate": 8.175913461538461e-06,
245518
+ "loss": 0.3354,
245519
+ "step": 114870
245520
+ },
245521
+ {
245522
+ "epoch": 926.11,
245523
+ "learning_rate": 8.175833333333334e-06,
245524
+ "loss": 0.2706,
245525
+ "step": 114875
245526
+ },
245527
+ {
245528
+ "epoch": 926.15,
245529
+ "learning_rate": 8.175753205128207e-06,
245530
+ "loss": 0.3976,
245531
+ "step": 114880
245532
+ },
245533
+ {
245534
+ "epoch": 926.19,
245535
+ "learning_rate": 8.175673076923077e-06,
245536
+ "loss": 0.9176,
245537
+ "step": 114885
245538
+ },
245539
+ {
245540
+ "epoch": 926.23,
245541
+ "learning_rate": 8.17559294871795e-06,
245542
+ "loss": 0.6236,
245543
+ "step": 114890
245544
+ },
245545
+ {
245546
+ "epoch": 926.27,
245547
+ "learning_rate": 8.175512820512821e-06,
245548
+ "loss": 0.2932,
245549
+ "step": 114895
245550
+ },
245551
+ {
245552
+ "epoch": 926.31,
245553
+ "learning_rate": 8.175432692307692e-06,
245554
+ "loss": 0.3139,
245555
+ "step": 114900
245556
+ },
245557
+ {
245558
+ "epoch": 926.35,
245559
+ "learning_rate": 8.175352564102564e-06,
245560
+ "loss": 0.4103,
245561
+ "step": 114905
245562
+ },
245563
+ {
245564
+ "epoch": 926.39,
245565
+ "learning_rate": 8.175272435897437e-06,
245566
+ "loss": 0.8761,
245567
+ "step": 114910
245568
+ },
245569
+ {
245570
+ "epoch": 926.43,
245571
+ "learning_rate": 8.175192307692308e-06,
245572
+ "loss": 0.6826,
245573
+ "step": 114915
245574
+ },
245575
+ {
245576
+ "epoch": 926.47,
245577
+ "learning_rate": 8.17511217948718e-06,
245578
+ "loss": 0.268,
245579
+ "step": 114920
245580
+ },
245581
+ {
245582
+ "epoch": 926.51,
245583
+ "learning_rate": 8.175032051282053e-06,
245584
+ "loss": 0.2962,
245585
+ "step": 114925
245586
+ },
245587
+ {
245588
+ "epoch": 926.55,
245589
+ "learning_rate": 8.174951923076924e-06,
245590
+ "loss": 0.4022,
245591
+ "step": 114930
245592
+ },
245593
+ {
245594
+ "epoch": 926.59,
245595
+ "learning_rate": 8.174871794871795e-06,
245596
+ "loss": 0.8668,
245597
+ "step": 114935
245598
+ },
245599
+ {
245600
+ "epoch": 926.63,
245601
+ "learning_rate": 8.174791666666667e-06,
245602
+ "loss": 0.6838,
245603
+ "step": 114940
245604
+ },
245605
+ {
245606
+ "epoch": 926.67,
245607
+ "learning_rate": 8.17471153846154e-06,
245608
+ "loss": 0.2747,
245609
+ "step": 114945
245610
+ },
245611
+ {
245612
+ "epoch": 926.71,
245613
+ "learning_rate": 8.174631410256411e-06,
245614
+ "loss": 0.364,
245615
+ "step": 114950
245616
+ },
245617
+ {
245618
+ "epoch": 926.76,
245619
+ "learning_rate": 8.174551282051282e-06,
245620
+ "loss": 0.3743,
245621
+ "step": 114955
245622
+ },
245623
+ {
245624
+ "epoch": 926.8,
245625
+ "learning_rate": 8.174471153846154e-06,
245626
+ "loss": 0.9189,
245627
+ "step": 114960
245628
+ },
245629
+ {
245630
+ "epoch": 926.84,
245631
+ "learning_rate": 8.174391025641027e-06,
245632
+ "loss": 0.6867,
245633
+ "step": 114965
245634
+ },
245635
+ {
245636
+ "epoch": 926.88,
245637
+ "learning_rate": 8.174310897435898e-06,
245638
+ "loss": 0.3038,
245639
+ "step": 114970
245640
+ },
245641
+ {
245642
+ "epoch": 926.92,
245643
+ "learning_rate": 8.17423076923077e-06,
245644
+ "loss": 0.2775,
245645
+ "step": 114975
245646
+ },
245647
+ {
245648
+ "epoch": 926.96,
245649
+ "learning_rate": 8.174150641025643e-06,
245650
+ "loss": 0.4685,
245651
+ "step": 114980
245652
+ },
245653
+ {
245654
+ "epoch": 927.0,
245655
+ "learning_rate": 8.174070512820514e-06,
245656
+ "loss": 1.0971,
245657
+ "step": 114985
245658
+ },
245659
+ {
245660
+ "epoch": 927.0,
245661
+ "eval_loss": 0.3635014295578003,
245662
+ "eval_runtime": 39.7417,
245663
+ "eval_samples_per_second": 21.212,
245664
+ "eval_steps_per_second": 0.679,
245665
+ "eval_wer": 0.18649314104818854,
245666
+ "step": 114985
245667
+ },
245668
+ {
245669
+ "epoch": 919.04,
245670
+ "learning_rate": 8.173990384615385e-06,
245671
+ "loss": 0.302,
245672
+ "step": 114990
245673
+ },
245674
+ {
245675
+ "epoch": 919.08,
245676
+ "learning_rate": 8.173910256410257e-06,
245677
+ "loss": 0.3467,
245678
+ "step": 114995
245679
+ },
245680
+ {
245681
+ "epoch": 919.12,
245682
+ "learning_rate": 8.17383012820513e-06,
245683
+ "loss": 0.339,
245684
+ "step": 115000
245685
+ },
245686
+ {
245687
+ "epoch": 919.16,
245688
+ "learning_rate": 8.17375e-06,
245689
+ "loss": 0.4937,
245690
+ "step": 115005
245691
+ },
245692
+ {
245693
+ "epoch": 919.2,
245694
+ "learning_rate": 8.173669871794873e-06,
245695
+ "loss": 1.1532,
245696
+ "step": 115010
245697
+ },
245698
+ {
245699
+ "epoch": 919.24,
245700
+ "learning_rate": 8.173589743589744e-06,
245701
+ "loss": 0.3483,
245702
+ "step": 115015
245703
+ },
245704
+ {
245705
+ "epoch": 919.28,
245706
+ "learning_rate": 8.173509615384615e-06,
245707
+ "loss": 0.3049,
245708
+ "step": 115020
245709
+ },
245710
+ {
245711
+ "epoch": 919.32,
245712
+ "learning_rate": 8.173429487179488e-06,
245713
+ "loss": 0.3562,
245714
+ "step": 115025
245715
+ },
245716
+ {
245717
+ "epoch": 919.36,
245718
+ "learning_rate": 8.17334935897436e-06,
245719
+ "loss": 0.4262,
245720
+ "step": 115030
245721
+ },
245722
+ {
245723
+ "epoch": 919.4,
245724
+ "learning_rate": 8.173269230769233e-06,
245725
+ "loss": 1.2422,
245726
+ "step": 115035
245727
+ },
245728
+ {
245729
+ "epoch": 919.44,
245730
+ "learning_rate": 8.173189102564102e-06,
245731
+ "loss": 0.341,
245732
+ "step": 115040
245733
+ },
245734
+ {
245735
+ "epoch": 919.48,
245736
+ "learning_rate": 8.173108974358975e-06,
245737
+ "loss": 0.2614,
245738
+ "step": 115045
245739
+ },
245740
+ {
245741
+ "epoch": 919.52,
245742
+ "learning_rate": 8.173028846153847e-06,
245743
+ "loss": 0.2816,
245744
+ "step": 115050
245745
+ },
245746
+ {
245747
+ "epoch": 919.56,
245748
+ "learning_rate": 8.172948717948718e-06,
245749
+ "loss": 0.3867,
245750
+ "step": 115055
245751
+ },
245752
+ {
245753
+ "epoch": 919.6,
245754
+ "learning_rate": 8.17286858974359e-06,
245755
+ "loss": 1.1399,
245756
+ "step": 115060
245757
+ },
245758
+ {
245759
+ "epoch": 919.64,
245760
+ "learning_rate": 8.172788461538463e-06,
245761
+ "loss": 0.278,
245762
+ "step": 115065
245763
+ },
245764
+ {
245765
+ "epoch": 919.68,
245766
+ "learning_rate": 8.172708333333334e-06,
245767
+ "loss": 0.2704,
245768
+ "step": 115070
245769
+ },
245770
+ {
245771
+ "epoch": 919.72,
245772
+ "learning_rate": 8.172628205128205e-06,
245773
+ "loss": 0.2866,
245774
+ "step": 115075
245775
+ },
245776
+ {
245777
+ "epoch": 919.76,
245778
+ "learning_rate": 8.172548076923078e-06,
245779
+ "loss": 0.4154,
245780
+ "step": 115080
245781
+ },
245782
+ {
245783
+ "epoch": 919.8,
245784
+ "learning_rate": 8.17246794871795e-06,
245785
+ "loss": 1.0377,
245786
+ "step": 115085
245787
+ },
245788
+ {
245789
+ "epoch": 919.84,
245790
+ "learning_rate": 8.172387820512821e-06,
245791
+ "loss": 0.3569,
245792
+ "step": 115090
245793
+ },
245794
+ {
245795
+ "epoch": 919.88,
245796
+ "learning_rate": 8.172307692307692e-06,
245797
+ "loss": 0.2438,
245798
+ "step": 115095
245799
+ },
245800
+ {
245801
+ "epoch": 919.92,
245802
+ "learning_rate": 8.172227564102565e-06,
245803
+ "loss": 0.3913,
245804
+ "step": 115100
245805
+ },
245806
+ {
245807
+ "epoch": 919.96,
245808
+ "learning_rate": 8.172147435897437e-06,
245809
+ "loss": 0.4283,
245810
+ "step": 115105
245811
+ },
245812
+ {
245813
+ "epoch": 920.0,
245814
+ "learning_rate": 8.172067307692308e-06,
245815
+ "loss": 1.2519,
245816
+ "step": 115110
245817
+ },
245818
+ {
245819
+ "epoch": 920.0,
245820
+ "eval_loss": 0.36768418550491333,
245821
+ "eval_runtime": 39.8638,
245822
+ "eval_samples_per_second": 21.147,
245823
+ "eval_steps_per_second": 0.677,
245824
+ "eval_wer": 0.18435032304379037,
245825
+ "step": 115110
245826
+ },
245827
+ {
245828
+ "epoch": 928.04,
245829
+ "learning_rate": 8.17198717948718e-06,
245830
+ "loss": 0.3685,
245831
+ "step": 115115
245832
+ },
245833
+ {
245834
+ "epoch": 928.08,
245835
+ "learning_rate": 8.171907051282053e-06,
245836
+ "loss": 0.3035,
245837
+ "step": 115120
245838
+ },
245839
+ {
245840
+ "epoch": 928.12,
245841
+ "learning_rate": 8.171826923076924e-06,
245842
+ "loss": 0.3222,
245843
+ "step": 115125
245844
+ },
245845
+ {
245846
+ "epoch": 928.16,
245847
+ "learning_rate": 8.171746794871795e-06,
245848
+ "loss": 0.4358,
245849
+ "step": 115130
245850
+ },
245851
+ {
245852
+ "epoch": 928.2,
245853
+ "learning_rate": 8.171666666666668e-06,
245854
+ "loss": 1.1252,
245855
+ "step": 115135
245856
+ },
245857
+ {
245858
+ "epoch": 928.24,
245859
+ "learning_rate": 8.17158653846154e-06,
245860
+ "loss": 0.3178,
245861
+ "step": 115140
245862
+ },
245863
+ {
245864
+ "epoch": 928.28,
245865
+ "learning_rate": 8.171506410256411e-06,
245866
+ "loss": 0.3091,
245867
+ "step": 115145
245868
+ },
245869
+ {
245870
+ "epoch": 928.32,
245871
+ "learning_rate": 8.171426282051282e-06,
245872
+ "loss": 0.3167,
245873
+ "step": 115150
245874
+ },
245875
+ {
245876
+ "epoch": 928.36,
245877
+ "learning_rate": 8.171346153846155e-06,
245878
+ "loss": 0.4127,
245879
+ "step": 115155
245880
+ },
245881
+ {
245882
+ "epoch": 928.4,
245883
+ "learning_rate": 8.171266025641025e-06,
245884
+ "loss": 1.0843,
245885
+ "step": 115160
245886
+ },
245887
+ {
245888
+ "epoch": 928.44,
245889
+ "learning_rate": 8.171185897435898e-06,
245890
+ "loss": 0.3319,
245891
+ "step": 115165
245892
+ },
245893
+ {
245894
+ "epoch": 928.48,
245895
+ "learning_rate": 8.17110576923077e-06,
245896
+ "loss": 0.2568,
245897
+ "step": 115170
245898
+ },
245899
+ {
245900
+ "epoch": 928.52,
245901
+ "learning_rate": 8.171025641025641e-06,
245902
+ "loss": 0.3007,
245903
+ "step": 115175
245904
+ },
245905
+ {
245906
+ "epoch": 928.56,
245907
+ "learning_rate": 8.170945512820514e-06,
245908
+ "loss": 0.4638,
245909
+ "step": 115180
245910
+ },
245911
+ {
245912
+ "epoch": 928.6,
245913
+ "learning_rate": 8.170865384615385e-06,
245914
+ "loss": 1.3013,
245915
+ "step": 115185
245916
+ },
245917
+ {
245918
+ "epoch": 928.64,
245919
+ "learning_rate": 8.170785256410257e-06,
245920
+ "loss": 0.3193,
245921
+ "step": 115190
245922
+ },
245923
+ {
245924
+ "epoch": 928.68,
245925
+ "learning_rate": 8.170705128205128e-06,
245926
+ "loss": 0.273,
245927
+ "step": 115195
245928
+ },
245929
+ {
245930
+ "epoch": 928.72,
245931
+ "learning_rate": 8.170625000000001e-06,
245932
+ "loss": 0.3355,
245933
+ "step": 115200
245934
+ },
245935
+ {
245936
+ "epoch": 928.76,
245937
+ "learning_rate": 8.170544871794872e-06,
245938
+ "loss": 0.4521,
245939
+ "step": 115205
245940
+ },
245941
+ {
245942
+ "epoch": 928.8,
245943
+ "learning_rate": 8.170464743589744e-06,
245944
+ "loss": 1.1896,
245945
+ "step": 115210
245946
+ },
245947
+ {
245948
+ "epoch": 928.84,
245949
+ "learning_rate": 8.170384615384615e-06,
245950
+ "loss": 0.2887,
245951
+ "step": 115215
245952
+ },
245953
+ {
245954
+ "epoch": 928.88,
245955
+ "learning_rate": 8.170304487179488e-06,
245956
+ "loss": 0.27,
245957
+ "step": 115220
245958
+ },
245959
+ {
245960
+ "epoch": 928.92,
245961
+ "learning_rate": 8.17022435897436e-06,
245962
+ "loss": 0.3528,
245963
+ "step": 115225
245964
+ },
245965
+ {
245966
+ "epoch": 928.96,
245967
+ "learning_rate": 8.170144230769231e-06,
245968
+ "loss": 0.4421,
245969
+ "step": 115230
245970
+ },
245971
+ {
245972
+ "epoch": 929.0,
245973
+ "eval_loss": 0.38033702969551086,
245974
+ "eval_runtime": 39.6678,
245975
+ "eval_samples_per_second": 21.277,
245976
+ "eval_steps_per_second": 0.681,
245977
+ "eval_wer": 0.18934867596075827,
245978
+ "step": 115234
245979
  }
245980
  ],
245981
  "max_steps": 620000,
245982
  "num_train_epochs": 5000,
245983
+ "total_flos": 3.243039409882843e+20,
245984
  "trial_name": null,
245985
  "trial_params": null
245986
  }
model-bin/finetune/base/{checkpoint-114613 β†’ checkpoint-115234}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630109431.8143556/events.out.tfevents.1630109431.86bb0ddabf9b.1042.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1752755bd47a1298986b325a70db8b95c0108718e4f7caea9973b49ff20223e
3
+ size 4194
model-bin/finetune/base/log/1630109864.1808882/events.out.tfevents.1630109864.86bb0ddabf9b.1042.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:492fcb0f04e16405c50032f9b9f601fbf6ccb916bc8c6f9c82b6cbf6072c2f51
3
+ size 4194
model-bin/finetune/base/log/1630110291.1942098/events.out.tfevents.1630110291.86bb0ddabf9b.1042.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9678335599331e6c8a5e95ac8ae114b537ca30f5aa61f90121c6ad0909316186
3
+ size 4194
model-bin/finetune/base/log/1630110708.1539922/events.out.tfevents.1630110708.86bb0ddabf9b.1042.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8789ebd4a0d049bcc6e7a6cfb131f3de072740ef62692b320e8d3b2457ea1f9
3
+ size 4194
model-bin/finetune/base/log/1630111124.0089736/events.out.tfevents.1630111124.86bb0ddabf9b.1042.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50220645914c1340a01a0938428ab7312dd5ae9ee446dcb2cf3f29d82df09f88
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630109431.86bb0ddabf9b.1042.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2f69acf780d6b84f0085f54961d3041b1ec46c0d7af1f152bfabb4b960ce1e9
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630109864.86bb0ddabf9b.1042.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbf005686af30527fa466f8f955ccb86be67ba4c19c6e2a7a603c53d9ffe89c7
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630110291.86bb0ddabf9b.1042.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0f307121ff5f987c2125b3494b029ead7320a9107cbbe1f662acbfb27ad23fa
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630110708.86bb0ddabf9b.1042.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8c9cc1c211c48db928e1b7c27c884da18dd9b5870bc8b09f9fbb05f683bdcf5
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630111124.86bb0ddabf9b.1042.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b51ed23240af71ded7356cae20da843c27a40e6f060bd5a294754f7398910155
3
+ size 8462