Check commited on
Commit
520d67b
Β·
1 Parent(s): c6a5092

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629879092.38512/events.out.tfevents.1629879092.7e498afd5545.905.83 +3 -0
  11. model-bin/finetune/base/log/1629879755.9543498/events.out.tfevents.1629879755.7e498afd5545.905.85 +3 -0
  12. model-bin/finetune/base/log/1629880403.5211415/events.out.tfevents.1629880403.7e498afd5545.905.87 +3 -0
  13. model-bin/finetune/base/log/1629881049.788205/events.out.tfevents.1629881049.7e498afd5545.905.89 +3 -0
  14. model-bin/finetune/base/log/1629881697.4184577/events.out.tfevents.1629881697.7e498afd5545.905.91 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629879092.7e498afd5545.905.82 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629879755.7e498afd5545.905.84 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629880403.7e498afd5545.905.86 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629881049.7e498afd5545.905.88 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629881697.7e498afd5545.905.90 +3 -0
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56aefc71d0980c995e93aee106a5ea7eb2b8637a7232ca5a4dcc1cdc4511335a
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54b35bdaaf4933af7fe3f9bbd9b1424d570c568987740f15bcd85cc870d83f51
3
  size 722165393
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bd62e5038f5a12dc25600aabe73f692b4ca8f39c6d252fb9e3be2bbc5ed1145
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f70ffbe9923a30dbf8b87abb1106958feef550de2620a2d2b80cc91fc8a098e2
3
  size 377909911
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a523a00f2b4a8aa255a532787f19739af80b95bca833548653c7ca44d7fe6056
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4be019330451dcdea1470053a969250ec2ac7591e9e422df7bf9734de62a6607
3
  size 14503
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89a798027e8a65425e8946906994aff28f6a07534f7b67f44af157705b99bf25
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3d8eee4042019854a091a06285230ead9805a530d0a1e8630cf2695e05d1d3b
3
  size 559
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5817db540e6b590b213245436521e08f3ff794616869a44c9a430a401dc99d6
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aae4596cc6445369fecda3c5539d240321af736664343a1c7e72481ab8edb80
3
  size 623
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
- "epoch": 600.995983935743,
5
- "global_step": 74539,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -194199,11 +194199,806 @@
194199
  "eval_steps_per_second": 0.666,
194200
  "eval_wer": 0.18913761467889909,
194201
  "step": 74539
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194202
  }
194203
  ],
194204
- "max_steps": 620000,
194205
  "num_train_epochs": 5000,
194206
- "total_flos": 2.097692339430875e+20,
194207
  "trial_name": null,
194208
  "trial_params": null
194209
  }
 
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
+ "epoch": 601.0,
5
+ "global_step": 75162,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
194199
  "eval_steps_per_second": 0.666,
194200
  "eval_wer": 0.18913761467889909,
194201
  "step": 74539
194202
+ },
194203
+ {
194204
+ "epoch": 596.01,
194205
+ "learning_rate": 8.821939102564104e-06,
194206
+ "loss": 0.4295,
194207
+ "step": 74540
194208
+ },
194209
+ {
194210
+ "epoch": 596.05,
194211
+ "learning_rate": 8.821858974358975e-06,
194212
+ "loss": 0.3124,
194213
+ "step": 74545
194214
+ },
194215
+ {
194216
+ "epoch": 596.09,
194217
+ "learning_rate": 8.821778846153846e-06,
194218
+ "loss": 0.3017,
194219
+ "step": 74550
194220
+ },
194221
+ {
194222
+ "epoch": 596.13,
194223
+ "learning_rate": 8.82169871794872e-06,
194224
+ "loss": 0.3852,
194225
+ "step": 74555
194226
+ },
194227
+ {
194228
+ "epoch": 596.17,
194229
+ "learning_rate": 8.82161858974359e-06,
194230
+ "loss": 0.5428,
194231
+ "step": 74560
194232
+ },
194233
+ {
194234
+ "epoch": 596.21,
194235
+ "learning_rate": 8.821538461538462e-06,
194236
+ "loss": 1.2103,
194237
+ "step": 74565
194238
+ },
194239
+ {
194240
+ "epoch": 596.25,
194241
+ "learning_rate": 8.821458333333333e-06,
194242
+ "loss": 0.3342,
194243
+ "step": 74570
194244
+ },
194245
+ {
194246
+ "epoch": 596.29,
194247
+ "learning_rate": 8.821378205128206e-06,
194248
+ "loss": 0.279,
194249
+ "step": 74575
194250
+ },
194251
+ {
194252
+ "epoch": 596.33,
194253
+ "learning_rate": 8.821298076923078e-06,
194254
+ "loss": 0.3352,
194255
+ "step": 74580
194256
+ },
194257
+ {
194258
+ "epoch": 596.37,
194259
+ "learning_rate": 8.82121794871795e-06,
194260
+ "loss": 0.6326,
194261
+ "step": 74585
194262
+ },
194263
+ {
194264
+ "epoch": 596.41,
194265
+ "learning_rate": 8.821137820512822e-06,
194266
+ "loss": 1.2738,
194267
+ "step": 74590
194268
+ },
194269
+ {
194270
+ "epoch": 596.45,
194271
+ "learning_rate": 8.821057692307694e-06,
194272
+ "loss": 0.2898,
194273
+ "step": 74595
194274
+ },
194275
+ {
194276
+ "epoch": 596.49,
194277
+ "learning_rate": 8.820977564102565e-06,
194278
+ "loss": 0.3831,
194279
+ "step": 74600
194280
+ },
194281
+ {
194282
+ "epoch": 596.53,
194283
+ "learning_rate": 8.820897435897436e-06,
194284
+ "loss": 0.3918,
194285
+ "step": 74605
194286
+ },
194287
+ {
194288
+ "epoch": 596.57,
194289
+ "learning_rate": 8.82081730769231e-06,
194290
+ "loss": 0.4952,
194291
+ "step": 74610
194292
+ },
194293
+ {
194294
+ "epoch": 596.61,
194295
+ "learning_rate": 8.82073717948718e-06,
194296
+ "loss": 1.0689,
194297
+ "step": 74615
194298
+ },
194299
+ {
194300
+ "epoch": 596.65,
194301
+ "learning_rate": 8.820657051282052e-06,
194302
+ "loss": 0.3247,
194303
+ "step": 74620
194304
+ },
194305
+ {
194306
+ "epoch": 596.69,
194307
+ "learning_rate": 8.820576923076923e-06,
194308
+ "loss": 0.3352,
194309
+ "step": 74625
194310
+ },
194311
+ {
194312
+ "epoch": 596.73,
194313
+ "learning_rate": 8.820496794871796e-06,
194314
+ "loss": 0.3389,
194315
+ "step": 74630
194316
+ },
194317
+ {
194318
+ "epoch": 596.76,
194319
+ "learning_rate": 8.820416666666666e-06,
194320
+ "loss": 0.6059,
194321
+ "step": 74635
194322
+ },
194323
+ {
194324
+ "epoch": 596.8,
194325
+ "learning_rate": 8.82033653846154e-06,
194326
+ "loss": 1.2196,
194327
+ "step": 74640
194328
+ },
194329
+ {
194330
+ "epoch": 596.84,
194331
+ "learning_rate": 8.820256410256412e-06,
194332
+ "loss": 0.3291,
194333
+ "step": 74645
194334
+ },
194335
+ {
194336
+ "epoch": 596.88,
194337
+ "learning_rate": 8.820176282051282e-06,
194338
+ "loss": 0.3028,
194339
+ "step": 74650
194340
+ },
194341
+ {
194342
+ "epoch": 596.92,
194343
+ "learning_rate": 8.820096153846155e-06,
194344
+ "loss": 0.3241,
194345
+ "step": 74655
194346
+ },
194347
+ {
194348
+ "epoch": 596.96,
194349
+ "learning_rate": 8.820016025641026e-06,
194350
+ "loss": 0.5501,
194351
+ "step": 74660
194352
+ },
194353
+ {
194354
+ "epoch": 597.0,
194355
+ "eval_loss": 0.3461139500141144,
194356
+ "eval_runtime": 40.6764,
194357
+ "eval_samples_per_second": 20.651,
194358
+ "eval_steps_per_second": 0.664,
194359
+ "eval_wer": 0.18884180790960453,
194360
+ "step": 74664
194361
+ },
194362
+ {
194363
+ "epoch": 602.01,
194364
+ "learning_rate": 8.819935897435898e-06,
194365
+ "loss": 0.4039,
194366
+ "step": 74665
194367
+ },
194368
+ {
194369
+ "epoch": 602.05,
194370
+ "learning_rate": 8.819855769230769e-06,
194371
+ "loss": 0.3075,
194372
+ "step": 74670
194373
+ },
194374
+ {
194375
+ "epoch": 602.09,
194376
+ "learning_rate": 8.819775641025642e-06,
194377
+ "loss": 0.2638,
194378
+ "step": 74675
194379
+ },
194380
+ {
194381
+ "epoch": 602.13,
194382
+ "learning_rate": 8.819695512820513e-06,
194383
+ "loss": 0.346,
194384
+ "step": 74680
194385
+ },
194386
+ {
194387
+ "epoch": 602.17,
194388
+ "learning_rate": 8.819615384615385e-06,
194389
+ "loss": 0.6652,
194390
+ "step": 74685
194391
+ },
194392
+ {
194393
+ "epoch": 602.21,
194394
+ "learning_rate": 8.819535256410258e-06,
194395
+ "loss": 1.2922,
194396
+ "step": 74690
194397
+ },
194398
+ {
194399
+ "epoch": 602.25,
194400
+ "learning_rate": 8.81945512820513e-06,
194401
+ "loss": 0.307,
194402
+ "step": 74695
194403
+ },
194404
+ {
194405
+ "epoch": 602.29,
194406
+ "learning_rate": 8.819375e-06,
194407
+ "loss": 0.2766,
194408
+ "step": 74700
194409
+ },
194410
+ {
194411
+ "epoch": 602.33,
194412
+ "learning_rate": 8.819294871794872e-06,
194413
+ "loss": 0.3401,
194414
+ "step": 74705
194415
+ },
194416
+ {
194417
+ "epoch": 602.37,
194418
+ "learning_rate": 8.819214743589745e-06,
194419
+ "loss": 0.5566,
194420
+ "step": 74710
194421
+ },
194422
+ {
194423
+ "epoch": 602.41,
194424
+ "learning_rate": 8.819134615384616e-06,
194425
+ "loss": 1.1703,
194426
+ "step": 74715
194427
+ },
194428
+ {
194429
+ "epoch": 602.45,
194430
+ "learning_rate": 8.819054487179488e-06,
194431
+ "loss": 0.3373,
194432
+ "step": 74720
194433
+ },
194434
+ {
194435
+ "epoch": 602.49,
194436
+ "learning_rate": 8.818974358974359e-06,
194437
+ "loss": 0.3019,
194438
+ "step": 74725
194439
+ },
194440
+ {
194441
+ "epoch": 602.53,
194442
+ "learning_rate": 8.818894230769232e-06,
194443
+ "loss": 0.351,
194444
+ "step": 74730
194445
+ },
194446
+ {
194447
+ "epoch": 602.57,
194448
+ "learning_rate": 8.818814102564103e-06,
194449
+ "loss": 0.5587,
194450
+ "step": 74735
194451
+ },
194452
+ {
194453
+ "epoch": 602.61,
194454
+ "learning_rate": 8.818733974358975e-06,
194455
+ "loss": 1.1618,
194456
+ "step": 74740
194457
+ },
194458
+ {
194459
+ "epoch": 602.65,
194460
+ "learning_rate": 8.818653846153848e-06,
194461
+ "loss": 0.327,
194462
+ "step": 74745
194463
+ },
194464
+ {
194465
+ "epoch": 602.69,
194466
+ "learning_rate": 8.81857371794872e-06,
194467
+ "loss": 0.3309,
194468
+ "step": 74750
194469
+ },
194470
+ {
194471
+ "epoch": 602.73,
194472
+ "learning_rate": 8.81849358974359e-06,
194473
+ "loss": 0.4235,
194474
+ "step": 74755
194475
+ },
194476
+ {
194477
+ "epoch": 602.77,
194478
+ "learning_rate": 8.818413461538462e-06,
194479
+ "loss": 0.5577,
194480
+ "step": 74760
194481
+ },
194482
+ {
194483
+ "epoch": 602.81,
194484
+ "learning_rate": 8.818333333333335e-06,
194485
+ "loss": 1.1446,
194486
+ "step": 74765
194487
+ },
194488
+ {
194489
+ "epoch": 602.85,
194490
+ "learning_rate": 8.818253205128205e-06,
194491
+ "loss": 0.3165,
194492
+ "step": 74770
194493
+ },
194494
+ {
194495
+ "epoch": 602.9,
194496
+ "learning_rate": 8.818173076923078e-06,
194497
+ "loss": 0.2766,
194498
+ "step": 74775
194499
+ },
194500
+ {
194501
+ "epoch": 602.94,
194502
+ "learning_rate": 8.818092948717949e-06,
194503
+ "loss": 0.3203,
194504
+ "step": 74780
194505
+ },
194506
+ {
194507
+ "epoch": 602.98,
194508
+ "learning_rate": 8.81801282051282e-06,
194509
+ "loss": 0.5715,
194510
+ "step": 74785
194511
+ },
194512
+ {
194513
+ "epoch": 603.0,
194514
+ "eval_loss": 0.38998129963874817,
194515
+ "eval_runtime": 40.9192,
194516
+ "eval_samples_per_second": 20.528,
194517
+ "eval_steps_per_second": 0.66,
194518
+ "eval_wer": 0.1910048391688016,
194519
+ "step": 74788
194520
+ },
194521
+ {
194522
+ "epoch": 603.02,
194523
+ "learning_rate": 8.817932692307694e-06,
194524
+ "loss": 0.3641,
194525
+ "step": 74790
194526
+ },
194527
+ {
194528
+ "epoch": 603.06,
194529
+ "learning_rate": 8.817852564102565e-06,
194530
+ "loss": 0.3938,
194531
+ "step": 74795
194532
+ },
194533
+ {
194534
+ "epoch": 603.1,
194535
+ "learning_rate": 8.817772435897436e-06,
194536
+ "loss": 0.3374,
194537
+ "step": 74800
194538
+ },
194539
+ {
194540
+ "epoch": 603.14,
194541
+ "learning_rate": 8.817692307692308e-06,
194542
+ "loss": 0.36,
194543
+ "step": 74805
194544
+ },
194545
+ {
194546
+ "epoch": 603.18,
194547
+ "learning_rate": 8.81761217948718e-06,
194548
+ "loss": 0.7182,
194549
+ "step": 74810
194550
+ },
194551
+ {
194552
+ "epoch": 603.22,
194553
+ "learning_rate": 8.817532051282052e-06,
194554
+ "loss": 1.0542,
194555
+ "step": 74815
194556
+ },
194557
+ {
194558
+ "epoch": 603.26,
194559
+ "learning_rate": 8.817451923076923e-06,
194560
+ "loss": 0.3317,
194561
+ "step": 74820
194562
+ },
194563
+ {
194564
+ "epoch": 603.3,
194565
+ "learning_rate": 8.817371794871795e-06,
194566
+ "loss": 0.4227,
194567
+ "step": 74825
194568
+ },
194569
+ {
194570
+ "epoch": 603.34,
194571
+ "learning_rate": 8.817291666666668e-06,
194572
+ "loss": 0.4196,
194573
+ "step": 74830
194574
+ },
194575
+ {
194576
+ "epoch": 603.38,
194577
+ "learning_rate": 8.817211538461539e-06,
194578
+ "loss": 0.6258,
194579
+ "step": 74835
194580
+ },
194581
+ {
194582
+ "epoch": 603.42,
194583
+ "learning_rate": 8.81713141025641e-06,
194584
+ "loss": 1.0404,
194585
+ "step": 74840
194586
+ },
194587
+ {
194588
+ "epoch": 603.46,
194589
+ "learning_rate": 8.817051282051284e-06,
194590
+ "loss": 0.3388,
194591
+ "step": 74845
194592
+ },
194593
+ {
194594
+ "epoch": 603.5,
194595
+ "learning_rate": 8.816971153846155e-06,
194596
+ "loss": 0.4682,
194597
+ "step": 74850
194598
+ },
194599
+ {
194600
+ "epoch": 603.54,
194601
+ "learning_rate": 8.816891025641026e-06,
194602
+ "loss": 0.3918,
194603
+ "step": 74855
194604
+ },
194605
+ {
194606
+ "epoch": 603.58,
194607
+ "learning_rate": 8.816810897435898e-06,
194608
+ "loss": 0.6384,
194609
+ "step": 74860
194610
+ },
194611
+ {
194612
+ "epoch": 603.62,
194613
+ "learning_rate": 8.81673076923077e-06,
194614
+ "loss": 1.0819,
194615
+ "step": 74865
194616
+ },
194617
+ {
194618
+ "epoch": 603.66,
194619
+ "learning_rate": 8.816650641025642e-06,
194620
+ "loss": 0.2616,
194621
+ "step": 74870
194622
+ },
194623
+ {
194624
+ "epoch": 603.7,
194625
+ "learning_rate": 8.816570512820513e-06,
194626
+ "loss": 0.3247,
194627
+ "step": 74875
194628
+ },
194629
+ {
194630
+ "epoch": 603.74,
194631
+ "learning_rate": 8.816490384615385e-06,
194632
+ "loss": 0.318,
194633
+ "step": 74880
194634
+ },
194635
+ {
194636
+ "epoch": 603.78,
194637
+ "learning_rate": 8.816410256410258e-06,
194638
+ "loss": 0.6705,
194639
+ "step": 74885
194640
+ },
194641
+ {
194642
+ "epoch": 603.82,
194643
+ "learning_rate": 8.816330128205129e-06,
194644
+ "loss": 1.2115,
194645
+ "step": 74890
194646
+ },
194647
+ {
194648
+ "epoch": 603.86,
194649
+ "learning_rate": 8.81625e-06,
194650
+ "loss": 0.3773,
194651
+ "step": 74895
194652
+ },
194653
+ {
194654
+ "epoch": 603.9,
194655
+ "learning_rate": 8.816169871794874e-06,
194656
+ "loss": 0.3811,
194657
+ "step": 74900
194658
+ },
194659
+ {
194660
+ "epoch": 603.94,
194661
+ "learning_rate": 8.816089743589745e-06,
194662
+ "loss": 0.388,
194663
+ "step": 74905
194664
+ },
194665
+ {
194666
+ "epoch": 603.98,
194667
+ "learning_rate": 8.816009615384616e-06,
194668
+ "loss": 0.8204,
194669
+ "step": 74910
194670
+ },
194671
+ {
194672
+ "epoch": 604.0,
194673
+ "eval_loss": 0.4254598617553711,
194674
+ "eval_runtime": 39.7945,
194675
+ "eval_samples_per_second": 21.108,
194676
+ "eval_steps_per_second": 0.678,
194677
+ "eval_wer": 0.19010397463313916,
194678
+ "step": 74912
194679
+ },
194680
+ {
194681
+ "epoch": 599.02,
194682
+ "learning_rate": 8.815929487179488e-06,
194683
+ "loss": 0.3721,
194684
+ "step": 74915
194685
+ },
194686
+ {
194687
+ "epoch": 599.06,
194688
+ "learning_rate": 8.81584935897436e-06,
194689
+ "loss": 0.2658,
194690
+ "step": 74920
194691
+ },
194692
+ {
194693
+ "epoch": 599.1,
194694
+ "learning_rate": 8.81576923076923e-06,
194695
+ "loss": 0.3926,
194696
+ "step": 74925
194697
+ },
194698
+ {
194699
+ "epoch": 599.14,
194700
+ "learning_rate": 8.815689102564103e-06,
194701
+ "loss": 0.3716,
194702
+ "step": 74930
194703
+ },
194704
+ {
194705
+ "epoch": 599.18,
194706
+ "learning_rate": 8.815608974358975e-06,
194707
+ "loss": 0.7772,
194708
+ "step": 74935
194709
+ },
194710
+ {
194711
+ "epoch": 599.22,
194712
+ "learning_rate": 8.815528846153846e-06,
194713
+ "loss": 0.8761,
194714
+ "step": 74940
194715
+ },
194716
+ {
194717
+ "epoch": 599.26,
194718
+ "learning_rate": 8.81544871794872e-06,
194719
+ "loss": 0.3565,
194720
+ "step": 74945
194721
+ },
194722
+ {
194723
+ "epoch": 599.3,
194724
+ "learning_rate": 8.81536858974359e-06,
194725
+ "loss": 0.363,
194726
+ "step": 74950
194727
+ },
194728
+ {
194729
+ "epoch": 599.34,
194730
+ "learning_rate": 8.815288461538462e-06,
194731
+ "loss": 0.361,
194732
+ "step": 74955
194733
+ },
194734
+ {
194735
+ "epoch": 599.38,
194736
+ "learning_rate": 8.815208333333333e-06,
194737
+ "loss": 0.7271,
194738
+ "step": 74960
194739
+ },
194740
+ {
194741
+ "epoch": 599.42,
194742
+ "learning_rate": 8.815128205128206e-06,
194743
+ "loss": 0.9404,
194744
+ "step": 74965
194745
+ },
194746
+ {
194747
+ "epoch": 599.46,
194748
+ "learning_rate": 8.815048076923078e-06,
194749
+ "loss": 0.3435,
194750
+ "step": 74970
194751
+ },
194752
+ {
194753
+ "epoch": 599.5,
194754
+ "learning_rate": 8.814967948717949e-06,
194755
+ "loss": 0.3344,
194756
+ "step": 74975
194757
+ },
194758
+ {
194759
+ "epoch": 599.54,
194760
+ "learning_rate": 8.81488782051282e-06,
194761
+ "loss": 0.4215,
194762
+ "step": 74980
194763
+ },
194764
+ {
194765
+ "epoch": 599.58,
194766
+ "learning_rate": 8.814807692307693e-06,
194767
+ "loss": 0.7642,
194768
+ "step": 74985
194769
+ },
194770
+ {
194771
+ "epoch": 599.62,
194772
+ "learning_rate": 8.814727564102565e-06,
194773
+ "loss": 0.9729,
194774
+ "step": 74990
194775
+ },
194776
+ {
194777
+ "epoch": 599.66,
194778
+ "learning_rate": 8.814647435897436e-06,
194779
+ "loss": 0.3428,
194780
+ "step": 74995
194781
+ },
194782
+ {
194783
+ "epoch": 599.7,
194784
+ "learning_rate": 8.81456730769231e-06,
194785
+ "loss": 0.2913,
194786
+ "step": 75000
194787
+ },
194788
+ {
194789
+ "epoch": 599.74,
194790
+ "learning_rate": 8.81448717948718e-06,
194791
+ "loss": 0.3823,
194792
+ "step": 75005
194793
+ },
194794
+ {
194795
+ "epoch": 599.78,
194796
+ "learning_rate": 8.814407051282052e-06,
194797
+ "loss": 0.6696,
194798
+ "step": 75010
194799
+ },
194800
+ {
194801
+ "epoch": 599.82,
194802
+ "learning_rate": 8.814326923076923e-06,
194803
+ "loss": 0.9848,
194804
+ "step": 75015
194805
+ },
194806
+ {
194807
+ "epoch": 599.86,
194808
+ "learning_rate": 8.814246794871796e-06,
194809
+ "loss": 0.2948,
194810
+ "step": 75020
194811
+ },
194812
+ {
194813
+ "epoch": 599.9,
194814
+ "learning_rate": 8.814166666666668e-06,
194815
+ "loss": 0.3141,
194816
+ "step": 75025
194817
+ },
194818
+ {
194819
+ "epoch": 599.94,
194820
+ "learning_rate": 8.814086538461539e-06,
194821
+ "loss": 0.4113,
194822
+ "step": 75030
194823
+ },
194824
+ {
194825
+ "epoch": 599.98,
194826
+ "learning_rate": 8.81400641025641e-06,
194827
+ "loss": 0.7153,
194828
+ "step": 75035
194829
+ },
194830
+ {
194831
+ "epoch": 600.0,
194832
+ "eval_loss": 0.5015895962715149,
194833
+ "eval_runtime": 39.7322,
194834
+ "eval_samples_per_second": 21.142,
194835
+ "eval_steps_per_second": 0.68,
194836
+ "eval_wer": 0.20653854649878506,
194837
+ "step": 75037
194838
+ },
194839
+ {
194840
+ "epoch": 600.02,
194841
+ "learning_rate": 8.813926282051283e-06,
194842
+ "loss": 0.3683,
194843
+ "step": 75040
194844
+ },
194845
+ {
194846
+ "epoch": 600.06,
194847
+ "learning_rate": 8.813846153846155e-06,
194848
+ "loss": 0.27,
194849
+ "step": 75045
194850
+ },
194851
+ {
194852
+ "epoch": 600.1,
194853
+ "learning_rate": 8.813766025641026e-06,
194854
+ "loss": 0.3321,
194855
+ "step": 75050
194856
+ },
194857
+ {
194858
+ "epoch": 600.14,
194859
+ "learning_rate": 8.8136858974359e-06,
194860
+ "loss": 0.4144,
194861
+ "step": 75055
194862
+ },
194863
+ {
194864
+ "epoch": 600.18,
194865
+ "learning_rate": 8.813605769230769e-06,
194866
+ "loss": 0.9552,
194867
+ "step": 75060
194868
+ },
194869
+ {
194870
+ "epoch": 600.22,
194871
+ "learning_rate": 8.813525641025642e-06,
194872
+ "loss": 0.9213,
194873
+ "step": 75065
194874
+ },
194875
+ {
194876
+ "epoch": 600.26,
194877
+ "learning_rate": 8.813445512820513e-06,
194878
+ "loss": 0.3018,
194879
+ "step": 75070
194880
+ },
194881
+ {
194882
+ "epoch": 600.3,
194883
+ "learning_rate": 8.813365384615385e-06,
194884
+ "loss": 0.3287,
194885
+ "step": 75075
194886
+ },
194887
+ {
194888
+ "epoch": 600.34,
194889
+ "learning_rate": 8.813285256410256e-06,
194890
+ "loss": 0.4535,
194891
+ "step": 75080
194892
+ },
194893
+ {
194894
+ "epoch": 600.38,
194895
+ "learning_rate": 8.813205128205129e-06,
194896
+ "loss": 0.8355,
194897
+ "step": 75085
194898
+ },
194899
+ {
194900
+ "epoch": 600.42,
194901
+ "learning_rate": 8.813125e-06,
194902
+ "loss": 0.8693,
194903
+ "step": 75090
194904
+ },
194905
+ {
194906
+ "epoch": 600.46,
194907
+ "learning_rate": 8.813044871794872e-06,
194908
+ "loss": 0.2923,
194909
+ "step": 75095
194910
+ },
194911
+ {
194912
+ "epoch": 600.5,
194913
+ "learning_rate": 8.812964743589745e-06,
194914
+ "loss": 0.3067,
194915
+ "step": 75100
194916
+ },
194917
+ {
194918
+ "epoch": 600.54,
194919
+ "learning_rate": 8.812884615384616e-06,
194920
+ "loss": 0.3881,
194921
+ "step": 75105
194922
+ },
194923
+ {
194924
+ "epoch": 600.58,
194925
+ "learning_rate": 8.812804487179488e-06,
194926
+ "loss": 0.7686,
194927
+ "step": 75110
194928
+ },
194929
+ {
194930
+ "epoch": 600.62,
194931
+ "learning_rate": 8.812724358974359e-06,
194932
+ "loss": 0.9086,
194933
+ "step": 75115
194934
+ },
194935
+ {
194936
+ "epoch": 600.66,
194937
+ "learning_rate": 8.812644230769232e-06,
194938
+ "loss": 0.2994,
194939
+ "step": 75120
194940
+ },
194941
+ {
194942
+ "epoch": 600.7,
194943
+ "learning_rate": 8.812564102564103e-06,
194944
+ "loss": 0.3495,
194945
+ "step": 75125
194946
+ },
194947
+ {
194948
+ "epoch": 600.74,
194949
+ "learning_rate": 8.812483974358975e-06,
194950
+ "loss": 0.3605,
194951
+ "step": 75130
194952
+ },
194953
+ {
194954
+ "epoch": 600.78,
194955
+ "learning_rate": 8.812403846153846e-06,
194956
+ "loss": 0.8099,
194957
+ "step": 75135
194958
+ },
194959
+ {
194960
+ "epoch": 600.82,
194961
+ "learning_rate": 8.812323717948719e-06,
194962
+ "loss": 1.0115,
194963
+ "step": 75140
194964
+ },
194965
+ {
194966
+ "epoch": 600.86,
194967
+ "learning_rate": 8.81224358974359e-06,
194968
+ "loss": 0.3424,
194969
+ "step": 75145
194970
+ },
194971
+ {
194972
+ "epoch": 600.9,
194973
+ "learning_rate": 8.812163461538462e-06,
194974
+ "loss": 0.2935,
194975
+ "step": 75150
194976
+ },
194977
+ {
194978
+ "epoch": 600.94,
194979
+ "learning_rate": 8.812083333333335e-06,
194980
+ "loss": 0.3522,
194981
+ "step": 75155
194982
+ },
194983
+ {
194984
+ "epoch": 600.98,
194985
+ "learning_rate": 8.812003205128206e-06,
194986
+ "loss": 0.9099,
194987
+ "step": 75160
194988
+ },
194989
+ {
194990
+ "epoch": 601.0,
194991
+ "eval_loss": 0.41707369685173035,
194992
+ "eval_runtime": 40.6126,
194993
+ "eval_samples_per_second": 20.708,
194994
+ "eval_steps_per_second": 0.665,
194995
+ "eval_wer": 0.2013764427557531,
194996
+ "step": 75162
194997
  }
194998
  ],
194999
+ "max_steps": 625000,
195000
  "num_train_epochs": 5000,
195001
+ "total_flos": 2.115183129640869e+20,
195002
  "trial_name": null,
195003
  "trial_params": null
195004
  }
model-bin/finetune/base/{checkpoint-74539 β†’ checkpoint-75162}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629879092.38512/events.out.tfevents.1629879092.7e498afd5545.905.83 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de16129103ae23782997a2f6e16821239d7b72f259380afebb6a9c2bafa80298
3
+ size 4194
model-bin/finetune/base/log/1629879755.9543498/events.out.tfevents.1629879755.7e498afd5545.905.85 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daf01e48c1f5fdaa4932cbf36a0e8f1a360a6c9d8d78d37ea17dbc37f9889fbc
3
+ size 4194
model-bin/finetune/base/log/1629880403.5211415/events.out.tfevents.1629880403.7e498afd5545.905.87 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73f895454c759316114bbc447d0826926107eacd8b1c68db68b79c05418dfdb3
3
+ size 4194
model-bin/finetune/base/log/1629881049.788205/events.out.tfevents.1629881049.7e498afd5545.905.89 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eb08ad02c8458e333fddd2e3bcab37c63f5cf5bd79fc60136fe6ab62c5abfad
3
+ size 4194
model-bin/finetune/base/log/1629881697.4184577/events.out.tfevents.1629881697.7e498afd5545.905.91 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9e7fb9b395276852faf078e20ad5747928b80f779c086d8d8d6e472d95dc267
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629879092.7e498afd5545.905.82 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f4d2778fc57e6d6438bb3f914aa609d397f723631fb70490c2bf77e9ff71107
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629879755.7e498afd5545.905.84 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e987a946247bf31d37e50ff0bec8308355b873abadef1ee8a0ea045987b7ecf
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629880403.7e498afd5545.905.86 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fca7d1a3a06e4fe4d2adbfbe329d1bff6334e8c455a0a4ef00b8cbd407248b0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629881049.7e498afd5545.905.88 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36541710ccff020bfe48255466aa1183422f6b931e5ec67e8803f8e2442a82a2
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629881697.7e498afd5545.905.90 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cee8924f68cd5f67852530719b67535c0badcd4e01529498c6b85fcc616fd300
3
+ size 8622