Check commited on
Commit
c6a5092
Β·
1 Parent(s): 9ccd666

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629875783.5501642/events.out.tfevents.1629875783.7e498afd5545.905.73 +3 -0
  11. model-bin/finetune/base/log/1629876453.8615327/events.out.tfevents.1629876453.7e498afd5545.905.75 +3 -0
  12. model-bin/finetune/base/log/1629877099.8360853/events.out.tfevents.1629877099.7e498afd5545.905.77 +3 -0
  13. model-bin/finetune/base/log/1629877746.4587185/events.out.tfevents.1629877746.7e498afd5545.905.79 +3 -0
  14. model-bin/finetune/base/log/1629878403.3794868/events.out.tfevents.1629878403.7e498afd5545.905.81 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629875783.7e498afd5545.905.72 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629876453.7e498afd5545.905.74 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629877099.7e498afd5545.905.76 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629877746.7e498afd5545.905.78 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629878403.7e498afd5545.905.80 +3 -0
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfd36b6ee499d562847b385dc3ea784f1e658ef42e11f2762942b6000097d6a6
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56aefc71d0980c995e93aee106a5ea7eb2b8637a7232ca5a4dcc1cdc4511335a
3
  size 722165393
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:633059da288f26b3e326a5988ea125d0c4538291da52f14c3625a68fb22478fa
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bd62e5038f5a12dc25600aabe73f692b4ca8f39c6d252fb9e3be2bbc5ed1145
3
  size 377909911
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e563717b5c623c6ea44320c2c602e0fe01a69436b866f3fa268004b71402a438
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a523a00f2b4a8aa255a532787f19739af80b95bca833548653c7ca44d7fe6056
3
  size 14503
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb98e84d72fc9445ed9adb6eb66f1ee1e64e6fd672c8cf163f23ed10f15dd0e7
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89a798027e8a65425e8946906994aff28f6a07534f7b67f44af157705b99bf25
3
  size 559
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5509a63f3d04d5bbd7c28685f03d0c9d6c508593d58909a296e0f060d7e0530
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5817db540e6b590b213245436521e08f3ff794616869a44c9a430a401dc99d6
3
  size 623
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
- "epoch": 595.995983935743,
5
- "global_step": 73916,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -193410,11 +193410,800 @@
193410
  "eval_steps_per_second": 0.69,
193411
  "eval_wer": 0.1899564928840056,
193412
  "step": 73916
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
193413
  }
193414
  ],
193415
  "max_steps": 620000,
193416
  "num_train_epochs": 5000,
193417
- "total_flos": 2.08014258346129e+20,
193418
  "trial_name": null,
193419
  "trial_params": null
193420
  }
 
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
+ "epoch": 600.995983935743,
5
+ "global_step": 74539,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
193410
  "eval_steps_per_second": 0.69,
193411
  "eval_wer": 0.1899564928840056,
193412
  "step": 73916
193413
+ },
193414
+ {
193415
+ "epoch": 596.03,
193416
+ "learning_rate": 8.831875000000001e-06,
193417
+ "loss": 0.4505,
193418
+ "step": 73920
193419
+ },
193420
+ {
193421
+ "epoch": 596.07,
193422
+ "learning_rate": 8.831794871794872e-06,
193423
+ "loss": 0.3225,
193424
+ "step": 73925
193425
+ },
193426
+ {
193427
+ "epoch": 596.11,
193428
+ "learning_rate": 8.831714743589744e-06,
193429
+ "loss": 0.3288,
193430
+ "step": 73930
193431
+ },
193432
+ {
193433
+ "epoch": 596.15,
193434
+ "learning_rate": 8.831634615384617e-06,
193435
+ "loss": 0.413,
193436
+ "step": 73935
193437
+ },
193438
+ {
193439
+ "epoch": 596.19,
193440
+ "learning_rate": 8.831554487179488e-06,
193441
+ "loss": 0.8731,
193442
+ "step": 73940
193443
+ },
193444
+ {
193445
+ "epoch": 596.23,
193446
+ "learning_rate": 8.83147435897436e-06,
193447
+ "loss": 0.6002,
193448
+ "step": 73945
193449
+ },
193450
+ {
193451
+ "epoch": 596.27,
193452
+ "learning_rate": 8.83139423076923e-06,
193453
+ "loss": 0.3424,
193454
+ "step": 73950
193455
+ },
193456
+ {
193457
+ "epoch": 596.31,
193458
+ "learning_rate": 8.831314102564104e-06,
193459
+ "loss": 0.3209,
193460
+ "step": 73955
193461
+ },
193462
+ {
193463
+ "epoch": 596.35,
193464
+ "learning_rate": 8.831233974358975e-06,
193465
+ "loss": 0.4331,
193466
+ "step": 73960
193467
+ },
193468
+ {
193469
+ "epoch": 596.39,
193470
+ "learning_rate": 8.831153846153846e-06,
193471
+ "loss": 0.8799,
193472
+ "step": 73965
193473
+ },
193474
+ {
193475
+ "epoch": 596.43,
193476
+ "learning_rate": 8.831073717948718e-06,
193477
+ "loss": 0.7031,
193478
+ "step": 73970
193479
+ },
193480
+ {
193481
+ "epoch": 596.47,
193482
+ "learning_rate": 8.830993589743591e-06,
193483
+ "loss": 0.3835,
193484
+ "step": 73975
193485
+ },
193486
+ {
193487
+ "epoch": 596.51,
193488
+ "learning_rate": 8.830913461538462e-06,
193489
+ "loss": 0.2978,
193490
+ "step": 73980
193491
+ },
193492
+ {
193493
+ "epoch": 596.55,
193494
+ "learning_rate": 8.830833333333334e-06,
193495
+ "loss": 0.3954,
193496
+ "step": 73985
193497
+ },
193498
+ {
193499
+ "epoch": 596.59,
193500
+ "learning_rate": 8.830753205128207e-06,
193501
+ "loss": 0.8978,
193502
+ "step": 73990
193503
+ },
193504
+ {
193505
+ "epoch": 596.63,
193506
+ "learning_rate": 8.830673076923078e-06,
193507
+ "loss": 0.6776,
193508
+ "step": 73995
193509
+ },
193510
+ {
193511
+ "epoch": 596.67,
193512
+ "learning_rate": 8.83059294871795e-06,
193513
+ "loss": 0.3265,
193514
+ "step": 74000
193515
+ },
193516
+ {
193517
+ "epoch": 596.71,
193518
+ "learning_rate": 8.83051282051282e-06,
193519
+ "loss": 0.3834,
193520
+ "step": 74005
193521
+ },
193522
+ {
193523
+ "epoch": 596.76,
193524
+ "learning_rate": 8.830432692307694e-06,
193525
+ "loss": 0.366,
193526
+ "step": 74010
193527
+ },
193528
+ {
193529
+ "epoch": 596.8,
193530
+ "learning_rate": 8.830352564102565e-06,
193531
+ "loss": 0.9071,
193532
+ "step": 74015
193533
+ },
193534
+ {
193535
+ "epoch": 596.84,
193536
+ "learning_rate": 8.830272435897437e-06,
193537
+ "loss": 0.6363,
193538
+ "step": 74020
193539
+ },
193540
+ {
193541
+ "epoch": 596.88,
193542
+ "learning_rate": 8.830192307692308e-06,
193543
+ "loss": 0.3411,
193544
+ "step": 74025
193545
+ },
193546
+ {
193547
+ "epoch": 596.92,
193548
+ "learning_rate": 8.830112179487181e-06,
193549
+ "loss": 0.3157,
193550
+ "step": 74030
193551
+ },
193552
+ {
193553
+ "epoch": 596.96,
193554
+ "learning_rate": 8.830032051282052e-06,
193555
+ "loss": 0.3928,
193556
+ "step": 74035
193557
+ },
193558
+ {
193559
+ "epoch": 597.0,
193560
+ "learning_rate": 8.829951923076924e-06,
193561
+ "loss": 1.1771,
193562
+ "step": 74040
193563
+ },
193564
+ {
193565
+ "epoch": 597.0,
193566
+ "eval_loss": 0.5054441690444946,
193567
+ "eval_runtime": 40.5036,
193568
+ "eval_samples_per_second": 20.739,
193569
+ "eval_steps_per_second": 0.667,
193570
+ "eval_wer": 0.20609675281643472,
193571
+ "step": 74040
193572
+ },
193573
+ {
193574
+ "epoch": 592.04,
193575
+ "learning_rate": 8.829871794871797e-06,
193576
+ "loss": 0.3384,
193577
+ "step": 74045
193578
+ },
193579
+ {
193580
+ "epoch": 592.08,
193581
+ "learning_rate": 8.829791666666666e-06,
193582
+ "loss": 0.3687,
193583
+ "step": 74050
193584
+ },
193585
+ {
193586
+ "epoch": 592.12,
193587
+ "learning_rate": 8.82971153846154e-06,
193588
+ "loss": 0.3776,
193589
+ "step": 74055
193590
+ },
193591
+ {
193592
+ "epoch": 592.16,
193593
+ "learning_rate": 8.82963141025641e-06,
193594
+ "loss": 0.4675,
193595
+ "step": 74060
193596
+ },
193597
+ {
193598
+ "epoch": 592.2,
193599
+ "learning_rate": 8.829551282051282e-06,
193600
+ "loss": 1.0907,
193601
+ "step": 74065
193602
+ },
193603
+ {
193604
+ "epoch": 592.24,
193605
+ "learning_rate": 8.829471153846153e-06,
193606
+ "loss": 0.4606,
193607
+ "step": 74070
193608
+ },
193609
+ {
193610
+ "epoch": 592.28,
193611
+ "learning_rate": 8.829391025641027e-06,
193612
+ "loss": 0.2879,
193613
+ "step": 74075
193614
+ },
193615
+ {
193616
+ "epoch": 592.32,
193617
+ "learning_rate": 8.829310897435898e-06,
193618
+ "loss": 0.3348,
193619
+ "step": 74080
193620
+ },
193621
+ {
193622
+ "epoch": 592.36,
193623
+ "learning_rate": 8.82923076923077e-06,
193624
+ "loss": 0.5071,
193625
+ "step": 74085
193626
+ },
193627
+ {
193628
+ "epoch": 592.4,
193629
+ "learning_rate": 8.829150641025642e-06,
193630
+ "loss": 1.1857,
193631
+ "step": 74090
193632
+ },
193633
+ {
193634
+ "epoch": 592.44,
193635
+ "learning_rate": 8.829070512820514e-06,
193636
+ "loss": 0.3519,
193637
+ "step": 74095
193638
+ },
193639
+ {
193640
+ "epoch": 592.48,
193641
+ "learning_rate": 8.828990384615385e-06,
193642
+ "loss": 0.2731,
193643
+ "step": 74100
193644
+ },
193645
+ {
193646
+ "epoch": 592.52,
193647
+ "learning_rate": 8.828910256410256e-06,
193648
+ "loss": 0.3082,
193649
+ "step": 74105
193650
+ },
193651
+ {
193652
+ "epoch": 592.56,
193653
+ "learning_rate": 8.82883012820513e-06,
193654
+ "loss": 0.4305,
193655
+ "step": 74110
193656
+ },
193657
+ {
193658
+ "epoch": 592.6,
193659
+ "learning_rate": 8.82875e-06,
193660
+ "loss": 1.1372,
193661
+ "step": 74115
193662
+ },
193663
+ {
193664
+ "epoch": 592.64,
193665
+ "learning_rate": 8.828669871794872e-06,
193666
+ "loss": 0.3594,
193667
+ "step": 74120
193668
+ },
193669
+ {
193670
+ "epoch": 592.68,
193671
+ "learning_rate": 8.828589743589744e-06,
193672
+ "loss": 0.3154,
193673
+ "step": 74125
193674
+ },
193675
+ {
193676
+ "epoch": 592.72,
193677
+ "learning_rate": 8.828509615384617e-06,
193678
+ "loss": 0.3338,
193679
+ "step": 74130
193680
+ },
193681
+ {
193682
+ "epoch": 592.76,
193683
+ "learning_rate": 8.828429487179488e-06,
193684
+ "loss": 0.5104,
193685
+ "step": 74135
193686
+ },
193687
+ {
193688
+ "epoch": 592.8,
193689
+ "learning_rate": 8.82834935897436e-06,
193690
+ "loss": 1.0787,
193691
+ "step": 74140
193692
+ },
193693
+ {
193694
+ "epoch": 592.84,
193695
+ "learning_rate": 8.828269230769232e-06,
193696
+ "loss": 0.3784,
193697
+ "step": 74145
193698
+ },
193699
+ {
193700
+ "epoch": 592.88,
193701
+ "learning_rate": 8.828189102564104e-06,
193702
+ "loss": 0.2921,
193703
+ "step": 74150
193704
+ },
193705
+ {
193706
+ "epoch": 592.92,
193707
+ "learning_rate": 8.828108974358975e-06,
193708
+ "loss": 0.3931,
193709
+ "step": 74155
193710
+ },
193711
+ {
193712
+ "epoch": 592.96,
193713
+ "learning_rate": 8.828028846153846e-06,
193714
+ "loss": 0.5299,
193715
+ "step": 74160
193716
+ },
193717
+ {
193718
+ "epoch": 593.0,
193719
+ "learning_rate": 8.82794871794872e-06,
193720
+ "loss": 1.2197,
193721
+ "step": 74165
193722
+ },
193723
+ {
193724
+ "epoch": 593.0,
193725
+ "eval_loss": 0.4171510636806488,
193726
+ "eval_runtime": 40.75,
193727
+ "eval_samples_per_second": 20.638,
193728
+ "eval_steps_per_second": 0.663,
193729
+ "eval_wer": 0.20073123521399383,
193730
+ "step": 74165
193731
+ },
193732
+ {
193733
+ "epoch": 593.04,
193734
+ "learning_rate": 8.82786858974359e-06,
193735
+ "loss": 0.3557,
193736
+ "step": 74170
193737
+ },
193738
+ {
193739
+ "epoch": 593.08,
193740
+ "learning_rate": 8.827788461538462e-06,
193741
+ "loss": 0.3218,
193742
+ "step": 74175
193743
+ },
193744
+ {
193745
+ "epoch": 593.12,
193746
+ "learning_rate": 8.827708333333334e-06,
193747
+ "loss": 0.3495,
193748
+ "step": 74180
193749
+ },
193750
+ {
193751
+ "epoch": 593.16,
193752
+ "learning_rate": 8.827628205128207e-06,
193753
+ "loss": 0.5121,
193754
+ "step": 74185
193755
+ },
193756
+ {
193757
+ "epoch": 593.2,
193758
+ "learning_rate": 8.827548076923078e-06,
193759
+ "loss": 1.1842,
193760
+ "step": 74190
193761
+ },
193762
+ {
193763
+ "epoch": 593.24,
193764
+ "learning_rate": 8.82746794871795e-06,
193765
+ "loss": 0.3834,
193766
+ "step": 74195
193767
+ },
193768
+ {
193769
+ "epoch": 593.28,
193770
+ "learning_rate": 8.827387820512822e-06,
193771
+ "loss": 0.3191,
193772
+ "step": 74200
193773
+ },
193774
+ {
193775
+ "epoch": 593.32,
193776
+ "learning_rate": 8.827307692307692e-06,
193777
+ "loss": 0.3646,
193778
+ "step": 74205
193779
+ },
193780
+ {
193781
+ "epoch": 593.36,
193782
+ "learning_rate": 8.827227564102565e-06,
193783
+ "loss": 0.4677,
193784
+ "step": 74210
193785
+ },
193786
+ {
193787
+ "epoch": 593.4,
193788
+ "learning_rate": 8.827147435897436e-06,
193789
+ "loss": 1.2346,
193790
+ "step": 74215
193791
+ },
193792
+ {
193793
+ "epoch": 593.44,
193794
+ "learning_rate": 8.827067307692308e-06,
193795
+ "loss": 0.3663,
193796
+ "step": 74220
193797
+ },
193798
+ {
193799
+ "epoch": 593.48,
193800
+ "learning_rate": 8.82698717948718e-06,
193801
+ "loss": 0.2893,
193802
+ "step": 74225
193803
+ },
193804
+ {
193805
+ "epoch": 593.52,
193806
+ "learning_rate": 8.826907051282052e-06,
193807
+ "loss": 0.334,
193808
+ "step": 74230
193809
+ },
193810
+ {
193811
+ "epoch": 593.56,
193812
+ "learning_rate": 8.826826923076924e-06,
193813
+ "loss": 0.5513,
193814
+ "step": 74235
193815
+ },
193816
+ {
193817
+ "epoch": 593.6,
193818
+ "learning_rate": 8.826746794871795e-06,
193819
+ "loss": 1.281,
193820
+ "step": 74240
193821
+ },
193822
+ {
193823
+ "epoch": 593.64,
193824
+ "learning_rate": 8.826666666666668e-06,
193825
+ "loss": 0.3202,
193826
+ "step": 74245
193827
+ },
193828
+ {
193829
+ "epoch": 593.68,
193830
+ "learning_rate": 8.82658653846154e-06,
193831
+ "loss": 0.3076,
193832
+ "step": 74250
193833
+ },
193834
+ {
193835
+ "epoch": 593.72,
193836
+ "learning_rate": 8.82650641025641e-06,
193837
+ "loss": 0.3823,
193838
+ "step": 74255
193839
+ },
193840
+ {
193841
+ "epoch": 593.76,
193842
+ "learning_rate": 8.826426282051282e-06,
193843
+ "loss": 0.4498,
193844
+ "step": 74260
193845
+ },
193846
+ {
193847
+ "epoch": 593.8,
193848
+ "learning_rate": 8.826346153846155e-06,
193849
+ "loss": 1.3628,
193850
+ "step": 74265
193851
+ },
193852
+ {
193853
+ "epoch": 593.84,
193854
+ "learning_rate": 8.826266025641026e-06,
193855
+ "loss": 0.3403,
193856
+ "step": 74270
193857
+ },
193858
+ {
193859
+ "epoch": 593.88,
193860
+ "learning_rate": 8.826185897435898e-06,
193861
+ "loss": 0.3188,
193862
+ "step": 74275
193863
+ },
193864
+ {
193865
+ "epoch": 593.92,
193866
+ "learning_rate": 8.82610576923077e-06,
193867
+ "loss": 0.3402,
193868
+ "step": 74280
193869
+ },
193870
+ {
193871
+ "epoch": 593.96,
193872
+ "learning_rate": 8.826025641025642e-06,
193873
+ "loss": 0.486,
193874
+ "step": 74285
193875
+ },
193876
+ {
193877
+ "epoch": 594.0,
193878
+ "learning_rate": 8.825945512820514e-06,
193879
+ "loss": 1.311,
193880
+ "step": 74290
193881
+ },
193882
+ {
193883
+ "epoch": 594.0,
193884
+ "eval_loss": 0.35683706402778625,
193885
+ "eval_runtime": 40.5555,
193886
+ "eval_samples_per_second": 20.737,
193887
+ "eval_steps_per_second": 0.666,
193888
+ "eval_wer": 0.19015243684248193,
193889
+ "step": 74290
193890
+ },
193891
+ {
193892
+ "epoch": 594.04,
193893
+ "learning_rate": 8.825865384615385e-06,
193894
+ "loss": 0.358,
193895
+ "step": 74295
193896
+ },
193897
+ {
193898
+ "epoch": 594.08,
193899
+ "learning_rate": 8.825785256410258e-06,
193900
+ "loss": 0.2652,
193901
+ "step": 74300
193902
+ },
193903
+ {
193904
+ "epoch": 594.12,
193905
+ "learning_rate": 8.82570512820513e-06,
193906
+ "loss": 0.3892,
193907
+ "step": 74305
193908
+ },
193909
+ {
193910
+ "epoch": 594.16,
193911
+ "learning_rate": 8.825625e-06,
193912
+ "loss": 0.4418,
193913
+ "step": 74310
193914
+ },
193915
+ {
193916
+ "epoch": 594.2,
193917
+ "learning_rate": 8.825544871794872e-06,
193918
+ "loss": 1.2462,
193919
+ "step": 74315
193920
+ },
193921
+ {
193922
+ "epoch": 594.24,
193923
+ "learning_rate": 8.825464743589745e-06,
193924
+ "loss": 0.3345,
193925
+ "step": 74320
193926
+ },
193927
+ {
193928
+ "epoch": 594.28,
193929
+ "learning_rate": 8.825384615384617e-06,
193930
+ "loss": 0.2987,
193931
+ "step": 74325
193932
+ },
193933
+ {
193934
+ "epoch": 594.32,
193935
+ "learning_rate": 8.825304487179488e-06,
193936
+ "loss": 0.2656,
193937
+ "step": 74330
193938
+ },
193939
+ {
193940
+ "epoch": 594.36,
193941
+ "learning_rate": 8.825224358974361e-06,
193942
+ "loss": 0.5242,
193943
+ "step": 74335
193944
+ },
193945
+ {
193946
+ "epoch": 594.4,
193947
+ "learning_rate": 8.825144230769232e-06,
193948
+ "loss": 1.2575,
193949
+ "step": 74340
193950
+ },
193951
+ {
193952
+ "epoch": 594.44,
193953
+ "learning_rate": 8.825064102564104e-06,
193954
+ "loss": 0.3369,
193955
+ "step": 74345
193956
+ },
193957
+ {
193958
+ "epoch": 594.48,
193959
+ "learning_rate": 8.824983974358975e-06,
193960
+ "loss": 0.305,
193961
+ "step": 74350
193962
+ },
193963
+ {
193964
+ "epoch": 594.52,
193965
+ "learning_rate": 8.824903846153848e-06,
193966
+ "loss": 0.3236,
193967
+ "step": 74355
193968
+ },
193969
+ {
193970
+ "epoch": 594.56,
193971
+ "learning_rate": 8.824823717948718e-06,
193972
+ "loss": 0.4855,
193973
+ "step": 74360
193974
+ },
193975
+ {
193976
+ "epoch": 594.6,
193977
+ "learning_rate": 8.82474358974359e-06,
193978
+ "loss": 1.2836,
193979
+ "step": 74365
193980
+ },
193981
+ {
193982
+ "epoch": 594.64,
193983
+ "learning_rate": 8.824663461538462e-06,
193984
+ "loss": 0.3867,
193985
+ "step": 74370
193986
+ },
193987
+ {
193988
+ "epoch": 594.68,
193989
+ "learning_rate": 8.824583333333333e-06,
193990
+ "loss": 0.3239,
193991
+ "step": 74375
193992
+ },
193993
+ {
193994
+ "epoch": 594.72,
193995
+ "learning_rate": 8.824503205128205e-06,
193996
+ "loss": 0.4218,
193997
+ "step": 74380
193998
+ },
193999
+ {
194000
+ "epoch": 594.76,
194001
+ "learning_rate": 8.824423076923078e-06,
194002
+ "loss": 0.5141,
194003
+ "step": 74385
194004
+ },
194005
+ {
194006
+ "epoch": 594.8,
194007
+ "learning_rate": 8.82434294871795e-06,
194008
+ "loss": 1.3074,
194009
+ "step": 74390
194010
+ },
194011
+ {
194012
+ "epoch": 594.84,
194013
+ "learning_rate": 8.82426282051282e-06,
194014
+ "loss": 0.3945,
194015
+ "step": 74395
194016
+ },
194017
+ {
194018
+ "epoch": 594.88,
194019
+ "learning_rate": 8.824182692307694e-06,
194020
+ "loss": 0.3776,
194021
+ "step": 74400
194022
+ },
194023
+ {
194024
+ "epoch": 594.92,
194025
+ "learning_rate": 8.824102564102565e-06,
194026
+ "loss": 0.3246,
194027
+ "step": 74405
194028
+ },
194029
+ {
194030
+ "epoch": 594.96,
194031
+ "learning_rate": 8.824022435897436e-06,
194032
+ "loss": 0.4568,
194033
+ "step": 74410
194034
+ },
194035
+ {
194036
+ "epoch": 595.0,
194037
+ "learning_rate": 8.823942307692308e-06,
194038
+ "loss": 1.5204,
194039
+ "step": 74415
194040
+ },
194041
+ {
194042
+ "epoch": 595.0,
194043
+ "eval_loss": 0.437023401260376,
194044
+ "eval_runtime": 39.7077,
194045
+ "eval_samples_per_second": 21.18,
194046
+ "eval_steps_per_second": 0.68,
194047
+ "eval_wer": 0.1853523587652883,
194048
+ "step": 74415
194049
+ },
194050
+ {
194051
+ "epoch": 600.04,
194052
+ "learning_rate": 8.82386217948718e-06,
194053
+ "loss": 0.2979,
194054
+ "step": 74420
194055
+ },
194056
+ {
194057
+ "epoch": 600.08,
194058
+ "learning_rate": 8.823782051282052e-06,
194059
+ "loss": 0.2777,
194060
+ "step": 74425
194061
+ },
194062
+ {
194063
+ "epoch": 600.12,
194064
+ "learning_rate": 8.823701923076924e-06,
194065
+ "loss": 0.3034,
194066
+ "step": 74430
194067
+ },
194068
+ {
194069
+ "epoch": 600.16,
194070
+ "learning_rate": 8.823621794871795e-06,
194071
+ "loss": 0.4698,
194072
+ "step": 74435
194073
+ },
194074
+ {
194075
+ "epoch": 600.2,
194076
+ "learning_rate": 8.823541666666668e-06,
194077
+ "loss": 1.2657,
194078
+ "step": 74440
194079
+ },
194080
+ {
194081
+ "epoch": 600.24,
194082
+ "learning_rate": 8.82346153846154e-06,
194083
+ "loss": 0.3598,
194084
+ "step": 74445
194085
+ },
194086
+ {
194087
+ "epoch": 600.28,
194088
+ "learning_rate": 8.82338141025641e-06,
194089
+ "loss": 0.3204,
194090
+ "step": 74450
194091
+ },
194092
+ {
194093
+ "epoch": 600.32,
194094
+ "learning_rate": 8.823301282051284e-06,
194095
+ "loss": 0.3404,
194096
+ "step": 74455
194097
+ },
194098
+ {
194099
+ "epoch": 600.36,
194100
+ "learning_rate": 8.823221153846155e-06,
194101
+ "loss": 0.4718,
194102
+ "step": 74460
194103
+ },
194104
+ {
194105
+ "epoch": 600.4,
194106
+ "learning_rate": 8.823141025641026e-06,
194107
+ "loss": 1.1487,
194108
+ "step": 74465
194109
+ },
194110
+ {
194111
+ "epoch": 600.44,
194112
+ "learning_rate": 8.823060897435898e-06,
194113
+ "loss": 0.337,
194114
+ "step": 74470
194115
+ },
194116
+ {
194117
+ "epoch": 600.48,
194118
+ "learning_rate": 8.82298076923077e-06,
194119
+ "loss": 0.2972,
194120
+ "step": 74475
194121
+ },
194122
+ {
194123
+ "epoch": 600.52,
194124
+ "learning_rate": 8.82290064102564e-06,
194125
+ "loss": 0.3133,
194126
+ "step": 74480
194127
+ },
194128
+ {
194129
+ "epoch": 600.56,
194130
+ "learning_rate": 8.822820512820514e-06,
194131
+ "loss": 0.5256,
194132
+ "step": 74485
194133
+ },
194134
+ {
194135
+ "epoch": 600.6,
194136
+ "learning_rate": 8.822740384615387e-06,
194137
+ "loss": 1.1288,
194138
+ "step": 74490
194139
+ },
194140
+ {
194141
+ "epoch": 600.64,
194142
+ "learning_rate": 8.822660256410256e-06,
194143
+ "loss": 0.3383,
194144
+ "step": 74495
194145
+ },
194146
+ {
194147
+ "epoch": 600.68,
194148
+ "learning_rate": 8.82258012820513e-06,
194149
+ "loss": 0.3663,
194150
+ "step": 74500
194151
+ },
194152
+ {
194153
+ "epoch": 600.72,
194154
+ "learning_rate": 8.8225e-06,
194155
+ "loss": 0.5953,
194156
+ "step": 74505
194157
+ },
194158
+ {
194159
+ "epoch": 600.76,
194160
+ "learning_rate": 8.822419871794872e-06,
194161
+ "loss": 0.4843,
194162
+ "step": 74510
194163
+ },
194164
+ {
194165
+ "epoch": 600.8,
194166
+ "learning_rate": 8.822339743589743e-06,
194167
+ "loss": 1.1774,
194168
+ "step": 74515
194169
+ },
194170
+ {
194171
+ "epoch": 600.84,
194172
+ "learning_rate": 8.822259615384616e-06,
194173
+ "loss": 0.3648,
194174
+ "step": 74520
194175
+ },
194176
+ {
194177
+ "epoch": 600.88,
194178
+ "learning_rate": 8.822179487179488e-06,
194179
+ "loss": 0.312,
194180
+ "step": 74525
194181
+ },
194182
+ {
194183
+ "epoch": 600.92,
194184
+ "learning_rate": 8.822099358974359e-06,
194185
+ "loss": 0.3538,
194186
+ "step": 74530
194187
+ },
194188
+ {
194189
+ "epoch": 600.96,
194190
+ "learning_rate": 8.82201923076923e-06,
194191
+ "loss": 0.4753,
194192
+ "step": 74535
194193
+ },
194194
+ {
194195
+ "epoch": 601.0,
194196
+ "eval_loss": 0.4208657741546631,
194197
+ "eval_runtime": 40.5395,
194198
+ "eval_samples_per_second": 20.745,
194199
+ "eval_steps_per_second": 0.666,
194200
+ "eval_wer": 0.18913761467889909,
194201
+ "step": 74539
194202
  }
194203
  ],
194204
  "max_steps": 620000,
194205
  "num_train_epochs": 5000,
194206
+ "total_flos": 2.097692339430875e+20,
194207
  "trial_name": null,
194208
  "trial_params": null
194209
  }
model-bin/finetune/base/{checkpoint-73916 β†’ checkpoint-74539}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629875783.5501642/events.out.tfevents.1629875783.7e498afd5545.905.73 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56464a81e08d7f1a82ade83b736f0756eb87693e0ddad64d889d382198e563fd
3
+ size 4194
model-bin/finetune/base/log/1629876453.8615327/events.out.tfevents.1629876453.7e498afd5545.905.75 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be4bfbaec62c29f516a7e9a9d3917649f8fe892178caed42353cf30289db3e17
3
+ size 4194
model-bin/finetune/base/log/1629877099.8360853/events.out.tfevents.1629877099.7e498afd5545.905.77 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:645beaf58ccf8e567b5e37c2c291607674188f19f50f3e7ee43210e6407638dd
3
+ size 4194
model-bin/finetune/base/log/1629877746.4587185/events.out.tfevents.1629877746.7e498afd5545.905.79 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f60338304f297ccf1a1c0d050fa5246f8e8acbd98bdea98b1eef75c28f7d1170
3
+ size 4194
model-bin/finetune/base/log/1629878403.3794868/events.out.tfevents.1629878403.7e498afd5545.905.81 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df1922471be3de7f9dd9e4f2a4714803af3f9197b17624b7f01328af60d370dc
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629875783.7e498afd5545.905.72 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77271c3153a042764056f6227bdbbc65ec6f3766b6800dfffce708dfdd9b7f0c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629876453.7e498afd5545.905.74 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30d0a63e279858ca5f738732cff0f1f32c7986714dad24fba8428eb4bd16a122
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629877099.7e498afd5545.905.76 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:709a794ebfbf29355911f836c4c1181d57b07ab751c07e59e543d2171a1932a6
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629877746.7e498afd5545.905.78 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32769bd5c161b9f9e3573e58a8963e15f99507303208c12bcc0e1745c936edc3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629878403.7e498afd5545.905.80 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9de9c1b83cc90e427f15e980270464cf53514719155c0e57b3b2b6f8df1b48f1
3
+ size 8462