Check commited on
Commit
bb2eb52
Β·
1 Parent(s): d11468a

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630115798.4136324/events.out.tfevents.1630115798.86bb0ddabf9b.1042.41 +3 -0
  11. model-bin/finetune/base/log/1630116207.9281766/events.out.tfevents.1630116207.86bb0ddabf9b.1042.43 +3 -0
  12. model-bin/finetune/base/log/1630116621.3491957/events.out.tfevents.1630116621.86bb0ddabf9b.1042.45 +3 -0
  13. model-bin/finetune/base/log/1630117036.3824139/events.out.tfevents.1630117036.86bb0ddabf9b.1042.47 +3 -0
  14. model-bin/finetune/base/log/1630117449.777196/events.out.tfevents.1630117449.86bb0ddabf9b.1042.49 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630115798.86bb0ddabf9b.1042.40 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630116207.86bb0ddabf9b.1042.42 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630116621.86bb0ddabf9b.1042.44 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630117036.86bb0ddabf9b.1042.46 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630117449.86bb0ddabf9b.1042.48 +3 -0
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed6a4a1b60032720f1c3b30726a6595b0807d4670f546ca8940edf524a4e7cd6
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce19a2e5b74de571f58c07d682570d2210d3daa6eb857675fefc28ee8ffe8cfe
3
  size 722165393
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2bf7f96504e1c0a6174b9f2d5915d9873c10b1506cc00abe352873e28517867
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe6f837af1e6098ca87218f9f6f489ef7326656822686625f997aa9e4da06051
3
  size 377909911
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:796e263a25f4a0a7434c18d8024159eaa3656ad1fa20d7a25b7eae15f6439f16
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:310447b7017194990c17a691e647cf4393cfa4912178125648656448f89be377
3
+ size 14567
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8ec6882b7592c0ac7b9cd7d7fe3da25ac3c6fd19100a89c73fb078e31a1ddc3
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1caec4dc20dfb31b5977c6031ec782c60b5f7b286f94be7ec1097ec7544efc9
3
  size 559
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1cd22da4f6bd733f8bb533d0e36af7533566eafa7ac974fe7b8b64dcc9e30a39
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:747fd4894013b519f10367353486f379ccd36aaffe81e92b6be8bd41396f7093
3
  size 623
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 931.0,
5
- "global_step": 116478,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -247560,11 +247560,806 @@
247560
  "eval_steps_per_second": 0.686,
247561
  "eval_wer": 0.18822154026011984,
247562
  "step": 116478
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
247563
  }
247564
  ],
247565
  "max_steps": 625000,
247566
  "num_train_epochs": 5000,
247567
- "total_flos": 3.2779879386977056e+20,
247568
  "trial_name": null,
247569
  "trial_params": null
247570
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 935.9960159362549,
5
+ "global_step": 117100,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
247560
  "eval_steps_per_second": 0.686,
247561
  "eval_wer": 0.18822154026011984,
247562
  "step": 116478
247563
+ },
247564
+ {
247565
+ "epoch": 931.02,
247566
+ "learning_rate": 8.15011217948718e-06,
247567
+ "loss": 0.2933,
247568
+ "step": 116480
247569
+ },
247570
+ {
247571
+ "epoch": 931.06,
247572
+ "learning_rate": 8.150032051282052e-06,
247573
+ "loss": 0.292,
247574
+ "step": 116485
247575
+ },
247576
+ {
247577
+ "epoch": 931.1,
247578
+ "learning_rate": 8.149951923076923e-06,
247579
+ "loss": 0.2725,
247580
+ "step": 116490
247581
+ },
247582
+ {
247583
+ "epoch": 931.14,
247584
+ "learning_rate": 8.149871794871795e-06,
247585
+ "loss": 0.383,
247586
+ "step": 116495
247587
+ },
247588
+ {
247589
+ "epoch": 931.18,
247590
+ "learning_rate": 8.149791666666668e-06,
247591
+ "loss": 0.6222,
247592
+ "step": 116500
247593
+ },
247594
+ {
247595
+ "epoch": 931.22,
247596
+ "learning_rate": 8.14971153846154e-06,
247597
+ "loss": 0.9348,
247598
+ "step": 116505
247599
+ },
247600
+ {
247601
+ "epoch": 931.26,
247602
+ "learning_rate": 8.14963141025641e-06,
247603
+ "loss": 0.2812,
247604
+ "step": 116510
247605
+ },
247606
+ {
247607
+ "epoch": 931.3,
247608
+ "learning_rate": 8.149551282051282e-06,
247609
+ "loss": 0.2711,
247610
+ "step": 116515
247611
+ },
247612
+ {
247613
+ "epoch": 931.34,
247614
+ "learning_rate": 8.149471153846155e-06,
247615
+ "loss": 0.4673,
247616
+ "step": 116520
247617
+ },
247618
+ {
247619
+ "epoch": 931.38,
247620
+ "learning_rate": 8.149391025641026e-06,
247621
+ "loss": 0.7297,
247622
+ "step": 116525
247623
+ },
247624
+ {
247625
+ "epoch": 931.42,
247626
+ "learning_rate": 8.149310897435898e-06,
247627
+ "loss": 1.0439,
247628
+ "step": 116530
247629
+ },
247630
+ {
247631
+ "epoch": 931.46,
247632
+ "learning_rate": 8.14923076923077e-06,
247633
+ "loss": 0.3004,
247634
+ "step": 116535
247635
+ },
247636
+ {
247637
+ "epoch": 931.5,
247638
+ "learning_rate": 8.149150641025642e-06,
247639
+ "loss": 0.289,
247640
+ "step": 116540
247641
+ },
247642
+ {
247643
+ "epoch": 931.54,
247644
+ "learning_rate": 8.149070512820513e-06,
247645
+ "loss": 0.532,
247646
+ "step": 116545
247647
+ },
247648
+ {
247649
+ "epoch": 931.58,
247650
+ "learning_rate": 8.148990384615385e-06,
247651
+ "loss": 0.6062,
247652
+ "step": 116550
247653
+ },
247654
+ {
247655
+ "epoch": 931.62,
247656
+ "learning_rate": 8.148910256410258e-06,
247657
+ "loss": 1.037,
247658
+ "step": 116555
247659
+ },
247660
+ {
247661
+ "epoch": 931.66,
247662
+ "learning_rate": 8.14883012820513e-06,
247663
+ "loss": 0.2608,
247664
+ "step": 116560
247665
+ },
247666
+ {
247667
+ "epoch": 931.7,
247668
+ "learning_rate": 8.14875e-06,
247669
+ "loss": 0.2704,
247670
+ "step": 116565
247671
+ },
247672
+ {
247673
+ "epoch": 931.74,
247674
+ "learning_rate": 8.148669871794872e-06,
247675
+ "loss": 0.3608,
247676
+ "step": 116570
247677
+ },
247678
+ {
247679
+ "epoch": 931.78,
247680
+ "learning_rate": 8.148589743589745e-06,
247681
+ "loss": 0.615,
247682
+ "step": 116575
247683
+ },
247684
+ {
247685
+ "epoch": 931.82,
247686
+ "learning_rate": 8.148509615384616e-06,
247687
+ "loss": 0.9766,
247688
+ "step": 116580
247689
+ },
247690
+ {
247691
+ "epoch": 931.86,
247692
+ "learning_rate": 8.148429487179488e-06,
247693
+ "loss": 0.3059,
247694
+ "step": 116585
247695
+ },
247696
+ {
247697
+ "epoch": 931.9,
247698
+ "learning_rate": 8.14834935897436e-06,
247699
+ "loss": 0.2919,
247700
+ "step": 116590
247701
+ },
247702
+ {
247703
+ "epoch": 931.94,
247704
+ "learning_rate": 8.148269230769232e-06,
247705
+ "loss": 0.3559,
247706
+ "step": 116595
247707
+ },
247708
+ {
247709
+ "epoch": 931.98,
247710
+ "learning_rate": 8.148189102564103e-06,
247711
+ "loss": 0.7633,
247712
+ "step": 116600
247713
+ },
247714
+ {
247715
+ "epoch": 932.0,
247716
+ "eval_loss": 0.35194557905197144,
247717
+ "eval_runtime": 39.5509,
247718
+ "eval_samples_per_second": 21.314,
247719
+ "eval_steps_per_second": 0.683,
247720
+ "eval_wer": 0.18437974134816848,
247721
+ "step": 116603
247722
+ },
247723
+ {
247724
+ "epoch": 940.02,
247725
+ "learning_rate": 8.148108974358975e-06,
247726
+ "loss": 0.3737,
247727
+ "step": 116605
247728
+ },
247729
+ {
247730
+ "epoch": 940.06,
247731
+ "learning_rate": 8.148028846153848e-06,
247732
+ "loss": 0.2789,
247733
+ "step": 116610
247734
+ },
247735
+ {
247736
+ "epoch": 940.1,
247737
+ "learning_rate": 8.147948717948718e-06,
247738
+ "loss": 0.2851,
247739
+ "step": 116615
247740
+ },
247741
+ {
247742
+ "epoch": 940.14,
247743
+ "learning_rate": 8.14786858974359e-06,
247744
+ "loss": 0.3367,
247745
+ "step": 116620
247746
+ },
247747
+ {
247748
+ "epoch": 940.18,
247749
+ "learning_rate": 8.147788461538462e-06,
247750
+ "loss": 0.679,
247751
+ "step": 116625
247752
+ },
247753
+ {
247754
+ "epoch": 940.22,
247755
+ "learning_rate": 8.147708333333333e-06,
247756
+ "loss": 0.9981,
247757
+ "step": 116630
247758
+ },
247759
+ {
247760
+ "epoch": 940.26,
247761
+ "learning_rate": 8.147628205128206e-06,
247762
+ "loss": 0.3318,
247763
+ "step": 116635
247764
+ },
247765
+ {
247766
+ "epoch": 940.3,
247767
+ "learning_rate": 8.147548076923078e-06,
247768
+ "loss": 0.3239,
247769
+ "step": 116640
247770
+ },
247771
+ {
247772
+ "epoch": 940.34,
247773
+ "learning_rate": 8.147467948717949e-06,
247774
+ "loss": 0.3073,
247775
+ "step": 116645
247776
+ },
247777
+ {
247778
+ "epoch": 940.38,
247779
+ "learning_rate": 8.14738782051282e-06,
247780
+ "loss": 0.6717,
247781
+ "step": 116650
247782
+ },
247783
+ {
247784
+ "epoch": 940.42,
247785
+ "learning_rate": 8.147307692307693e-06,
247786
+ "loss": 1.1175,
247787
+ "step": 116655
247788
+ },
247789
+ {
247790
+ "epoch": 940.46,
247791
+ "learning_rate": 8.147227564102565e-06,
247792
+ "loss": 0.2822,
247793
+ "step": 116660
247794
+ },
247795
+ {
247796
+ "epoch": 940.5,
247797
+ "learning_rate": 8.147147435897436e-06,
247798
+ "loss": 0.2605,
247799
+ "step": 116665
247800
+ },
247801
+ {
247802
+ "epoch": 940.54,
247803
+ "learning_rate": 8.147067307692308e-06,
247804
+ "loss": 0.3236,
247805
+ "step": 116670
247806
+ },
247807
+ {
247808
+ "epoch": 940.58,
247809
+ "learning_rate": 8.14698717948718e-06,
247810
+ "loss": 0.6434,
247811
+ "step": 116675
247812
+ },
247813
+ {
247814
+ "epoch": 940.62,
247815
+ "learning_rate": 8.146907051282052e-06,
247816
+ "loss": 1.0242,
247817
+ "step": 116680
247818
+ },
247819
+ {
247820
+ "epoch": 940.66,
247821
+ "learning_rate": 8.146826923076923e-06,
247822
+ "loss": 0.3286,
247823
+ "step": 116685
247824
+ },
247825
+ {
247826
+ "epoch": 940.7,
247827
+ "learning_rate": 8.146746794871796e-06,
247828
+ "loss": 0.2694,
247829
+ "step": 116690
247830
+ },
247831
+ {
247832
+ "epoch": 940.74,
247833
+ "learning_rate": 8.146666666666668e-06,
247834
+ "loss": 0.2618,
247835
+ "step": 116695
247836
+ },
247837
+ {
247838
+ "epoch": 940.78,
247839
+ "learning_rate": 8.146586538461539e-06,
247840
+ "loss": 0.6336,
247841
+ "step": 116700
247842
+ },
247843
+ {
247844
+ "epoch": 940.82,
247845
+ "learning_rate": 8.14650641025641e-06,
247846
+ "loss": 0.9335,
247847
+ "step": 116705
247848
+ },
247849
+ {
247850
+ "epoch": 940.86,
247851
+ "learning_rate": 8.146426282051283e-06,
247852
+ "loss": 0.2617,
247853
+ "step": 116710
247854
+ },
247855
+ {
247856
+ "epoch": 940.9,
247857
+ "learning_rate": 8.146346153846155e-06,
247858
+ "loss": 0.3125,
247859
+ "step": 116715
247860
+ },
247861
+ {
247862
+ "epoch": 940.94,
247863
+ "learning_rate": 8.146266025641026e-06,
247864
+ "loss": 0.3524,
247865
+ "step": 116720
247866
+ },
247867
+ {
247868
+ "epoch": 940.98,
247869
+ "learning_rate": 8.146185897435898e-06,
247870
+ "loss": 0.6945,
247871
+ "step": 116725
247872
+ },
247873
+ {
247874
+ "epoch": 941.0,
247875
+ "eval_loss": 0.3652080297470093,
247876
+ "eval_runtime": 39.2072,
247877
+ "eval_samples_per_second": 21.501,
247878
+ "eval_steps_per_second": 0.689,
247879
+ "eval_wer": 0.18559156861329676,
247880
+ "step": 116727
247881
+ },
247882
+ {
247883
+ "epoch": 941.02,
247884
+ "learning_rate": 8.14610576923077e-06,
247885
+ "loss": 0.3172,
247886
+ "step": 116730
247887
+ },
247888
+ {
247889
+ "epoch": 941.06,
247890
+ "learning_rate": 8.146025641025642e-06,
247891
+ "loss": 0.2391,
247892
+ "step": 116735
247893
+ },
247894
+ {
247895
+ "epoch": 941.1,
247896
+ "learning_rate": 8.145945512820513e-06,
247897
+ "loss": 0.2877,
247898
+ "step": 116740
247899
+ },
247900
+ {
247901
+ "epoch": 941.14,
247902
+ "learning_rate": 8.145865384615386e-06,
247903
+ "loss": 0.3587,
247904
+ "step": 116745
247905
+ },
247906
+ {
247907
+ "epoch": 941.18,
247908
+ "learning_rate": 8.145785256410256e-06,
247909
+ "loss": 0.7811,
247910
+ "step": 116750
247911
+ },
247912
+ {
247913
+ "epoch": 941.22,
247914
+ "learning_rate": 8.145705128205129e-06,
247915
+ "loss": 0.7006,
247916
+ "step": 116755
247917
+ },
247918
+ {
247919
+ "epoch": 941.27,
247920
+ "learning_rate": 8.145625e-06,
247921
+ "loss": 0.327,
247922
+ "step": 116760
247923
+ },
247924
+ {
247925
+ "epoch": 941.31,
247926
+ "learning_rate": 8.145544871794872e-06,
247927
+ "loss": 0.3295,
247928
+ "step": 116765
247929
+ },
247930
+ {
247931
+ "epoch": 941.35,
247932
+ "learning_rate": 8.145464743589743e-06,
247933
+ "loss": 0.3501,
247934
+ "step": 116770
247935
+ },
247936
+ {
247937
+ "epoch": 941.39,
247938
+ "learning_rate": 8.145384615384616e-06,
247939
+ "loss": 0.9432,
247940
+ "step": 116775
247941
+ },
247942
+ {
247943
+ "epoch": 941.43,
247944
+ "learning_rate": 8.145304487179488e-06,
247945
+ "loss": 0.8745,
247946
+ "step": 116780
247947
+ },
247948
+ {
247949
+ "epoch": 941.47,
247950
+ "learning_rate": 8.145224358974359e-06,
247951
+ "loss": 0.2931,
247952
+ "step": 116785
247953
+ },
247954
+ {
247955
+ "epoch": 941.51,
247956
+ "learning_rate": 8.145144230769232e-06,
247957
+ "loss": 0.2615,
247958
+ "step": 116790
247959
+ },
247960
+ {
247961
+ "epoch": 941.55,
247962
+ "learning_rate": 8.145064102564103e-06,
247963
+ "loss": 0.3783,
247964
+ "step": 116795
247965
+ },
247966
+ {
247967
+ "epoch": 941.59,
247968
+ "learning_rate": 8.144983974358975e-06,
247969
+ "loss": 0.6995,
247970
+ "step": 116800
247971
+ },
247972
+ {
247973
+ "epoch": 941.63,
247974
+ "learning_rate": 8.144903846153846e-06,
247975
+ "loss": 0.8687,
247976
+ "step": 116805
247977
+ },
247978
+ {
247979
+ "epoch": 941.67,
247980
+ "learning_rate": 8.144823717948719e-06,
247981
+ "loss": 0.2578,
247982
+ "step": 116810
247983
+ },
247984
+ {
247985
+ "epoch": 941.71,
247986
+ "learning_rate": 8.14474358974359e-06,
247987
+ "loss": 0.2837,
247988
+ "step": 116815
247989
+ },
247990
+ {
247991
+ "epoch": 941.75,
247992
+ "learning_rate": 8.144663461538462e-06,
247993
+ "loss": 0.3907,
247994
+ "step": 116820
247995
+ },
247996
+ {
247997
+ "epoch": 941.79,
247998
+ "learning_rate": 8.144583333333333e-06,
247999
+ "loss": 0.6813,
248000
+ "step": 116825
248001
+ },
248002
+ {
248003
+ "epoch": 941.83,
248004
+ "learning_rate": 8.144503205128206e-06,
248005
+ "loss": 0.8454,
248006
+ "step": 116830
248007
+ },
248008
+ {
248009
+ "epoch": 941.87,
248010
+ "learning_rate": 8.144423076923078e-06,
248011
+ "loss": 0.2834,
248012
+ "step": 116835
248013
+ },
248014
+ {
248015
+ "epoch": 941.91,
248016
+ "learning_rate": 8.144342948717949e-06,
248017
+ "loss": 0.3404,
248018
+ "step": 116840
248019
+ },
248020
+ {
248021
+ "epoch": 941.95,
248022
+ "learning_rate": 8.144262820512822e-06,
248023
+ "loss": 0.3596,
248024
+ "step": 116845
248025
+ },
248026
+ {
248027
+ "epoch": 941.99,
248028
+ "learning_rate": 8.144182692307693e-06,
248029
+ "loss": 0.8048,
248030
+ "step": 116850
248031
+ },
248032
+ {
248033
+ "epoch": 942.0,
248034
+ "eval_loss": 0.37358829379081726,
248035
+ "eval_runtime": 38.8227,
248036
+ "eval_samples_per_second": 21.714,
248037
+ "eval_steps_per_second": 0.695,
248038
+ "eval_wer": 0.181710652768537,
248039
+ "step": 116851
248040
+ },
248041
+ {
248042
+ "epoch": 942.03,
248043
+ "learning_rate": 8.144102564102565e-06,
248044
+ "loss": 0.3532,
248045
+ "step": 116855
248046
+ },
248047
+ {
248048
+ "epoch": 942.07,
248049
+ "learning_rate": 8.144022435897436e-06,
248050
+ "loss": 0.2954,
248051
+ "step": 116860
248052
+ },
248053
+ {
248054
+ "epoch": 942.11,
248055
+ "learning_rate": 8.14394230769231e-06,
248056
+ "loss": 0.321,
248057
+ "step": 116865
248058
+ },
248059
+ {
248060
+ "epoch": 942.15,
248061
+ "learning_rate": 8.14386217948718e-06,
248062
+ "loss": 0.4422,
248063
+ "step": 116870
248064
+ },
248065
+ {
248066
+ "epoch": 942.19,
248067
+ "learning_rate": 8.143782051282052e-06,
248068
+ "loss": 0.8507,
248069
+ "step": 116875
248070
+ },
248071
+ {
248072
+ "epoch": 942.23,
248073
+ "learning_rate": 8.143701923076923e-06,
248074
+ "loss": 0.6194,
248075
+ "step": 116880
248076
+ },
248077
+ {
248078
+ "epoch": 942.27,
248079
+ "learning_rate": 8.143621794871796e-06,
248080
+ "loss": 0.3064,
248081
+ "step": 116885
248082
+ },
248083
+ {
248084
+ "epoch": 942.31,
248085
+ "learning_rate": 8.143541666666668e-06,
248086
+ "loss": 0.2845,
248087
+ "step": 116890
248088
+ },
248089
+ {
248090
+ "epoch": 942.35,
248091
+ "learning_rate": 8.143461538461539e-06,
248092
+ "loss": 0.3395,
248093
+ "step": 116895
248094
+ },
248095
+ {
248096
+ "epoch": 942.39,
248097
+ "learning_rate": 8.143381410256412e-06,
248098
+ "loss": 0.8709,
248099
+ "step": 116900
248100
+ },
248101
+ {
248102
+ "epoch": 942.43,
248103
+ "learning_rate": 8.143301282051282e-06,
248104
+ "loss": 0.5708,
248105
+ "step": 116905
248106
+ },
248107
+ {
248108
+ "epoch": 942.47,
248109
+ "learning_rate": 8.143221153846155e-06,
248110
+ "loss": 0.257,
248111
+ "step": 116910
248112
+ },
248113
+ {
248114
+ "epoch": 942.51,
248115
+ "learning_rate": 8.143141025641026e-06,
248116
+ "loss": 0.2881,
248117
+ "step": 116915
248118
+ },
248119
+ {
248120
+ "epoch": 942.55,
248121
+ "learning_rate": 8.143060897435897e-06,
248122
+ "loss": 0.3739,
248123
+ "step": 116920
248124
+ },
248125
+ {
248126
+ "epoch": 942.59,
248127
+ "learning_rate": 8.142980769230769e-06,
248128
+ "loss": 1.091,
248129
+ "step": 116925
248130
+ },
248131
+ {
248132
+ "epoch": 942.63,
248133
+ "learning_rate": 8.142900641025642e-06,
248134
+ "loss": 0.571,
248135
+ "step": 116930
248136
+ },
248137
+ {
248138
+ "epoch": 942.67,
248139
+ "learning_rate": 8.142820512820513e-06,
248140
+ "loss": 0.2523,
248141
+ "step": 116935
248142
+ },
248143
+ {
248144
+ "epoch": 942.71,
248145
+ "learning_rate": 8.142740384615385e-06,
248146
+ "loss": 0.3116,
248147
+ "step": 116940
248148
+ },
248149
+ {
248150
+ "epoch": 942.76,
248151
+ "learning_rate": 8.142660256410258e-06,
248152
+ "loss": 0.4294,
248153
+ "step": 116945
248154
+ },
248155
+ {
248156
+ "epoch": 942.8,
248157
+ "learning_rate": 8.142580128205129e-06,
248158
+ "loss": 0.8474,
248159
+ "step": 116950
248160
+ },
248161
+ {
248162
+ "epoch": 942.84,
248163
+ "learning_rate": 8.1425e-06,
248164
+ "loss": 0.6085,
248165
+ "step": 116955
248166
+ },
248167
+ {
248168
+ "epoch": 942.88,
248169
+ "learning_rate": 8.142419871794872e-06,
248170
+ "loss": 0.4619,
248171
+ "step": 116960
248172
+ },
248173
+ {
248174
+ "epoch": 942.92,
248175
+ "learning_rate": 8.142339743589745e-06,
248176
+ "loss": 0.4268,
248177
+ "step": 116965
248178
+ },
248179
+ {
248180
+ "epoch": 942.96,
248181
+ "learning_rate": 8.142259615384616e-06,
248182
+ "loss": 0.4165,
248183
+ "step": 116970
248184
+ },
248185
+ {
248186
+ "epoch": 943.0,
248187
+ "learning_rate": 8.142179487179488e-06,
248188
+ "loss": 1.148,
248189
+ "step": 116975
248190
+ },
248191
+ {
248192
+ "epoch": 943.0,
248193
+ "eval_loss": 0.3594907820224762,
248194
+ "eval_runtime": 39.3165,
248195
+ "eval_samples_per_second": 21.441,
248196
+ "eval_steps_per_second": 0.687,
248197
+ "eval_wer": 0.18600070348223707,
248198
+ "step": 116975
248199
+ },
248200
+ {
248201
+ "epoch": 935.04,
248202
+ "learning_rate": 8.142099358974359e-06,
248203
+ "loss": 0.323,
248204
+ "step": 116980
248205
+ },
248206
+ {
248207
+ "epoch": 935.08,
248208
+ "learning_rate": 8.142019230769232e-06,
248209
+ "loss": 0.284,
248210
+ "step": 116985
248211
+ },
248212
+ {
248213
+ "epoch": 935.12,
248214
+ "learning_rate": 8.141939102564103e-06,
248215
+ "loss": 0.2959,
248216
+ "step": 116990
248217
+ },
248218
+ {
248219
+ "epoch": 935.16,
248220
+ "learning_rate": 8.141858974358975e-06,
248221
+ "loss": 0.4164,
248222
+ "step": 116995
248223
+ },
248224
+ {
248225
+ "epoch": 935.2,
248226
+ "learning_rate": 8.141778846153848e-06,
248227
+ "loss": 1.1566,
248228
+ "step": 117000
248229
+ },
248230
+ {
248231
+ "epoch": 935.24,
248232
+ "learning_rate": 8.141698717948719e-06,
248233
+ "loss": 0.3514,
248234
+ "step": 117005
248235
+ },
248236
+ {
248237
+ "epoch": 935.28,
248238
+ "learning_rate": 8.14161858974359e-06,
248239
+ "loss": 0.2887,
248240
+ "step": 117010
248241
+ },
248242
+ {
248243
+ "epoch": 935.32,
248244
+ "learning_rate": 8.141538461538462e-06,
248245
+ "loss": 0.3248,
248246
+ "step": 117015
248247
+ },
248248
+ {
248249
+ "epoch": 935.36,
248250
+ "learning_rate": 8.141458333333335e-06,
248251
+ "loss": 0.5457,
248252
+ "step": 117020
248253
+ },
248254
+ {
248255
+ "epoch": 935.4,
248256
+ "learning_rate": 8.141378205128204e-06,
248257
+ "loss": 1.2681,
248258
+ "step": 117025
248259
+ },
248260
+ {
248261
+ "epoch": 935.44,
248262
+ "learning_rate": 8.141298076923078e-06,
248263
+ "loss": 0.2972,
248264
+ "step": 117030
248265
+ },
248266
+ {
248267
+ "epoch": 935.48,
248268
+ "learning_rate": 8.14121794871795e-06,
248269
+ "loss": 0.2916,
248270
+ "step": 117035
248271
+ },
248272
+ {
248273
+ "epoch": 935.52,
248274
+ "learning_rate": 8.14113782051282e-06,
248275
+ "loss": 0.3855,
248276
+ "step": 117040
248277
+ },
248278
+ {
248279
+ "epoch": 935.56,
248280
+ "learning_rate": 8.141057692307693e-06,
248281
+ "loss": 0.3866,
248282
+ "step": 117045
248283
+ },
248284
+ {
248285
+ "epoch": 935.6,
248286
+ "learning_rate": 8.140977564102565e-06,
248287
+ "loss": 1.0996,
248288
+ "step": 117050
248289
+ },
248290
+ {
248291
+ "epoch": 935.64,
248292
+ "learning_rate": 8.140897435897436e-06,
248293
+ "loss": 0.3514,
248294
+ "step": 117055
248295
+ },
248296
+ {
248297
+ "epoch": 935.68,
248298
+ "learning_rate": 8.140817307692307e-06,
248299
+ "loss": 0.2344,
248300
+ "step": 117060
248301
+ },
248302
+ {
248303
+ "epoch": 935.72,
248304
+ "learning_rate": 8.14073717948718e-06,
248305
+ "loss": 0.3085,
248306
+ "step": 117065
248307
+ },
248308
+ {
248309
+ "epoch": 935.76,
248310
+ "learning_rate": 8.140657051282052e-06,
248311
+ "loss": 0.413,
248312
+ "step": 117070
248313
+ },
248314
+ {
248315
+ "epoch": 935.8,
248316
+ "learning_rate": 8.140576923076923e-06,
248317
+ "loss": 1.332,
248318
+ "step": 117075
248319
+ },
248320
+ {
248321
+ "epoch": 935.84,
248322
+ "learning_rate": 8.140496794871795e-06,
248323
+ "loss": 0.301,
248324
+ "step": 117080
248325
+ },
248326
+ {
248327
+ "epoch": 935.88,
248328
+ "learning_rate": 8.140416666666668e-06,
248329
+ "loss": 0.2786,
248330
+ "step": 117085
248331
+ },
248332
+ {
248333
+ "epoch": 935.92,
248334
+ "learning_rate": 8.140336538461539e-06,
248335
+ "loss": 0.3301,
248336
+ "step": 117090
248337
+ },
248338
+ {
248339
+ "epoch": 935.96,
248340
+ "learning_rate": 8.14025641025641e-06,
248341
+ "loss": 0.557,
248342
+ "step": 117095
248343
+ },
248344
+ {
248345
+ "epoch": 936.0,
248346
+ "learning_rate": 8.140176282051283e-06,
248347
+ "loss": 1.4011,
248348
+ "step": 117100
248349
+ },
248350
+ {
248351
+ "epoch": 936.0,
248352
+ "eval_loss": 0.34906426072120667,
248353
+ "eval_runtime": 39.7868,
248354
+ "eval_samples_per_second": 21.188,
248355
+ "eval_steps_per_second": 0.679,
248356
+ "eval_wer": 0.1851399856424982,
248357
+ "step": 117100
248358
  }
248359
  ],
248360
  "max_steps": 625000,
248361
  "num_train_epochs": 5000,
248362
+ "total_flos": 3.29550208819677e+20,
248363
  "trial_name": null,
248364
  "trial_params": null
248365
  }
model-bin/finetune/base/{checkpoint-116478 β†’ checkpoint-117100}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630115798.4136324/events.out.tfevents.1630115798.86bb0ddabf9b.1042.41 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1b074a755f2daa10761fb515413ac8941c898fbb3c07592e5fda37594331551
3
+ size 4194
model-bin/finetune/base/log/1630116207.9281766/events.out.tfevents.1630116207.86bb0ddabf9b.1042.43 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68f3715bd0ab3d6f1ef9528a3b33c373e6085f271f727d16032a3c8f800f5c32
3
+ size 4194
model-bin/finetune/base/log/1630116621.3491957/events.out.tfevents.1630116621.86bb0ddabf9b.1042.45 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b61e75b6b5eedd92c5c3ec7d2c6390e778c14598efdc20e5064d0427e336885a
3
+ size 4194
model-bin/finetune/base/log/1630117036.3824139/events.out.tfevents.1630117036.86bb0ddabf9b.1042.47 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ab210bdc5285d925d2e47a8308f6851a592c9f596915f05ffec0a75f9de165
3
+ size 4194
model-bin/finetune/base/log/1630117449.777196/events.out.tfevents.1630117449.86bb0ddabf9b.1042.49 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8abf16a78934443be86f9165efeb04aef6c69ed81bc56bc9d7f82f4abe26ba9
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630115798.86bb0ddabf9b.1042.40 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11e43862968b4f8a128db20b0bdfb5c095f3a586a6dd95906ead603ea7a306a3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630116207.86bb0ddabf9b.1042.42 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b0d26ec2fa4bd9832fa8d496a62c8dded8ab637811b0ff8e9731ec6b08accc1
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630116621.86bb0ddabf9b.1042.44 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe9d89f47418ab426a2df3d71b21fbae52017b55817c17fbc00d6fd3dfbf7344
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630117036.86bb0ddabf9b.1042.46 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ace9ba494ce37d956652650695ce1ffb4ec148d656c10db037ae03ec311edb80
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630117449.86bb0ddabf9b.1042.48 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:251d71865605240570d7ac2db4f9cd67a49e2ad8ffbe1d36f89a65c721298fec
3
+ size 8622