Check commited on
Commit
d11468a
Β·
1 Parent(s): 99d73e3

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630113694.5899777/events.out.tfevents.1630113694.86bb0ddabf9b.1042.31 +3 -0
  11. model-bin/finetune/base/log/1630114108.4351602/events.out.tfevents.1630114108.86bb0ddabf9b.1042.33 +3 -0
  12. model-bin/finetune/base/log/1630114521.4023952/events.out.tfevents.1630114521.86bb0ddabf9b.1042.35 +3 -0
  13. model-bin/finetune/base/log/1630114935.8797095/events.out.tfevents.1630114935.86bb0ddabf9b.1042.37 +3 -0
  14. model-bin/finetune/base/log/1630115344.0735083/events.out.tfevents.1630115344.86bb0ddabf9b.1042.39 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630113694.86bb0ddabf9b.1042.30 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630114108.86bb0ddabf9b.1042.32 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630114521.86bb0ddabf9b.1042.34 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630114935.86bb0ddabf9b.1042.36 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630115344.86bb0ddabf9b.1042.38 +3 -0
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f58f1dcc9bfb4099e2ea22006b4aa19dc101120c78d4e27ed04418e2aa3131c
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed6a4a1b60032720f1c3b30726a6595b0807d4670f546ca8940edf524a4e7cd6
3
  size 722165393
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c3434b3bbed70f641bae6fda0466c25c030ac8b24f7e064b84c605b9ad0e2f2
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2bf7f96504e1c0a6174b9f2d5915d9873c10b1506cc00abe352873e28517867
3
  size 377909911
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75bdff2a7c9917d8d2929d23e1aa6358f86a7a2889a1c230ec6a8044b485696e
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:796e263a25f4a0a7434c18d8024159eaa3656ad1fa20d7a25b7eae15f6439f16
3
  size 14503
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:afc512941b00b221a55e7420c802cd31625bebe00c214122aedc10efbd598877
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8ec6882b7592c0ac7b9cd7d7fe3da25ac3c6fd19100a89c73fb078e31a1ddc3
3
  size 559
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee060aa484196bb13ded0b787197087803b23a7d6619688fd62bf478fba823e4
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cd22da4f6bd733f8bb533d0e36af7533566eafa7ac974fe7b8b64dcc9e30a39
3
  size 623
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 933.995983935743,
5
- "global_step": 115856,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -246771,11 +246771,800 @@
246771
  "eval_steps_per_second": 0.696,
246772
  "eval_wer": 0.18200635765506026,
246773
  "step": 115856
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
246774
  }
246775
  ],
246776
- "max_steps": 620000,
246777
  "num_train_epochs": 5000,
246778
- "total_flos": 3.260533347890663e+20,
246779
  "trial_name": null,
246780
  "trial_params": null
246781
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 931.0,
5
+ "global_step": 116478,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
246771
  "eval_steps_per_second": 0.696,
246772
  "eval_wer": 0.18200635765506026,
246773
  "step": 115856
246774
+ },
246775
+ {
246776
+ "epoch": 934.03,
246777
+ "learning_rate": 8.160048076923078e-06,
246778
+ "loss": 0.3382,
246779
+ "step": 115860
246780
+ },
246781
+ {
246782
+ "epoch": 934.07,
246783
+ "learning_rate": 8.15996794871795e-06,
246784
+ "loss": 0.2679,
246785
+ "step": 115865
246786
+ },
246787
+ {
246788
+ "epoch": 934.11,
246789
+ "learning_rate": 8.15988782051282e-06,
246790
+ "loss": 0.277,
246791
+ "step": 115870
246792
+ },
246793
+ {
246794
+ "epoch": 934.15,
246795
+ "learning_rate": 8.159807692307694e-06,
246796
+ "loss": 0.4102,
246797
+ "step": 115875
246798
+ },
246799
+ {
246800
+ "epoch": 934.19,
246801
+ "learning_rate": 8.159727564102565e-06,
246802
+ "loss": 1.1364,
246803
+ "step": 115880
246804
+ },
246805
+ {
246806
+ "epoch": 934.23,
246807
+ "learning_rate": 8.159647435897436e-06,
246808
+ "loss": 0.627,
246809
+ "step": 115885
246810
+ },
246811
+ {
246812
+ "epoch": 934.27,
246813
+ "learning_rate": 8.15956730769231e-06,
246814
+ "loss": 0.2331,
246815
+ "step": 115890
246816
+ },
246817
+ {
246818
+ "epoch": 934.31,
246819
+ "learning_rate": 8.15948717948718e-06,
246820
+ "loss": 0.2426,
246821
+ "step": 115895
246822
+ },
246823
+ {
246824
+ "epoch": 934.35,
246825
+ "learning_rate": 8.159407051282052e-06,
246826
+ "loss": 0.3757,
246827
+ "step": 115900
246828
+ },
246829
+ {
246830
+ "epoch": 934.39,
246831
+ "learning_rate": 8.159326923076924e-06,
246832
+ "loss": 0.8871,
246833
+ "step": 115905
246834
+ },
246835
+ {
246836
+ "epoch": 934.43,
246837
+ "learning_rate": 8.159246794871795e-06,
246838
+ "loss": 0.6067,
246839
+ "step": 115910
246840
+ },
246841
+ {
246842
+ "epoch": 934.47,
246843
+ "learning_rate": 8.159166666666666e-06,
246844
+ "loss": 0.3254,
246845
+ "step": 115915
246846
+ },
246847
+ {
246848
+ "epoch": 934.51,
246849
+ "learning_rate": 8.15908653846154e-06,
246850
+ "loss": 0.3118,
246851
+ "step": 115920
246852
+ },
246853
+ {
246854
+ "epoch": 934.55,
246855
+ "learning_rate": 8.15900641025641e-06,
246856
+ "loss": 0.425,
246857
+ "step": 115925
246858
+ },
246859
+ {
246860
+ "epoch": 934.59,
246861
+ "learning_rate": 8.158926282051282e-06,
246862
+ "loss": 0.9399,
246863
+ "step": 115930
246864
+ },
246865
+ {
246866
+ "epoch": 934.63,
246867
+ "learning_rate": 8.158846153846155e-06,
246868
+ "loss": 0.7605,
246869
+ "step": 115935
246870
+ },
246871
+ {
246872
+ "epoch": 934.67,
246873
+ "learning_rate": 8.158766025641027e-06,
246874
+ "loss": 0.2649,
246875
+ "step": 115940
246876
+ },
246877
+ {
246878
+ "epoch": 934.71,
246879
+ "learning_rate": 8.158685897435898e-06,
246880
+ "loss": 0.3038,
246881
+ "step": 115945
246882
+ },
246883
+ {
246884
+ "epoch": 934.76,
246885
+ "learning_rate": 8.15860576923077e-06,
246886
+ "loss": 0.3894,
246887
+ "step": 115950
246888
+ },
246889
+ {
246890
+ "epoch": 934.8,
246891
+ "learning_rate": 8.158525641025642e-06,
246892
+ "loss": 0.8665,
246893
+ "step": 115955
246894
+ },
246895
+ {
246896
+ "epoch": 934.84,
246897
+ "learning_rate": 8.158445512820514e-06,
246898
+ "loss": 0.6454,
246899
+ "step": 115960
246900
+ },
246901
+ {
246902
+ "epoch": 934.88,
246903
+ "learning_rate": 8.158365384615385e-06,
246904
+ "loss": 0.2806,
246905
+ "step": 115965
246906
+ },
246907
+ {
246908
+ "epoch": 934.92,
246909
+ "learning_rate": 8.158285256410256e-06,
246910
+ "loss": 0.2804,
246911
+ "step": 115970
246912
+ },
246913
+ {
246914
+ "epoch": 934.96,
246915
+ "learning_rate": 8.15820512820513e-06,
246916
+ "loss": 0.4341,
246917
+ "step": 115975
246918
+ },
246919
+ {
246920
+ "epoch": 935.0,
246921
+ "learning_rate": 8.158125e-06,
246922
+ "loss": 0.9677,
246923
+ "step": 115980
246924
+ },
246925
+ {
246926
+ "epoch": 935.0,
246927
+ "eval_loss": 0.3683865964412689,
246928
+ "eval_runtime": 39.1683,
246929
+ "eval_samples_per_second": 21.523,
246930
+ "eval_steps_per_second": 0.689,
246931
+ "eval_wer": 0.18747801618009144,
246932
+ "step": 115980
246933
+ },
246934
+ {
246935
+ "epoch": 935.04,
246936
+ "learning_rate": 8.158044871794872e-06,
246937
+ "loss": 0.2787,
246938
+ "step": 115985
246939
+ },
246940
+ {
246941
+ "epoch": 935.08,
246942
+ "learning_rate": 8.157964743589745e-06,
246943
+ "loss": 0.292,
246944
+ "step": 115990
246945
+ },
246946
+ {
246947
+ "epoch": 935.12,
246948
+ "learning_rate": 8.157884615384617e-06,
246949
+ "loss": 0.3054,
246950
+ "step": 115995
246951
+ },
246952
+ {
246953
+ "epoch": 935.16,
246954
+ "learning_rate": 8.157804487179488e-06,
246955
+ "loss": 0.5644,
246956
+ "step": 116000
246957
+ },
246958
+ {
246959
+ "epoch": 935.2,
246960
+ "learning_rate": 8.15772435897436e-06,
246961
+ "loss": 1.306,
246962
+ "step": 116005
246963
+ },
246964
+ {
246965
+ "epoch": 935.24,
246966
+ "learning_rate": 8.157644230769232e-06,
246967
+ "loss": 0.3061,
246968
+ "step": 116010
246969
+ },
246970
+ {
246971
+ "epoch": 935.28,
246972
+ "learning_rate": 8.157564102564102e-06,
246973
+ "loss": 0.2666,
246974
+ "step": 116015
246975
+ },
246976
+ {
246977
+ "epoch": 935.32,
246978
+ "learning_rate": 8.157483974358975e-06,
246979
+ "loss": 0.3729,
246980
+ "step": 116020
246981
+ },
246982
+ {
246983
+ "epoch": 935.36,
246984
+ "learning_rate": 8.157403846153846e-06,
246985
+ "loss": 0.4313,
246986
+ "step": 116025
246987
+ },
246988
+ {
246989
+ "epoch": 935.4,
246990
+ "learning_rate": 8.157323717948718e-06,
246991
+ "loss": 1.1394,
246992
+ "step": 116030
246993
+ },
246994
+ {
246995
+ "epoch": 935.44,
246996
+ "learning_rate": 8.15724358974359e-06,
246997
+ "loss": 0.3986,
246998
+ "step": 116035
246999
+ },
247000
+ {
247001
+ "epoch": 935.48,
247002
+ "learning_rate": 8.157163461538462e-06,
247003
+ "loss": 0.2467,
247004
+ "step": 116040
247005
+ },
247006
+ {
247007
+ "epoch": 935.52,
247008
+ "learning_rate": 8.157083333333334e-06,
247009
+ "loss": 0.2822,
247010
+ "step": 116045
247011
+ },
247012
+ {
247013
+ "epoch": 935.56,
247014
+ "learning_rate": 8.157003205128205e-06,
247015
+ "loss": 0.4465,
247016
+ "step": 116050
247017
+ },
247018
+ {
247019
+ "epoch": 935.6,
247020
+ "learning_rate": 8.156923076923078e-06,
247021
+ "loss": 1.1917,
247022
+ "step": 116055
247023
+ },
247024
+ {
247025
+ "epoch": 935.64,
247026
+ "learning_rate": 8.15684294871795e-06,
247027
+ "loss": 0.3459,
247028
+ "step": 116060
247029
+ },
247030
+ {
247031
+ "epoch": 935.68,
247032
+ "learning_rate": 8.15676282051282e-06,
247033
+ "loss": 0.2845,
247034
+ "step": 116065
247035
+ },
247036
+ {
247037
+ "epoch": 935.72,
247038
+ "learning_rate": 8.156682692307692e-06,
247039
+ "loss": 0.2771,
247040
+ "step": 116070
247041
+ },
247042
+ {
247043
+ "epoch": 935.76,
247044
+ "learning_rate": 8.156602564102565e-06,
247045
+ "loss": 0.4098,
247046
+ "step": 116075
247047
+ },
247048
+ {
247049
+ "epoch": 935.8,
247050
+ "learning_rate": 8.156522435897436e-06,
247051
+ "loss": 1.3603,
247052
+ "step": 116080
247053
+ },
247054
+ {
247055
+ "epoch": 935.84,
247056
+ "learning_rate": 8.156442307692308e-06,
247057
+ "loss": 0.301,
247058
+ "step": 116085
247059
+ },
247060
+ {
247061
+ "epoch": 935.88,
247062
+ "learning_rate": 8.15636217948718e-06,
247063
+ "loss": 0.2724,
247064
+ "step": 116090
247065
+ },
247066
+ {
247067
+ "epoch": 935.92,
247068
+ "learning_rate": 8.156282051282052e-06,
247069
+ "loss": 0.3323,
247070
+ "step": 116095
247071
+ },
247072
+ {
247073
+ "epoch": 935.96,
247074
+ "learning_rate": 8.156201923076924e-06,
247075
+ "loss": 0.4345,
247076
+ "step": 116100
247077
+ },
247078
+ {
247079
+ "epoch": 936.0,
247080
+ "eval_loss": 0.3591098189353943,
247081
+ "eval_runtime": 40.0141,
247082
+ "eval_samples_per_second": 21.068,
247083
+ "eval_steps_per_second": 0.675,
247084
+ "eval_wer": 0.1825556353194544,
247085
+ "step": 116104
247086
+ },
247087
+ {
247088
+ "epoch": 936.01,
247089
+ "learning_rate": 8.156121794871795e-06,
247090
+ "loss": 0.3954,
247091
+ "step": 116105
247092
+ },
247093
+ {
247094
+ "epoch": 936.05,
247095
+ "learning_rate": 8.156041666666668e-06,
247096
+ "loss": 0.3009,
247097
+ "step": 116110
247098
+ },
247099
+ {
247100
+ "epoch": 936.09,
247101
+ "learning_rate": 8.15596153846154e-06,
247102
+ "loss": 0.3602,
247103
+ "step": 116115
247104
+ },
247105
+ {
247106
+ "epoch": 936.13,
247107
+ "learning_rate": 8.15588141025641e-06,
247108
+ "loss": 0.279,
247109
+ "step": 116120
247110
+ },
247111
+ {
247112
+ "epoch": 936.17,
247113
+ "learning_rate": 8.155801282051282e-06,
247114
+ "loss": 0.5056,
247115
+ "step": 116125
247116
+ },
247117
+ {
247118
+ "epoch": 936.21,
247119
+ "learning_rate": 8.155721153846155e-06,
247120
+ "loss": 1.0783,
247121
+ "step": 116130
247122
+ },
247123
+ {
247124
+ "epoch": 936.25,
247125
+ "learning_rate": 8.155641025641026e-06,
247126
+ "loss": 0.3115,
247127
+ "step": 116135
247128
+ },
247129
+ {
247130
+ "epoch": 936.29,
247131
+ "learning_rate": 8.155560897435898e-06,
247132
+ "loss": 0.2519,
247133
+ "step": 116140
247134
+ },
247135
+ {
247136
+ "epoch": 936.33,
247137
+ "learning_rate": 8.155480769230771e-06,
247138
+ "loss": 0.361,
247139
+ "step": 116145
247140
+ },
247141
+ {
247142
+ "epoch": 936.37,
247143
+ "learning_rate": 8.155400641025642e-06,
247144
+ "loss": 0.5359,
247145
+ "step": 116150
247146
+ },
247147
+ {
247148
+ "epoch": 936.41,
247149
+ "learning_rate": 8.155320512820514e-06,
247150
+ "loss": 1.0703,
247151
+ "step": 116155
247152
+ },
247153
+ {
247154
+ "epoch": 936.45,
247155
+ "learning_rate": 8.155240384615385e-06,
247156
+ "loss": 0.3052,
247157
+ "step": 116160
247158
+ },
247159
+ {
247160
+ "epoch": 936.49,
247161
+ "learning_rate": 8.155160256410258e-06,
247162
+ "loss": 0.2976,
247163
+ "step": 116165
247164
+ },
247165
+ {
247166
+ "epoch": 936.53,
247167
+ "learning_rate": 8.155080128205128e-06,
247168
+ "loss": 0.316,
247169
+ "step": 116170
247170
+ },
247171
+ {
247172
+ "epoch": 936.57,
247173
+ "learning_rate": 8.155e-06,
247174
+ "loss": 0.5374,
247175
+ "step": 116175
247176
+ },
247177
+ {
247178
+ "epoch": 936.61,
247179
+ "learning_rate": 8.154919871794874e-06,
247180
+ "loss": 0.9842,
247181
+ "step": 116180
247182
+ },
247183
+ {
247184
+ "epoch": 936.65,
247185
+ "learning_rate": 8.154839743589743e-06,
247186
+ "loss": 0.2637,
247187
+ "step": 116185
247188
+ },
247189
+ {
247190
+ "epoch": 936.69,
247191
+ "learning_rate": 8.154759615384616e-06,
247192
+ "loss": 0.2658,
247193
+ "step": 116190
247194
+ },
247195
+ {
247196
+ "epoch": 936.73,
247197
+ "learning_rate": 8.154679487179488e-06,
247198
+ "loss": 0.4132,
247199
+ "step": 116195
247200
+ },
247201
+ {
247202
+ "epoch": 936.77,
247203
+ "learning_rate": 8.15459935897436e-06,
247204
+ "loss": 0.4989,
247205
+ "step": 116200
247206
+ },
247207
+ {
247208
+ "epoch": 936.81,
247209
+ "learning_rate": 8.15451923076923e-06,
247210
+ "loss": 1.1565,
247211
+ "step": 116205
247212
+ },
247213
+ {
247214
+ "epoch": 936.85,
247215
+ "learning_rate": 8.154439102564104e-06,
247216
+ "loss": 0.37,
247217
+ "step": 116210
247218
+ },
247219
+ {
247220
+ "epoch": 936.89,
247221
+ "learning_rate": 8.154358974358975e-06,
247222
+ "loss": 0.4002,
247223
+ "step": 116215
247224
+ },
247225
+ {
247226
+ "epoch": 936.93,
247227
+ "learning_rate": 8.154278846153846e-06,
247228
+ "loss": 1.0027,
247229
+ "step": 116220
247230
+ },
247231
+ {
247232
+ "epoch": 936.97,
247233
+ "learning_rate": 8.154198717948718e-06,
247234
+ "loss": 0.5068,
247235
+ "step": 116225
247236
+ },
247237
+ {
247238
+ "epoch": 937.0,
247239
+ "eval_loss": 0.37429821491241455,
247240
+ "eval_runtime": 38.8986,
247241
+ "eval_samples_per_second": 21.697,
247242
+ "eval_steps_per_second": 0.694,
247243
+ "eval_wer": 0.18772589806004278,
247244
+ "step": 116228
247245
+ },
247246
+ {
247247
+ "epoch": 929.02,
247248
+ "learning_rate": 8.15411858974359e-06,
247249
+ "loss": 0.3236,
247250
+ "step": 116230
247251
+ },
247252
+ {
247253
+ "epoch": 929.06,
247254
+ "learning_rate": 8.154038461538462e-06,
247255
+ "loss": 0.3169,
247256
+ "step": 116235
247257
+ },
247258
+ {
247259
+ "epoch": 929.1,
247260
+ "learning_rate": 8.153958333333333e-06,
247261
+ "loss": 0.303,
247262
+ "step": 116240
247263
+ },
247264
+ {
247265
+ "epoch": 929.14,
247266
+ "learning_rate": 8.153878205128206e-06,
247267
+ "loss": 0.3253,
247268
+ "step": 116245
247269
+ },
247270
+ {
247271
+ "epoch": 929.18,
247272
+ "learning_rate": 8.153798076923078e-06,
247273
+ "loss": 0.6763,
247274
+ "step": 116250
247275
+ },
247276
+ {
247277
+ "epoch": 929.22,
247278
+ "learning_rate": 8.15371794871795e-06,
247279
+ "loss": 1.0281,
247280
+ "step": 116255
247281
+ },
247282
+ {
247283
+ "epoch": 929.26,
247284
+ "learning_rate": 8.15363782051282e-06,
247285
+ "loss": 0.2937,
247286
+ "step": 116260
247287
+ },
247288
+ {
247289
+ "epoch": 929.3,
247290
+ "learning_rate": 8.153557692307694e-06,
247291
+ "loss": 0.28,
247292
+ "step": 116265
247293
+ },
247294
+ {
247295
+ "epoch": 929.34,
247296
+ "learning_rate": 8.153477564102565e-06,
247297
+ "loss": 0.282,
247298
+ "step": 116270
247299
+ },
247300
+ {
247301
+ "epoch": 929.38,
247302
+ "learning_rate": 8.153397435897436e-06,
247303
+ "loss": 0.6766,
247304
+ "step": 116275
247305
+ },
247306
+ {
247307
+ "epoch": 929.42,
247308
+ "learning_rate": 8.15331730769231e-06,
247309
+ "loss": 0.8845,
247310
+ "step": 116280
247311
+ },
247312
+ {
247313
+ "epoch": 929.46,
247314
+ "learning_rate": 8.15323717948718e-06,
247315
+ "loss": 0.2686,
247316
+ "step": 116285
247317
+ },
247318
+ {
247319
+ "epoch": 929.5,
247320
+ "learning_rate": 8.153157051282052e-06,
247321
+ "loss": 0.2538,
247322
+ "step": 116290
247323
+ },
247324
+ {
247325
+ "epoch": 929.54,
247326
+ "learning_rate": 8.153076923076923e-06,
247327
+ "loss": 0.2814,
247328
+ "step": 116295
247329
+ },
247330
+ {
247331
+ "epoch": 929.58,
247332
+ "learning_rate": 8.152996794871797e-06,
247333
+ "loss": 0.5815,
247334
+ "step": 116300
247335
+ },
247336
+ {
247337
+ "epoch": 929.62,
247338
+ "learning_rate": 8.152916666666668e-06,
247339
+ "loss": 1.0262,
247340
+ "step": 116305
247341
+ },
247342
+ {
247343
+ "epoch": 929.66,
247344
+ "learning_rate": 8.15283653846154e-06,
247345
+ "loss": 0.2782,
247346
+ "step": 116310
247347
+ },
247348
+ {
247349
+ "epoch": 929.7,
247350
+ "learning_rate": 8.15275641025641e-06,
247351
+ "loss": 0.3347,
247352
+ "step": 116315
247353
+ },
247354
+ {
247355
+ "epoch": 929.74,
247356
+ "learning_rate": 8.152676282051284e-06,
247357
+ "loss": 0.3231,
247358
+ "step": 116320
247359
+ },
247360
+ {
247361
+ "epoch": 929.78,
247362
+ "learning_rate": 8.152596153846153e-06,
247363
+ "loss": 0.5842,
247364
+ "step": 116325
247365
+ },
247366
+ {
247367
+ "epoch": 929.82,
247368
+ "learning_rate": 8.152516025641026e-06,
247369
+ "loss": 1.0521,
247370
+ "step": 116330
247371
+ },
247372
+ {
247373
+ "epoch": 929.86,
247374
+ "learning_rate": 8.1524358974359e-06,
247375
+ "loss": 0.2646,
247376
+ "step": 116335
247377
+ },
247378
+ {
247379
+ "epoch": 929.9,
247380
+ "learning_rate": 8.152355769230769e-06,
247381
+ "loss": 0.3342,
247382
+ "step": 116340
247383
+ },
247384
+ {
247385
+ "epoch": 929.94,
247386
+ "learning_rate": 8.152275641025642e-06,
247387
+ "loss": 0.4099,
247388
+ "step": 116345
247389
+ },
247390
+ {
247391
+ "epoch": 929.98,
247392
+ "learning_rate": 8.152195512820513e-06,
247393
+ "loss": 0.6846,
247394
+ "step": 116350
247395
+ },
247396
+ {
247397
+ "epoch": 930.0,
247398
+ "eval_loss": 0.42502668499946594,
247399
+ "eval_runtime": 39.0001,
247400
+ "eval_samples_per_second": 21.615,
247401
+ "eval_steps_per_second": 0.692,
247402
+ "eval_wer": 0.1844075759800323,
247403
+ "step": 116353
247404
+ },
247405
+ {
247406
+ "epoch": 930.02,
247407
+ "learning_rate": 8.152115384615385e-06,
247408
+ "loss": 0.4367,
247409
+ "step": 116355
247410
+ },
247411
+ {
247412
+ "epoch": 930.06,
247413
+ "learning_rate": 8.152035256410256e-06,
247414
+ "loss": 0.2788,
247415
+ "step": 116360
247416
+ },
247417
+ {
247418
+ "epoch": 930.1,
247419
+ "learning_rate": 8.15195512820513e-06,
247420
+ "loss": 0.289,
247421
+ "step": 116365
247422
+ },
247423
+ {
247424
+ "epoch": 930.14,
247425
+ "learning_rate": 8.151875e-06,
247426
+ "loss": 0.364,
247427
+ "step": 116370
247428
+ },
247429
+ {
247430
+ "epoch": 930.18,
247431
+ "learning_rate": 8.151794871794872e-06,
247432
+ "loss": 0.5893,
247433
+ "step": 116375
247434
+ },
247435
+ {
247436
+ "epoch": 930.22,
247437
+ "learning_rate": 8.151714743589745e-06,
247438
+ "loss": 1.0691,
247439
+ "step": 116380
247440
+ },
247441
+ {
247442
+ "epoch": 930.26,
247443
+ "learning_rate": 8.151634615384616e-06,
247444
+ "loss": 0.2525,
247445
+ "step": 116385
247446
+ },
247447
+ {
247448
+ "epoch": 930.3,
247449
+ "learning_rate": 8.151554487179488e-06,
247450
+ "loss": 0.2818,
247451
+ "step": 116390
247452
+ },
247453
+ {
247454
+ "epoch": 930.34,
247455
+ "learning_rate": 8.151474358974359e-06,
247456
+ "loss": 0.3698,
247457
+ "step": 116395
247458
+ },
247459
+ {
247460
+ "epoch": 930.38,
247461
+ "learning_rate": 8.151394230769232e-06,
247462
+ "loss": 0.6021,
247463
+ "step": 116400
247464
+ },
247465
+ {
247466
+ "epoch": 930.42,
247467
+ "learning_rate": 8.151314102564104e-06,
247468
+ "loss": 1.0208,
247469
+ "step": 116405
247470
+ },
247471
+ {
247472
+ "epoch": 930.46,
247473
+ "learning_rate": 8.151233974358975e-06,
247474
+ "loss": 0.2736,
247475
+ "step": 116410
247476
+ },
247477
+ {
247478
+ "epoch": 930.5,
247479
+ "learning_rate": 8.151153846153846e-06,
247480
+ "loss": 0.3401,
247481
+ "step": 116415
247482
+ },
247483
+ {
247484
+ "epoch": 930.54,
247485
+ "learning_rate": 8.15107371794872e-06,
247486
+ "loss": 0.3492,
247487
+ "step": 116420
247488
+ },
247489
+ {
247490
+ "epoch": 930.58,
247491
+ "learning_rate": 8.15099358974359e-06,
247492
+ "loss": 0.7202,
247493
+ "step": 116425
247494
+ },
247495
+ {
247496
+ "epoch": 930.62,
247497
+ "learning_rate": 8.150913461538462e-06,
247498
+ "loss": 1.0942,
247499
+ "step": 116430
247500
+ },
247501
+ {
247502
+ "epoch": 930.66,
247503
+ "learning_rate": 8.150833333333335e-06,
247504
+ "loss": 0.2699,
247505
+ "step": 116435
247506
+ },
247507
+ {
247508
+ "epoch": 930.7,
247509
+ "learning_rate": 8.150753205128206e-06,
247510
+ "loss": 0.2856,
247511
+ "step": 116440
247512
+ },
247513
+ {
247514
+ "epoch": 930.74,
247515
+ "learning_rate": 8.150673076923078e-06,
247516
+ "loss": 0.3806,
247517
+ "step": 116445
247518
+ },
247519
+ {
247520
+ "epoch": 930.78,
247521
+ "learning_rate": 8.150592948717949e-06,
247522
+ "loss": 0.6902,
247523
+ "step": 116450
247524
+ },
247525
+ {
247526
+ "epoch": 930.82,
247527
+ "learning_rate": 8.150512820512822e-06,
247528
+ "loss": 1.0085,
247529
+ "step": 116455
247530
+ },
247531
+ {
247532
+ "epoch": 930.86,
247533
+ "learning_rate": 8.150432692307692e-06,
247534
+ "loss": 0.2629,
247535
+ "step": 116460
247536
+ },
247537
+ {
247538
+ "epoch": 930.9,
247539
+ "learning_rate": 8.150352564102565e-06,
247540
+ "loss": 0.2615,
247541
+ "step": 116465
247542
+ },
247543
+ {
247544
+ "epoch": 930.94,
247545
+ "learning_rate": 8.150272435897436e-06,
247546
+ "loss": 0.3081,
247547
+ "step": 116470
247548
+ },
247549
+ {
247550
+ "epoch": 930.98,
247551
+ "learning_rate": 8.150192307692308e-06,
247552
+ "loss": 0.5409,
247553
+ "step": 116475
247554
+ },
247555
+ {
247556
+ "epoch": 931.0,
247557
+ "eval_loss": 0.427347868680954,
247558
+ "eval_runtime": 39.3615,
247559
+ "eval_samples_per_second": 21.417,
247560
+ "eval_steps_per_second": 0.686,
247561
+ "eval_wer": 0.18822154026011984,
247562
+ "step": 116478
247563
  }
247564
  ],
247565
+ "max_steps": 625000,
247566
  "num_train_epochs": 5000,
247567
+ "total_flos": 3.2779879386977056e+20,
247568
  "trial_name": null,
247569
  "trial_params": null
247570
  }
model-bin/finetune/base/{checkpoint-115856 β†’ checkpoint-116478}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630113694.5899777/events.out.tfevents.1630113694.86bb0ddabf9b.1042.31 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2bafd466c630fd1f651cfc6bbaac47320578e8cbe124912388e189cdd02f369
3
+ size 4194
model-bin/finetune/base/log/1630114108.4351602/events.out.tfevents.1630114108.86bb0ddabf9b.1042.33 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07a0a4953561629052e2d0bc1801579b24c34391dc125a4716a5f56a2a3045fa
3
+ size 4194
model-bin/finetune/base/log/1630114521.4023952/events.out.tfevents.1630114521.86bb0ddabf9b.1042.35 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f29c80132da81c5a6dd7313800425f0fb049072d5f15e5425c2f62b5e3e9619b
3
+ size 4194
model-bin/finetune/base/log/1630114935.8797095/events.out.tfevents.1630114935.86bb0ddabf9b.1042.37 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:840305c14ee3814d1fdf3db5aca2eb7cc8afa2665aa999949f65e2e0cb1dee27
3
+ size 4194
model-bin/finetune/base/log/1630115344.0735083/events.out.tfevents.1630115344.86bb0ddabf9b.1042.39 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:755d20a6b48427f9c481163078b8315b0f5e80a84035f6c92eaf75a877a3c1bf
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630113694.86bb0ddabf9b.1042.30 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0bce1d69ec0c9a328677eb09f0ad0770a749ad9c3fea932b09d28be9fbd1a70
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630114108.86bb0ddabf9b.1042.32 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba95ea0c440eb81e806d98f8334fc2cca8bd634a1c8b8264996852ee87ee21d0
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630114521.86bb0ddabf9b.1042.34 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cebd1f50853e7c5b4a05ca51952ae5a5a5f3c76be45e7ddf63b91f9739d919c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630114935.86bb0ddabf9b.1042.36 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:683c8af391b89ef17e13a05e99e3522534eda937abf38257a90c3d8c661bdd3e
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630115344.86bb0ddabf9b.1042.38 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:830c5154d8a6e363532f21d0a69517dce85a978e3fa9e0cb2b35a98412204cb1
3
+ size 8622