Check commited on
Commit
19f8f70
Β·
1 Parent(s): 8de933c

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630082884.969348/events.out.tfevents.1630082884.dea8124df033.1391.7 +3 -0
  11. model-bin/finetune/base/log/1630083307.1487231/events.out.tfevents.1630083307.dea8124df033.1391.9 +3 -0
  12. model-bin/finetune/base/log/1630083735.6256015/events.out.tfevents.1630083735.dea8124df033.1391.11 +3 -0
  13. model-bin/finetune/base/log/1630084174.1170754/events.out.tfevents.1630084174.dea8124df033.1391.13 +3 -0
  14. model-bin/finetune/base/log/1630084598.2761755/events.out.tfevents.1630084598.dea8124df033.1391.15 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630082884.dea8124df033.1391.6 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630083307.dea8124df033.1391.8 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630083735.dea8124df033.1391.10 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630084174.dea8124df033.1391.12 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630084598.dea8124df033.1391.14 +3 -0
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f3ce461a41ebb478a3033a35cb4191908dec25dc3b09f575609fc6aa146dd60
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d348482d17ba9e738c11c659795d83ff2bcf3cfd11baf8797570f6c9b26411a
3
  size 722165393
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed4c4bc6da00d916aba5b2523194212a682bc5e8e07fee75fc5d3fcb170a87ce
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc2b6659e86d951e27ce212a257a0b916a94a565fd3d84e430bfde3718539cb4
3
  size 377909911
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8135104c63b5165310558e47d8ba777edf4c3b459df9a523c789c28dcc1e549
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df743e69dcda2e4cf4a9f3f8ef82556dde39a400b2c9ae8b01c8bdc7894d03fb
3
  size 14503
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:531a84460dc36ad58f317625a8b2df924c497e322bd18a71440cc0d51bac941b
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fe07346d923a1da088bea107f5ab09f94c498e2865fabad8db6fc04e90f2d85
3
  size 559
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57fccb27f6be770c60514fd2f134fe1b6e0a8965992c964c9bd1421740a9ae29
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1905b108e5460ced8e245231c20dc93cf00ce38aab3458596eda343809ebdd48
3
  size 623
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 902.0,
5
- "global_step": 112750,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -242820,11 +242820,800 @@
242820
  "eval_steps_per_second": 0.674,
242821
  "eval_wer": 0.19159501927067893,
242822
  "step": 112750
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
242823
  }
242824
  ],
242825
- "max_steps": 625000,
242826
  "num_train_epochs": 5000,
242827
- "total_flos": 3.1730977775534506e+20,
242828
  "trial_name": null,
242829
  "trial_params": null
242830
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 913.995983935743,
5
+ "global_step": 113370,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
242820
  "eval_steps_per_second": 0.674,
242821
  "eval_wer": 0.19159501927067893,
242822
  "step": 112750
242823
+ },
242824
+ {
242825
+ "epoch": 909.04,
242826
+ "learning_rate": 8.19529886914378e-06,
242827
+ "loss": 0.3194,
242828
+ "step": 112755
242829
+ },
242830
+ {
242831
+ "epoch": 909.08,
242832
+ "learning_rate": 8.195218093699516e-06,
242833
+ "loss": 0.2594,
242834
+ "step": 112760
242835
+ },
242836
+ {
242837
+ "epoch": 909.12,
242838
+ "learning_rate": 8.19513731825525e-06,
242839
+ "loss": 0.323,
242840
+ "step": 112765
242841
+ },
242842
+ {
242843
+ "epoch": 909.16,
242844
+ "learning_rate": 8.195056542810986e-06,
242845
+ "loss": 0.3869,
242846
+ "step": 112770
242847
+ },
242848
+ {
242849
+ "epoch": 909.2,
242850
+ "learning_rate": 8.19497576736672e-06,
242851
+ "loss": 1.2147,
242852
+ "step": 112775
242853
+ },
242854
+ {
242855
+ "epoch": 909.24,
242856
+ "learning_rate": 8.194894991922456e-06,
242857
+ "loss": 0.2554,
242858
+ "step": 112780
242859
+ },
242860
+ {
242861
+ "epoch": 909.28,
242862
+ "learning_rate": 8.19481421647819e-06,
242863
+ "loss": 0.2766,
242864
+ "step": 112785
242865
+ },
242866
+ {
242867
+ "epoch": 909.32,
242868
+ "learning_rate": 8.194733441033926e-06,
242869
+ "loss": 0.3144,
242870
+ "step": 112790
242871
+ },
242872
+ {
242873
+ "epoch": 909.36,
242874
+ "learning_rate": 8.194652665589662e-06,
242875
+ "loss": 0.4586,
242876
+ "step": 112795
242877
+ },
242878
+ {
242879
+ "epoch": 909.4,
242880
+ "learning_rate": 8.194571890145396e-06,
242881
+ "loss": 1.1297,
242882
+ "step": 112800
242883
+ },
242884
+ {
242885
+ "epoch": 909.44,
242886
+ "learning_rate": 8.194491114701132e-06,
242887
+ "loss": 0.3203,
242888
+ "step": 112805
242889
+ },
242890
+ {
242891
+ "epoch": 909.48,
242892
+ "learning_rate": 8.194410339256866e-06,
242893
+ "loss": 0.2947,
242894
+ "step": 112810
242895
+ },
242896
+ {
242897
+ "epoch": 909.52,
242898
+ "learning_rate": 8.194329563812602e-06,
242899
+ "loss": 0.3166,
242900
+ "step": 112815
242901
+ },
242902
+ {
242903
+ "epoch": 909.56,
242904
+ "learning_rate": 8.194248788368336e-06,
242905
+ "loss": 0.4555,
242906
+ "step": 112820
242907
+ },
242908
+ {
242909
+ "epoch": 909.6,
242910
+ "learning_rate": 8.194168012924072e-06,
242911
+ "loss": 1.2846,
242912
+ "step": 112825
242913
+ },
242914
+ {
242915
+ "epoch": 909.64,
242916
+ "learning_rate": 8.194087237479806e-06,
242917
+ "loss": 0.3921,
242918
+ "step": 112830
242919
+ },
242920
+ {
242921
+ "epoch": 909.68,
242922
+ "learning_rate": 8.194006462035542e-06,
242923
+ "loss": 0.3258,
242924
+ "step": 112835
242925
+ },
242926
+ {
242927
+ "epoch": 909.72,
242928
+ "learning_rate": 8.193925686591276e-06,
242929
+ "loss": 0.3415,
242930
+ "step": 112840
242931
+ },
242932
+ {
242933
+ "epoch": 909.76,
242934
+ "learning_rate": 8.193844911147012e-06,
242935
+ "loss": 0.4425,
242936
+ "step": 112845
242937
+ },
242938
+ {
242939
+ "epoch": 909.8,
242940
+ "learning_rate": 8.193764135702748e-06,
242941
+ "loss": 1.1291,
242942
+ "step": 112850
242943
+ },
242944
+ {
242945
+ "epoch": 909.84,
242946
+ "learning_rate": 8.193683360258482e-06,
242947
+ "loss": 0.3087,
242948
+ "step": 112855
242949
+ },
242950
+ {
242951
+ "epoch": 909.88,
242952
+ "learning_rate": 8.193602584814218e-06,
242953
+ "loss": 0.3314,
242954
+ "step": 112860
242955
+ },
242956
+ {
242957
+ "epoch": 909.92,
242958
+ "learning_rate": 8.193521809369952e-06,
242959
+ "loss": 0.329,
242960
+ "step": 112865
242961
+ },
242962
+ {
242963
+ "epoch": 909.96,
242964
+ "learning_rate": 8.193441033925688e-06,
242965
+ "loss": 0.4559,
242966
+ "step": 112870
242967
+ },
242968
+ {
242969
+ "epoch": 910.0,
242970
+ "eval_loss": 0.3853907585144043,
242971
+ "eval_runtime": 39.1333,
242972
+ "eval_samples_per_second": 21.363,
242973
+ "eval_steps_per_second": 0.69,
242974
+ "eval_wer": 0.18328861061419202,
242975
+ "step": 112874
242976
+ },
242977
+ {
242978
+ "epoch": 910.01,
242979
+ "learning_rate": 8.193360258481422e-06,
242980
+ "loss": 0.3418,
242981
+ "step": 112875
242982
+ },
242983
+ {
242984
+ "epoch": 910.05,
242985
+ "learning_rate": 8.193279483037158e-06,
242986
+ "loss": 0.3218,
242987
+ "step": 112880
242988
+ },
242989
+ {
242990
+ "epoch": 910.09,
242991
+ "learning_rate": 8.193198707592892e-06,
242992
+ "loss": 0.2755,
242993
+ "step": 112885
242994
+ },
242995
+ {
242996
+ "epoch": 910.13,
242997
+ "learning_rate": 8.193117932148628e-06,
242998
+ "loss": 0.3143,
242999
+ "step": 112890
243000
+ },
243001
+ {
243002
+ "epoch": 910.17,
243003
+ "learning_rate": 8.193037156704362e-06,
243004
+ "loss": 0.4879,
243005
+ "step": 112895
243006
+ },
243007
+ {
243008
+ "epoch": 910.21,
243009
+ "learning_rate": 8.192956381260098e-06,
243010
+ "loss": 1.1204,
243011
+ "step": 112900
243012
+ },
243013
+ {
243014
+ "epoch": 910.25,
243015
+ "learning_rate": 8.192875605815834e-06,
243016
+ "loss": 0.3345,
243017
+ "step": 112905
243018
+ },
243019
+ {
243020
+ "epoch": 910.29,
243021
+ "learning_rate": 8.192794830371568e-06,
243022
+ "loss": 0.2979,
243023
+ "step": 112910
243024
+ },
243025
+ {
243026
+ "epoch": 910.33,
243027
+ "learning_rate": 8.192714054927304e-06,
243028
+ "loss": 0.3332,
243029
+ "step": 112915
243030
+ },
243031
+ {
243032
+ "epoch": 910.37,
243033
+ "learning_rate": 8.192633279483038e-06,
243034
+ "loss": 0.5393,
243035
+ "step": 112920
243036
+ },
243037
+ {
243038
+ "epoch": 910.41,
243039
+ "learning_rate": 8.192552504038774e-06,
243040
+ "loss": 1.2886,
243041
+ "step": 112925
243042
+ },
243043
+ {
243044
+ "epoch": 910.45,
243045
+ "learning_rate": 8.192471728594508e-06,
243046
+ "loss": 0.3061,
243047
+ "step": 112930
243048
+ },
243049
+ {
243050
+ "epoch": 910.49,
243051
+ "learning_rate": 8.192390953150244e-06,
243052
+ "loss": 0.2628,
243053
+ "step": 112935
243054
+ },
243055
+ {
243056
+ "epoch": 910.53,
243057
+ "learning_rate": 8.192310177705978e-06,
243058
+ "loss": 0.2704,
243059
+ "step": 112940
243060
+ },
243061
+ {
243062
+ "epoch": 910.57,
243063
+ "learning_rate": 8.192229402261714e-06,
243064
+ "loss": 0.5995,
243065
+ "step": 112945
243066
+ },
243067
+ {
243068
+ "epoch": 910.61,
243069
+ "learning_rate": 8.192148626817448e-06,
243070
+ "loss": 1.1335,
243071
+ "step": 112950
243072
+ },
243073
+ {
243074
+ "epoch": 910.65,
243075
+ "learning_rate": 8.192067851373183e-06,
243076
+ "loss": 0.3048,
243077
+ "step": 112955
243078
+ },
243079
+ {
243080
+ "epoch": 910.69,
243081
+ "learning_rate": 8.191987075928918e-06,
243082
+ "loss": 0.3209,
243083
+ "step": 112960
243084
+ },
243085
+ {
243086
+ "epoch": 910.73,
243087
+ "learning_rate": 8.191906300484653e-06,
243088
+ "loss": 0.3437,
243089
+ "step": 112965
243090
+ },
243091
+ {
243092
+ "epoch": 910.77,
243093
+ "learning_rate": 8.19182552504039e-06,
243094
+ "loss": 0.6269,
243095
+ "step": 112970
243096
+ },
243097
+ {
243098
+ "epoch": 910.81,
243099
+ "learning_rate": 8.191744749596123e-06,
243100
+ "loss": 1.0891,
243101
+ "step": 112975
243102
+ },
243103
+ {
243104
+ "epoch": 910.85,
243105
+ "learning_rate": 8.19166397415186e-06,
243106
+ "loss": 0.3427,
243107
+ "step": 112980
243108
+ },
243109
+ {
243110
+ "epoch": 910.89,
243111
+ "learning_rate": 8.191583198707593e-06,
243112
+ "loss": 0.2962,
243113
+ "step": 112985
243114
+ },
243115
+ {
243116
+ "epoch": 910.93,
243117
+ "learning_rate": 8.19150242326333e-06,
243118
+ "loss": 0.3852,
243119
+ "step": 112990
243120
+ },
243121
+ {
243122
+ "epoch": 910.97,
243123
+ "learning_rate": 8.191421647819063e-06,
243124
+ "loss": 0.663,
243125
+ "step": 112995
243126
+ },
243127
+ {
243128
+ "epoch": 911.0,
243129
+ "eval_loss": 0.3766191005706787,
243130
+ "eval_runtime": 41.0289,
243131
+ "eval_samples_per_second": 20.376,
243132
+ "eval_steps_per_second": 0.658,
243133
+ "eval_wer": 0.18356084809758932,
243134
+ "step": 112998
243135
+ },
243136
+ {
243137
+ "epoch": 911.02,
243138
+ "learning_rate": 8.1913408723748e-06,
243139
+ "loss": 0.3573,
243140
+ "step": 113000
243141
+ },
243142
+ {
243143
+ "epoch": 911.06,
243144
+ "learning_rate": 8.191260096930533e-06,
243145
+ "loss": 0.2916,
243146
+ "step": 113005
243147
+ },
243148
+ {
243149
+ "epoch": 911.1,
243150
+ "learning_rate": 8.19117932148627e-06,
243151
+ "loss": 0.2908,
243152
+ "step": 113010
243153
+ },
243154
+ {
243155
+ "epoch": 911.14,
243156
+ "learning_rate": 8.191098546042003e-06,
243157
+ "loss": 0.322,
243158
+ "step": 113015
243159
+ },
243160
+ {
243161
+ "epoch": 911.18,
243162
+ "learning_rate": 8.19101777059774e-06,
243163
+ "loss": 0.5831,
243164
+ "step": 113020
243165
+ },
243166
+ {
243167
+ "epoch": 911.22,
243168
+ "learning_rate": 8.190936995153475e-06,
243169
+ "loss": 0.9917,
243170
+ "step": 113025
243171
+ },
243172
+ {
243173
+ "epoch": 911.26,
243174
+ "learning_rate": 8.19085621970921e-06,
243175
+ "loss": 0.2945,
243176
+ "step": 113030
243177
+ },
243178
+ {
243179
+ "epoch": 911.3,
243180
+ "learning_rate": 8.190775444264945e-06,
243181
+ "loss": 0.2512,
243182
+ "step": 113035
243183
+ },
243184
+ {
243185
+ "epoch": 911.34,
243186
+ "learning_rate": 8.190694668820679e-06,
243187
+ "loss": 0.3935,
243188
+ "step": 113040
243189
+ },
243190
+ {
243191
+ "epoch": 911.38,
243192
+ "learning_rate": 8.190613893376415e-06,
243193
+ "loss": 0.6559,
243194
+ "step": 113045
243195
+ },
243196
+ {
243197
+ "epoch": 911.42,
243198
+ "learning_rate": 8.190533117932149e-06,
243199
+ "loss": 1.014,
243200
+ "step": 113050
243201
+ },
243202
+ {
243203
+ "epoch": 911.46,
243204
+ "learning_rate": 8.190452342487885e-06,
243205
+ "loss": 0.3467,
243206
+ "step": 113055
243207
+ },
243208
+ {
243209
+ "epoch": 911.5,
243210
+ "learning_rate": 8.190371567043619e-06,
243211
+ "loss": 0.3123,
243212
+ "step": 113060
243213
+ },
243214
+ {
243215
+ "epoch": 911.54,
243216
+ "learning_rate": 8.190290791599355e-06,
243217
+ "loss": 0.4067,
243218
+ "step": 113065
243219
+ },
243220
+ {
243221
+ "epoch": 911.58,
243222
+ "learning_rate": 8.190210016155089e-06,
243223
+ "loss": 0.6399,
243224
+ "step": 113070
243225
+ },
243226
+ {
243227
+ "epoch": 911.62,
243228
+ "learning_rate": 8.190129240710825e-06,
243229
+ "loss": 0.9487,
243230
+ "step": 113075
243231
+ },
243232
+ {
243233
+ "epoch": 911.66,
243234
+ "learning_rate": 8.19004846526656e-06,
243235
+ "loss": 0.2608,
243236
+ "step": 113080
243237
+ },
243238
+ {
243239
+ "epoch": 911.7,
243240
+ "learning_rate": 8.189967689822295e-06,
243241
+ "loss": 0.229,
243242
+ "step": 113085
243243
+ },
243244
+ {
243245
+ "epoch": 911.74,
243246
+ "learning_rate": 8.18988691437803e-06,
243247
+ "loss": 0.3772,
243248
+ "step": 113090
243249
+ },
243250
+ {
243251
+ "epoch": 911.78,
243252
+ "learning_rate": 8.189806138933765e-06,
243253
+ "loss": 0.6131,
243254
+ "step": 113095
243255
+ },
243256
+ {
243257
+ "epoch": 911.82,
243258
+ "learning_rate": 8.189741518578352e-06,
243259
+ "loss": 1.0935,
243260
+ "step": 113100
243261
+ },
243262
+ {
243263
+ "epoch": 911.86,
243264
+ "learning_rate": 8.189660743134088e-06,
243265
+ "loss": 0.3042,
243266
+ "step": 113105
243267
+ },
243268
+ {
243269
+ "epoch": 911.9,
243270
+ "learning_rate": 8.189579967689822e-06,
243271
+ "loss": 0.2531,
243272
+ "step": 113110
243273
+ },
243274
+ {
243275
+ "epoch": 911.94,
243276
+ "learning_rate": 8.189499192245558e-06,
243277
+ "loss": 0.4252,
243278
+ "step": 113115
243279
+ },
243280
+ {
243281
+ "epoch": 911.98,
243282
+ "learning_rate": 8.189418416801294e-06,
243283
+ "loss": 0.6596,
243284
+ "step": 113120
243285
+ },
243286
+ {
243287
+ "epoch": 912.0,
243288
+ "eval_loss": 0.444323867559433,
243289
+ "eval_runtime": 40.7673,
243290
+ "eval_samples_per_second": 20.507,
243291
+ "eval_steps_per_second": 0.662,
243292
+ "eval_wer": 0.18970503181029497,
243293
+ "step": 113122
243294
+ },
243295
+ {
243296
+ "epoch": 912.02,
243297
+ "learning_rate": 8.189337641357028e-06,
243298
+ "loss": 0.3245,
243299
+ "step": 113125
243300
+ },
243301
+ {
243302
+ "epoch": 912.06,
243303
+ "learning_rate": 8.189256865912764e-06,
243304
+ "loss": 0.2808,
243305
+ "step": 113130
243306
+ },
243307
+ {
243308
+ "epoch": 912.1,
243309
+ "learning_rate": 8.189176090468498e-06,
243310
+ "loss": 0.3132,
243311
+ "step": 113135
243312
+ },
243313
+ {
243314
+ "epoch": 912.14,
243315
+ "learning_rate": 8.189095315024234e-06,
243316
+ "loss": 0.3814,
243317
+ "step": 113140
243318
+ },
243319
+ {
243320
+ "epoch": 912.18,
243321
+ "learning_rate": 8.189014539579968e-06,
243322
+ "loss": 0.7656,
243323
+ "step": 113145
243324
+ },
243325
+ {
243326
+ "epoch": 912.22,
243327
+ "learning_rate": 8.188933764135704e-06,
243328
+ "loss": 0.9321,
243329
+ "step": 113150
243330
+ },
243331
+ {
243332
+ "epoch": 912.27,
243333
+ "learning_rate": 8.188852988691438e-06,
243334
+ "loss": 0.2926,
243335
+ "step": 113155
243336
+ },
243337
+ {
243338
+ "epoch": 912.31,
243339
+ "learning_rate": 8.188772213247174e-06,
243340
+ "loss": 0.3008,
243341
+ "step": 113160
243342
+ },
243343
+ {
243344
+ "epoch": 912.35,
243345
+ "learning_rate": 8.188691437802908e-06,
243346
+ "loss": 0.3474,
243347
+ "step": 113165
243348
+ },
243349
+ {
243350
+ "epoch": 912.39,
243351
+ "learning_rate": 8.188610662358644e-06,
243352
+ "loss": 0.8603,
243353
+ "step": 113170
243354
+ },
243355
+ {
243356
+ "epoch": 912.43,
243357
+ "learning_rate": 8.18852988691438e-06,
243358
+ "loss": 0.9023,
243359
+ "step": 113175
243360
+ },
243361
+ {
243362
+ "epoch": 912.47,
243363
+ "learning_rate": 8.188449111470114e-06,
243364
+ "loss": 0.2997,
243365
+ "step": 113180
243366
+ },
243367
+ {
243368
+ "epoch": 912.51,
243369
+ "learning_rate": 8.18836833602585e-06,
243370
+ "loss": 0.3067,
243371
+ "step": 113185
243372
+ },
243373
+ {
243374
+ "epoch": 912.55,
243375
+ "learning_rate": 8.188287560581584e-06,
243376
+ "loss": 0.36,
243377
+ "step": 113190
243378
+ },
243379
+ {
243380
+ "epoch": 912.59,
243381
+ "learning_rate": 8.18820678513732e-06,
243382
+ "loss": 0.7879,
243383
+ "step": 113195
243384
+ },
243385
+ {
243386
+ "epoch": 912.63,
243387
+ "learning_rate": 8.188126009693054e-06,
243388
+ "loss": 0.9157,
243389
+ "step": 113200
243390
+ },
243391
+ {
243392
+ "epoch": 912.67,
243393
+ "learning_rate": 8.18804523424879e-06,
243394
+ "loss": 0.3224,
243395
+ "step": 113205
243396
+ },
243397
+ {
243398
+ "epoch": 912.71,
243399
+ "learning_rate": 8.187964458804524e-06,
243400
+ "loss": 0.2828,
243401
+ "step": 113210
243402
+ },
243403
+ {
243404
+ "epoch": 912.75,
243405
+ "learning_rate": 8.18788368336026e-06,
243406
+ "loss": 0.3505,
243407
+ "step": 113215
243408
+ },
243409
+ {
243410
+ "epoch": 912.79,
243411
+ "learning_rate": 8.187802907915994e-06,
243412
+ "loss": 0.7458,
243413
+ "step": 113220
243414
+ },
243415
+ {
243416
+ "epoch": 912.83,
243417
+ "learning_rate": 8.18772213247173e-06,
243418
+ "loss": 0.793,
243419
+ "step": 113225
243420
+ },
243421
+ {
243422
+ "epoch": 912.87,
243423
+ "learning_rate": 8.187641357027464e-06,
243424
+ "loss": 0.3386,
243425
+ "step": 113230
243426
+ },
243427
+ {
243428
+ "epoch": 912.91,
243429
+ "learning_rate": 8.1875605815832e-06,
243430
+ "loss": 0.3076,
243431
+ "step": 113235
243432
+ },
243433
+ {
243434
+ "epoch": 912.95,
243435
+ "learning_rate": 8.187479806138935e-06,
243436
+ "loss": 0.3701,
243437
+ "step": 113240
243438
+ },
243439
+ {
243440
+ "epoch": 912.99,
243441
+ "learning_rate": 8.18739903069467e-06,
243442
+ "loss": 0.807,
243443
+ "step": 113245
243444
+ },
243445
+ {
243446
+ "epoch": 913.0,
243447
+ "eval_loss": 0.4701347053050995,
243448
+ "eval_runtime": 41.2757,
243449
+ "eval_samples_per_second": 20.278,
243450
+ "eval_steps_per_second": 0.654,
243451
+ "eval_wer": 0.17971326164874551,
243452
+ "step": 113246
243453
+ },
243454
+ {
243455
+ "epoch": 913.03,
243456
+ "learning_rate": 8.187318255250405e-06,
243457
+ "loss": 0.3049,
243458
+ "step": 113250
243459
+ },
243460
+ {
243461
+ "epoch": 913.07,
243462
+ "learning_rate": 8.18723747980614e-06,
243463
+ "loss": 0.3326,
243464
+ "step": 113255
243465
+ },
243466
+ {
243467
+ "epoch": 913.11,
243468
+ "learning_rate": 8.187156704361875e-06,
243469
+ "loss": 0.2772,
243470
+ "step": 113260
243471
+ },
243472
+ {
243473
+ "epoch": 913.15,
243474
+ "learning_rate": 8.18707592891761e-06,
243475
+ "loss": 0.3754,
243476
+ "step": 113265
243477
+ },
243478
+ {
243479
+ "epoch": 913.19,
243480
+ "learning_rate": 8.186995153473345e-06,
243481
+ "loss": 0.8271,
243482
+ "step": 113270
243483
+ },
243484
+ {
243485
+ "epoch": 913.23,
243486
+ "learning_rate": 8.18691437802908e-06,
243487
+ "loss": 0.7523,
243488
+ "step": 113275
243489
+ },
243490
+ {
243491
+ "epoch": 913.27,
243492
+ "learning_rate": 8.186833602584815e-06,
243493
+ "loss": 0.2855,
243494
+ "step": 113280
243495
+ },
243496
+ {
243497
+ "epoch": 913.31,
243498
+ "learning_rate": 8.18675282714055e-06,
243499
+ "loss": 0.3099,
243500
+ "step": 113285
243501
+ },
243502
+ {
243503
+ "epoch": 913.35,
243504
+ "learning_rate": 8.186672051696285e-06,
243505
+ "loss": 0.3889,
243506
+ "step": 113290
243507
+ },
243508
+ {
243509
+ "epoch": 913.39,
243510
+ "learning_rate": 8.186591276252021e-06,
243511
+ "loss": 0.8476,
243512
+ "step": 113295
243513
+ },
243514
+ {
243515
+ "epoch": 913.43,
243516
+ "learning_rate": 8.186510500807755e-06,
243517
+ "loss": 0.6524,
243518
+ "step": 113300
243519
+ },
243520
+ {
243521
+ "epoch": 913.47,
243522
+ "learning_rate": 8.186429725363491e-06,
243523
+ "loss": 0.2498,
243524
+ "step": 113305
243525
+ },
243526
+ {
243527
+ "epoch": 913.51,
243528
+ "learning_rate": 8.186348949919225e-06,
243529
+ "loss": 0.3264,
243530
+ "step": 113310
243531
+ },
243532
+ {
243533
+ "epoch": 913.55,
243534
+ "learning_rate": 8.186268174474961e-06,
243535
+ "loss": 0.3934,
243536
+ "step": 113315
243537
+ },
243538
+ {
243539
+ "epoch": 913.59,
243540
+ "learning_rate": 8.186187399030695e-06,
243541
+ "loss": 0.9911,
243542
+ "step": 113320
243543
+ },
243544
+ {
243545
+ "epoch": 913.63,
243546
+ "learning_rate": 8.18610662358643e-06,
243547
+ "loss": 0.6645,
243548
+ "step": 113325
243549
+ },
243550
+ {
243551
+ "epoch": 913.67,
243552
+ "learning_rate": 8.186025848142165e-06,
243553
+ "loss": 0.2985,
243554
+ "step": 113330
243555
+ },
243556
+ {
243557
+ "epoch": 913.71,
243558
+ "learning_rate": 8.1859450726979e-06,
243559
+ "loss": 0.3302,
243560
+ "step": 113335
243561
+ },
243562
+ {
243563
+ "epoch": 913.76,
243564
+ "learning_rate": 8.185864297253635e-06,
243565
+ "loss": 0.3933,
243566
+ "step": 113340
243567
+ },
243568
+ {
243569
+ "epoch": 913.8,
243570
+ "learning_rate": 8.18578352180937e-06,
243571
+ "loss": 0.9581,
243572
+ "step": 113345
243573
+ },
243574
+ {
243575
+ "epoch": 913.84,
243576
+ "learning_rate": 8.185702746365107e-06,
243577
+ "loss": 0.7659,
243578
+ "step": 113350
243579
+ },
243580
+ {
243581
+ "epoch": 913.88,
243582
+ "learning_rate": 8.18562197092084e-06,
243583
+ "loss": 0.2852,
243584
+ "step": 113355
243585
+ },
243586
+ {
243587
+ "epoch": 913.92,
243588
+ "learning_rate": 8.185541195476577e-06,
243589
+ "loss": 0.2891,
243590
+ "step": 113360
243591
+ },
243592
+ {
243593
+ "epoch": 913.96,
243594
+ "learning_rate": 8.18546042003231e-06,
243595
+ "loss": 0.4714,
243596
+ "step": 113365
243597
+ },
243598
+ {
243599
+ "epoch": 914.0,
243600
+ "learning_rate": 8.185379644588047e-06,
243601
+ "loss": 1.1086,
243602
+ "step": 113370
243603
+ },
243604
+ {
243605
+ "epoch": 914.0,
243606
+ "eval_loss": 0.39893263578414917,
243607
+ "eval_runtime": 39.8088,
243608
+ "eval_samples_per_second": 21.025,
243609
+ "eval_steps_per_second": 0.678,
243610
+ "eval_wer": 0.18508635386553998,
243611
+ "step": 113370
243612
  }
243613
  ],
243614
+ "max_steps": 620000,
243615
  "num_train_epochs": 5000,
243616
+ "total_flos": 3.190555567482095e+20,
243617
  "trial_name": null,
243618
  "trial_params": null
243619
  }
model-bin/finetune/base/{checkpoint-112750 β†’ checkpoint-113370}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630082884.969348/events.out.tfevents.1630082884.dea8124df033.1391.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41fd86e6e7c1a5c77593c5bbda325bd5c975686ea34b46f2f3a6e958dbca7861
3
+ size 4194
model-bin/finetune/base/log/1630083307.1487231/events.out.tfevents.1630083307.dea8124df033.1391.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11391acdb58354ca2cac2c60fd34dff94d1f18b0bcbdfb0572682728e5a86432
3
+ size 4194
model-bin/finetune/base/log/1630083735.6256015/events.out.tfevents.1630083735.dea8124df033.1391.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e25a50c37d80004082e7bc62597d5bf40a647914e074c4a409928874ead615a
3
+ size 4194
model-bin/finetune/base/log/1630084174.1170754/events.out.tfevents.1630084174.dea8124df033.1391.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6063e6cb2ea6a249fc018a71d2bc14207d6991ca21c0e26094c96758b8d8827a
3
+ size 4194
model-bin/finetune/base/log/1630084598.2761755/events.out.tfevents.1630084598.dea8124df033.1391.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f620b324894ddc530420ee660ac32ef67a80a90e8eaf65637db8ee2548cf5cd
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630082884.dea8124df033.1391.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f9dc3f5f57f07b97c5848164fa8a44cf6407370c74df678c5318e16f5e9bd6e
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630083307.dea8124df033.1391.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81129d21b671588c0030205e0ce3ee50b437dcd581b30365dc937c7d279d6f2b
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630083735.dea8124df033.1391.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9203203993faf14718631ba75e93c0624f6aa871e966f57dfaa38274522d3aea
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630084174.dea8124df033.1391.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1059774e54b65b2bea6167c90283c2debf4d7672d04144c264476154ca1eedc
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630084598.dea8124df033.1391.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c1ce76c497b5a7a4d19715c80fc0b5e0948b520e68b45153bb07b93b8e9a7f4
3
+ size 8622