Check commited on
Commit
abe7973
Β·
1 Parent(s): f458085

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630250753.668995/events.out.tfevents.1630250753.cc93b136ebf5.1086.209 +3 -0
  11. model-bin/finetune/base/log/1630251188.88167/events.out.tfevents.1630251188.cc93b136ebf5.1086.211 +3 -0
  12. model-bin/finetune/base/log/1630251628.1797204/events.out.tfevents.1630251628.cc93b136ebf5.1086.213 +3 -0
  13. model-bin/finetune/base/log/1630252065.7514985/events.out.tfevents.1630252065.cc93b136ebf5.1086.215 +3 -0
  14. model-bin/finetune/base/log/1630252500.5741549/events.out.tfevents.1630252500.cc93b136ebf5.1086.217 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630250753.cc93b136ebf5.1086.208 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630251188.cc93b136ebf5.1086.210 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630251628.cc93b136ebf5.1086.212 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630252065.cc93b136ebf5.1086.214 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630252500.cc93b136ebf5.1086.216 +3 -0
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88106383b07c091dad16598681c368eb1f7d5dd6754f75dcdede18a251a76193
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:013e7596c9833c0f02cb6d858875c8b42145994a60410ff7acc22c55ed3a3f78
3
  size 722165393
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e90c5a61e1081391cac20c2749dbcf1cb5ad34446b8a2074676f8de81385b325
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:358bbb0a31f3eada3dcf3518c389f01e7260b174069b7a039e533c23adcbc695
3
  size 377909911
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e5c8178a08780b7e94358c9559057df556e07dd698f9fa5dc769d640159e51e
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c64726b931c7ed69c0aa3b500fe65f0f56bb20ca55efd46d255f359a9d38051f
3
  size 14503
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fe336fce1f24fb6ba716a3ec77454f1b6f380bbf0069997841f401a2fb77c2a
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e38d467b638b9bb9004c2051e9035f041fc23da780dcb2db9c2dbc1bb4f1e23d
3
  size 559
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d1c8fd19316f91f2cc4123c75c86425f3af495dca75999776dc0193fa6dd50cb
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1740c53bf7ef796ec727899064cf449f45e9e8026922bf22f64ed1a0b45bd32d
3
  size 623
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1225.995983935743,
5
- "global_step": 152068,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -292842,11 +292842,806 @@
292842
  "eval_steps_per_second": 0.667,
292843
  "eval_wer": 0.1764970733903647,
292844
  "step": 152068
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
292845
  }
292846
  ],
292847
- "max_steps": 620000,
292848
  "num_train_epochs": 5000,
292849
- "total_flos": 4.2793273628046655e+20,
292850
  "trial_name": null,
292851
  "trial_params": null
292852
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1221.0,
5
+ "global_step": 152690,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
292842
  "eval_steps_per_second": 0.667,
292843
  "eval_wer": 0.1764970733903647,
292844
  "step": 152068
292845
+ },
292846
+ {
292847
+ "epoch": 1226.02,
292848
+ "learning_rate": 7.560403877221325e-06,
292849
+ "loss": 0.3566,
292850
+ "step": 152070
292851
+ },
292852
+ {
292853
+ "epoch": 1226.06,
292854
+ "learning_rate": 7.5603231017770605e-06,
292855
+ "loss": 0.2873,
292856
+ "step": 152075
292857
+ },
292858
+ {
292859
+ "epoch": 1226.1,
292860
+ "learning_rate": 7.5602423263327955e-06,
292861
+ "loss": 0.2621,
292862
+ "step": 152080
292863
+ },
292864
+ {
292865
+ "epoch": 1226.14,
292866
+ "learning_rate": 7.5601615508885305e-06,
292867
+ "loss": 0.3207,
292868
+ "step": 152085
292869
+ },
292870
+ {
292871
+ "epoch": 1226.18,
292872
+ "learning_rate": 7.5600807754442655e-06,
292873
+ "loss": 0.7188,
292874
+ "step": 152090
292875
+ },
292876
+ {
292877
+ "epoch": 1226.22,
292878
+ "learning_rate": 7.5600000000000005e-06,
292879
+ "loss": 1.0223,
292880
+ "step": 152095
292881
+ },
292882
+ {
292883
+ "epoch": 1226.26,
292884
+ "learning_rate": 7.5599192245557355e-06,
292885
+ "loss": 0.2622,
292886
+ "step": 152100
292887
+ },
292888
+ {
292889
+ "epoch": 1226.3,
292890
+ "learning_rate": 7.5598384491114704e-06,
292891
+ "loss": 0.2759,
292892
+ "step": 152105
292893
+ },
292894
+ {
292895
+ "epoch": 1226.34,
292896
+ "learning_rate": 7.559757673667206e-06,
292897
+ "loss": 0.3349,
292898
+ "step": 152110
292899
+ },
292900
+ {
292901
+ "epoch": 1226.38,
292902
+ "learning_rate": 7.55967689822294e-06,
292903
+ "loss": 0.5856,
292904
+ "step": 152115
292905
+ },
292906
+ {
292907
+ "epoch": 1226.42,
292908
+ "learning_rate": 7.559596122778676e-06,
292909
+ "loss": 0.9452,
292910
+ "step": 152120
292911
+ },
292912
+ {
292913
+ "epoch": 1226.46,
292914
+ "learning_rate": 7.55951534733441e-06,
292915
+ "loss": 0.2844,
292916
+ "step": 152125
292917
+ },
292918
+ {
292919
+ "epoch": 1226.5,
292920
+ "learning_rate": 7.559434571890146e-06,
292921
+ "loss": 0.2637,
292922
+ "step": 152130
292923
+ },
292924
+ {
292925
+ "epoch": 1226.54,
292926
+ "learning_rate": 7.55935379644588e-06,
292927
+ "loss": 0.2986,
292928
+ "step": 152135
292929
+ },
292930
+ {
292931
+ "epoch": 1226.58,
292932
+ "learning_rate": 7.559273021001616e-06,
292933
+ "loss": 0.6334,
292934
+ "step": 152140
292935
+ },
292936
+ {
292937
+ "epoch": 1226.62,
292938
+ "learning_rate": 7.55919224555735e-06,
292939
+ "loss": 0.9182,
292940
+ "step": 152145
292941
+ },
292942
+ {
292943
+ "epoch": 1226.66,
292944
+ "learning_rate": 7.559111470113086e-06,
292945
+ "loss": 0.2254,
292946
+ "step": 152150
292947
+ },
292948
+ {
292949
+ "epoch": 1226.7,
292950
+ "learning_rate": 7.559030694668821e-06,
292951
+ "loss": 0.2388,
292952
+ "step": 152155
292953
+ },
292954
+ {
292955
+ "epoch": 1226.74,
292956
+ "learning_rate": 7.558949919224556e-06,
292957
+ "loss": 0.3118,
292958
+ "step": 152160
292959
+ },
292960
+ {
292961
+ "epoch": 1226.78,
292962
+ "learning_rate": 7.558869143780292e-06,
292963
+ "loss": 0.6253,
292964
+ "step": 152165
292965
+ },
292966
+ {
292967
+ "epoch": 1226.82,
292968
+ "learning_rate": 7.558788368336026e-06,
292969
+ "loss": 0.9901,
292970
+ "step": 152170
292971
+ },
292972
+ {
292973
+ "epoch": 1226.86,
292974
+ "learning_rate": 7.558707592891762e-06,
292975
+ "loss": 0.2865,
292976
+ "step": 152175
292977
+ },
292978
+ {
292979
+ "epoch": 1226.9,
292980
+ "learning_rate": 7.558626817447496e-06,
292981
+ "loss": 0.3002,
292982
+ "step": 152180
292983
+ },
292984
+ {
292985
+ "epoch": 1226.94,
292986
+ "learning_rate": 7.558546042003232e-06,
292987
+ "loss": 0.3956,
292988
+ "step": 152185
292989
+ },
292990
+ {
292991
+ "epoch": 1226.98,
292992
+ "learning_rate": 7.558465266558966e-06,
292993
+ "loss": 0.687,
292994
+ "step": 152190
292995
+ },
292996
+ {
292997
+ "epoch": 1227.0,
292998
+ "eval_loss": 0.39228349924087524,
292999
+ "eval_runtime": 41.5829,
293000
+ "eval_samples_per_second": 20.104,
293001
+ "eval_steps_per_second": 0.649,
293002
+ "eval_wer": 0.18111669214530102,
293003
+ "step": 152192
293004
+ },
293005
+ {
293006
+ "epoch": 1227.02,
293007
+ "learning_rate": 7.558384491114702e-06,
293008
+ "loss": 0.2944,
293009
+ "step": 152195
293010
+ },
293011
+ {
293012
+ "epoch": 1227.06,
293013
+ "learning_rate": 7.558303715670436e-06,
293014
+ "loss": 0.2883,
293015
+ "step": 152200
293016
+ },
293017
+ {
293018
+ "epoch": 1227.1,
293019
+ "learning_rate": 7.558222940226172e-06,
293020
+ "loss": 0.299,
293021
+ "step": 152205
293022
+ },
293023
+ {
293024
+ "epoch": 1227.14,
293025
+ "learning_rate": 7.558142164781906e-06,
293026
+ "loss": 0.2715,
293027
+ "step": 152210
293028
+ },
293029
+ {
293030
+ "epoch": 1227.18,
293031
+ "learning_rate": 7.558061389337642e-06,
293032
+ "loss": 0.7032,
293033
+ "step": 152215
293034
+ },
293035
+ {
293036
+ "epoch": 1227.22,
293037
+ "learning_rate": 7.557980613893378e-06,
293038
+ "loss": 0.889,
293039
+ "step": 152220
293040
+ },
293041
+ {
293042
+ "epoch": 1227.27,
293043
+ "learning_rate": 7.557899838449112e-06,
293044
+ "loss": 0.2777,
293045
+ "step": 152225
293046
+ },
293047
+ {
293048
+ "epoch": 1227.31,
293049
+ "learning_rate": 7.557819063004848e-06,
293050
+ "loss": 0.3006,
293051
+ "step": 152230
293052
+ },
293053
+ {
293054
+ "epoch": 1227.35,
293055
+ "learning_rate": 7.557738287560582e-06,
293056
+ "loss": 0.3308,
293057
+ "step": 152235
293058
+ },
293059
+ {
293060
+ "epoch": 1227.39,
293061
+ "learning_rate": 7.557657512116318e-06,
293062
+ "loss": 0.7281,
293063
+ "step": 152240
293064
+ },
293065
+ {
293066
+ "epoch": 1227.43,
293067
+ "learning_rate": 7.557576736672052e-06,
293068
+ "loss": 0.8714,
293069
+ "step": 152245
293070
+ },
293071
+ {
293072
+ "epoch": 1227.47,
293073
+ "learning_rate": 7.557495961227788e-06,
293074
+ "loss": 0.2962,
293075
+ "step": 152250
293076
+ },
293077
+ {
293078
+ "epoch": 1227.51,
293079
+ "learning_rate": 7.557415185783522e-06,
293080
+ "loss": 0.2503,
293081
+ "step": 152255
293082
+ },
293083
+ {
293084
+ "epoch": 1227.55,
293085
+ "learning_rate": 7.557334410339258e-06,
293086
+ "loss": 0.3374,
293087
+ "step": 152260
293088
+ },
293089
+ {
293090
+ "epoch": 1227.59,
293091
+ "learning_rate": 7.557253634894992e-06,
293092
+ "loss": 0.6683,
293093
+ "step": 152265
293094
+ },
293095
+ {
293096
+ "epoch": 1227.63,
293097
+ "learning_rate": 7.5571728594507276e-06,
293098
+ "loss": 0.7008,
293099
+ "step": 152270
293100
+ },
293101
+ {
293102
+ "epoch": 1227.67,
293103
+ "learning_rate": 7.557092084006462e-06,
293104
+ "loss": 0.3346,
293105
+ "step": 152275
293106
+ },
293107
+ {
293108
+ "epoch": 1227.71,
293109
+ "learning_rate": 7.5570113085621975e-06,
293110
+ "loss": 0.2314,
293111
+ "step": 152280
293112
+ },
293113
+ {
293114
+ "epoch": 1227.75,
293115
+ "learning_rate": 7.556930533117933e-06,
293116
+ "loss": 0.4819,
293117
+ "step": 152285
293118
+ },
293119
+ {
293120
+ "epoch": 1227.79,
293121
+ "learning_rate": 7.5568497576736675e-06,
293122
+ "loss": 0.5945,
293123
+ "step": 152290
293124
+ },
293125
+ {
293126
+ "epoch": 1227.83,
293127
+ "learning_rate": 7.556768982229403e-06,
293128
+ "loss": 0.8223,
293129
+ "step": 152295
293130
+ },
293131
+ {
293132
+ "epoch": 1227.87,
293133
+ "learning_rate": 7.5566882067851375e-06,
293134
+ "loss": 0.2747,
293135
+ "step": 152300
293136
+ },
293137
+ {
293138
+ "epoch": 1227.91,
293139
+ "learning_rate": 7.556607431340873e-06,
293140
+ "loss": 0.3093,
293141
+ "step": 152305
293142
+ },
293143
+ {
293144
+ "epoch": 1227.95,
293145
+ "learning_rate": 7.5565266558966075e-06,
293146
+ "loss": 0.4192,
293147
+ "step": 152310
293148
+ },
293149
+ {
293150
+ "epoch": 1227.99,
293151
+ "learning_rate": 7.556445880452343e-06,
293152
+ "loss": 0.7994,
293153
+ "step": 152315
293154
+ },
293155
+ {
293156
+ "epoch": 1228.0,
293157
+ "eval_loss": 0.4167693257331848,
293158
+ "eval_runtime": 41.5016,
293159
+ "eval_samples_per_second": 20.144,
293160
+ "eval_steps_per_second": 0.651,
293161
+ "eval_wer": 0.18405892219062203,
293162
+ "step": 152316
293163
+ },
293164
+ {
293165
+ "epoch": 1218.03,
293166
+ "learning_rate": 7.5563651050080774e-06,
293167
+ "loss": 0.2815,
293168
+ "step": 152320
293169
+ },
293170
+ {
293171
+ "epoch": 1218.07,
293172
+ "learning_rate": 7.556284329563813e-06,
293173
+ "loss": 0.2471,
293174
+ "step": 152325
293175
+ },
293176
+ {
293177
+ "epoch": 1218.11,
293178
+ "learning_rate": 7.556203554119547e-06,
293179
+ "loss": 0.2546,
293180
+ "step": 152330
293181
+ },
293182
+ {
293183
+ "epoch": 1218.15,
293184
+ "learning_rate": 7.556122778675283e-06,
293185
+ "loss": 0.339,
293186
+ "step": 152335
293187
+ },
293188
+ {
293189
+ "epoch": 1218.19,
293190
+ "learning_rate": 7.556042003231019e-06,
293191
+ "loss": 0.7478,
293192
+ "step": 152340
293193
+ },
293194
+ {
293195
+ "epoch": 1218.23,
293196
+ "learning_rate": 7.555961227786753e-06,
293197
+ "loss": 0.6217,
293198
+ "step": 152345
293199
+ },
293200
+ {
293201
+ "epoch": 1218.27,
293202
+ "learning_rate": 7.555880452342489e-06,
293203
+ "loss": 0.2577,
293204
+ "step": 152350
293205
+ },
293206
+ {
293207
+ "epoch": 1218.31,
293208
+ "learning_rate": 7.555799676898223e-06,
293209
+ "loss": 0.2436,
293210
+ "step": 152355
293211
+ },
293212
+ {
293213
+ "epoch": 1218.35,
293214
+ "learning_rate": 7.555718901453959e-06,
293215
+ "loss": 0.437,
293216
+ "step": 152360
293217
+ },
293218
+ {
293219
+ "epoch": 1218.39,
293220
+ "learning_rate": 7.555638126009693e-06,
293221
+ "loss": 0.829,
293222
+ "step": 152365
293223
+ },
293224
+ {
293225
+ "epoch": 1218.43,
293226
+ "learning_rate": 7.555557350565429e-06,
293227
+ "loss": 0.5973,
293228
+ "step": 152370
293229
+ },
293230
+ {
293231
+ "epoch": 1218.47,
293232
+ "learning_rate": 7.555476575121163e-06,
293233
+ "loss": 0.2641,
293234
+ "step": 152375
293235
+ },
293236
+ {
293237
+ "epoch": 1218.51,
293238
+ "learning_rate": 7.555395799676899e-06,
293239
+ "loss": 0.2739,
293240
+ "step": 152380
293241
+ },
293242
+ {
293243
+ "epoch": 1218.55,
293244
+ "learning_rate": 7.555315024232633e-06,
293245
+ "loss": 0.4165,
293246
+ "step": 152385
293247
+ },
293248
+ {
293249
+ "epoch": 1218.59,
293250
+ "learning_rate": 7.555234248788369e-06,
293251
+ "loss": 0.826,
293252
+ "step": 152390
293253
+ },
293254
+ {
293255
+ "epoch": 1218.63,
293256
+ "learning_rate": 7.555153473344105e-06,
293257
+ "loss": 0.6447,
293258
+ "step": 152395
293259
+ },
293260
+ {
293261
+ "epoch": 1218.67,
293262
+ "learning_rate": 7.555072697899839e-06,
293263
+ "loss": 0.2597,
293264
+ "step": 152400
293265
+ },
293266
+ {
293267
+ "epoch": 1218.71,
293268
+ "learning_rate": 7.554991922455575e-06,
293269
+ "loss": 0.2887,
293270
+ "step": 152405
293271
+ },
293272
+ {
293273
+ "epoch": 1218.75,
293274
+ "learning_rate": 7.554911147011309e-06,
293275
+ "loss": 0.4304,
293276
+ "step": 152410
293277
+ },
293278
+ {
293279
+ "epoch": 1218.79,
293280
+ "learning_rate": 7.554830371567045e-06,
293281
+ "loss": 0.9143,
293282
+ "step": 152415
293283
+ },
293284
+ {
293285
+ "epoch": 1218.83,
293286
+ "learning_rate": 7.554749596122779e-06,
293287
+ "loss": 0.6371,
293288
+ "step": 152420
293289
+ },
293290
+ {
293291
+ "epoch": 1218.87,
293292
+ "learning_rate": 7.554668820678515e-06,
293293
+ "loss": 0.2591,
293294
+ "step": 152425
293295
+ },
293296
+ {
293297
+ "epoch": 1218.91,
293298
+ "learning_rate": 7.554588045234249e-06,
293299
+ "loss": 0.3687,
293300
+ "step": 152430
293301
+ },
293302
+ {
293303
+ "epoch": 1218.95,
293304
+ "learning_rate": 7.554507269789985e-06,
293305
+ "loss": 0.4046,
293306
+ "step": 152435
293307
+ },
293308
+ {
293309
+ "epoch": 1218.99,
293310
+ "learning_rate": 7.554426494345719e-06,
293311
+ "loss": 0.902,
293312
+ "step": 152440
293313
+ },
293314
+ {
293315
+ "epoch": 1219.0,
293316
+ "eval_loss": 0.35546669363975525,
293317
+ "eval_runtime": 42.7299,
293318
+ "eval_samples_per_second": 19.565,
293319
+ "eval_steps_per_second": 0.632,
293320
+ "eval_wer": 0.17738750176329524,
293321
+ "step": 152441
293322
+ },
293323
+ {
293324
+ "epoch": 1229.03,
293325
+ "learning_rate": 7.554345718901455e-06,
293326
+ "loss": 0.2988,
293327
+ "step": 152445
293328
+ },
293329
+ {
293330
+ "epoch": 1229.07,
293331
+ "learning_rate": 7.554264943457189e-06,
293332
+ "loss": 0.2575,
293333
+ "step": 152450
293334
+ },
293335
+ {
293336
+ "epoch": 1229.11,
293337
+ "learning_rate": 7.554184168012925e-06,
293338
+ "loss": 0.3322,
293339
+ "step": 152455
293340
+ },
293341
+ {
293342
+ "epoch": 1229.15,
293343
+ "learning_rate": 7.5541033925686605e-06,
293344
+ "loss": 0.4137,
293345
+ "step": 152460
293346
+ },
293347
+ {
293348
+ "epoch": 1229.19,
293349
+ "learning_rate": 7.554022617124395e-06,
293350
+ "loss": 0.8767,
293351
+ "step": 152465
293352
+ },
293353
+ {
293354
+ "epoch": 1229.23,
293355
+ "learning_rate": 7.5539418416801305e-06,
293356
+ "loss": 0.625,
293357
+ "step": 152470
293358
+ },
293359
+ {
293360
+ "epoch": 1229.27,
293361
+ "learning_rate": 7.553861066235865e-06,
293362
+ "loss": 0.272,
293363
+ "step": 152475
293364
+ },
293365
+ {
293366
+ "epoch": 1229.31,
293367
+ "learning_rate": 7.5537802907916004e-06,
293368
+ "loss": 0.3351,
293369
+ "step": 152480
293370
+ },
293371
+ {
293372
+ "epoch": 1229.35,
293373
+ "learning_rate": 7.5536995153473346e-06,
293374
+ "loss": 0.356,
293375
+ "step": 152485
293376
+ },
293377
+ {
293378
+ "epoch": 1229.39,
293379
+ "learning_rate": 7.55361873990307e-06,
293380
+ "loss": 0.8533,
293381
+ "step": 152490
293382
+ },
293383
+ {
293384
+ "epoch": 1229.43,
293385
+ "learning_rate": 7.5535379644588045e-06,
293386
+ "loss": 0.6315,
293387
+ "step": 152495
293388
+ },
293389
+ {
293390
+ "epoch": 1229.47,
293391
+ "learning_rate": 7.55345718901454e-06,
293392
+ "loss": 0.2547,
293393
+ "step": 152500
293394
+ },
293395
+ {
293396
+ "epoch": 1229.51,
293397
+ "learning_rate": 7.5533764135702745e-06,
293398
+ "loss": 0.2782,
293399
+ "step": 152505
293400
+ },
293401
+ {
293402
+ "epoch": 1229.55,
293403
+ "learning_rate": 7.55329563812601e-06,
293404
+ "loss": 0.3844,
293405
+ "step": 152510
293406
+ },
293407
+ {
293408
+ "epoch": 1229.59,
293409
+ "learning_rate": 7.553214862681746e-06,
293410
+ "loss": 1.0582,
293411
+ "step": 152515
293412
+ },
293413
+ {
293414
+ "epoch": 1229.63,
293415
+ "learning_rate": 7.55313408723748e-06,
293416
+ "loss": 0.6986,
293417
+ "step": 152520
293418
+ },
293419
+ {
293420
+ "epoch": 1229.67,
293421
+ "learning_rate": 7.553053311793216e-06,
293422
+ "loss": 0.2519,
293423
+ "step": 152525
293424
+ },
293425
+ {
293426
+ "epoch": 1229.71,
293427
+ "learning_rate": 7.55297253634895e-06,
293428
+ "loss": 0.3056,
293429
+ "step": 152530
293430
+ },
293431
+ {
293432
+ "epoch": 1229.76,
293433
+ "learning_rate": 7.552891760904686e-06,
293434
+ "loss": 0.3928,
293435
+ "step": 152535
293436
+ },
293437
+ {
293438
+ "epoch": 1229.8,
293439
+ "learning_rate": 7.55281098546042e-06,
293440
+ "loss": 0.8044,
293441
+ "step": 152540
293442
+ },
293443
+ {
293444
+ "epoch": 1229.84,
293445
+ "learning_rate": 7.552730210016156e-06,
293446
+ "loss": 0.651,
293447
+ "step": 152545
293448
+ },
293449
+ {
293450
+ "epoch": 1229.88,
293451
+ "learning_rate": 7.55264943457189e-06,
293452
+ "loss": 0.2421,
293453
+ "step": 152550
293454
+ },
293455
+ {
293456
+ "epoch": 1229.92,
293457
+ "learning_rate": 7.552568659127626e-06,
293458
+ "loss": 0.3304,
293459
+ "step": 152555
293460
+ },
293461
+ {
293462
+ "epoch": 1229.96,
293463
+ "learning_rate": 7.55248788368336e-06,
293464
+ "loss": 0.4315,
293465
+ "step": 152560
293466
+ },
293467
+ {
293468
+ "epoch": 1230.0,
293469
+ "learning_rate": 7.552407108239096e-06,
293470
+ "loss": 1.0926,
293471
+ "step": 152565
293472
+ },
293473
+ {
293474
+ "epoch": 1230.0,
293475
+ "eval_loss": 0.5449385643005371,
293476
+ "eval_runtime": 40.681,
293477
+ "eval_samples_per_second": 20.55,
293478
+ "eval_steps_per_second": 0.664,
293479
+ "eval_wer": 0.18418858485560524,
293480
+ "step": 152565
293481
+ },
293482
+ {
293483
+ "epoch": 1220.04,
293484
+ "learning_rate": 7.552326332794832e-06,
293485
+ "loss": 0.2789,
293486
+ "step": 152570
293487
+ },
293488
+ {
293489
+ "epoch": 1220.08,
293490
+ "learning_rate": 7.552245557350566e-06,
293491
+ "loss": 0.271,
293492
+ "step": 152575
293493
+ },
293494
+ {
293495
+ "epoch": 1220.12,
293496
+ "learning_rate": 7.552164781906302e-06,
293497
+ "loss": 0.243,
293498
+ "step": 152580
293499
+ },
293500
+ {
293501
+ "epoch": 1220.16,
293502
+ "learning_rate": 7.552084006462036e-06,
293503
+ "loss": 0.4346,
293504
+ "step": 152585
293505
+ },
293506
+ {
293507
+ "epoch": 1220.2,
293508
+ "learning_rate": 7.552003231017772e-06,
293509
+ "loss": 1.2121,
293510
+ "step": 152590
293511
+ },
293512
+ {
293513
+ "epoch": 1220.24,
293514
+ "learning_rate": 7.551922455573506e-06,
293515
+ "loss": 0.373,
293516
+ "step": 152595
293517
+ },
293518
+ {
293519
+ "epoch": 1220.28,
293520
+ "learning_rate": 7.551841680129242e-06,
293521
+ "loss": 0.2537,
293522
+ "step": 152600
293523
+ },
293524
+ {
293525
+ "epoch": 1220.32,
293526
+ "learning_rate": 7.551760904684976e-06,
293527
+ "loss": 0.2496,
293528
+ "step": 152605
293529
+ },
293530
+ {
293531
+ "epoch": 1220.36,
293532
+ "learning_rate": 7.551680129240712e-06,
293533
+ "loss": 0.5073,
293534
+ "step": 152610
293535
+ },
293536
+ {
293537
+ "epoch": 1220.4,
293538
+ "learning_rate": 7.551599353796446e-06,
293539
+ "loss": 1.2567,
293540
+ "step": 152615
293541
+ },
293542
+ {
293543
+ "epoch": 1220.44,
293544
+ "learning_rate": 7.551518578352182e-06,
293545
+ "loss": 0.3609,
293546
+ "step": 152620
293547
+ },
293548
+ {
293549
+ "epoch": 1220.48,
293550
+ "learning_rate": 7.551437802907916e-06,
293551
+ "loss": 0.2748,
293552
+ "step": 152625
293553
+ },
293554
+ {
293555
+ "epoch": 1220.52,
293556
+ "learning_rate": 7.551357027463652e-06,
293557
+ "loss": 0.3177,
293558
+ "step": 152630
293559
+ },
293560
+ {
293561
+ "epoch": 1220.56,
293562
+ "learning_rate": 7.551276252019388e-06,
293563
+ "loss": 0.437,
293564
+ "step": 152635
293565
+ },
293566
+ {
293567
+ "epoch": 1220.6,
293568
+ "learning_rate": 7.551195476575122e-06,
293569
+ "loss": 1.3232,
293570
+ "step": 152640
293571
+ },
293572
+ {
293573
+ "epoch": 1220.64,
293574
+ "learning_rate": 7.5511147011308576e-06,
293575
+ "loss": 0.2868,
293576
+ "step": 152645
293577
+ },
293578
+ {
293579
+ "epoch": 1220.68,
293580
+ "learning_rate": 7.551033925686592e-06,
293581
+ "loss": 0.3602,
293582
+ "step": 152650
293583
+ },
293584
+ {
293585
+ "epoch": 1220.72,
293586
+ "learning_rate": 7.5509531502423275e-06,
293587
+ "loss": 0.2706,
293588
+ "step": 152655
293589
+ },
293590
+ {
293591
+ "epoch": 1220.76,
293592
+ "learning_rate": 7.550872374798062e-06,
293593
+ "loss": 0.4301,
293594
+ "step": 152660
293595
+ },
293596
+ {
293597
+ "epoch": 1220.8,
293598
+ "learning_rate": 7.5507915993537975e-06,
293599
+ "loss": 1.0362,
293600
+ "step": 152665
293601
+ },
293602
+ {
293603
+ "epoch": 1220.84,
293604
+ "learning_rate": 7.550710823909532e-06,
293605
+ "loss": 0.2926,
293606
+ "step": 152670
293607
+ },
293608
+ {
293609
+ "epoch": 1220.88,
293610
+ "learning_rate": 7.5506300484652675e-06,
293611
+ "loss": 0.2753,
293612
+ "step": 152675
293613
+ },
293614
+ {
293615
+ "epoch": 1220.92,
293616
+ "learning_rate": 7.550549273021002e-06,
293617
+ "loss": 0.3452,
293618
+ "step": 152680
293619
+ },
293620
+ {
293621
+ "epoch": 1220.96,
293622
+ "learning_rate": 7.5504684975767375e-06,
293623
+ "loss": 0.4678,
293624
+ "step": 152685
293625
+ },
293626
+ {
293627
+ "epoch": 1221.0,
293628
+ "learning_rate": 7.5503877221324724e-06,
293629
+ "loss": 1.3258,
293630
+ "step": 152690
293631
+ },
293632
+ {
293633
+ "epoch": 1221.0,
293634
+ "eval_loss": 0.3559703230857849,
293635
+ "eval_runtime": 43.0859,
293636
+ "eval_samples_per_second": 19.403,
293637
+ "eval_steps_per_second": 0.627,
293638
+ "eval_wer": 0.17048071387449626,
293639
+ "step": 152690
293640
  }
293641
  ],
293642
+ "max_steps": 625000,
293643
  "num_train_epochs": 5000,
293644
+ "total_flos": 4.2968328943638807e+20,
293645
  "trial_name": null,
293646
  "trial_params": null
293647
  }
model-bin/finetune/base/{checkpoint-152068 β†’ checkpoint-152690}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630250753.668995/events.out.tfevents.1630250753.cc93b136ebf5.1086.209 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c0aee1ce7801ac75359895fcb1668e7c7233a293e1e85b17034353ff0003ee
3
+ size 4194
model-bin/finetune/base/log/1630251188.88167/events.out.tfevents.1630251188.cc93b136ebf5.1086.211 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc10d5878cb756554085f94d5e325de11d0130822fcaaad5d044e2996b03f139
3
+ size 4194
model-bin/finetune/base/log/1630251628.1797204/events.out.tfevents.1630251628.cc93b136ebf5.1086.213 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a75e79f95a32dbd01c2a774a76e9bb2824f0d5b1551051b1f8085c6c884e1d3
3
+ size 4194
model-bin/finetune/base/log/1630252065.7514985/events.out.tfevents.1630252065.cc93b136ebf5.1086.215 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e323b0936deda44bf91875918182857d58ba5428f8afdebac01580255cb0e269
3
+ size 4194
model-bin/finetune/base/log/1630252500.5741549/events.out.tfevents.1630252500.cc93b136ebf5.1086.217 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb9ebf7e8137afa957d4d2026869e20b8bd962c5bce171c25618696ce1316d6e
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630250753.cc93b136ebf5.1086.208 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c6b329615b997a29dc2441fdc952ec848b8f5de5513d43042f86ea38b90097d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630251188.cc93b136ebf5.1086.210 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28f4b9a76a7866c02a50de0f72d113f2770ef0fb3e160da1f4087186466043da
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630251628.cc93b136ebf5.1086.212 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c24d8523270a58fc6c0a96845b48c37ed90191d876506302e6a8a7e1c204bf0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630252065.cc93b136ebf5.1086.214 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ad901e5a01c00724eaa3e2c6d1064f0725f8468b8af7d9f4449a542c3c3cd99
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630252500.cc93b136ebf5.1086.216 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1c9e6fc6b8beb6d7bda1c768e2664d3a7732f3851f11540d7cf4be9cc49d64f
3
+ size 8622