Check commited on
Commit
c68fa3e
Β·
1 Parent(s): e1f3883

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630228288.1449773/events.out.tfevents.1630228288.cc93b136ebf5.1086.109 +3 -0
  11. model-bin/finetune/base/log/1630228721.9317095/events.out.tfevents.1630228721.cc93b136ebf5.1086.111 +3 -0
  12. model-bin/finetune/base/log/1630229152.9104557/events.out.tfevents.1630229152.cc93b136ebf5.1086.113 +3 -0
  13. model-bin/finetune/base/log/1630229586.5053954/events.out.tfevents.1630229586.cc93b136ebf5.1086.115 +3 -0
  14. model-bin/finetune/base/log/1630230014.08416/events.out.tfevents.1630230014.cc93b136ebf5.1086.117 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630228288.cc93b136ebf5.1086.108 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630228721.cc93b136ebf5.1086.110 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630229152.cc93b136ebf5.1086.112 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630229586.cc93b136ebf5.1086.114 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630230014.cc93b136ebf5.1086.116 +3 -0
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c3e8284682b82934af7c2cb7a47d0aad1afbbab74b9a20b0bdf99e1f1b28b90
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f7568a2befca944be90026427116813896320af24320ff0d6d6cce3345d3e75
3
  size 722165393
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7400c5774686da0c05e6115d2e42b7e425eb92bed75c7039ed458db7dc841243
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a22ed2c49e0003da5e3bff5a28ee7e00afcf8ecb8920666d8355ae042273f4b
3
  size 377909911
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:791f31e71e6884d8a8929ef220193ccb17e2cfec3b0e299f7e8687cba4af3ce4
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49db60bb7be014c03e20f099f8b5b41640dd5b10966a8d382a375eaacf0dba1f
3
  size 14503
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28a0b76f68ffd1c2d412704c5679e6f1d5de953fb95b5cacc4cb5d077d0d04b0
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e32e4b8d64538adb2e3a82fff0984d5e19137de747f31f2ccfe398a635069ee
3
  size 559
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7dff24b5bd65b8a22b77667e8a8418682532690368c32e2fbb3b1785624a9976
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75a35ac11ff6fdc1ff80807c5620a77c06912969c11e90901d0e650569700174
3
  size 623
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.16716790141568252,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
4
- "epoch": 1166.0,
5
- "global_step": 145852,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -284934,11 +284934,800 @@
284934
  "eval_steps_per_second": 0.655,
284935
  "eval_wer": 0.17012599228024178,
284936
  "step": 145852
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
284937
  }
284938
  ],
284939
- "max_steps": 625000,
284940
  "num_train_epochs": 5000,
284941
- "total_flos": 4.104335476145737e+20,
284942
  "trial_name": null,
284943
  "trial_params": null
284944
  }
 
1
  {
2
  "best_metric": 0.16716790141568252,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
4
+ "epoch": 1180.995983935743,
5
+ "global_step": 146472,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
284934
  "eval_steps_per_second": 0.655,
284935
  "eval_wer": 0.17012599228024178,
284936
  "step": 145852
284937
+ },
284938
+ {
284939
+ "epoch": 1176.02,
284940
+ "learning_rate": 7.660775444264943e-06,
284941
+ "loss": 0.4127,
284942
+ "step": 145855
284943
+ },
284944
+ {
284945
+ "epoch": 1176.06,
284946
+ "learning_rate": 7.660694668820679e-06,
284947
+ "loss": 0.2818,
284948
+ "step": 145860
284949
+ },
284950
+ {
284951
+ "epoch": 1176.1,
284952
+ "learning_rate": 7.660613893376413e-06,
284953
+ "loss": 0.2966,
284954
+ "step": 145865
284955
+ },
284956
+ {
284957
+ "epoch": 1176.15,
284958
+ "learning_rate": 7.660533117932149e-06,
284959
+ "loss": 0.3262,
284960
+ "step": 145870
284961
+ },
284962
+ {
284963
+ "epoch": 1176.19,
284964
+ "learning_rate": 7.660452342487883e-06,
284965
+ "loss": 0.744,
284966
+ "step": 145875
284967
+ },
284968
+ {
284969
+ "epoch": 1176.23,
284970
+ "learning_rate": 7.660371567043619e-06,
284971
+ "loss": 0.8761,
284972
+ "step": 145880
284973
+ },
284974
+ {
284975
+ "epoch": 1176.27,
284976
+ "learning_rate": 7.660290791599355e-06,
284977
+ "loss": 0.2513,
284978
+ "step": 145885
284979
+ },
284980
+ {
284981
+ "epoch": 1176.31,
284982
+ "learning_rate": 7.660210016155089e-06,
284983
+ "loss": 0.2938,
284984
+ "step": 145890
284985
+ },
284986
+ {
284987
+ "epoch": 1176.35,
284988
+ "learning_rate": 7.660129240710825e-06,
284989
+ "loss": 0.3237,
284990
+ "step": 145895
284991
+ },
284992
+ {
284993
+ "epoch": 1176.39,
284994
+ "learning_rate": 7.660048465266559e-06,
284995
+ "loss": 0.7375,
284996
+ "step": 145900
284997
+ },
284998
+ {
284999
+ "epoch": 1176.43,
285000
+ "learning_rate": 7.659967689822295e-06,
285001
+ "loss": 0.8132,
285002
+ "step": 145905
285003
+ },
285004
+ {
285005
+ "epoch": 1176.47,
285006
+ "learning_rate": 7.659886914378029e-06,
285007
+ "loss": 0.3004,
285008
+ "step": 145910
285009
+ },
285010
+ {
285011
+ "epoch": 1176.51,
285012
+ "learning_rate": 7.659806138933765e-06,
285013
+ "loss": 0.2726,
285014
+ "step": 145915
285015
+ },
285016
+ {
285017
+ "epoch": 1176.55,
285018
+ "learning_rate": 7.659725363489499e-06,
285019
+ "loss": 0.3235,
285020
+ "step": 145920
285021
+ },
285022
+ {
285023
+ "epoch": 1176.59,
285024
+ "learning_rate": 7.659644588045235e-06,
285025
+ "loss": 0.7467,
285026
+ "step": 145925
285027
+ },
285028
+ {
285029
+ "epoch": 1176.63,
285030
+ "learning_rate": 7.659563812600969e-06,
285031
+ "loss": 0.7942,
285032
+ "step": 145930
285033
+ },
285034
+ {
285035
+ "epoch": 1176.67,
285036
+ "learning_rate": 7.659483037156705e-06,
285037
+ "loss": 0.2774,
285038
+ "step": 145935
285039
+ },
285040
+ {
285041
+ "epoch": 1176.71,
285042
+ "learning_rate": 7.659402261712439e-06,
285043
+ "loss": 0.2752,
285044
+ "step": 145940
285045
+ },
285046
+ {
285047
+ "epoch": 1176.75,
285048
+ "learning_rate": 7.659321486268175e-06,
285049
+ "loss": 0.3578,
285050
+ "step": 145945
285051
+ },
285052
+ {
285053
+ "epoch": 1176.79,
285054
+ "learning_rate": 7.659240710823909e-06,
285055
+ "loss": 0.6986,
285056
+ "step": 145950
285057
+ },
285058
+ {
285059
+ "epoch": 1176.83,
285060
+ "learning_rate": 7.659159935379645e-06,
285061
+ "loss": 0.7739,
285062
+ "step": 145955
285063
+ },
285064
+ {
285065
+ "epoch": 1176.87,
285066
+ "learning_rate": 7.65907915993538e-06,
285067
+ "loss": 0.2683,
285068
+ "step": 145960
285069
+ },
285070
+ {
285071
+ "epoch": 1176.91,
285072
+ "learning_rate": 7.658998384491115e-06,
285073
+ "loss": 0.3132,
285074
+ "step": 145965
285075
+ },
285076
+ {
285077
+ "epoch": 1176.95,
285078
+ "learning_rate": 7.65891760904685e-06,
285079
+ "loss": 0.373,
285080
+ "step": 145970
285081
+ },
285082
+ {
285083
+ "epoch": 1176.99,
285084
+ "learning_rate": 7.658836833602585e-06,
285085
+ "loss": 1.049,
285086
+ "step": 145975
285087
+ },
285088
+ {
285089
+ "epoch": 1177.0,
285090
+ "eval_loss": 0.3793235719203949,
285091
+ "eval_runtime": 40.3575,
285092
+ "eval_samples_per_second": 20.74,
285093
+ "eval_steps_per_second": 0.669,
285094
+ "eval_wer": 0.17174677608440797,
285095
+ "step": 145976
285096
+ },
285097
+ {
285098
+ "epoch": 1177.03,
285099
+ "learning_rate": 7.65875605815832e-06,
285100
+ "loss": 0.3211,
285101
+ "step": 145980
285102
+ },
285103
+ {
285104
+ "epoch": 1177.07,
285105
+ "learning_rate": 7.658675282714055e-06,
285106
+ "loss": 0.2587,
285107
+ "step": 145985
285108
+ },
285109
+ {
285110
+ "epoch": 1177.11,
285111
+ "learning_rate": 7.65859450726979e-06,
285112
+ "loss": 0.3043,
285113
+ "step": 145990
285114
+ },
285115
+ {
285116
+ "epoch": 1177.15,
285117
+ "learning_rate": 7.658513731825525e-06,
285118
+ "loss": 0.374,
285119
+ "step": 145995
285120
+ },
285121
+ {
285122
+ "epoch": 1177.19,
285123
+ "learning_rate": 7.65843295638126e-06,
285124
+ "loss": 0.9153,
285125
+ "step": 146000
285126
+ },
285127
+ {
285128
+ "epoch": 1177.23,
285129
+ "learning_rate": 7.658352180936995e-06,
285130
+ "loss": 0.7194,
285131
+ "step": 146005
285132
+ },
285133
+ {
285134
+ "epoch": 1177.27,
285135
+ "learning_rate": 7.65827140549273e-06,
285136
+ "loss": 0.2684,
285137
+ "step": 146010
285138
+ },
285139
+ {
285140
+ "epoch": 1177.31,
285141
+ "learning_rate": 7.658190630048465e-06,
285142
+ "loss": 0.2866,
285143
+ "step": 146015
285144
+ },
285145
+ {
285146
+ "epoch": 1177.35,
285147
+ "learning_rate": 7.6581098546042e-06,
285148
+ "loss": 0.3972,
285149
+ "step": 146020
285150
+ },
285151
+ {
285152
+ "epoch": 1177.39,
285153
+ "learning_rate": 7.658029079159936e-06,
285154
+ "loss": 1.0221,
285155
+ "step": 146025
285156
+ },
285157
+ {
285158
+ "epoch": 1177.43,
285159
+ "learning_rate": 7.65794830371567e-06,
285160
+ "loss": 0.5779,
285161
+ "step": 146030
285162
+ },
285163
+ {
285164
+ "epoch": 1177.47,
285165
+ "learning_rate": 7.657867528271406e-06,
285166
+ "loss": 0.3161,
285167
+ "step": 146035
285168
+ },
285169
+ {
285170
+ "epoch": 1177.51,
285171
+ "learning_rate": 7.65778675282714e-06,
285172
+ "loss": 0.2923,
285173
+ "step": 146040
285174
+ },
285175
+ {
285176
+ "epoch": 1177.55,
285177
+ "learning_rate": 7.657705977382876e-06,
285178
+ "loss": 0.4056,
285179
+ "step": 146045
285180
+ },
285181
+ {
285182
+ "epoch": 1177.59,
285183
+ "learning_rate": 7.65762520193861e-06,
285184
+ "loss": 0.8064,
285185
+ "step": 146050
285186
+ },
285187
+ {
285188
+ "epoch": 1177.63,
285189
+ "learning_rate": 7.657544426494346e-06,
285190
+ "loss": 0.6026,
285191
+ "step": 146055
285192
+ },
285193
+ {
285194
+ "epoch": 1177.67,
285195
+ "learning_rate": 7.65746365105008e-06,
285196
+ "loss": 0.2731,
285197
+ "step": 146060
285198
+ },
285199
+ {
285200
+ "epoch": 1177.71,
285201
+ "learning_rate": 7.657382875605816e-06,
285202
+ "loss": 0.2761,
285203
+ "step": 146065
285204
+ },
285205
+ {
285206
+ "epoch": 1177.76,
285207
+ "learning_rate": 7.65730210016155e-06,
285208
+ "loss": 0.3563,
285209
+ "step": 146070
285210
+ },
285211
+ {
285212
+ "epoch": 1177.8,
285213
+ "learning_rate": 7.657221324717286e-06,
285214
+ "loss": 0.8102,
285215
+ "step": 146075
285216
+ },
285217
+ {
285218
+ "epoch": 1177.84,
285219
+ "learning_rate": 7.65714054927302e-06,
285220
+ "loss": 0.5813,
285221
+ "step": 146080
285222
+ },
285223
+ {
285224
+ "epoch": 1177.88,
285225
+ "learning_rate": 7.657059773828756e-06,
285226
+ "loss": 0.3342,
285227
+ "step": 146085
285228
+ },
285229
+ {
285230
+ "epoch": 1177.92,
285231
+ "learning_rate": 7.656978998384492e-06,
285232
+ "loss": 0.3355,
285233
+ "step": 146090
285234
+ },
285235
+ {
285236
+ "epoch": 1177.96,
285237
+ "learning_rate": 7.656898222940226e-06,
285238
+ "loss": 0.4923,
285239
+ "step": 146095
285240
+ },
285241
+ {
285242
+ "epoch": 1178.0,
285243
+ "learning_rate": 7.656817447495962e-06,
285244
+ "loss": 1.2601,
285245
+ "step": 146100
285246
+ },
285247
+ {
285248
+ "epoch": 1178.0,
285249
+ "eval_loss": 0.4012053608894348,
285250
+ "eval_runtime": 40.577,
285251
+ "eval_samples_per_second": 20.627,
285252
+ "eval_steps_per_second": 0.665,
285253
+ "eval_wer": 0.1786023678891135,
285254
+ "step": 146100
285255
+ },
285256
+ {
285257
+ "epoch": 1178.04,
285258
+ "learning_rate": 7.656736672051696e-06,
285259
+ "loss": 0.2932,
285260
+ "step": 146105
285261
+ },
285262
+ {
285263
+ "epoch": 1178.08,
285264
+ "learning_rate": 7.656655896607432e-06,
285265
+ "loss": 0.2603,
285266
+ "step": 146110
285267
+ },
285268
+ {
285269
+ "epoch": 1178.12,
285270
+ "learning_rate": 7.656575121163166e-06,
285271
+ "loss": 0.2993,
285272
+ "step": 146115
285273
+ },
285274
+ {
285275
+ "epoch": 1178.16,
285276
+ "learning_rate": 7.656494345718902e-06,
285277
+ "loss": 0.4776,
285278
+ "step": 146120
285279
+ },
285280
+ {
285281
+ "epoch": 1178.2,
285282
+ "learning_rate": 7.656413570274636e-06,
285283
+ "loss": 1.1354,
285284
+ "step": 146125
285285
+ },
285286
+ {
285287
+ "epoch": 1178.24,
285288
+ "learning_rate": 7.656332794830372e-06,
285289
+ "loss": 0.2897,
285290
+ "step": 146130
285291
+ },
285292
+ {
285293
+ "epoch": 1178.28,
285294
+ "learning_rate": 7.656252019386106e-06,
285295
+ "loss": 0.2256,
285296
+ "step": 146135
285297
+ },
285298
+ {
285299
+ "epoch": 1178.32,
285300
+ "learning_rate": 7.656171243941842e-06,
285301
+ "loss": 0.2845,
285302
+ "step": 146140
285303
+ },
285304
+ {
285305
+ "epoch": 1178.36,
285306
+ "learning_rate": 7.656090468497578e-06,
285307
+ "loss": 0.4063,
285308
+ "step": 146145
285309
+ },
285310
+ {
285311
+ "epoch": 1178.4,
285312
+ "learning_rate": 7.656009693053312e-06,
285313
+ "loss": 1.1799,
285314
+ "step": 146150
285315
+ },
285316
+ {
285317
+ "epoch": 1178.44,
285318
+ "learning_rate": 7.655928917609048e-06,
285319
+ "loss": 0.3141,
285320
+ "step": 146155
285321
+ },
285322
+ {
285323
+ "epoch": 1178.48,
285324
+ "learning_rate": 7.655848142164782e-06,
285325
+ "loss": 0.2586,
285326
+ "step": 146160
285327
+ },
285328
+ {
285329
+ "epoch": 1178.52,
285330
+ "learning_rate": 7.655767366720518e-06,
285331
+ "loss": 0.3502,
285332
+ "step": 146165
285333
+ },
285334
+ {
285335
+ "epoch": 1178.56,
285336
+ "learning_rate": 7.655686591276252e-06,
285337
+ "loss": 0.4228,
285338
+ "step": 146170
285339
+ },
285340
+ {
285341
+ "epoch": 1178.6,
285342
+ "learning_rate": 7.655605815831988e-06,
285343
+ "loss": 1.2601,
285344
+ "step": 146175
285345
+ },
285346
+ {
285347
+ "epoch": 1178.64,
285348
+ "learning_rate": 7.655525040387722e-06,
285349
+ "loss": 0.2679,
285350
+ "step": 146180
285351
+ },
285352
+ {
285353
+ "epoch": 1178.68,
285354
+ "learning_rate": 7.655444264943458e-06,
285355
+ "loss": 0.2357,
285356
+ "step": 146185
285357
+ },
285358
+ {
285359
+ "epoch": 1178.72,
285360
+ "learning_rate": 7.655363489499192e-06,
285361
+ "loss": 0.3045,
285362
+ "step": 146190
285363
+ },
285364
+ {
285365
+ "epoch": 1178.76,
285366
+ "learning_rate": 7.655282714054928e-06,
285367
+ "loss": 0.5178,
285368
+ "step": 146195
285369
+ },
285370
+ {
285371
+ "epoch": 1178.8,
285372
+ "learning_rate": 7.655201938610663e-06,
285373
+ "loss": 1.2614,
285374
+ "step": 146200
285375
+ },
285376
+ {
285377
+ "epoch": 1178.84,
285378
+ "learning_rate": 7.655121163166397e-06,
285379
+ "loss": 0.3759,
285380
+ "step": 146205
285381
+ },
285382
+ {
285383
+ "epoch": 1178.88,
285384
+ "learning_rate": 7.655040387722133e-06,
285385
+ "loss": 0.2394,
285386
+ "step": 146210
285387
+ },
285388
+ {
285389
+ "epoch": 1178.92,
285390
+ "learning_rate": 7.654959612277867e-06,
285391
+ "loss": 0.275,
285392
+ "step": 146215
285393
+ },
285394
+ {
285395
+ "epoch": 1178.96,
285396
+ "learning_rate": 7.654878836833603e-06,
285397
+ "loss": 0.5791,
285398
+ "step": 146220
285399
+ },
285400
+ {
285401
+ "epoch": 1179.0,
285402
+ "eval_loss": 0.358010470867157,
285403
+ "eval_runtime": 40.5201,
285404
+ "eval_samples_per_second": 20.656,
285405
+ "eval_steps_per_second": 0.666,
285406
+ "eval_wer": 0.176530911221799,
285407
+ "step": 146224
285408
+ },
285409
+ {
285410
+ "epoch": 1179.01,
285411
+ "learning_rate": 7.654798061389337e-06,
285412
+ "loss": 0.5013,
285413
+ "step": 146225
285414
+ },
285415
+ {
285416
+ "epoch": 1179.05,
285417
+ "learning_rate": 7.654717285945073e-06,
285418
+ "loss": 0.2654,
285419
+ "step": 146230
285420
+ },
285421
+ {
285422
+ "epoch": 1179.09,
285423
+ "learning_rate": 7.654636510500807e-06,
285424
+ "loss": 0.2647,
285425
+ "step": 146235
285426
+ },
285427
+ {
285428
+ "epoch": 1179.13,
285429
+ "learning_rate": 7.654555735056543e-06,
285430
+ "loss": 0.3248,
285431
+ "step": 146240
285432
+ },
285433
+ {
285434
+ "epoch": 1179.17,
285435
+ "learning_rate": 7.654474959612277e-06,
285436
+ "loss": 0.4694,
285437
+ "step": 146245
285438
+ },
285439
+ {
285440
+ "epoch": 1179.21,
285441
+ "learning_rate": 7.654394184168013e-06,
285442
+ "loss": 1.0148,
285443
+ "step": 146250
285444
+ },
285445
+ {
285446
+ "epoch": 1179.25,
285447
+ "learning_rate": 7.654313408723747e-06,
285448
+ "loss": 0.2833,
285449
+ "step": 146255
285450
+ },
285451
+ {
285452
+ "epoch": 1179.29,
285453
+ "learning_rate": 7.654232633279483e-06,
285454
+ "loss": 0.2983,
285455
+ "step": 146260
285456
+ },
285457
+ {
285458
+ "epoch": 1179.33,
285459
+ "learning_rate": 7.654151857835219e-06,
285460
+ "loss": 0.3522,
285461
+ "step": 146265
285462
+ },
285463
+ {
285464
+ "epoch": 1179.37,
285465
+ "learning_rate": 7.654071082390953e-06,
285466
+ "loss": 0.5271,
285467
+ "step": 146270
285468
+ },
285469
+ {
285470
+ "epoch": 1179.41,
285471
+ "learning_rate": 7.653990306946689e-06,
285472
+ "loss": 1.1747,
285473
+ "step": 146275
285474
+ },
285475
+ {
285476
+ "epoch": 1179.45,
285477
+ "learning_rate": 7.653909531502423e-06,
285478
+ "loss": 0.2796,
285479
+ "step": 146280
285480
+ },
285481
+ {
285482
+ "epoch": 1179.49,
285483
+ "learning_rate": 7.653828756058159e-06,
285484
+ "loss": 0.2518,
285485
+ "step": 146285
285486
+ },
285487
+ {
285488
+ "epoch": 1179.53,
285489
+ "learning_rate": 7.653747980613893e-06,
285490
+ "loss": 0.307,
285491
+ "step": 146290
285492
+ },
285493
+ {
285494
+ "epoch": 1179.57,
285495
+ "learning_rate": 7.653667205169629e-06,
285496
+ "loss": 0.4902,
285497
+ "step": 146295
285498
+ },
285499
+ {
285500
+ "epoch": 1179.61,
285501
+ "learning_rate": 7.653586429725363e-06,
285502
+ "loss": 1.0019,
285503
+ "step": 146300
285504
+ },
285505
+ {
285506
+ "epoch": 1179.65,
285507
+ "learning_rate": 7.653505654281099e-06,
285508
+ "loss": 0.278,
285509
+ "step": 146305
285510
+ },
285511
+ {
285512
+ "epoch": 1179.69,
285513
+ "learning_rate": 7.653424878836833e-06,
285514
+ "loss": 0.3384,
285515
+ "step": 146310
285516
+ },
285517
+ {
285518
+ "epoch": 1179.73,
285519
+ "learning_rate": 7.653344103392569e-06,
285520
+ "loss": 0.3353,
285521
+ "step": 146315
285522
+ },
285523
+ {
285524
+ "epoch": 1179.77,
285525
+ "learning_rate": 7.653263327948305e-06,
285526
+ "loss": 0.5559,
285527
+ "step": 146320
285528
+ },
285529
+ {
285530
+ "epoch": 1179.81,
285531
+ "learning_rate": 7.653182552504039e-06,
285532
+ "loss": 1.4385,
285533
+ "step": 146325
285534
+ },
285535
+ {
285536
+ "epoch": 1179.85,
285537
+ "learning_rate": 7.653101777059775e-06,
285538
+ "loss": 0.2435,
285539
+ "step": 146330
285540
+ },
285541
+ {
285542
+ "epoch": 1179.9,
285543
+ "learning_rate": 7.653021001615509e-06,
285544
+ "loss": 0.2579,
285545
+ "step": 146335
285546
+ },
285547
+ {
285548
+ "epoch": 1179.94,
285549
+ "learning_rate": 7.652940226171245e-06,
285550
+ "loss": 0.345,
285551
+ "step": 146340
285552
+ },
285553
+ {
285554
+ "epoch": 1179.98,
285555
+ "learning_rate": 7.652859450726979e-06,
285556
+ "loss": 0.5621,
285557
+ "step": 146345
285558
+ },
285559
+ {
285560
+ "epoch": 1180.0,
285561
+ "eval_loss": 0.449453204870224,
285562
+ "eval_runtime": 40.0278,
285563
+ "eval_samples_per_second": 20.91,
285564
+ "eval_steps_per_second": 0.675,
285565
+ "eval_wer": 0.18728573557907288,
285566
+ "step": 146348
285567
+ },
285568
+ {
285569
+ "epoch": 1180.02,
285570
+ "learning_rate": 7.652778675282715e-06,
285571
+ "loss": 0.2882,
285572
+ "step": 146350
285573
+ },
285574
+ {
285575
+ "epoch": 1180.06,
285576
+ "learning_rate": 7.652697899838449e-06,
285577
+ "loss": 0.2908,
285578
+ "step": 146355
285579
+ },
285580
+ {
285581
+ "epoch": 1180.1,
285582
+ "learning_rate": 7.652617124394185e-06,
285583
+ "loss": 0.2317,
285584
+ "step": 146360
285585
+ },
285586
+ {
285587
+ "epoch": 1180.14,
285588
+ "learning_rate": 7.652536348949919e-06,
285589
+ "loss": 0.3416,
285590
+ "step": 146365
285591
+ },
285592
+ {
285593
+ "epoch": 1180.18,
285594
+ "learning_rate": 7.652455573505655e-06,
285595
+ "loss": 0.6812,
285596
+ "step": 146370
285597
+ },
285598
+ {
285599
+ "epoch": 1180.22,
285600
+ "learning_rate": 7.65237479806139e-06,
285601
+ "loss": 0.8339,
285602
+ "step": 146375
285603
+ },
285604
+ {
285605
+ "epoch": 1180.26,
285606
+ "learning_rate": 7.652294022617125e-06,
285607
+ "loss": 0.431,
285608
+ "step": 146380
285609
+ },
285610
+ {
285611
+ "epoch": 1180.3,
285612
+ "learning_rate": 7.65221324717286e-06,
285613
+ "loss": 0.3809,
285614
+ "step": 146385
285615
+ },
285616
+ {
285617
+ "epoch": 1180.34,
285618
+ "learning_rate": 7.652132471728595e-06,
285619
+ "loss": 0.3004,
285620
+ "step": 146390
285621
+ },
285622
+ {
285623
+ "epoch": 1180.38,
285624
+ "learning_rate": 7.65205169628433e-06,
285625
+ "loss": 0.578,
285626
+ "step": 146395
285627
+ },
285628
+ {
285629
+ "epoch": 1180.42,
285630
+ "learning_rate": 7.651970920840065e-06,
285631
+ "loss": 0.9797,
285632
+ "step": 146400
285633
+ },
285634
+ {
285635
+ "epoch": 1180.46,
285636
+ "learning_rate": 7.6518901453958e-06,
285637
+ "loss": 0.2545,
285638
+ "step": 146405
285639
+ },
285640
+ {
285641
+ "epoch": 1180.5,
285642
+ "learning_rate": 7.651809369951535e-06,
285643
+ "loss": 0.2621,
285644
+ "step": 146410
285645
+ },
285646
+ {
285647
+ "epoch": 1180.54,
285648
+ "learning_rate": 7.65172859450727e-06,
285649
+ "loss": 0.3193,
285650
+ "step": 146415
285651
+ },
285652
+ {
285653
+ "epoch": 1180.58,
285654
+ "learning_rate": 7.651647819063005e-06,
285655
+ "loss": 0.6539,
285656
+ "step": 146420
285657
+ },
285658
+ {
285659
+ "epoch": 1180.62,
285660
+ "learning_rate": 7.65156704361874e-06,
285661
+ "loss": 1.0436,
285662
+ "step": 146425
285663
+ },
285664
+ {
285665
+ "epoch": 1180.66,
285666
+ "learning_rate": 7.651486268174474e-06,
285667
+ "loss": 0.2563,
285668
+ "step": 146430
285669
+ },
285670
+ {
285671
+ "epoch": 1180.7,
285672
+ "learning_rate": 7.65140549273021e-06,
285673
+ "loss": 0.3142,
285674
+ "step": 146435
285675
+ },
285676
+ {
285677
+ "epoch": 1180.74,
285678
+ "learning_rate": 7.651324717285946e-06,
285679
+ "loss": 0.3735,
285680
+ "step": 146440
285681
+ },
285682
+ {
285683
+ "epoch": 1180.78,
285684
+ "learning_rate": 7.65124394184168e-06,
285685
+ "loss": 0.6211,
285686
+ "step": 146445
285687
+ },
285688
+ {
285689
+ "epoch": 1180.82,
285690
+ "learning_rate": 7.651163166397416e-06,
285691
+ "loss": 1.0659,
285692
+ "step": 146450
285693
+ },
285694
+ {
285695
+ "epoch": 1180.86,
285696
+ "learning_rate": 7.65108239095315e-06,
285697
+ "loss": 0.3007,
285698
+ "step": 146455
285699
+ },
285700
+ {
285701
+ "epoch": 1180.9,
285702
+ "learning_rate": 7.651001615508886e-06,
285703
+ "loss": 0.2761,
285704
+ "step": 146460
285705
+ },
285706
+ {
285707
+ "epoch": 1180.94,
285708
+ "learning_rate": 7.65092084006462e-06,
285709
+ "loss": 0.3487,
285710
+ "step": 146465
285711
+ },
285712
+ {
285713
+ "epoch": 1180.98,
285714
+ "learning_rate": 7.650840064620356e-06,
285715
+ "loss": 0.7013,
285716
+ "step": 146470
285717
+ },
285718
+ {
285719
+ "epoch": 1181.0,
285720
+ "eval_loss": 0.34985774755477905,
285721
+ "eval_runtime": 41.8541,
285722
+ "eval_samples_per_second": 19.998,
285723
+ "eval_steps_per_second": 0.645,
285724
+ "eval_wer": 0.1758169934640523,
285725
+ "step": 146472
285726
  }
285727
  ],
285728
+ "max_steps": 620000,
285729
  "num_train_epochs": 5000,
285730
+ "total_flos": 4.1218134700506854e+20,
285731
  "trial_name": null,
285732
  "trial_params": null
285733
  }
model-bin/finetune/base/{checkpoint-145852 β†’ checkpoint-146472}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630228288.1449773/events.out.tfevents.1630228288.cc93b136ebf5.1086.109 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfe38cec38c3d2adbc09d27a46a1e57a1f70fa126623f4f52627be155109468e
3
+ size 4194
model-bin/finetune/base/log/1630228721.9317095/events.out.tfevents.1630228721.cc93b136ebf5.1086.111 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28112e380a1179c13caf1bfbf7221311ccc001134b60d565f8f56c89d707889c
3
+ size 4194
model-bin/finetune/base/log/1630229152.9104557/events.out.tfevents.1630229152.cc93b136ebf5.1086.113 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6748d316e70868842a9f08d249dc07f834ef93882afb2c7c7fd28f72375b2422
3
+ size 4194
model-bin/finetune/base/log/1630229586.5053954/events.out.tfevents.1630229586.cc93b136ebf5.1086.115 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feb8a302b12fa256e6d48cbf4c25ff505338ca01049c4f58a77b6d4f4b35098b
3
+ size 4194
model-bin/finetune/base/log/1630230014.08416/events.out.tfevents.1630230014.cc93b136ebf5.1086.117 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e6b0961bc142db049474b69ea745d4c3a673f78a77e8b18e020138d7fee8b38
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630228288.cc93b136ebf5.1086.108 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:630bfa7f41e9ac1296d8c100a479675d8878d9a89021acbe1d1e3fd54480eaec
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630228721.cc93b136ebf5.1086.110 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc375cfeb2958a56e334f99323d9a111298cd5b84f7f763f7c9f2e5c1b031302
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630229152.cc93b136ebf5.1086.112 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:716525fcb37cbd537930bceaeec45c754a919f6732b733a07571f7ff0a0ea616
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630229586.cc93b136ebf5.1086.114 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:933a727cbd0e343f0d73aec1a0317b5741e8e77cd32fdda735bd47490f98c084
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630230014.cc93b136ebf5.1086.116 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:931819ecb40a82e206200184d2b3d92970a599c9a49de08dcace11c58b3ff38f
3
+ size 8622