Check commited on
Commit
c79f680
Β·
1 Parent(s): 520d67b

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629882392.1236827/events.out.tfevents.1629882392.7e498afd5545.905.93 +3 -0
  11. model-bin/finetune/base/log/1629883056.4519293/events.out.tfevents.1629883056.7e498afd5545.905.95 +3 -0
  12. model-bin/finetune/base/log/1629883697.070274/events.out.tfevents.1629883697.7e498afd5545.905.97 +3 -0
  13. model-bin/finetune/base/log/1629884340.6409886/events.out.tfevents.1629884340.7e498afd5545.905.99 +3 -0
  14. model-bin/finetune/base/log/1629885006.4519074/events.out.tfevents.1629885006.7e498afd5545.905.101 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629882392.7e498afd5545.905.92 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629883056.7e498afd5545.905.94 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629883697.7e498afd5545.905.96 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629884340.7e498afd5545.905.98 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629885006.7e498afd5545.905.100 +3 -0
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54b35bdaaf4933af7fe3f9bbd9b1424d570c568987740f15bcd85cc870d83f51
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:241cf9a0d25f6dfe6c4b09f0250d75c7b8556e0708ad5a758a4ae728e99ed97f
3
  size 722165393
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f70ffbe9923a30dbf8b87abb1106958feef550de2620a2d2b80cc91fc8a098e2
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07eddd7e3b2c639376f65debd5c6e71a40abb39191deddef03737b5becdd9cb4
3
  size 377909911
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4be019330451dcdea1470053a969250ec2ac7591e9e422df7bf9734de62a6607
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff20bb0fbeb061a0d474c60bb2f68154a4d0944df8867df4e634564652ff6a60
3
  size 14503
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3d8eee4042019854a091a06285230ead9805a530d0a1e8630cf2695e05d1d3b
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6aaef4b9249c4bb80c33d2e747e6a2aee3a93d57119debf58fe887e8e98126a
3
  size 559
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8aae4596cc6445369fecda3c5539d240321af736664343a1c7e72481ab8edb80
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d36509b0709cb1c6a514d0b59c68cd5a3a2275d7fb28dc5813d880c7d4a4f0e
3
  size 623
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
- "epoch": 601.0,
5
- "global_step": 75162,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -194994,11 +194994,806 @@
194994
  "eval_steps_per_second": 0.665,
194995
  "eval_wer": 0.2013764427557531,
194996
  "step": 75162
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194997
  }
194998
  ],
194999
  "max_steps": 625000,
195000
  "num_train_epochs": 5000,
195001
- "total_flos": 2.115183129640869e+20,
195002
  "trial_name": null,
195003
  "trial_params": null
195004
  }
 
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
+ "epoch": 605.9960159362549,
5
+ "global_step": 75787,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
194994
  "eval_steps_per_second": 0.665,
194995
  "eval_wer": 0.2013764427557531,
194996
  "step": 75162
194997
+ },
194998
+ {
194999
+ "epoch": 601.02,
195000
+ "learning_rate": 8.811923076923078e-06,
195001
+ "loss": 0.3253,
195002
+ "step": 75165
195003
+ },
195004
+ {
195005
+ "epoch": 601.06,
195006
+ "learning_rate": 8.811842948717949e-06,
195007
+ "loss": 0.3342,
195008
+ "step": 75170
195009
+ },
195010
+ {
195011
+ "epoch": 601.1,
195012
+ "learning_rate": 8.811762820512822e-06,
195013
+ "loss": 0.383,
195014
+ "step": 75175
195015
+ },
195016
+ {
195017
+ "epoch": 601.14,
195018
+ "learning_rate": 8.811682692307693e-06,
195019
+ "loss": 0.4493,
195020
+ "step": 75180
195021
+ },
195022
+ {
195023
+ "epoch": 601.18,
195024
+ "learning_rate": 8.811602564102565e-06,
195025
+ "loss": 0.7843,
195026
+ "step": 75185
195027
+ },
195028
+ {
195029
+ "epoch": 601.22,
195030
+ "learning_rate": 8.811522435897438e-06,
195031
+ "loss": 0.8807,
195032
+ "step": 75190
195033
+ },
195034
+ {
195035
+ "epoch": 601.26,
195036
+ "learning_rate": 8.811442307692309e-06,
195037
+ "loss": 0.2823,
195038
+ "step": 75195
195039
+ },
195040
+ {
195041
+ "epoch": 601.3,
195042
+ "learning_rate": 8.81136217948718e-06,
195043
+ "loss": 0.387,
195044
+ "step": 75200
195045
+ },
195046
+ {
195047
+ "epoch": 601.34,
195048
+ "learning_rate": 8.811282051282052e-06,
195049
+ "loss": 0.3841,
195050
+ "step": 75205
195051
+ },
195052
+ {
195053
+ "epoch": 601.38,
195054
+ "learning_rate": 8.811201923076925e-06,
195055
+ "loss": 0.7324,
195056
+ "step": 75210
195057
+ },
195058
+ {
195059
+ "epoch": 601.42,
195060
+ "learning_rate": 8.811121794871795e-06,
195061
+ "loss": 0.8773,
195062
+ "step": 75215
195063
+ },
195064
+ {
195065
+ "epoch": 601.46,
195066
+ "learning_rate": 8.811041666666668e-06,
195067
+ "loss": 0.2424,
195068
+ "step": 75220
195069
+ },
195070
+ {
195071
+ "epoch": 601.5,
195072
+ "learning_rate": 8.810961538461539e-06,
195073
+ "loss": 0.366,
195074
+ "step": 75225
195075
+ },
195076
+ {
195077
+ "epoch": 601.54,
195078
+ "learning_rate": 8.81088141025641e-06,
195079
+ "loss": 0.5086,
195080
+ "step": 75230
195081
+ },
195082
+ {
195083
+ "epoch": 601.58,
195084
+ "learning_rate": 8.810801282051282e-06,
195085
+ "loss": 0.8468,
195086
+ "step": 75235
195087
+ },
195088
+ {
195089
+ "epoch": 601.62,
195090
+ "learning_rate": 8.810721153846155e-06,
195091
+ "loss": 1.0401,
195092
+ "step": 75240
195093
+ },
195094
+ {
195095
+ "epoch": 601.66,
195096
+ "learning_rate": 8.810641025641026e-06,
195097
+ "loss": 0.3034,
195098
+ "step": 75245
195099
+ },
195100
+ {
195101
+ "epoch": 601.7,
195102
+ "learning_rate": 8.810560897435897e-06,
195103
+ "loss": 0.3393,
195104
+ "step": 75250
195105
+ },
195106
+ {
195107
+ "epoch": 601.74,
195108
+ "learning_rate": 8.81048076923077e-06,
195109
+ "loss": 0.3378,
195110
+ "step": 75255
195111
+ },
195112
+ {
195113
+ "epoch": 601.78,
195114
+ "learning_rate": 8.810400641025642e-06,
195115
+ "loss": 0.7611,
195116
+ "step": 75260
195117
+ },
195118
+ {
195119
+ "epoch": 601.82,
195120
+ "learning_rate": 8.810320512820513e-06,
195121
+ "loss": 1.0786,
195122
+ "step": 75265
195123
+ },
195124
+ {
195125
+ "epoch": 601.86,
195126
+ "learning_rate": 8.810240384615385e-06,
195127
+ "loss": 0.3302,
195128
+ "step": 75270
195129
+ },
195130
+ {
195131
+ "epoch": 601.9,
195132
+ "learning_rate": 8.810160256410258e-06,
195133
+ "loss": 0.2704,
195134
+ "step": 75275
195135
+ },
195136
+ {
195137
+ "epoch": 601.94,
195138
+ "learning_rate": 8.810080128205129e-06,
195139
+ "loss": 0.4022,
195140
+ "step": 75280
195141
+ },
195142
+ {
195143
+ "epoch": 601.98,
195144
+ "learning_rate": 8.81e-06,
195145
+ "loss": 0.8502,
195146
+ "step": 75285
195147
+ },
195148
+ {
195149
+ "epoch": 602.0,
195150
+ "eval_loss": 0.3613823354244232,
195151
+ "eval_runtime": 40.6158,
195152
+ "eval_samples_per_second": 20.706,
195153
+ "eval_steps_per_second": 0.665,
195154
+ "eval_wer": 0.1880054390610463,
195155
+ "step": 75287
195156
+ },
195157
+ {
195158
+ "epoch": 602.02,
195159
+ "learning_rate": 8.809919871794873e-06,
195160
+ "loss": 0.3835,
195161
+ "step": 75290
195162
+ },
195163
+ {
195164
+ "epoch": 602.06,
195165
+ "learning_rate": 8.809839743589745e-06,
195166
+ "loss": 0.2873,
195167
+ "step": 75295
195168
+ },
195169
+ {
195170
+ "epoch": 602.1,
195171
+ "learning_rate": 8.809759615384616e-06,
195172
+ "loss": 0.2944,
195173
+ "step": 75300
195174
+ },
195175
+ {
195176
+ "epoch": 602.14,
195177
+ "learning_rate": 8.809679487179487e-06,
195178
+ "loss": 0.3795,
195179
+ "step": 75305
195180
+ },
195181
+ {
195182
+ "epoch": 602.18,
195183
+ "learning_rate": 8.80959935897436e-06,
195184
+ "loss": 0.8076,
195185
+ "step": 75310
195186
+ },
195187
+ {
195188
+ "epoch": 602.22,
195189
+ "learning_rate": 8.809519230769232e-06,
195190
+ "loss": 0.8927,
195191
+ "step": 75315
195192
+ },
195193
+ {
195194
+ "epoch": 602.26,
195195
+ "learning_rate": 8.809439102564103e-06,
195196
+ "loss": 0.3271,
195197
+ "step": 75320
195198
+ },
195199
+ {
195200
+ "epoch": 602.3,
195201
+ "learning_rate": 8.809358974358975e-06,
195202
+ "loss": 0.3014,
195203
+ "step": 75325
195204
+ },
195205
+ {
195206
+ "epoch": 602.34,
195207
+ "learning_rate": 8.809278846153848e-06,
195208
+ "loss": 0.449,
195209
+ "step": 75330
195210
+ },
195211
+ {
195212
+ "epoch": 602.38,
195213
+ "learning_rate": 8.809198717948717e-06,
195214
+ "loss": 0.9804,
195215
+ "step": 75335
195216
+ },
195217
+ {
195218
+ "epoch": 602.42,
195219
+ "learning_rate": 8.80911858974359e-06,
195220
+ "loss": 0.8483,
195221
+ "step": 75340
195222
+ },
195223
+ {
195224
+ "epoch": 602.46,
195225
+ "learning_rate": 8.809038461538463e-06,
195226
+ "loss": 0.307,
195227
+ "step": 75345
195228
+ },
195229
+ {
195230
+ "epoch": 602.5,
195231
+ "learning_rate": 8.808958333333335e-06,
195232
+ "loss": 0.3112,
195233
+ "step": 75350
195234
+ },
195235
+ {
195236
+ "epoch": 602.54,
195237
+ "learning_rate": 8.808878205128206e-06,
195238
+ "loss": 0.4685,
195239
+ "step": 75355
195240
+ },
195241
+ {
195242
+ "epoch": 602.58,
195243
+ "learning_rate": 8.808798076923078e-06,
195244
+ "loss": 0.8196,
195245
+ "step": 75360
195246
+ },
195247
+ {
195248
+ "epoch": 602.62,
195249
+ "learning_rate": 8.80871794871795e-06,
195250
+ "loss": 0.7176,
195251
+ "step": 75365
195252
+ },
195253
+ {
195254
+ "epoch": 602.66,
195255
+ "learning_rate": 8.80863782051282e-06,
195256
+ "loss": 0.3299,
195257
+ "step": 75370
195258
+ },
195259
+ {
195260
+ "epoch": 602.7,
195261
+ "learning_rate": 8.808557692307693e-06,
195262
+ "loss": 0.3395,
195263
+ "step": 75375
195264
+ },
195265
+ {
195266
+ "epoch": 602.74,
195267
+ "learning_rate": 8.808477564102565e-06,
195268
+ "loss": 0.482,
195269
+ "step": 75380
195270
+ },
195271
+ {
195272
+ "epoch": 602.78,
195273
+ "learning_rate": 8.808397435897436e-06,
195274
+ "loss": 0.7871,
195275
+ "step": 75385
195276
+ },
195277
+ {
195278
+ "epoch": 602.82,
195279
+ "learning_rate": 8.808317307692307e-06,
195280
+ "loss": 1.0059,
195281
+ "step": 75390
195282
+ },
195283
+ {
195284
+ "epoch": 602.86,
195285
+ "learning_rate": 8.80823717948718e-06,
195286
+ "loss": 0.289,
195287
+ "step": 75395
195288
+ },
195289
+ {
195290
+ "epoch": 602.9,
195291
+ "learning_rate": 8.808157051282052e-06,
195292
+ "loss": 0.3244,
195293
+ "step": 75400
195294
+ },
195295
+ {
195296
+ "epoch": 602.94,
195297
+ "learning_rate": 8.808076923076923e-06,
195298
+ "loss": 0.4663,
195299
+ "step": 75405
195300
+ },
195301
+ {
195302
+ "epoch": 602.98,
195303
+ "learning_rate": 8.807996794871796e-06,
195304
+ "loss": 0.7288,
195305
+ "step": 75410
195306
+ },
195307
+ {
195308
+ "epoch": 603.0,
195309
+ "eval_loss": 0.4494644105434418,
195310
+ "eval_runtime": 40.3969,
195311
+ "eval_samples_per_second": 20.818,
195312
+ "eval_steps_per_second": 0.668,
195313
+ "eval_wer": 0.18498835177635412,
195314
+ "step": 75412
195315
+ },
195316
+ {
195317
+ "epoch": 603.02,
195318
+ "learning_rate": 8.807916666666668e-06,
195319
+ "loss": 0.4818,
195320
+ "step": 75415
195321
+ },
195322
+ {
195323
+ "epoch": 603.06,
195324
+ "learning_rate": 8.807836538461539e-06,
195325
+ "loss": 0.2858,
195326
+ "step": 75420
195327
+ },
195328
+ {
195329
+ "epoch": 603.1,
195330
+ "learning_rate": 8.80775641025641e-06,
195331
+ "loss": 0.3195,
195332
+ "step": 75425
195333
+ },
195334
+ {
195335
+ "epoch": 603.14,
195336
+ "learning_rate": 8.807676282051283e-06,
195337
+ "loss": 0.3992,
195338
+ "step": 75430
195339
+ },
195340
+ {
195341
+ "epoch": 603.18,
195342
+ "learning_rate": 8.807596153846155e-06,
195343
+ "loss": 0.8451,
195344
+ "step": 75435
195345
+ },
195346
+ {
195347
+ "epoch": 603.22,
195348
+ "learning_rate": 8.807516025641026e-06,
195349
+ "loss": 0.8592,
195350
+ "step": 75440
195351
+ },
195352
+ {
195353
+ "epoch": 603.26,
195354
+ "learning_rate": 8.807435897435899e-06,
195355
+ "loss": 0.3101,
195356
+ "step": 75445
195357
+ },
195358
+ {
195359
+ "epoch": 603.3,
195360
+ "learning_rate": 8.80735576923077e-06,
195361
+ "loss": 0.3108,
195362
+ "step": 75450
195363
+ },
195364
+ {
195365
+ "epoch": 603.34,
195366
+ "learning_rate": 8.807275641025642e-06,
195367
+ "loss": 0.3803,
195368
+ "step": 75455
195369
+ },
195370
+ {
195371
+ "epoch": 603.38,
195372
+ "learning_rate": 8.807195512820513e-06,
195373
+ "loss": 0.8631,
195374
+ "step": 75460
195375
+ },
195376
+ {
195377
+ "epoch": 603.42,
195378
+ "learning_rate": 8.807115384615386e-06,
195379
+ "loss": 0.9733,
195380
+ "step": 75465
195381
+ },
195382
+ {
195383
+ "epoch": 603.46,
195384
+ "learning_rate": 8.807035256410258e-06,
195385
+ "loss": 0.3768,
195386
+ "step": 75470
195387
+ },
195388
+ {
195389
+ "epoch": 603.5,
195390
+ "learning_rate": 8.806955128205129e-06,
195391
+ "loss": 0.3623,
195392
+ "step": 75475
195393
+ },
195394
+ {
195395
+ "epoch": 603.54,
195396
+ "learning_rate": 8.806875e-06,
195397
+ "loss": 0.3788,
195398
+ "step": 75480
195399
+ },
195400
+ {
195401
+ "epoch": 603.58,
195402
+ "learning_rate": 8.806794871794873e-06,
195403
+ "loss": 0.7772,
195404
+ "step": 75485
195405
+ },
195406
+ {
195407
+ "epoch": 603.62,
195408
+ "learning_rate": 8.806714743589743e-06,
195409
+ "loss": 0.9261,
195410
+ "step": 75490
195411
+ },
195412
+ {
195413
+ "epoch": 603.66,
195414
+ "learning_rate": 8.806634615384616e-06,
195415
+ "loss": 0.352,
195416
+ "step": 75495
195417
+ },
195418
+ {
195419
+ "epoch": 603.7,
195420
+ "learning_rate": 8.806554487179489e-06,
195421
+ "loss": 0.3595,
195422
+ "step": 75500
195423
+ },
195424
+ {
195425
+ "epoch": 603.74,
195426
+ "learning_rate": 8.806474358974359e-06,
195427
+ "loss": 0.3123,
195428
+ "step": 75505
195429
+ },
195430
+ {
195431
+ "epoch": 603.78,
195432
+ "learning_rate": 8.806394230769232e-06,
195433
+ "loss": 0.7048,
195434
+ "step": 75510
195435
+ },
195436
+ {
195437
+ "epoch": 603.82,
195438
+ "learning_rate": 8.806314102564103e-06,
195439
+ "loss": 0.9546,
195440
+ "step": 75515
195441
+ },
195442
+ {
195443
+ "epoch": 603.86,
195444
+ "learning_rate": 8.806233974358975e-06,
195445
+ "loss": 0.29,
195446
+ "step": 75520
195447
+ },
195448
+ {
195449
+ "epoch": 603.9,
195450
+ "learning_rate": 8.806153846153846e-06,
195451
+ "loss": 0.3054,
195452
+ "step": 75525
195453
+ },
195454
+ {
195455
+ "epoch": 603.94,
195456
+ "learning_rate": 8.806073717948719e-06,
195457
+ "loss": 0.3936,
195458
+ "step": 75530
195459
+ },
195460
+ {
195461
+ "epoch": 603.98,
195462
+ "learning_rate": 8.80599358974359e-06,
195463
+ "loss": 0.9524,
195464
+ "step": 75535
195465
+ },
195466
+ {
195467
+ "epoch": 604.0,
195468
+ "eval_loss": 0.4324324429035187,
195469
+ "eval_runtime": 41.0453,
195470
+ "eval_samples_per_second": 20.49,
195471
+ "eval_steps_per_second": 0.658,
195472
+ "eval_wer": 0.18972477064220183,
195473
+ "step": 75537
195474
+ },
195475
+ {
195476
+ "epoch": 604.02,
195477
+ "learning_rate": 8.805913461538462e-06,
195478
+ "loss": 0.3299,
195479
+ "step": 75540
195480
+ },
195481
+ {
195482
+ "epoch": 604.06,
195483
+ "learning_rate": 8.805833333333335e-06,
195484
+ "loss": 0.3726,
195485
+ "step": 75545
195486
+ },
195487
+ {
195488
+ "epoch": 604.1,
195489
+ "learning_rate": 8.805753205128206e-06,
195490
+ "loss": 0.3474,
195491
+ "step": 75550
195492
+ },
195493
+ {
195494
+ "epoch": 604.14,
195495
+ "learning_rate": 8.805673076923077e-06,
195496
+ "loss": 0.3935,
195497
+ "step": 75555
195498
+ },
195499
+ {
195500
+ "epoch": 604.18,
195501
+ "learning_rate": 8.805592948717949e-06,
195502
+ "loss": 0.7502,
195503
+ "step": 75560
195504
+ },
195505
+ {
195506
+ "epoch": 604.22,
195507
+ "learning_rate": 8.805512820512822e-06,
195508
+ "loss": 0.8715,
195509
+ "step": 75565
195510
+ },
195511
+ {
195512
+ "epoch": 604.26,
195513
+ "learning_rate": 8.805432692307693e-06,
195514
+ "loss": 0.4304,
195515
+ "step": 75570
195516
+ },
195517
+ {
195518
+ "epoch": 604.3,
195519
+ "learning_rate": 8.805352564102565e-06,
195520
+ "loss": 0.3175,
195521
+ "step": 75575
195522
+ },
195523
+ {
195524
+ "epoch": 604.34,
195525
+ "learning_rate": 8.805272435897436e-06,
195526
+ "loss": 0.3441,
195527
+ "step": 75580
195528
+ },
195529
+ {
195530
+ "epoch": 604.38,
195531
+ "learning_rate": 8.805192307692309e-06,
195532
+ "loss": 0.7054,
195533
+ "step": 75585
195534
+ },
195535
+ {
195536
+ "epoch": 604.42,
195537
+ "learning_rate": 8.80511217948718e-06,
195538
+ "loss": 0.8604,
195539
+ "step": 75590
195540
+ },
195541
+ {
195542
+ "epoch": 604.46,
195543
+ "learning_rate": 8.805032051282052e-06,
195544
+ "loss": 0.2919,
195545
+ "step": 75595
195546
+ },
195547
+ {
195548
+ "epoch": 604.5,
195549
+ "learning_rate": 8.804951923076925e-06,
195550
+ "loss": 0.3158,
195551
+ "step": 75600
195552
+ },
195553
+ {
195554
+ "epoch": 604.54,
195555
+ "learning_rate": 8.804871794871796e-06,
195556
+ "loss": 0.4358,
195557
+ "step": 75605
195558
+ },
195559
+ {
195560
+ "epoch": 604.58,
195561
+ "learning_rate": 8.804791666666667e-06,
195562
+ "loss": 0.8265,
195563
+ "step": 75610
195564
+ },
195565
+ {
195566
+ "epoch": 604.62,
195567
+ "learning_rate": 8.804711538461539e-06,
195568
+ "loss": 0.8654,
195569
+ "step": 75615
195570
+ },
195571
+ {
195572
+ "epoch": 604.66,
195573
+ "learning_rate": 8.804631410256412e-06,
195574
+ "loss": 0.272,
195575
+ "step": 75620
195576
+ },
195577
+ {
195578
+ "epoch": 604.7,
195579
+ "learning_rate": 8.804551282051283e-06,
195580
+ "loss": 0.2909,
195581
+ "step": 75625
195582
+ },
195583
+ {
195584
+ "epoch": 604.74,
195585
+ "learning_rate": 8.804471153846155e-06,
195586
+ "loss": 0.4165,
195587
+ "step": 75630
195588
+ },
195589
+ {
195590
+ "epoch": 604.78,
195591
+ "learning_rate": 8.804391025641026e-06,
195592
+ "loss": 0.8363,
195593
+ "step": 75635
195594
+ },
195595
+ {
195596
+ "epoch": 604.82,
195597
+ "learning_rate": 8.804310897435899e-06,
195598
+ "loss": 0.8427,
195599
+ "step": 75640
195600
+ },
195601
+ {
195602
+ "epoch": 604.86,
195603
+ "learning_rate": 8.80423076923077e-06,
195604
+ "loss": 0.3652,
195605
+ "step": 75645
195606
+ },
195607
+ {
195608
+ "epoch": 604.9,
195609
+ "learning_rate": 8.804150641025642e-06,
195610
+ "loss": 0.3643,
195611
+ "step": 75650
195612
+ },
195613
+ {
195614
+ "epoch": 604.94,
195615
+ "learning_rate": 8.804070512820515e-06,
195616
+ "loss": 0.3866,
195617
+ "step": 75655
195618
+ },
195619
+ {
195620
+ "epoch": 604.98,
195621
+ "learning_rate": 8.803990384615384e-06,
195622
+ "loss": 0.8886,
195623
+ "step": 75660
195624
+ },
195625
+ {
195626
+ "epoch": 605.0,
195627
+ "eval_loss": 0.3445565700531006,
195628
+ "eval_runtime": 40.5466,
195629
+ "eval_samples_per_second": 20.717,
195630
+ "eval_steps_per_second": 0.666,
195631
+ "eval_wer": 0.18799435028248587,
195632
+ "step": 75662
195633
+ },
195634
+ {
195635
+ "epoch": 605.02,
195636
+ "learning_rate": 8.803910256410257e-06,
195637
+ "loss": 0.3486,
195638
+ "step": 75665
195639
+ },
195640
+ {
195641
+ "epoch": 605.06,
195642
+ "learning_rate": 8.803830128205129e-06,
195643
+ "loss": 0.3072,
195644
+ "step": 75670
195645
+ },
195646
+ {
195647
+ "epoch": 605.1,
195648
+ "learning_rate": 8.80375e-06,
195649
+ "loss": 0.2955,
195650
+ "step": 75675
195651
+ },
195652
+ {
195653
+ "epoch": 605.14,
195654
+ "learning_rate": 8.803669871794872e-06,
195655
+ "loss": 0.3452,
195656
+ "step": 75680
195657
+ },
195658
+ {
195659
+ "epoch": 605.18,
195660
+ "learning_rate": 8.803589743589745e-06,
195661
+ "loss": 0.716,
195662
+ "step": 75685
195663
+ },
195664
+ {
195665
+ "epoch": 605.22,
195666
+ "learning_rate": 8.803509615384616e-06,
195667
+ "loss": 0.9065,
195668
+ "step": 75690
195669
+ },
195670
+ {
195671
+ "epoch": 605.26,
195672
+ "learning_rate": 8.803429487179487e-06,
195673
+ "loss": 0.2987,
195674
+ "step": 75695
195675
+ },
195676
+ {
195677
+ "epoch": 605.3,
195678
+ "learning_rate": 8.80334935897436e-06,
195679
+ "loss": 0.3722,
195680
+ "step": 75700
195681
+ },
195682
+ {
195683
+ "epoch": 605.34,
195684
+ "learning_rate": 8.803269230769232e-06,
195685
+ "loss": 0.4089,
195686
+ "step": 75705
195687
+ },
195688
+ {
195689
+ "epoch": 605.38,
195690
+ "learning_rate": 8.803189102564103e-06,
195691
+ "loss": 0.7393,
195692
+ "step": 75710
195693
+ },
195694
+ {
195695
+ "epoch": 605.42,
195696
+ "learning_rate": 8.803108974358974e-06,
195697
+ "loss": 1.0048,
195698
+ "step": 75715
195699
+ },
195700
+ {
195701
+ "epoch": 605.46,
195702
+ "learning_rate": 8.803028846153848e-06,
195703
+ "loss": 0.3025,
195704
+ "step": 75720
195705
+ },
195706
+ {
195707
+ "epoch": 605.5,
195708
+ "learning_rate": 8.802948717948719e-06,
195709
+ "loss": 0.2752,
195710
+ "step": 75725
195711
+ },
195712
+ {
195713
+ "epoch": 605.54,
195714
+ "learning_rate": 8.80286858974359e-06,
195715
+ "loss": 0.4407,
195716
+ "step": 75730
195717
+ },
195718
+ {
195719
+ "epoch": 605.58,
195720
+ "learning_rate": 8.802788461538462e-06,
195721
+ "loss": 0.7918,
195722
+ "step": 75735
195723
+ },
195724
+ {
195725
+ "epoch": 605.62,
195726
+ "learning_rate": 8.802708333333335e-06,
195727
+ "loss": 0.8038,
195728
+ "step": 75740
195729
+ },
195730
+ {
195731
+ "epoch": 605.66,
195732
+ "learning_rate": 8.802628205128206e-06,
195733
+ "loss": 0.2619,
195734
+ "step": 75745
195735
+ },
195736
+ {
195737
+ "epoch": 605.7,
195738
+ "learning_rate": 8.802548076923077e-06,
195739
+ "loss": 0.2717,
195740
+ "step": 75750
195741
+ },
195742
+ {
195743
+ "epoch": 605.74,
195744
+ "learning_rate": 8.80246794871795e-06,
195745
+ "loss": 0.3758,
195746
+ "step": 75755
195747
+ },
195748
+ {
195749
+ "epoch": 605.78,
195750
+ "learning_rate": 8.802387820512822e-06,
195751
+ "loss": 0.862,
195752
+ "step": 75760
195753
+ },
195754
+ {
195755
+ "epoch": 605.82,
195756
+ "learning_rate": 8.802307692307693e-06,
195757
+ "loss": 0.8831,
195758
+ "step": 75765
195759
+ },
195760
+ {
195761
+ "epoch": 605.86,
195762
+ "learning_rate": 8.802227564102564e-06,
195763
+ "loss": 0.2855,
195764
+ "step": 75770
195765
+ },
195766
+ {
195767
+ "epoch": 605.9,
195768
+ "learning_rate": 8.802147435897438e-06,
195769
+ "loss": 0.2964,
195770
+ "step": 75775
195771
+ },
195772
+ {
195773
+ "epoch": 605.94,
195774
+ "learning_rate": 8.802067307692307e-06,
195775
+ "loss": 0.3788,
195776
+ "step": 75780
195777
+ },
195778
+ {
195779
+ "epoch": 605.98,
195780
+ "learning_rate": 8.80198717948718e-06,
195781
+ "loss": 0.7917,
195782
+ "step": 75785
195783
+ },
195784
+ {
195785
+ "epoch": 606.0,
195786
+ "eval_loss": 0.40229326486587524,
195787
+ "eval_runtime": 40.811,
195788
+ "eval_samples_per_second": 20.583,
195789
+ "eval_steps_per_second": 0.662,
195790
+ "eval_wer": 0.1924281241104469,
195791
+ "step": 75787
195792
  }
195793
  ],
195794
  "max_steps": 625000,
195795
  "num_train_epochs": 5000,
195796
+ "total_flos": 2.1327406074811505e+20,
195797
  "trial_name": null,
195798
  "trial_params": null
195799
  }
model-bin/finetune/base/{checkpoint-75162 β†’ checkpoint-75787}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629882392.1236827/events.out.tfevents.1629882392.7e498afd5545.905.93 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40faae71c44c577de7af7caf69225cd70fa91b8106891c9da59aeafbc4240253
3
+ size 4194
model-bin/finetune/base/log/1629883056.4519293/events.out.tfevents.1629883056.7e498afd5545.905.95 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:939b06a3739cc17a7e4511ffb94973599bbcb123803382e8d793b74decdad891
3
+ size 4194
model-bin/finetune/base/log/1629883697.070274/events.out.tfevents.1629883697.7e498afd5545.905.97 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd4a1c43c8d0f8393356223e62609d12fe99feeffdaa48864cd0affc5cd80930
3
+ size 4194
model-bin/finetune/base/log/1629884340.6409886/events.out.tfevents.1629884340.7e498afd5545.905.99 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3047e66a47c520ca0d38a0885389812710a8a99fe0f4c1eee86b29e1ed75e104
3
+ size 4194
model-bin/finetune/base/log/1629885006.4519074/events.out.tfevents.1629885006.7e498afd5545.905.101 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:082cef36521c91b17fff53cd50db092c8004bc2f4c943a8df79734e4914e2dbf
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629882392.7e498afd5545.905.92 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e16d61233274ee5c7a30f50978aade662d5dab45638811ad3f304ccc8670268a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629883056.7e498afd5545.905.94 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90b55c10715a3dd6e440f8c50d728d3f72f21cd59abdfaebc6a4eb38fc95b3af
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629883697.7e498afd5545.905.96 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9f668ebd3dee0e05fcda258dc1e05f0bfad91436085cd5c6757bb21b1221574
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629884340.7e498afd5545.905.98 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e943e824b462a7e1503eebe32d0a77c274a8edcba1936795e22fd327cc802000
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629885006.7e498afd5545.905.100 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45aa3d1626d97711f20adbacc8607e2484e99e05409cd57762d9b59f5f3c47cb
3
+ size 8622