Check commited on
Commit
51bc235
Β·
1 Parent(s): c327d38

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630246352.8300264/events.out.tfevents.1630246352.cc93b136ebf5.1086.189 +3 -0
  11. model-bin/finetune/base/log/1630246785.4372547/events.out.tfevents.1630246785.cc93b136ebf5.1086.191 +3 -0
  12. model-bin/finetune/base/log/1630247219.004471/events.out.tfevents.1630247219.cc93b136ebf5.1086.193 +3 -0
  13. model-bin/finetune/base/log/1630247655.9469938/events.out.tfevents.1630247655.cc93b136ebf5.1086.195 +3 -0
  14. model-bin/finetune/base/log/1630248088.0553622/events.out.tfevents.1630248088.cc93b136ebf5.1086.197 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630246352.cc93b136ebf5.1086.188 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630246785.cc93b136ebf5.1086.190 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630247218.cc93b136ebf5.1086.192 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630247655.cc93b136ebf5.1086.194 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630248088.cc93b136ebf5.1086.196 +3 -0
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9665c6e1c6f2987ac982b4f02510d642ccf49cb18e475af98729715643f2744
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e9eede5251c3f12c63b46d66a179f1dbd2b6d9f12ab83956ee5715976561602
3
  size 722165393
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83712bb0bb146694098ef77fead19cec258379ae2b0f648e4edb18652a55565a
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6bd337b8f0e2f3b46fc1c4ecb5f3bafbfc576e1e6bde2071ce0f670336e2879
3
  size 377909911
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f44a922c3d45f1ec6984b47ad3b841b6c8eca8288bc65f7fa757f180153fd40
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a1d6790720c3684b5fab1b80fa056d808968462ad941960785ea5dec6ae7387
3
  size 14503
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:adf29ee28b173423b576c771747dd93d436f5c3f2a3af9a31f91c508ff5dd1b2
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:884cb1f5ec43d0bdfcfad8769f8573a767f703fd8562f4c240659e0fdd8266d6
3
  size 559
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee134784171ae4f4005cf971653fad3f3e37c1a6c7b565c91f04852ab760d0e4
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be40c7eaf300a7ee9457562db278d81160ffb7d793ceebd4250c4e629248949f
3
  size 623
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1206.0,
5
- "global_step": 150827,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -291264,11 +291264,800 @@
291264
  "eval_steps_per_second": 0.671,
291265
  "eval_wer": 0.182829413960433,
291266
  "step": 150827
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
291267
  }
291268
  ],
291269
- "max_steps": 625000,
291270
  "num_train_epochs": 5000,
291271
- "total_flos": 4.2444315493114164e+20,
291272
  "trial_name": null,
291273
  "trial_params": null
291274
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1220.995983935743,
5
+ "global_step": 151448,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
291264
  "eval_steps_per_second": 0.671,
291265
  "eval_wer": 0.182829413960433,
291266
  "step": 150827
291267
+ },
291268
+ {
291269
+ "epoch": 1216.02,
291270
+ "learning_rate": 7.580436187399032e-06,
291271
+ "loss": 0.2669,
291272
+ "step": 150830
291273
+ },
291274
+ {
291275
+ "epoch": 1216.06,
291276
+ "learning_rate": 7.580355411954766e-06,
291277
+ "loss": 0.2552,
291278
+ "step": 150835
291279
+ },
291280
+ {
291281
+ "epoch": 1216.1,
291282
+ "learning_rate": 7.580274636510502e-06,
291283
+ "loss": 0.2963,
291284
+ "step": 150840
291285
+ },
291286
+ {
291287
+ "epoch": 1216.14,
291288
+ "learning_rate": 7.580193861066237e-06,
291289
+ "loss": 0.3486,
291290
+ "step": 150845
291291
+ },
291292
+ {
291293
+ "epoch": 1216.18,
291294
+ "learning_rate": 7.580113085621972e-06,
291295
+ "loss": 0.7076,
291296
+ "step": 150850
291297
+ },
291298
+ {
291299
+ "epoch": 1216.22,
291300
+ "learning_rate": 7.580032310177707e-06,
291301
+ "loss": 0.7936,
291302
+ "step": 150855
291303
+ },
291304
+ {
291305
+ "epoch": 1216.27,
291306
+ "learning_rate": 7.579951534733442e-06,
291307
+ "loss": 0.2483,
291308
+ "step": 150860
291309
+ },
291310
+ {
291311
+ "epoch": 1216.31,
291312
+ "learning_rate": 7.579870759289177e-06,
291313
+ "loss": 0.3237,
291314
+ "step": 150865
291315
+ },
291316
+ {
291317
+ "epoch": 1216.35,
291318
+ "learning_rate": 7.579789983844912e-06,
291319
+ "loss": 0.3846,
291320
+ "step": 150870
291321
+ },
291322
+ {
291323
+ "epoch": 1216.39,
291324
+ "learning_rate": 7.5797092084006466e-06,
291325
+ "loss": 0.673,
291326
+ "step": 150875
291327
+ },
291328
+ {
291329
+ "epoch": 1216.43,
291330
+ "learning_rate": 7.5796284329563816e-06,
291331
+ "loss": 0.897,
291332
+ "step": 150880
291333
+ },
291334
+ {
291335
+ "epoch": 1216.47,
291336
+ "learning_rate": 7.579547657512117e-06,
291337
+ "loss": 0.2648,
291338
+ "step": 150885
291339
+ },
291340
+ {
291341
+ "epoch": 1216.51,
291342
+ "learning_rate": 7.5794668820678515e-06,
291343
+ "loss": 0.3112,
291344
+ "step": 150890
291345
+ },
291346
+ {
291347
+ "epoch": 1216.55,
291348
+ "learning_rate": 7.579386106623587e-06,
291349
+ "loss": 0.367,
291350
+ "step": 150895
291351
+ },
291352
+ {
291353
+ "epoch": 1216.59,
291354
+ "learning_rate": 7.5793053311793215e-06,
291355
+ "loss": 0.7596,
291356
+ "step": 150900
291357
+ },
291358
+ {
291359
+ "epoch": 1216.63,
291360
+ "learning_rate": 7.579224555735057e-06,
291361
+ "loss": 0.8846,
291362
+ "step": 150905
291363
+ },
291364
+ {
291365
+ "epoch": 1216.67,
291366
+ "learning_rate": 7.5791437802907915e-06,
291367
+ "loss": 0.2878,
291368
+ "step": 150910
291369
+ },
291370
+ {
291371
+ "epoch": 1216.71,
291372
+ "learning_rate": 7.579063004846527e-06,
291373
+ "loss": 0.2911,
291374
+ "step": 150915
291375
+ },
291376
+ {
291377
+ "epoch": 1216.75,
291378
+ "learning_rate": 7.578982229402262e-06,
291379
+ "loss": 0.4213,
291380
+ "step": 150920
291381
+ },
291382
+ {
291383
+ "epoch": 1216.79,
291384
+ "learning_rate": 7.578901453957997e-06,
291385
+ "loss": 0.6366,
291386
+ "step": 150925
291387
+ },
291388
+ {
291389
+ "epoch": 1216.83,
291390
+ "learning_rate": 7.578820678513732e-06,
291391
+ "loss": 0.886,
291392
+ "step": 150930
291393
+ },
291394
+ {
291395
+ "epoch": 1216.87,
291396
+ "learning_rate": 7.578739903069467e-06,
291397
+ "loss": 0.2832,
291398
+ "step": 150935
291399
+ },
291400
+ {
291401
+ "epoch": 1216.91,
291402
+ "learning_rate": 7.578659127625203e-06,
291403
+ "loss": 0.2781,
291404
+ "step": 150940
291405
+ },
291406
+ {
291407
+ "epoch": 1216.95,
291408
+ "learning_rate": 7.578578352180937e-06,
291409
+ "loss": 0.3466,
291410
+ "step": 150945
291411
+ },
291412
+ {
291413
+ "epoch": 1216.99,
291414
+ "learning_rate": 7.578497576736673e-06,
291415
+ "loss": 0.7638,
291416
+ "step": 150950
291417
+ },
291418
+ {
291419
+ "epoch": 1217.0,
291420
+ "eval_loss": 0.3567521274089813,
291421
+ "eval_runtime": 40.4597,
291422
+ "eval_samples_per_second": 20.687,
291423
+ "eval_steps_per_second": 0.667,
291424
+ "eval_wer": 0.17730812013348166,
291425
+ "step": 150951
291426
+ },
291427
+ {
291428
+ "epoch": 1207.03,
291429
+ "learning_rate": 7.578416801292407e-06,
291430
+ "loss": 0.2737,
291431
+ "step": 150955
291432
+ },
291433
+ {
291434
+ "epoch": 1207.07,
291435
+ "learning_rate": 7.578336025848143e-06,
291436
+ "loss": 0.261,
291437
+ "step": 150960
291438
+ },
291439
+ {
291440
+ "epoch": 1207.11,
291441
+ "learning_rate": 7.578255250403877e-06,
291442
+ "loss": 0.2653,
291443
+ "step": 150965
291444
+ },
291445
+ {
291446
+ "epoch": 1207.15,
291447
+ "learning_rate": 7.578174474959613e-06,
291448
+ "loss": 0.4096,
291449
+ "step": 150970
291450
+ },
291451
+ {
291452
+ "epoch": 1207.19,
291453
+ "learning_rate": 7.578093699515347e-06,
291454
+ "loss": 0.8663,
291455
+ "step": 150975
291456
+ },
291457
+ {
291458
+ "epoch": 1207.23,
291459
+ "learning_rate": 7.578012924071083e-06,
291460
+ "loss": 0.5652,
291461
+ "step": 150980
291462
+ },
291463
+ {
291464
+ "epoch": 1207.27,
291465
+ "learning_rate": 7.577932148626818e-06,
291466
+ "loss": 0.2107,
291467
+ "step": 150985
291468
+ },
291469
+ {
291470
+ "epoch": 1207.31,
291471
+ "learning_rate": 7.577851373182553e-06,
291472
+ "loss": 0.2782,
291473
+ "step": 150990
291474
+ },
291475
+ {
291476
+ "epoch": 1207.35,
291477
+ "learning_rate": 7.577770597738289e-06,
291478
+ "loss": 0.4066,
291479
+ "step": 150995
291480
+ },
291481
+ {
291482
+ "epoch": 1207.39,
291483
+ "learning_rate": 7.577689822294023e-06,
291484
+ "loss": 0.9056,
291485
+ "step": 151000
291486
+ },
291487
+ {
291488
+ "epoch": 1207.43,
291489
+ "learning_rate": 7.577609046849759e-06,
291490
+ "loss": 0.6806,
291491
+ "step": 151005
291492
+ },
291493
+ {
291494
+ "epoch": 1207.47,
291495
+ "learning_rate": 7.577528271405493e-06,
291496
+ "loss": 0.3028,
291497
+ "step": 151010
291498
+ },
291499
+ {
291500
+ "epoch": 1207.51,
291501
+ "learning_rate": 7.577447495961229e-06,
291502
+ "loss": 0.3177,
291503
+ "step": 151015
291504
+ },
291505
+ {
291506
+ "epoch": 1207.55,
291507
+ "learning_rate": 7.577366720516963e-06,
291508
+ "loss": 0.37,
291509
+ "step": 151020
291510
+ },
291511
+ {
291512
+ "epoch": 1207.59,
291513
+ "learning_rate": 7.577285945072699e-06,
291514
+ "loss": 0.8868,
291515
+ "step": 151025
291516
+ },
291517
+ {
291518
+ "epoch": 1207.63,
291519
+ "learning_rate": 7.577205169628433e-06,
291520
+ "loss": 0.588,
291521
+ "step": 151030
291522
+ },
291523
+ {
291524
+ "epoch": 1207.67,
291525
+ "learning_rate": 7.577124394184169e-06,
291526
+ "loss": 0.277,
291527
+ "step": 151035
291528
+ },
291529
+ {
291530
+ "epoch": 1207.71,
291531
+ "learning_rate": 7.577043618739903e-06,
291532
+ "loss": 0.2769,
291533
+ "step": 151040
291534
+ },
291535
+ {
291536
+ "epoch": 1207.75,
291537
+ "learning_rate": 7.576962843295639e-06,
291538
+ "loss": 0.3973,
291539
+ "step": 151045
291540
+ },
291541
+ {
291542
+ "epoch": 1207.79,
291543
+ "learning_rate": 7.576882067851374e-06,
291544
+ "loss": 0.9731,
291545
+ "step": 151050
291546
+ },
291547
+ {
291548
+ "epoch": 1207.83,
291549
+ "learning_rate": 7.576801292407109e-06,
291550
+ "loss": 0.6487,
291551
+ "step": 151055
291552
+ },
291553
+ {
291554
+ "epoch": 1207.87,
291555
+ "learning_rate": 7.5767205169628445e-06,
291556
+ "loss": 0.2917,
291557
+ "step": 151060
291558
+ },
291559
+ {
291560
+ "epoch": 1207.91,
291561
+ "learning_rate": 7.576639741518579e-06,
291562
+ "loss": 0.2741,
291563
+ "step": 151065
291564
+ },
291565
+ {
291566
+ "epoch": 1207.95,
291567
+ "learning_rate": 7.5765589660743145e-06,
291568
+ "loss": 0.3645,
291569
+ "step": 151070
291570
+ },
291571
+ {
291572
+ "epoch": 1207.99,
291573
+ "learning_rate": 7.576478190630049e-06,
291574
+ "loss": 0.825,
291575
+ "step": 151075
291576
+ },
291577
+ {
291578
+ "epoch": 1208.0,
291579
+ "eval_loss": 0.3887929916381836,
291580
+ "eval_runtime": 39.8424,
291581
+ "eval_samples_per_second": 21.008,
291582
+ "eval_steps_per_second": 0.678,
291583
+ "eval_wer": 0.17792285757166443,
291584
+ "step": 151076
291585
+ },
291586
+ {
291587
+ "epoch": 1218.03,
291588
+ "learning_rate": 7.5763974151857844e-06,
291589
+ "loss": 0.347,
291590
+ "step": 151080
291591
+ },
291592
+ {
291593
+ "epoch": 1218.07,
291594
+ "learning_rate": 7.576316639741519e-06,
291595
+ "loss": 0.2568,
291596
+ "step": 151085
291597
+ },
291598
+ {
291599
+ "epoch": 1218.11,
291600
+ "learning_rate": 7.576235864297254e-06,
291601
+ "loss": 0.3279,
291602
+ "step": 151090
291603
+ },
291604
+ {
291605
+ "epoch": 1218.15,
291606
+ "learning_rate": 7.5761550888529886e-06,
291607
+ "loss": 0.3822,
291608
+ "step": 151095
291609
+ },
291610
+ {
291611
+ "epoch": 1218.19,
291612
+ "learning_rate": 7.576074313408724e-06,
291613
+ "loss": 0.8134,
291614
+ "step": 151100
291615
+ },
291616
+ {
291617
+ "epoch": 1218.23,
291618
+ "learning_rate": 7.5759935379644585e-06,
291619
+ "loss": 0.6241,
291620
+ "step": 151105
291621
+ },
291622
+ {
291623
+ "epoch": 1218.27,
291624
+ "learning_rate": 7.575912762520194e-06,
291625
+ "loss": 0.2843,
291626
+ "step": 151110
291627
+ },
291628
+ {
291629
+ "epoch": 1218.31,
291630
+ "learning_rate": 7.57583198707593e-06,
291631
+ "loss": 0.2978,
291632
+ "step": 151115
291633
+ },
291634
+ {
291635
+ "epoch": 1218.35,
291636
+ "learning_rate": 7.575751211631664e-06,
291637
+ "loss": 0.3903,
291638
+ "step": 151120
291639
+ },
291640
+ {
291641
+ "epoch": 1218.39,
291642
+ "learning_rate": 7.5756704361874e-06,
291643
+ "loss": 0.8314,
291644
+ "step": 151125
291645
+ },
291646
+ {
291647
+ "epoch": 1218.43,
291648
+ "learning_rate": 7.575589660743134e-06,
291649
+ "loss": 0.6191,
291650
+ "step": 151130
291651
+ },
291652
+ {
291653
+ "epoch": 1218.47,
291654
+ "learning_rate": 7.57550888529887e-06,
291655
+ "loss": 0.2565,
291656
+ "step": 151135
291657
+ },
291658
+ {
291659
+ "epoch": 1218.51,
291660
+ "learning_rate": 7.575428109854604e-06,
291661
+ "loss": 0.2411,
291662
+ "step": 151140
291663
+ },
291664
+ {
291665
+ "epoch": 1218.55,
291666
+ "learning_rate": 7.57534733441034e-06,
291667
+ "loss": 0.3642,
291668
+ "step": 151145
291669
+ },
291670
+ {
291671
+ "epoch": 1218.59,
291672
+ "learning_rate": 7.575266558966074e-06,
291673
+ "loss": 0.9774,
291674
+ "step": 151150
291675
+ },
291676
+ {
291677
+ "epoch": 1218.63,
291678
+ "learning_rate": 7.57518578352181e-06,
291679
+ "loss": 0.6026,
291680
+ "step": 151155
291681
+ },
291682
+ {
291683
+ "epoch": 1218.67,
291684
+ "learning_rate": 7.575105008077544e-06,
291685
+ "loss": 0.254,
291686
+ "step": 151160
291687
+ },
291688
+ {
291689
+ "epoch": 1218.71,
291690
+ "learning_rate": 7.57502423263328e-06,
291691
+ "loss": 0.3303,
291692
+ "step": 151165
291693
+ },
291694
+ {
291695
+ "epoch": 1218.76,
291696
+ "learning_rate": 7.574943457189016e-06,
291697
+ "loss": 0.3565,
291698
+ "step": 151170
291699
+ },
291700
+ {
291701
+ "epoch": 1218.8,
291702
+ "learning_rate": 7.57486268174475e-06,
291703
+ "loss": 0.7702,
291704
+ "step": 151175
291705
+ },
291706
+ {
291707
+ "epoch": 1218.84,
291708
+ "learning_rate": 7.574781906300486e-06,
291709
+ "loss": 0.5369,
291710
+ "step": 151180
291711
+ },
291712
+ {
291713
+ "epoch": 1218.88,
291714
+ "learning_rate": 7.57470113085622e-06,
291715
+ "loss": 0.2883,
291716
+ "step": 151185
291717
+ },
291718
+ {
291719
+ "epoch": 1218.92,
291720
+ "learning_rate": 7.574620355411956e-06,
291721
+ "loss": 0.2753,
291722
+ "step": 151190
291723
+ },
291724
+ {
291725
+ "epoch": 1218.96,
291726
+ "learning_rate": 7.57453957996769e-06,
291727
+ "loss": 0.4346,
291728
+ "step": 151195
291729
+ },
291730
+ {
291731
+ "epoch": 1219.0,
291732
+ "learning_rate": 7.574458804523426e-06,
291733
+ "loss": 0.9678,
291734
+ "step": 151200
291735
+ },
291736
+ {
291737
+ "epoch": 1219.0,
291738
+ "eval_loss": 0.40293607115745544,
291739
+ "eval_runtime": 40.725,
291740
+ "eval_samples_per_second": 20.528,
291741
+ "eval_steps_per_second": 0.663,
291742
+ "eval_wer": 0.17951517798645994,
291743
+ "step": 151200
291744
+ },
291745
+ {
291746
+ "epoch": 1219.04,
291747
+ "learning_rate": 7.57437802907916e-06,
291748
+ "loss": 0.3041,
291749
+ "step": 151205
291750
+ },
291751
+ {
291752
+ "epoch": 1219.08,
291753
+ "learning_rate": 7.574297253634896e-06,
291754
+ "loss": 0.2819,
291755
+ "step": 151210
291756
+ },
291757
+ {
291758
+ "epoch": 1219.12,
291759
+ "learning_rate": 7.57421647819063e-06,
291760
+ "loss": 0.2779,
291761
+ "step": 151215
291762
+ },
291763
+ {
291764
+ "epoch": 1219.16,
291765
+ "learning_rate": 7.574135702746366e-06,
291766
+ "loss": 0.4908,
291767
+ "step": 151220
291768
+ },
291769
+ {
291770
+ "epoch": 1219.2,
291771
+ "learning_rate": 7.5740549273021e-06,
291772
+ "loss": 1.049,
291773
+ "step": 151225
291774
+ },
291775
+ {
291776
+ "epoch": 1219.24,
291777
+ "learning_rate": 7.573974151857836e-06,
291778
+ "loss": 0.3181,
291779
+ "step": 151230
291780
+ },
291781
+ {
291782
+ "epoch": 1219.28,
291783
+ "learning_rate": 7.573893376413572e-06,
291784
+ "loss": 0.2837,
291785
+ "step": 151235
291786
+ },
291787
+ {
291788
+ "epoch": 1219.32,
291789
+ "learning_rate": 7.573812600969306e-06,
291790
+ "loss": 0.3345,
291791
+ "step": 151240
291792
+ },
291793
+ {
291794
+ "epoch": 1219.36,
291795
+ "learning_rate": 7.5737318255250416e-06,
291796
+ "loss": 0.4296,
291797
+ "step": 151245
291798
+ },
291799
+ {
291800
+ "epoch": 1219.4,
291801
+ "learning_rate": 7.573651050080776e-06,
291802
+ "loss": 1.2842,
291803
+ "step": 151250
291804
+ },
291805
+ {
291806
+ "epoch": 1219.44,
291807
+ "learning_rate": 7.5735702746365115e-06,
291808
+ "loss": 0.3158,
291809
+ "step": 151255
291810
+ },
291811
+ {
291812
+ "epoch": 1219.48,
291813
+ "learning_rate": 7.573489499192246e-06,
291814
+ "loss": 0.2343,
291815
+ "step": 151260
291816
+ },
291817
+ {
291818
+ "epoch": 1219.52,
291819
+ "learning_rate": 7.5734087237479815e-06,
291820
+ "loss": 0.2876,
291821
+ "step": 151265
291822
+ },
291823
+ {
291824
+ "epoch": 1219.56,
291825
+ "learning_rate": 7.573327948303716e-06,
291826
+ "loss": 0.4508,
291827
+ "step": 151270
291828
+ },
291829
+ {
291830
+ "epoch": 1219.6,
291831
+ "learning_rate": 7.5732471728594515e-06,
291832
+ "loss": 1.1773,
291833
+ "step": 151275
291834
+ },
291835
+ {
291836
+ "epoch": 1219.64,
291837
+ "learning_rate": 7.573166397415186e-06,
291838
+ "loss": 0.27,
291839
+ "step": 151280
291840
+ },
291841
+ {
291842
+ "epoch": 1219.68,
291843
+ "learning_rate": 7.5730856219709215e-06,
291844
+ "loss": 0.2492,
291845
+ "step": 151285
291846
+ },
291847
+ {
291848
+ "epoch": 1219.72,
291849
+ "learning_rate": 7.573004846526657e-06,
291850
+ "loss": 0.3389,
291851
+ "step": 151290
291852
+ },
291853
+ {
291854
+ "epoch": 1219.76,
291855
+ "learning_rate": 7.5729240710823914e-06,
291856
+ "loss": 0.39,
291857
+ "step": 151295
291858
+ },
291859
+ {
291860
+ "epoch": 1219.8,
291861
+ "learning_rate": 7.572843295638127e-06,
291862
+ "loss": 1.2323,
291863
+ "step": 151300
291864
+ },
291865
+ {
291866
+ "epoch": 1219.84,
291867
+ "learning_rate": 7.572762520193861e-06,
291868
+ "loss": 0.2676,
291869
+ "step": 151305
291870
+ },
291871
+ {
291872
+ "epoch": 1219.88,
291873
+ "learning_rate": 7.572681744749597e-06,
291874
+ "loss": 0.2777,
291875
+ "step": 151310
291876
+ },
291877
+ {
291878
+ "epoch": 1219.92,
291879
+ "learning_rate": 7.572600969305331e-06,
291880
+ "loss": 0.3009,
291881
+ "step": 151315
291882
+ },
291883
+ {
291884
+ "epoch": 1219.96,
291885
+ "learning_rate": 7.572520193861067e-06,
291886
+ "loss": 0.4905,
291887
+ "step": 151320
291888
+ },
291889
+ {
291890
+ "epoch": 1220.0,
291891
+ "eval_loss": 0.40317270159721375,
291892
+ "eval_runtime": 40.9503,
291893
+ "eval_samples_per_second": 20.415,
291894
+ "eval_steps_per_second": 0.659,
291895
+ "eval_wer": 0.18318384015168088,
291896
+ "step": 151324
291897
+ },
291898
+ {
291899
+ "epoch": 1220.01,
291900
+ "learning_rate": 7.572439418416801e-06,
291901
+ "loss": 0.2942,
291902
+ "step": 151325
291903
+ },
291904
+ {
291905
+ "epoch": 1220.05,
291906
+ "learning_rate": 7.572358642972537e-06,
291907
+ "loss": 0.2593,
291908
+ "step": 151330
291909
+ },
291910
+ {
291911
+ "epoch": 1220.09,
291912
+ "learning_rate": 7.572277867528271e-06,
291913
+ "loss": 0.2962,
291914
+ "step": 151335
291915
+ },
291916
+ {
291917
+ "epoch": 1220.13,
291918
+ "learning_rate": 7.572197092084007e-06,
291919
+ "loss": 0.3793,
291920
+ "step": 151340
291921
+ },
291922
+ {
291923
+ "epoch": 1220.17,
291924
+ "learning_rate": 7.572116316639743e-06,
291925
+ "loss": 0.473,
291926
+ "step": 151345
291927
+ },
291928
+ {
291929
+ "epoch": 1220.21,
291930
+ "learning_rate": 7.572035541195477e-06,
291931
+ "loss": 1.116,
291932
+ "step": 151350
291933
+ },
291934
+ {
291935
+ "epoch": 1220.25,
291936
+ "learning_rate": 7.571954765751213e-06,
291937
+ "loss": 0.2653,
291938
+ "step": 151355
291939
+ },
291940
+ {
291941
+ "epoch": 1220.29,
291942
+ "learning_rate": 7.571873990306947e-06,
291943
+ "loss": 0.2706,
291944
+ "step": 151360
291945
+ },
291946
+ {
291947
+ "epoch": 1220.33,
291948
+ "learning_rate": 7.571793214862683e-06,
291949
+ "loss": 0.2761,
291950
+ "step": 151365
291951
+ },
291952
+ {
291953
+ "epoch": 1220.37,
291954
+ "learning_rate": 7.571712439418417e-06,
291955
+ "loss": 0.5144,
291956
+ "step": 151370
291957
+ },
291958
+ {
291959
+ "epoch": 1220.41,
291960
+ "learning_rate": 7.571631663974153e-06,
291961
+ "loss": 1.0997,
291962
+ "step": 151375
291963
+ },
291964
+ {
291965
+ "epoch": 1220.45,
291966
+ "learning_rate": 7.571550888529887e-06,
291967
+ "loss": 0.2801,
291968
+ "step": 151380
291969
+ },
291970
+ {
291971
+ "epoch": 1220.49,
291972
+ "learning_rate": 7.571470113085623e-06,
291973
+ "loss": 0.3037,
291974
+ "step": 151385
291975
+ },
291976
+ {
291977
+ "epoch": 1220.53,
291978
+ "learning_rate": 7.571389337641357e-06,
291979
+ "loss": 0.3857,
291980
+ "step": 151390
291981
+ },
291982
+ {
291983
+ "epoch": 1220.57,
291984
+ "learning_rate": 7.571308562197093e-06,
291985
+ "loss": 0.554,
291986
+ "step": 151395
291987
+ },
291988
+ {
291989
+ "epoch": 1220.61,
291990
+ "learning_rate": 7.571227786752827e-06,
291991
+ "loss": 1.0511,
291992
+ "step": 151400
291993
+ },
291994
+ {
291995
+ "epoch": 1220.65,
291996
+ "learning_rate": 7.571147011308563e-06,
291997
+ "loss": 0.2562,
291998
+ "step": 151405
291999
+ },
292000
+ {
292001
+ "epoch": 1220.69,
292002
+ "learning_rate": 7.571066235864299e-06,
292003
+ "loss": 0.2393,
292004
+ "step": 151410
292005
+ },
292006
+ {
292007
+ "epoch": 1220.73,
292008
+ "learning_rate": 7.570985460420033e-06,
292009
+ "loss": 0.3112,
292010
+ "step": 151415
292011
+ },
292012
+ {
292013
+ "epoch": 1220.77,
292014
+ "learning_rate": 7.570904684975769e-06,
292015
+ "loss": 0.5436,
292016
+ "step": 151420
292017
+ },
292018
+ {
292019
+ "epoch": 1220.81,
292020
+ "learning_rate": 7.570823909531503e-06,
292021
+ "loss": 0.9745,
292022
+ "step": 151425
292023
+ },
292024
+ {
292025
+ "epoch": 1220.85,
292026
+ "learning_rate": 7.570743134087239e-06,
292027
+ "loss": 0.2797,
292028
+ "step": 151430
292029
+ },
292030
+ {
292031
+ "epoch": 1220.89,
292032
+ "learning_rate": 7.570662358642973e-06,
292033
+ "loss": 0.2683,
292034
+ "step": 151435
292035
+ },
292036
+ {
292037
+ "epoch": 1220.93,
292038
+ "learning_rate": 7.570581583198709e-06,
292039
+ "loss": 0.3428,
292040
+ "step": 151440
292041
+ },
292042
+ {
292043
+ "epoch": 1220.97,
292044
+ "learning_rate": 7.570500807754443e-06,
292045
+ "loss": 0.4689,
292046
+ "step": 151445
292047
+ },
292048
+ {
292049
+ "epoch": 1221.0,
292050
+ "eval_loss": 0.3577888607978821,
292051
+ "eval_runtime": 42.2225,
292052
+ "eval_samples_per_second": 19.8,
292053
+ "eval_steps_per_second": 0.639,
292054
+ "eval_wer": 0.17625899280575538,
292055
+ "step": 151448
292056
  }
292057
  ],
292058
+ "max_steps": 620000,
292059
  "num_train_epochs": 5000,
292060
+ "total_flos": 4.2618930913164073e+20,
292061
  "trial_name": null,
292062
  "trial_params": null
292063
  }
model-bin/finetune/base/{checkpoint-150827 β†’ checkpoint-151448}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630246352.8300264/events.out.tfevents.1630246352.cc93b136ebf5.1086.189 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d2e006a76411ca66fef0f0601b536f525ad6d1bf56071220648ca0968c5ce39
3
+ size 4194
model-bin/finetune/base/log/1630246785.4372547/events.out.tfevents.1630246785.cc93b136ebf5.1086.191 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88c677975dcb5799d881c0a06fd4f929de0324135c035871197a803122b7e182
3
+ size 4194
model-bin/finetune/base/log/1630247219.004471/events.out.tfevents.1630247219.cc93b136ebf5.1086.193 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba94dba6d83ae89638f94a50b5ed007dd6fa5fff4a9ef3c1d84e599ce33866a2
3
+ size 4194
model-bin/finetune/base/log/1630247655.9469938/events.out.tfevents.1630247655.cc93b136ebf5.1086.195 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b8bb97898f6233ed672cf04852addd9a7e44ee8bfe0ec7376dc5ec8454e6735
3
+ size 4194
model-bin/finetune/base/log/1630248088.0553622/events.out.tfevents.1630248088.cc93b136ebf5.1086.197 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b42f4a6bdb41edd1853e3d18e435253edac9dffde47c25b93a644f5ef02d2e20
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630246352.cc93b136ebf5.1086.188 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d09b4ca59b66bc0c0684dee2fe769e86548e33094e265be58139e3f0a4d37830
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630246785.cc93b136ebf5.1086.190 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0056a189953ac9b863ab9640ac16ddada116751dc2c5670f09d12244155c7710
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630247218.cc93b136ebf5.1086.192 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b7cc40664c5864ee6af5d3191ef612920626590925ef149aacf562c7b185192
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630247655.cc93b136ebf5.1086.194 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5e892ecadb74ebdc66ca0fcae41c145b395b336ec9b9b70d9c3394abfcfe60f
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630248088.cc93b136ebf5.1086.196 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:222a21717f9e387720acc83eac9a3017b439135e5baedab94209e89526ba9fa0
3
+ size 8622