Check commited on
Commit
23775cb
Β·
1 Parent(s): c300449

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630045620.5655968/events.out.tfevents.1630045620.52f5c7e305a3.886.71 +3 -0
  11. model-bin/finetune/base/log/1630046118.6803894/events.out.tfevents.1630046118.52f5c7e305a3.886.73 +3 -0
  12. model-bin/finetune/base/log/1630046681.8092928/events.out.tfevents.1630046681.52f5c7e305a3.886.75 +3 -0
  13. model-bin/finetune/base/log/1630047172.2078328/events.out.tfevents.1630047172.52f5c7e305a3.886.77 +3 -0
  14. model-bin/finetune/base/log/1630047639.1210005/events.out.tfevents.1630047639.52f5c7e305a3.886.79 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630045620.52f5c7e305a3.886.70 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630046118.52f5c7e305a3.886.72 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630046681.52f5c7e305a3.886.74 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630047172.52f5c7e305a3.886.76 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630047639.52f5c7e305a3.886.78 +3 -0
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ad06371735a046dd7c9ef9132035682f9c5de2a72ee91eadd07f7a4003b5e3e
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1b8197993eb37d9ee36acbb12a1f9605ec9f468a36950f0b19cfe128de2b5f0
3
  size 722165393
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a8eb7a6778d47cb003b93cbbdd60345b9fb33fdffcedb8fb466ca7f8f337a22
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8232bc1150a76590269afa478f75b4236d9220bb172b83e409418781d8efeebd
3
  size 377909911
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1bab8a8b45fe33a8814faa874970aaa38052e7827b5642f8165984321edc61a
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a14f61f7dddaa53dbcb08f44b28b663a43e59c31dc4d069de7af3154f39d5215
3
  size 14503
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01a18ea4d20203ec9fd1ae2b3ac1f3ea23edcb584eb282e1b66ac38040ce9b63
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f60fd12782841de6f909cf814170025073e913f168ae702ff0cdd93272700668
3
  size 559
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6819449b28fc88148155c2c7a3ee8e6199cb0db9b034e04f8a104713b88847e3
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e9a559e52c3b4b08e9232cdf54f10d4fa2d1e564b84998252aef70d42eaf820
3
  size 623
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 892.995983935743,
5
- "global_step": 110758,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -240282,11 +240282,800 @@
240282
  "eval_steps_per_second": 0.631,
240283
  "eval_wer": 0.19120699881376038,
240284
  "step": 110758
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
240285
  }
240286
  ],
240287
  "max_steps": 620000,
240288
  "num_train_epochs": 5000,
240289
- "total_flos": 3.116997433357764e+20,
240290
  "trial_name": null,
240291
  "trial_params": null
240292
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 898.0,
5
+ "global_step": 111379,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
240282
  "eval_steps_per_second": 0.631,
240283
  "eval_wer": 0.19120699881376038,
240284
  "step": 110758
240285
+ },
240286
+ {
240287
+ "epoch": 900.02,
240288
+ "learning_rate": 8.227528271405494e-06,
240289
+ "loss": 0.3431,
240290
+ "step": 110760
240291
+ },
240292
+ {
240293
+ "epoch": 900.06,
240294
+ "learning_rate": 8.227447495961228e-06,
240295
+ "loss": 0.2264,
240296
+ "step": 110765
240297
+ },
240298
+ {
240299
+ "epoch": 900.1,
240300
+ "learning_rate": 8.227366720516964e-06,
240301
+ "loss": 0.3079,
240302
+ "step": 110770
240303
+ },
240304
+ {
240305
+ "epoch": 900.14,
240306
+ "learning_rate": 8.227285945072698e-06,
240307
+ "loss": 0.4818,
240308
+ "step": 110775
240309
+ },
240310
+ {
240311
+ "epoch": 900.18,
240312
+ "learning_rate": 8.227205169628434e-06,
240313
+ "loss": 0.7211,
240314
+ "step": 110780
240315
+ },
240316
+ {
240317
+ "epoch": 900.22,
240318
+ "learning_rate": 8.227124394184168e-06,
240319
+ "loss": 1.0533,
240320
+ "step": 110785
240321
+ },
240322
+ {
240323
+ "epoch": 900.26,
240324
+ "learning_rate": 8.227043618739904e-06,
240325
+ "loss": 0.3337,
240326
+ "step": 110790
240327
+ },
240328
+ {
240329
+ "epoch": 900.3,
240330
+ "learning_rate": 8.226962843295638e-06,
240331
+ "loss": 0.3322,
240332
+ "step": 110795
240333
+ },
240334
+ {
240335
+ "epoch": 900.34,
240336
+ "learning_rate": 8.226882067851374e-06,
240337
+ "loss": 0.3738,
240338
+ "step": 110800
240339
+ },
240340
+ {
240341
+ "epoch": 900.38,
240342
+ "learning_rate": 8.22680129240711e-06,
240343
+ "loss": 0.6992,
240344
+ "step": 110805
240345
+ },
240346
+ {
240347
+ "epoch": 900.42,
240348
+ "learning_rate": 8.226720516962844e-06,
240349
+ "loss": 0.9494,
240350
+ "step": 110810
240351
+ },
240352
+ {
240353
+ "epoch": 900.46,
240354
+ "learning_rate": 8.22663974151858e-06,
240355
+ "loss": 0.2978,
240356
+ "step": 110815
240357
+ },
240358
+ {
240359
+ "epoch": 900.5,
240360
+ "learning_rate": 8.226558966074314e-06,
240361
+ "loss": 0.3104,
240362
+ "step": 110820
240363
+ },
240364
+ {
240365
+ "epoch": 900.54,
240366
+ "learning_rate": 8.22647819063005e-06,
240367
+ "loss": 0.3612,
240368
+ "step": 110825
240369
+ },
240370
+ {
240371
+ "epoch": 900.58,
240372
+ "learning_rate": 8.226397415185784e-06,
240373
+ "loss": 0.6377,
240374
+ "step": 110830
240375
+ },
240376
+ {
240377
+ "epoch": 900.62,
240378
+ "learning_rate": 8.22631663974152e-06,
240379
+ "loss": 0.8597,
240380
+ "step": 110835
240381
+ },
240382
+ {
240383
+ "epoch": 900.66,
240384
+ "learning_rate": 8.226235864297254e-06,
240385
+ "loss": 0.3028,
240386
+ "step": 110840
240387
+ },
240388
+ {
240389
+ "epoch": 900.7,
240390
+ "learning_rate": 8.22615508885299e-06,
240391
+ "loss": 0.3288,
240392
+ "step": 110845
240393
+ },
240394
+ {
240395
+ "epoch": 900.74,
240396
+ "learning_rate": 8.226074313408724e-06,
240397
+ "loss": 0.3319,
240398
+ "step": 110850
240399
+ },
240400
+ {
240401
+ "epoch": 900.79,
240402
+ "learning_rate": 8.22599353796446e-06,
240403
+ "loss": 0.5342,
240404
+ "step": 110855
240405
+ },
240406
+ {
240407
+ "epoch": 900.83,
240408
+ "learning_rate": 8.225912762520194e-06,
240409
+ "loss": 0.994,
240410
+ "step": 110860
240411
+ },
240412
+ {
240413
+ "epoch": 900.87,
240414
+ "learning_rate": 8.22583198707593e-06,
240415
+ "loss": 0.3007,
240416
+ "step": 110865
240417
+ },
240418
+ {
240419
+ "epoch": 900.91,
240420
+ "learning_rate": 8.225751211631666e-06,
240421
+ "loss": 0.3091,
240422
+ "step": 110870
240423
+ },
240424
+ {
240425
+ "epoch": 900.95,
240426
+ "learning_rate": 8.2256704361874e-06,
240427
+ "loss": 0.3254,
240428
+ "step": 110875
240429
+ },
240430
+ {
240431
+ "epoch": 900.99,
240432
+ "learning_rate": 8.225589660743136e-06,
240433
+ "loss": 0.7219,
240434
+ "step": 110880
240435
+ },
240436
+ {
240437
+ "epoch": 901.0,
240438
+ "eval_loss": 0.36234134435653687,
240439
+ "eval_runtime": 43.3119,
240440
+ "eval_samples_per_second": 19.325,
240441
+ "eval_steps_per_second": 0.623,
240442
+ "eval_wer": 0.18013960819635216,
240443
+ "step": 110881
240444
+ },
240445
+ {
240446
+ "epoch": 894.03,
240447
+ "learning_rate": 8.22550888529887e-06,
240448
+ "loss": 0.4134,
240449
+ "step": 110885
240450
+ },
240451
+ {
240452
+ "epoch": 894.07,
240453
+ "learning_rate": 8.225428109854606e-06,
240454
+ "loss": 0.2736,
240455
+ "step": 110890
240456
+ },
240457
+ {
240458
+ "epoch": 894.11,
240459
+ "learning_rate": 8.22534733441034e-06,
240460
+ "loss": 0.3003,
240461
+ "step": 110895
240462
+ },
240463
+ {
240464
+ "epoch": 894.15,
240465
+ "learning_rate": 8.225266558966076e-06,
240466
+ "loss": 0.457,
240467
+ "step": 110900
240468
+ },
240469
+ {
240470
+ "epoch": 894.19,
240471
+ "learning_rate": 8.22518578352181e-06,
240472
+ "loss": 0.8826,
240473
+ "step": 110905
240474
+ },
240475
+ {
240476
+ "epoch": 894.23,
240477
+ "learning_rate": 8.225105008077546e-06,
240478
+ "loss": 0.6164,
240479
+ "step": 110910
240480
+ },
240481
+ {
240482
+ "epoch": 894.27,
240483
+ "learning_rate": 8.22502423263328e-06,
240484
+ "loss": 0.2989,
240485
+ "step": 110915
240486
+ },
240487
+ {
240488
+ "epoch": 894.31,
240489
+ "learning_rate": 8.224943457189016e-06,
240490
+ "loss": 0.3474,
240491
+ "step": 110920
240492
+ },
240493
+ {
240494
+ "epoch": 894.35,
240495
+ "learning_rate": 8.224862681744751e-06,
240496
+ "loss": 0.3896,
240497
+ "step": 110925
240498
+ },
240499
+ {
240500
+ "epoch": 894.4,
240501
+ "learning_rate": 8.224781906300486e-06,
240502
+ "loss": 0.8943,
240503
+ "step": 110930
240504
+ },
240505
+ {
240506
+ "epoch": 894.44,
240507
+ "learning_rate": 8.224701130856221e-06,
240508
+ "loss": 0.6775,
240509
+ "step": 110935
240510
+ },
240511
+ {
240512
+ "epoch": 894.48,
240513
+ "learning_rate": 8.224620355411956e-06,
240514
+ "loss": 0.3509,
240515
+ "step": 110940
240516
+ },
240517
+ {
240518
+ "epoch": 894.52,
240519
+ "learning_rate": 8.224539579967691e-06,
240520
+ "loss": 0.2667,
240521
+ "step": 110945
240522
+ },
240523
+ {
240524
+ "epoch": 894.56,
240525
+ "learning_rate": 8.224458804523426e-06,
240526
+ "loss": 0.3828,
240527
+ "step": 110950
240528
+ },
240529
+ {
240530
+ "epoch": 894.6,
240531
+ "learning_rate": 8.224378029079161e-06,
240532
+ "loss": 0.8308,
240533
+ "step": 110955
240534
+ },
240535
+ {
240536
+ "epoch": 894.64,
240537
+ "learning_rate": 8.224297253634896e-06,
240538
+ "loss": 0.6964,
240539
+ "step": 110960
240540
+ },
240541
+ {
240542
+ "epoch": 894.68,
240543
+ "learning_rate": 8.224216478190631e-06,
240544
+ "loss": 0.3014,
240545
+ "step": 110965
240546
+ },
240547
+ {
240548
+ "epoch": 894.72,
240549
+ "learning_rate": 8.224135702746365e-06,
240550
+ "loss": 0.3124,
240551
+ "step": 110970
240552
+ },
240553
+ {
240554
+ "epoch": 894.76,
240555
+ "learning_rate": 8.224054927302101e-06,
240556
+ "loss": 0.3854,
240557
+ "step": 110975
240558
+ },
240559
+ {
240560
+ "epoch": 894.8,
240561
+ "learning_rate": 8.223974151857837e-06,
240562
+ "loss": 0.9649,
240563
+ "step": 110980
240564
+ },
240565
+ {
240566
+ "epoch": 894.84,
240567
+ "learning_rate": 8.223893376413571e-06,
240568
+ "loss": 0.7675,
240569
+ "step": 110985
240570
+ },
240571
+ {
240572
+ "epoch": 894.88,
240573
+ "learning_rate": 8.223812600969307e-06,
240574
+ "loss": 0.3051,
240575
+ "step": 110990
240576
+ },
240577
+ {
240578
+ "epoch": 894.92,
240579
+ "learning_rate": 8.223731825525041e-06,
240580
+ "loss": 0.2799,
240581
+ "step": 110995
240582
+ },
240583
+ {
240584
+ "epoch": 894.96,
240585
+ "learning_rate": 8.223651050080777e-06,
240586
+ "loss": 0.411,
240587
+ "step": 111000
240588
+ },
240589
+ {
240590
+ "epoch": 895.0,
240591
+ "learning_rate": 8.223570274636511e-06,
240592
+ "loss": 1.2268,
240593
+ "step": 111005
240594
+ },
240595
+ {
240596
+ "epoch": 895.0,
240597
+ "eval_loss": 0.4037691652774811,
240598
+ "eval_runtime": 44.1709,
240599
+ "eval_samples_per_second": 18.949,
240600
+ "eval_steps_per_second": 0.611,
240601
+ "eval_wer": 0.1897884018198888,
240602
+ "step": 111005
240603
+ },
240604
+ {
240605
+ "epoch": 888.04,
240606
+ "learning_rate": 8.223489499192247e-06,
240607
+ "loss": 0.2999,
240608
+ "step": 111010
240609
+ },
240610
+ {
240611
+ "epoch": 888.08,
240612
+ "learning_rate": 8.223408723747981e-06,
240613
+ "loss": 0.3209,
240614
+ "step": 111015
240615
+ },
240616
+ {
240617
+ "epoch": 888.12,
240618
+ "learning_rate": 8.223327948303717e-06,
240619
+ "loss": 0.398,
240620
+ "step": 111020
240621
+ },
240622
+ {
240623
+ "epoch": 888.16,
240624
+ "learning_rate": 8.223247172859451e-06,
240625
+ "loss": 0.4157,
240626
+ "step": 111025
240627
+ },
240628
+ {
240629
+ "epoch": 888.2,
240630
+ "learning_rate": 8.223166397415187e-06,
240631
+ "loss": 1.1022,
240632
+ "step": 111030
240633
+ },
240634
+ {
240635
+ "epoch": 888.24,
240636
+ "learning_rate": 8.223085621970921e-06,
240637
+ "loss": 0.4123,
240638
+ "step": 111035
240639
+ },
240640
+ {
240641
+ "epoch": 888.28,
240642
+ "learning_rate": 8.223004846526657e-06,
240643
+ "loss": 0.2387,
240644
+ "step": 111040
240645
+ },
240646
+ {
240647
+ "epoch": 888.32,
240648
+ "learning_rate": 8.222924071082393e-06,
240649
+ "loss": 0.3189,
240650
+ "step": 111045
240651
+ },
240652
+ {
240653
+ "epoch": 888.36,
240654
+ "learning_rate": 8.222843295638127e-06,
240655
+ "loss": 0.5308,
240656
+ "step": 111050
240657
+ },
240658
+ {
240659
+ "epoch": 888.4,
240660
+ "learning_rate": 8.222762520193863e-06,
240661
+ "loss": 1.1537,
240662
+ "step": 111055
240663
+ },
240664
+ {
240665
+ "epoch": 888.44,
240666
+ "learning_rate": 8.222681744749597e-06,
240667
+ "loss": 0.3102,
240668
+ "step": 111060
240669
+ },
240670
+ {
240671
+ "epoch": 888.48,
240672
+ "learning_rate": 8.222600969305333e-06,
240673
+ "loss": 0.2456,
240674
+ "step": 111065
240675
+ },
240676
+ {
240677
+ "epoch": 888.52,
240678
+ "learning_rate": 8.222520193861067e-06,
240679
+ "loss": 0.2704,
240680
+ "step": 111070
240681
+ },
240682
+ {
240683
+ "epoch": 888.56,
240684
+ "learning_rate": 8.222439418416803e-06,
240685
+ "loss": 0.3987,
240686
+ "step": 111075
240687
+ },
240688
+ {
240689
+ "epoch": 888.6,
240690
+ "learning_rate": 8.222358642972537e-06,
240691
+ "loss": 1.1706,
240692
+ "step": 111080
240693
+ },
240694
+ {
240695
+ "epoch": 888.64,
240696
+ "learning_rate": 8.222277867528273e-06,
240697
+ "loss": 0.3518,
240698
+ "step": 111085
240699
+ },
240700
+ {
240701
+ "epoch": 888.68,
240702
+ "learning_rate": 8.222197092084007e-06,
240703
+ "loss": 0.2963,
240704
+ "step": 111090
240705
+ },
240706
+ {
240707
+ "epoch": 888.72,
240708
+ "learning_rate": 8.222116316639743e-06,
240709
+ "loss": 0.3069,
240710
+ "step": 111095
240711
+ },
240712
+ {
240713
+ "epoch": 888.76,
240714
+ "learning_rate": 8.222035541195479e-06,
240715
+ "loss": 0.5053,
240716
+ "step": 111100
240717
+ },
240718
+ {
240719
+ "epoch": 888.8,
240720
+ "learning_rate": 8.221954765751213e-06,
240721
+ "loss": 1.1867,
240722
+ "step": 111105
240723
+ },
240724
+ {
240725
+ "epoch": 888.84,
240726
+ "learning_rate": 8.221873990306949e-06,
240727
+ "loss": 0.3344,
240728
+ "step": 111110
240729
+ },
240730
+ {
240731
+ "epoch": 888.88,
240732
+ "learning_rate": 8.221793214862683e-06,
240733
+ "loss": 0.2685,
240734
+ "step": 111115
240735
+ },
240736
+ {
240737
+ "epoch": 888.92,
240738
+ "learning_rate": 8.221712439418419e-06,
240739
+ "loss": 0.346,
240740
+ "step": 111120
240741
+ },
240742
+ {
240743
+ "epoch": 888.96,
240744
+ "learning_rate": 8.221631663974153e-06,
240745
+ "loss": 0.413,
240746
+ "step": 111125
240747
+ },
240748
+ {
240749
+ "epoch": 889.0,
240750
+ "learning_rate": 8.221550888529888e-06,
240751
+ "loss": 1.43,
240752
+ "step": 111130
240753
+ },
240754
+ {
240755
+ "epoch": 889.0,
240756
+ "eval_loss": 0.4563208818435669,
240757
+ "eval_runtime": 45.6982,
240758
+ "eval_samples_per_second": 18.316,
240759
+ "eval_steps_per_second": 0.591,
240760
+ "eval_wer": 0.17906810035842294,
240761
+ "step": 111130
240762
+ },
240763
+ {
240764
+ "epoch": 889.04,
240765
+ "learning_rate": 8.221470113085623e-06,
240766
+ "loss": 0.3338,
240767
+ "step": 111135
240768
+ },
240769
+ {
240770
+ "epoch": 889.08,
240771
+ "learning_rate": 8.221389337641358e-06,
240772
+ "loss": 0.2645,
240773
+ "step": 111140
240774
+ },
240775
+ {
240776
+ "epoch": 889.12,
240777
+ "learning_rate": 8.221308562197093e-06,
240778
+ "loss": 0.3281,
240779
+ "step": 111145
240780
+ },
240781
+ {
240782
+ "epoch": 889.16,
240783
+ "learning_rate": 8.221227786752828e-06,
240784
+ "loss": 0.4546,
240785
+ "step": 111150
240786
+ },
240787
+ {
240788
+ "epoch": 889.2,
240789
+ "learning_rate": 8.221147011308563e-06,
240790
+ "loss": 1.1137,
240791
+ "step": 111155
240792
+ },
240793
+ {
240794
+ "epoch": 889.24,
240795
+ "learning_rate": 8.221066235864298e-06,
240796
+ "loss": 0.3458,
240797
+ "step": 111160
240798
+ },
240799
+ {
240800
+ "epoch": 889.28,
240801
+ "learning_rate": 8.220985460420034e-06,
240802
+ "loss": 0.2908,
240803
+ "step": 111165
240804
+ },
240805
+ {
240806
+ "epoch": 889.32,
240807
+ "learning_rate": 8.220904684975768e-06,
240808
+ "loss": 0.3467,
240809
+ "step": 111170
240810
+ },
240811
+ {
240812
+ "epoch": 889.36,
240813
+ "learning_rate": 8.220823909531504e-06,
240814
+ "loss": 0.4733,
240815
+ "step": 111175
240816
+ },
240817
+ {
240818
+ "epoch": 889.4,
240819
+ "learning_rate": 8.220743134087238e-06,
240820
+ "loss": 1.119,
240821
+ "step": 111180
240822
+ },
240823
+ {
240824
+ "epoch": 889.44,
240825
+ "learning_rate": 8.220662358642974e-06,
240826
+ "loss": 0.3293,
240827
+ "step": 111185
240828
+ },
240829
+ {
240830
+ "epoch": 889.48,
240831
+ "learning_rate": 8.220581583198708e-06,
240832
+ "loss": 0.2844,
240833
+ "step": 111190
240834
+ },
240835
+ {
240836
+ "epoch": 889.52,
240837
+ "learning_rate": 8.220500807754444e-06,
240838
+ "loss": 0.294,
240839
+ "step": 111195
240840
+ },
240841
+ {
240842
+ "epoch": 889.56,
240843
+ "learning_rate": 8.220420032310178e-06,
240844
+ "loss": 0.4952,
240845
+ "step": 111200
240846
+ },
240847
+ {
240848
+ "epoch": 889.6,
240849
+ "learning_rate": 8.220339256865914e-06,
240850
+ "loss": 1.2199,
240851
+ "step": 111205
240852
+ },
240853
+ {
240854
+ "epoch": 889.64,
240855
+ "learning_rate": 8.220258481421648e-06,
240856
+ "loss": 0.329,
240857
+ "step": 111210
240858
+ },
240859
+ {
240860
+ "epoch": 889.68,
240861
+ "learning_rate": 8.220177705977384e-06,
240862
+ "loss": 0.3095,
240863
+ "step": 111215
240864
+ },
240865
+ {
240866
+ "epoch": 889.72,
240867
+ "learning_rate": 8.220096930533118e-06,
240868
+ "loss": 0.3582,
240869
+ "step": 111220
240870
+ },
240871
+ {
240872
+ "epoch": 889.76,
240873
+ "learning_rate": 8.220016155088854e-06,
240874
+ "loss": 0.4658,
240875
+ "step": 111225
240876
+ },
240877
+ {
240878
+ "epoch": 889.8,
240879
+ "learning_rate": 8.21993537964459e-06,
240880
+ "loss": 1.2943,
240881
+ "step": 111230
240882
+ },
240883
+ {
240884
+ "epoch": 889.84,
240885
+ "learning_rate": 8.219854604200324e-06,
240886
+ "loss": 0.3334,
240887
+ "step": 111235
240888
+ },
240889
+ {
240890
+ "epoch": 889.88,
240891
+ "learning_rate": 8.21977382875606e-06,
240892
+ "loss": 0.2945,
240893
+ "step": 111240
240894
+ },
240895
+ {
240896
+ "epoch": 889.92,
240897
+ "learning_rate": 8.219693053311794e-06,
240898
+ "loss": 0.2881,
240899
+ "step": 111245
240900
+ },
240901
+ {
240902
+ "epoch": 889.96,
240903
+ "learning_rate": 8.21961227786753e-06,
240904
+ "loss": 0.4068,
240905
+ "step": 111250
240906
+ },
240907
+ {
240908
+ "epoch": 890.0,
240909
+ "learning_rate": 8.219531502423264e-06,
240910
+ "loss": 1.2953,
240911
+ "step": 111255
240912
+ },
240913
+ {
240914
+ "epoch": 890.0,
240915
+ "eval_loss": 0.39579546451568604,
240916
+ "eval_runtime": 41.6212,
240917
+ "eval_samples_per_second": 20.11,
240918
+ "eval_steps_per_second": 0.649,
240919
+ "eval_wer": 0.18293677266325697,
240920
+ "step": 111255
240921
+ },
240922
+ {
240923
+ "epoch": 897.04,
240924
+ "learning_rate": 8.219450726979e-06,
240925
+ "loss": 0.3568,
240926
+ "step": 111260
240927
+ },
240928
+ {
240929
+ "epoch": 897.08,
240930
+ "learning_rate": 8.219369951534734e-06,
240931
+ "loss": 0.2539,
240932
+ "step": 111265
240933
+ },
240934
+ {
240935
+ "epoch": 897.12,
240936
+ "learning_rate": 8.21928917609047e-06,
240937
+ "loss": 0.3165,
240938
+ "step": 111270
240939
+ },
240940
+ {
240941
+ "epoch": 897.16,
240942
+ "learning_rate": 8.219208400646204e-06,
240943
+ "loss": 0.4533,
240944
+ "step": 111275
240945
+ },
240946
+ {
240947
+ "epoch": 897.2,
240948
+ "learning_rate": 8.21912762520194e-06,
240949
+ "loss": 1.0925,
240950
+ "step": 111280
240951
+ },
240952
+ {
240953
+ "epoch": 897.24,
240954
+ "learning_rate": 8.219046849757674e-06,
240955
+ "loss": 0.3357,
240956
+ "step": 111285
240957
+ },
240958
+ {
240959
+ "epoch": 897.28,
240960
+ "learning_rate": 8.21896607431341e-06,
240961
+ "loss": 0.2692,
240962
+ "step": 111290
240963
+ },
240964
+ {
240965
+ "epoch": 897.32,
240966
+ "learning_rate": 8.218885298869146e-06,
240967
+ "loss": 0.3194,
240968
+ "step": 111295
240969
+ },
240970
+ {
240971
+ "epoch": 897.36,
240972
+ "learning_rate": 8.21880452342488e-06,
240973
+ "loss": 0.555,
240974
+ "step": 111300
240975
+ },
240976
+ {
240977
+ "epoch": 897.4,
240978
+ "learning_rate": 8.218723747980616e-06,
240979
+ "loss": 1.1612,
240980
+ "step": 111305
240981
+ },
240982
+ {
240983
+ "epoch": 897.44,
240984
+ "learning_rate": 8.21864297253635e-06,
240985
+ "loss": 0.3347,
240986
+ "step": 111310
240987
+ },
240988
+ {
240989
+ "epoch": 897.48,
240990
+ "learning_rate": 8.218562197092086e-06,
240991
+ "loss": 0.2972,
240992
+ "step": 111315
240993
+ },
240994
+ {
240995
+ "epoch": 897.52,
240996
+ "learning_rate": 8.21848142164782e-06,
240997
+ "loss": 0.3082,
240998
+ "step": 111320
240999
+ },
241000
+ {
241001
+ "epoch": 897.56,
241002
+ "learning_rate": 8.218400646203556e-06,
241003
+ "loss": 0.4182,
241004
+ "step": 111325
241005
+ },
241006
+ {
241007
+ "epoch": 897.6,
241008
+ "learning_rate": 8.21831987075929e-06,
241009
+ "loss": 1.192,
241010
+ "step": 111330
241011
+ },
241012
+ {
241013
+ "epoch": 897.65,
241014
+ "learning_rate": 8.218239095315026e-06,
241015
+ "loss": 0.3286,
241016
+ "step": 111335
241017
+ },
241018
+ {
241019
+ "epoch": 897.69,
241020
+ "learning_rate": 8.21815831987076e-06,
241021
+ "loss": 0.2918,
241022
+ "step": 111340
241023
+ },
241024
+ {
241025
+ "epoch": 897.73,
241026
+ "learning_rate": 8.218077544426495e-06,
241027
+ "loss": 0.3403,
241028
+ "step": 111345
241029
+ },
241030
+ {
241031
+ "epoch": 897.77,
241032
+ "learning_rate": 8.21799676898223e-06,
241033
+ "loss": 0.3825,
241034
+ "step": 111350
241035
+ },
241036
+ {
241037
+ "epoch": 897.81,
241038
+ "learning_rate": 8.217915993537965e-06,
241039
+ "loss": 1.0865,
241040
+ "step": 111355
241041
+ },
241042
+ {
241043
+ "epoch": 897.85,
241044
+ "learning_rate": 8.2178352180937e-06,
241045
+ "loss": 0.3358,
241046
+ "step": 111360
241047
+ },
241048
+ {
241049
+ "epoch": 897.89,
241050
+ "learning_rate": 8.217754442649435e-06,
241051
+ "loss": 0.3336,
241052
+ "step": 111365
241053
+ },
241054
+ {
241055
+ "epoch": 897.93,
241056
+ "learning_rate": 8.217673667205171e-06,
241057
+ "loss": 0.3484,
241058
+ "step": 111370
241059
+ },
241060
+ {
241061
+ "epoch": 897.97,
241062
+ "learning_rate": 8.217592891760905e-06,
241063
+ "loss": 0.4983,
241064
+ "step": 111375
241065
+ },
241066
+ {
241067
+ "epoch": 898.0,
241068
+ "eval_loss": 0.3549356162548065,
241069
+ "eval_runtime": 42.074,
241070
+ "eval_samples_per_second": 19.894,
241071
+ "eval_steps_per_second": 0.642,
241072
+ "eval_wer": 0.18447376051029285,
241073
+ "step": 111379
241074
  }
241075
  ],
241076
  "max_steps": 620000,
241077
  "num_train_epochs": 5000,
241078
+ "total_flos": 3.134467323546747e+20,
241079
  "trial_name": null,
241080
  "trial_params": null
241081
  }
model-bin/finetune/base/{checkpoint-110758 β†’ checkpoint-111379}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630045620.5655968/events.out.tfevents.1630045620.52f5c7e305a3.886.71 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:789077e8a431041de909256226a12d9fd795a843e189968fcbd73a7d940c1bb0
3
+ size 4194
model-bin/finetune/base/log/1630046118.6803894/events.out.tfevents.1630046118.52f5c7e305a3.886.73 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d4196b10fc377834886dc0d987e6817a8ea8f93225248f6d0c19c43600641b2
3
+ size 4194
model-bin/finetune/base/log/1630046681.8092928/events.out.tfevents.1630046681.52f5c7e305a3.886.75 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abdd13375aec82b2ef9f339c6eb574e37141831c3243c001d07c32e4ab0cf46f
3
+ size 4194
model-bin/finetune/base/log/1630047172.2078328/events.out.tfevents.1630047172.52f5c7e305a3.886.77 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:232390e2831e3b474ff01088741dca913da5c1a45a34e06d55f1e7af7f05c211
3
+ size 4194
model-bin/finetune/base/log/1630047639.1210005/events.out.tfevents.1630047639.52f5c7e305a3.886.79 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63970938dab9c44f0940f5ad235a4c8e66651ae8540d2f66744893ea940f0ac2
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630045620.52f5c7e305a3.886.70 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8314966b5c7b8b998ed76cc6f32d38a8a2fd310b0f566a0ba90b2151c0e7b1f
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630046118.52f5c7e305a3.886.72 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2fca3e324b891d951182c535e721c6721fa0e5b2649c8e079560f44cf322a59
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630046681.52f5c7e305a3.886.74 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f79dfe9c06a70a4a2dc57702c07e9552251702809224aa774e01719c63bd929a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630047172.52f5c7e305a3.886.76 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86cc4daafbb29ca0d18efe9c264f9c5feb271b6fcd28a84322fbf266ffb68542
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630047639.52f5c7e305a3.886.78 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f6abc8f2020bcd704cade369621eba3e292108e363a1184c6f8b5a5f47ca085
3
+ size 8462