Check commited on
Commit
c327d38
Β·
1 Parent(s): 52ffbad

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630244012.7530751/events.out.tfevents.1630244012.cc93b136ebf5.1086.179 +3 -0
  11. model-bin/finetune/base/log/1630244448.5792234/events.out.tfevents.1630244448.cc93b136ebf5.1086.181 +3 -0
  12. model-bin/finetune/base/log/1630244993.356596/events.out.tfevents.1630244993.cc93b136ebf5.1086.183 +3 -0
  13. model-bin/finetune/base/log/1630245427.5586958/events.out.tfevents.1630245427.cc93b136ebf5.1086.185 +3 -0
  14. model-bin/finetune/base/log/1630245858.8737807/events.out.tfevents.1630245858.cc93b136ebf5.1086.187 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630244012.cc93b136ebf5.1086.178 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630244448.cc93b136ebf5.1086.180 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630244993.cc93b136ebf5.1086.182 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630245427.cc93b136ebf5.1086.184 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630245858.cc93b136ebf5.1086.186 +3 -0
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5341661b50f4e596b6fe02c28dfc7ba32504ab70fca5178ea2952aac4a6cbdf1
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9665c6e1c6f2987ac982b4f02510d642ccf49cb18e475af98729715643f2744
3
  size 722165393
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2831ef8aa83687ecb9c8fa7338b9766386e38377b04dd35c04a8c8c06f8148f2
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83712bb0bb146694098ef77fead19cec258379ae2b0f648e4edb18652a55565a
3
  size 377909911
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0626ef614583f5ab57c8ac6ae463c7a0f18875e587996d7e4c49f829145e7e88
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f44a922c3d45f1ec6984b47ad3b841b6c8eca8288bc65f7fa757f180153fd40
3
  size 14503
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3540207233bcabcb86e1d4ad4f38a1c87a6c2cdf020124389fd687b8b31fe0dd
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adf29ee28b173423b576c771747dd93d436f5c3f2a3af9a31f91c508ff5dd1b2
3
  size 559
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63f3f86c60dbd72222ca372243fb135bc1a49b12c2bc13b3077d16a8db5290be
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee134784171ae4f4005cf971653fad3f3e37c1a6c7b565c91f04852ab760d0e4
3
  size 623
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1211.0,
5
- "global_step": 150205,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -290475,11 +290475,800 @@
290475
  "eval_steps_per_second": 0.644,
290476
  "eval_wer": 0.17267648552564754,
290477
  "step": 150205
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
290478
  }
290479
  ],
290480
- "max_steps": 620000,
290481
  "num_train_epochs": 5000,
290482
- "total_flos": 4.226898230914657e+20,
290483
  "trial_name": null,
290484
  "trial_params": null
290485
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1206.0,
5
+ "global_step": 150827,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
290475
  "eval_steps_per_second": 0.644,
290476
  "eval_wer": 0.17267648552564754,
290477
  "step": 150205
290478
+ },
290479
+ {
290480
+ "epoch": 1211.04,
290481
+ "learning_rate": 7.590452342487884e-06,
290482
+ "loss": 0.3201,
290483
+ "step": 150210
290484
+ },
290485
+ {
290486
+ "epoch": 1211.08,
290487
+ "learning_rate": 7.59037156704362e-06,
290488
+ "loss": 0.2736,
290489
+ "step": 150215
290490
+ },
290491
+ {
290492
+ "epoch": 1211.12,
290493
+ "learning_rate": 7.590290791599354e-06,
290494
+ "loss": 0.2423,
290495
+ "step": 150220
290496
+ },
290497
+ {
290498
+ "epoch": 1211.16,
290499
+ "learning_rate": 7.59021001615509e-06,
290500
+ "loss": 0.4379,
290501
+ "step": 150225
290502
+ },
290503
+ {
290504
+ "epoch": 1211.2,
290505
+ "learning_rate": 7.590129240710824e-06,
290506
+ "loss": 0.9755,
290507
+ "step": 150230
290508
+ },
290509
+ {
290510
+ "epoch": 1211.24,
290511
+ "learning_rate": 7.59004846526656e-06,
290512
+ "loss": 0.3417,
290513
+ "step": 150235
290514
+ },
290515
+ {
290516
+ "epoch": 1211.28,
290517
+ "learning_rate": 7.5899676898222955e-06,
290518
+ "loss": 0.2895,
290519
+ "step": 150240
290520
+ },
290521
+ {
290522
+ "epoch": 1211.32,
290523
+ "learning_rate": 7.58988691437803e-06,
290524
+ "loss": 0.2984,
290525
+ "step": 150245
290526
+ },
290527
+ {
290528
+ "epoch": 1211.36,
290529
+ "learning_rate": 7.5898061389337655e-06,
290530
+ "loss": 0.3888,
290531
+ "step": 150250
290532
+ },
290533
+ {
290534
+ "epoch": 1211.4,
290535
+ "learning_rate": 7.5897253634895e-06,
290536
+ "loss": 1.1877,
290537
+ "step": 150255
290538
+ },
290539
+ {
290540
+ "epoch": 1211.44,
290541
+ "learning_rate": 7.5896445880452355e-06,
290542
+ "loss": 0.2986,
290543
+ "step": 150260
290544
+ },
290545
+ {
290546
+ "epoch": 1211.48,
290547
+ "learning_rate": 7.58956381260097e-06,
290548
+ "loss": 0.2978,
290549
+ "step": 150265
290550
+ },
290551
+ {
290552
+ "epoch": 1211.52,
290553
+ "learning_rate": 7.5894830371567055e-06,
290554
+ "loss": 0.3136,
290555
+ "step": 150270
290556
+ },
290557
+ {
290558
+ "epoch": 1211.56,
290559
+ "learning_rate": 7.58940226171244e-06,
290560
+ "loss": 0.4057,
290561
+ "step": 150275
290562
+ },
290563
+ {
290564
+ "epoch": 1211.6,
290565
+ "learning_rate": 7.5893214862681754e-06,
290566
+ "loss": 1.0684,
290567
+ "step": 150280
290568
+ },
290569
+ {
290570
+ "epoch": 1211.64,
290571
+ "learning_rate": 7.5892407108239096e-06,
290572
+ "loss": 0.3111,
290573
+ "step": 150285
290574
+ },
290575
+ {
290576
+ "epoch": 1211.68,
290577
+ "learning_rate": 7.589159935379645e-06,
290578
+ "loss": 0.3245,
290579
+ "step": 150290
290580
+ },
290581
+ {
290582
+ "epoch": 1211.72,
290583
+ "learning_rate": 7.58907915993538e-06,
290584
+ "loss": 0.3207,
290585
+ "step": 150295
290586
+ },
290587
+ {
290588
+ "epoch": 1211.76,
290589
+ "learning_rate": 7.588998384491115e-06,
290590
+ "loss": 0.3932,
290591
+ "step": 150300
290592
+ },
290593
+ {
290594
+ "epoch": 1211.8,
290595
+ "learning_rate": 7.588917609046851e-06,
290596
+ "loss": 1.1473,
290597
+ "step": 150305
290598
+ },
290599
+ {
290600
+ "epoch": 1211.84,
290601
+ "learning_rate": 7.588836833602585e-06,
290602
+ "loss": 0.2849,
290603
+ "step": 150310
290604
+ },
290605
+ {
290606
+ "epoch": 1211.88,
290607
+ "learning_rate": 7.588756058158321e-06,
290608
+ "loss": 0.3008,
290609
+ "step": 150315
290610
+ },
290611
+ {
290612
+ "epoch": 1211.92,
290613
+ "learning_rate": 7.588675282714055e-06,
290614
+ "loss": 0.3498,
290615
+ "step": 150320
290616
+ },
290617
+ {
290618
+ "epoch": 1211.96,
290619
+ "learning_rate": 7.588594507269791e-06,
290620
+ "loss": 0.528,
290621
+ "step": 150325
290622
+ },
290623
+ {
290624
+ "epoch": 1212.0,
290625
+ "eval_loss": 0.38405805826187134,
290626
+ "eval_runtime": 42.6893,
290627
+ "eval_samples_per_second": 19.7,
290628
+ "eval_steps_per_second": 0.632,
290629
+ "eval_wer": 0.17788736420811893,
290630
+ "step": 150329
290631
+ },
290632
+ {
290633
+ "epoch": 1212.01,
290634
+ "learning_rate": 7.588513731825525e-06,
290635
+ "loss": 0.4414,
290636
+ "step": 150330
290637
+ },
290638
+ {
290639
+ "epoch": 1212.05,
290640
+ "learning_rate": 7.588432956381261e-06,
290641
+ "loss": 0.3031,
290642
+ "step": 150335
290643
+ },
290644
+ {
290645
+ "epoch": 1212.09,
290646
+ "learning_rate": 7.588352180936995e-06,
290647
+ "loss": 0.2945,
290648
+ "step": 150340
290649
+ },
290650
+ {
290651
+ "epoch": 1212.13,
290652
+ "learning_rate": 7.588271405492731e-06,
290653
+ "loss": 0.3029,
290654
+ "step": 150345
290655
+ },
290656
+ {
290657
+ "epoch": 1212.17,
290658
+ "learning_rate": 7.588190630048465e-06,
290659
+ "loss": 0.4425,
290660
+ "step": 150350
290661
+ },
290662
+ {
290663
+ "epoch": 1212.21,
290664
+ "learning_rate": 7.588109854604201e-06,
290665
+ "loss": 1.0489,
290666
+ "step": 150355
290667
+ },
290668
+ {
290669
+ "epoch": 1212.25,
290670
+ "learning_rate": 7.588029079159936e-06,
290671
+ "loss": 0.2773,
290672
+ "step": 150360
290673
+ },
290674
+ {
290675
+ "epoch": 1212.29,
290676
+ "learning_rate": 7.587948303715671e-06,
290677
+ "loss": 0.267,
290678
+ "step": 150365
290679
+ },
290680
+ {
290681
+ "epoch": 1212.33,
290682
+ "learning_rate": 7.587867528271407e-06,
290683
+ "loss": 0.3697,
290684
+ "step": 150370
290685
+ },
290686
+ {
290687
+ "epoch": 1212.37,
290688
+ "learning_rate": 7.587786752827141e-06,
290689
+ "loss": 0.4697,
290690
+ "step": 150375
290691
+ },
290692
+ {
290693
+ "epoch": 1212.41,
290694
+ "learning_rate": 7.587705977382877e-06,
290695
+ "loss": 1.0624,
290696
+ "step": 150380
290697
+ },
290698
+ {
290699
+ "epoch": 1212.45,
290700
+ "learning_rate": 7.587625201938611e-06,
290701
+ "loss": 0.2572,
290702
+ "step": 150385
290703
+ },
290704
+ {
290705
+ "epoch": 1212.49,
290706
+ "learning_rate": 7.587544426494347e-06,
290707
+ "loss": 0.3007,
290708
+ "step": 150390
290709
+ },
290710
+ {
290711
+ "epoch": 1212.53,
290712
+ "learning_rate": 7.587463651050081e-06,
290713
+ "loss": 0.2912,
290714
+ "step": 150395
290715
+ },
290716
+ {
290717
+ "epoch": 1212.57,
290718
+ "learning_rate": 7.587382875605817e-06,
290719
+ "loss": 0.5251,
290720
+ "step": 150400
290721
+ },
290722
+ {
290723
+ "epoch": 1212.61,
290724
+ "learning_rate": 7.587302100161551e-06,
290725
+ "loss": 1.0472,
290726
+ "step": 150405
290727
+ },
290728
+ {
290729
+ "epoch": 1212.65,
290730
+ "learning_rate": 7.587221324717287e-06,
290731
+ "loss": 0.3116,
290732
+ "step": 150410
290733
+ },
290734
+ {
290735
+ "epoch": 1212.69,
290736
+ "learning_rate": 7.587140549273022e-06,
290737
+ "loss": 0.2427,
290738
+ "step": 150415
290739
+ },
290740
+ {
290741
+ "epoch": 1212.73,
290742
+ "learning_rate": 7.587059773828757e-06,
290743
+ "loss": 0.3504,
290744
+ "step": 150420
290745
+ },
290746
+ {
290747
+ "epoch": 1212.77,
290748
+ "learning_rate": 7.586978998384492e-06,
290749
+ "loss": 0.4429,
290750
+ "step": 150425
290751
+ },
290752
+ {
290753
+ "epoch": 1212.81,
290754
+ "learning_rate": 7.586898222940227e-06,
290755
+ "loss": 0.945,
290756
+ "step": 150430
290757
+ },
290758
+ {
290759
+ "epoch": 1212.85,
290760
+ "learning_rate": 7.586817447495963e-06,
290761
+ "loss": 0.3042,
290762
+ "step": 150435
290763
+ },
290764
+ {
290765
+ "epoch": 1212.9,
290766
+ "learning_rate": 7.586736672051697e-06,
290767
+ "loss": 0.2662,
290768
+ "step": 150440
290769
+ },
290770
+ {
290771
+ "epoch": 1212.94,
290772
+ "learning_rate": 7.5866558966074326e-06,
290773
+ "loss": 0.3849,
290774
+ "step": 150445
290775
+ },
290776
+ {
290777
+ "epoch": 1212.98,
290778
+ "learning_rate": 7.586575121163167e-06,
290779
+ "loss": 0.6568,
290780
+ "step": 150450
290781
+ },
290782
+ {
290783
+ "epoch": 1213.0,
290784
+ "eval_loss": 0.3243919909000397,
290785
+ "eval_runtime": 45.4766,
290786
+ "eval_samples_per_second": 18.493,
290787
+ "eval_steps_per_second": 0.594,
290788
+ "eval_wer": 0.17089605734767024,
290789
+ "step": 150453
290790
+ },
290791
+ {
290792
+ "epoch": 1203.02,
290793
+ "learning_rate": 7.5864943457189025e-06,
290794
+ "loss": 0.2572,
290795
+ "step": 150455
290796
+ },
290797
+ {
290798
+ "epoch": 1203.06,
290799
+ "learning_rate": 7.586413570274637e-06,
290800
+ "loss": 0.2823,
290801
+ "step": 150460
290802
+ },
290803
+ {
290804
+ "epoch": 1203.1,
290805
+ "learning_rate": 7.5863327948303725e-06,
290806
+ "loss": 0.2366,
290807
+ "step": 150465
290808
+ },
290809
+ {
290810
+ "epoch": 1203.14,
290811
+ "learning_rate": 7.5862520193861075e-06,
290812
+ "loss": 0.3375,
290813
+ "step": 150470
290814
+ },
290815
+ {
290816
+ "epoch": 1203.18,
290817
+ "learning_rate": 7.5861712439418425e-06,
290818
+ "loss": 0.6214,
290819
+ "step": 150475
290820
+ },
290821
+ {
290822
+ "epoch": 1203.22,
290823
+ "learning_rate": 7.5860904684975775e-06,
290824
+ "loss": 0.9851,
290825
+ "step": 150480
290826
+ },
290827
+ {
290828
+ "epoch": 1203.26,
290829
+ "learning_rate": 7.5860096930533125e-06,
290830
+ "loss": 0.2788,
290831
+ "step": 150485
290832
+ },
290833
+ {
290834
+ "epoch": 1203.3,
290835
+ "learning_rate": 7.5859289176090474e-06,
290836
+ "loss": 0.2879,
290837
+ "step": 150490
290838
+ },
290839
+ {
290840
+ "epoch": 1203.34,
290841
+ "learning_rate": 7.5858481421647824e-06,
290842
+ "loss": 0.3532,
290843
+ "step": 150495
290844
+ },
290845
+ {
290846
+ "epoch": 1203.38,
290847
+ "learning_rate": 7.585767366720517e-06,
290848
+ "loss": 0.5532,
290849
+ "step": 150500
290850
+ },
290851
+ {
290852
+ "epoch": 1203.42,
290853
+ "learning_rate": 7.585686591276252e-06,
290854
+ "loss": 1.016,
290855
+ "step": 150505
290856
+ },
290857
+ {
290858
+ "epoch": 1203.46,
290859
+ "learning_rate": 7.585605815831988e-06,
290860
+ "loss": 0.303,
290861
+ "step": 150510
290862
+ },
290863
+ {
290864
+ "epoch": 1203.5,
290865
+ "learning_rate": 7.585525040387722e-06,
290866
+ "loss": 0.2581,
290867
+ "step": 150515
290868
+ },
290869
+ {
290870
+ "epoch": 1203.54,
290871
+ "learning_rate": 7.585444264943458e-06,
290872
+ "loss": 0.3192,
290873
+ "step": 150520
290874
+ },
290875
+ {
290876
+ "epoch": 1203.58,
290877
+ "learning_rate": 7.585363489499192e-06,
290878
+ "loss": 0.5658,
290879
+ "step": 150525
290880
+ },
290881
+ {
290882
+ "epoch": 1203.62,
290883
+ "learning_rate": 7.585282714054928e-06,
290884
+ "loss": 1.1043,
290885
+ "step": 150530
290886
+ },
290887
+ {
290888
+ "epoch": 1203.66,
290889
+ "learning_rate": 7.585201938610663e-06,
290890
+ "loss": 0.3109,
290891
+ "step": 150535
290892
+ },
290893
+ {
290894
+ "epoch": 1203.7,
290895
+ "learning_rate": 7.585121163166398e-06,
290896
+ "loss": 0.3193,
290897
+ "step": 150540
290898
+ },
290899
+ {
290900
+ "epoch": 1203.74,
290901
+ "learning_rate": 7.585040387722133e-06,
290902
+ "loss": 0.2706,
290903
+ "step": 150545
290904
+ },
290905
+ {
290906
+ "epoch": 1203.78,
290907
+ "learning_rate": 7.584959612277868e-06,
290908
+ "loss": 0.6651,
290909
+ "step": 150550
290910
+ },
290911
+ {
290912
+ "epoch": 1203.82,
290913
+ "learning_rate": 7.584878836833603e-06,
290914
+ "loss": 0.9646,
290915
+ "step": 150555
290916
+ },
290917
+ {
290918
+ "epoch": 1203.86,
290919
+ "learning_rate": 7.584798061389338e-06,
290920
+ "loss": 0.284,
290921
+ "step": 150560
290922
+ },
290923
+ {
290924
+ "epoch": 1203.9,
290925
+ "learning_rate": 7.584717285945073e-06,
290926
+ "loss": 0.2926,
290927
+ "step": 150565
290928
+ },
290929
+ {
290930
+ "epoch": 1203.94,
290931
+ "learning_rate": 7.584636510500808e-06,
290932
+ "loss": 0.427,
290933
+ "step": 150570
290934
+ },
290935
+ {
290936
+ "epoch": 1203.98,
290937
+ "learning_rate": 7.584555735056544e-06,
290938
+ "loss": 0.6464,
290939
+ "step": 150575
290940
+ },
290941
+ {
290942
+ "epoch": 1204.0,
290943
+ "eval_loss": 0.5391775965690613,
290944
+ "eval_runtime": 39.7033,
290945
+ "eval_samples_per_second": 21.056,
290946
+ "eval_steps_per_second": 0.68,
290947
+ "eval_wer": 0.1843401803986963,
290948
+ "step": 150578
290949
+ },
290950
+ {
290951
+ "epoch": 1214.02,
290952
+ "learning_rate": 7.584474959612278e-06,
290953
+ "loss": 0.3288,
290954
+ "step": 150580
290955
+ },
290956
+ {
290957
+ "epoch": 1214.06,
290958
+ "learning_rate": 7.584394184168014e-06,
290959
+ "loss": 0.9793,
290960
+ "step": 150585
290961
+ },
290962
+ {
290963
+ "epoch": 1214.1,
290964
+ "learning_rate": 7.584313408723749e-06,
290965
+ "loss": 0.2777,
290966
+ "step": 150590
290967
+ },
290968
+ {
290969
+ "epoch": 1214.14,
290970
+ "learning_rate": 7.584232633279484e-06,
290971
+ "loss": 0.3273,
290972
+ "step": 150595
290973
+ },
290974
+ {
290975
+ "epoch": 1214.18,
290976
+ "learning_rate": 7.584151857835219e-06,
290977
+ "loss": 0.5917,
290978
+ "step": 150600
290979
+ },
290980
+ {
290981
+ "epoch": 1214.22,
290982
+ "learning_rate": 7.584071082390954e-06,
290983
+ "loss": 0.8979,
290984
+ "step": 150605
290985
+ },
290986
+ {
290987
+ "epoch": 1214.26,
290988
+ "learning_rate": 7.583990306946689e-06,
290989
+ "loss": 0.3145,
290990
+ "step": 150610
290991
+ },
290992
+ {
290993
+ "epoch": 1214.3,
290994
+ "learning_rate": 7.583909531502424e-06,
290995
+ "loss": 0.2725,
290996
+ "step": 150615
290997
+ },
290998
+ {
290999
+ "epoch": 1214.34,
291000
+ "learning_rate": 7.583828756058159e-06,
291001
+ "loss": 0.3622,
291002
+ "step": 150620
291003
+ },
291004
+ {
291005
+ "epoch": 1214.38,
291006
+ "learning_rate": 7.583747980613894e-06,
291007
+ "loss": 0.5381,
291008
+ "step": 150625
291009
+ },
291010
+ {
291011
+ "epoch": 1214.42,
291012
+ "learning_rate": 7.583667205169629e-06,
291013
+ "loss": 1.0317,
291014
+ "step": 150630
291015
+ },
291016
+ {
291017
+ "epoch": 1214.46,
291018
+ "learning_rate": 7.583586429725364e-06,
291019
+ "loss": 0.3174,
291020
+ "step": 150635
291021
+ },
291022
+ {
291023
+ "epoch": 1214.5,
291024
+ "learning_rate": 7.5835056542811e-06,
291025
+ "loss": 1.059,
291026
+ "step": 150640
291027
+ },
291028
+ {
291029
+ "epoch": 1214.54,
291030
+ "learning_rate": 7.583424878836835e-06,
291031
+ "loss": 0.355,
291032
+ "step": 150645
291033
+ },
291034
+ {
291035
+ "epoch": 1214.58,
291036
+ "learning_rate": 7.58334410339257e-06,
291037
+ "loss": 0.6424,
291038
+ "step": 150650
291039
+ },
291040
+ {
291041
+ "epoch": 1214.62,
291042
+ "learning_rate": 7.5832633279483046e-06,
291043
+ "loss": 1.0002,
291044
+ "step": 150655
291045
+ },
291046
+ {
291047
+ "epoch": 1214.66,
291048
+ "learning_rate": 7.5831825525040396e-06,
291049
+ "loss": 0.2473,
291050
+ "step": 150660
291051
+ },
291052
+ {
291053
+ "epoch": 1214.7,
291054
+ "learning_rate": 7.5831017770597745e-06,
291055
+ "loss": 0.2639,
291056
+ "step": 150665
291057
+ },
291058
+ {
291059
+ "epoch": 1214.74,
291060
+ "learning_rate": 7.5830210016155095e-06,
291061
+ "loss": 0.3395,
291062
+ "step": 150670
291063
+ },
291064
+ {
291065
+ "epoch": 1214.78,
291066
+ "learning_rate": 7.5829402261712445e-06,
291067
+ "loss": 0.5267,
291068
+ "step": 150675
291069
+ },
291070
+ {
291071
+ "epoch": 1214.82,
291072
+ "learning_rate": 7.5828594507269795e-06,
291073
+ "loss": 1.1767,
291074
+ "step": 150680
291075
+ },
291076
+ {
291077
+ "epoch": 1214.86,
291078
+ "learning_rate": 7.5827786752827145e-06,
291079
+ "loss": 0.2499,
291080
+ "step": 150685
291081
+ },
291082
+ {
291083
+ "epoch": 1214.9,
291084
+ "learning_rate": 7.5826978998384495e-06,
291085
+ "loss": 0.3152,
291086
+ "step": 150690
291087
+ },
291088
+ {
291089
+ "epoch": 1214.94,
291090
+ "learning_rate": 7.5826171243941845e-06,
291091
+ "loss": 0.3273,
291092
+ "step": 150695
291093
+ },
291094
+ {
291095
+ "epoch": 1214.98,
291096
+ "learning_rate": 7.5825363489499195e-06,
291097
+ "loss": 0.5964,
291098
+ "step": 150700
291099
+ },
291100
+ {
291101
+ "epoch": 1215.0,
291102
+ "eval_loss": 0.3669835031032562,
291103
+ "eval_runtime": 42.421,
291104
+ "eval_samples_per_second": 19.707,
291105
+ "eval_steps_per_second": 0.636,
291106
+ "eval_wer": 0.16976108232584916,
291107
+ "step": 150702
291108
+ },
291109
+ {
291110
+ "epoch": 1205.02,
291111
+ "learning_rate": 7.5824555735056544e-06,
291112
+ "loss": 0.3077,
291113
+ "step": 150705
291114
+ },
291115
+ {
291116
+ "epoch": 1205.06,
291117
+ "learning_rate": 7.58237479806139e-06,
291118
+ "loss": 0.294,
291119
+ "step": 150710
291120
+ },
291121
+ {
291122
+ "epoch": 1205.1,
291123
+ "learning_rate": 7.582294022617125e-06,
291124
+ "loss": 0.3353,
291125
+ "step": 150715
291126
+ },
291127
+ {
291128
+ "epoch": 1205.14,
291129
+ "learning_rate": 7.58221324717286e-06,
291130
+ "loss": 0.3076,
291131
+ "step": 150720
291132
+ },
291133
+ {
291134
+ "epoch": 1205.18,
291135
+ "learning_rate": 7.582132471728595e-06,
291136
+ "loss": 0.6049,
291137
+ "step": 150725
291138
+ },
291139
+ {
291140
+ "epoch": 1205.22,
291141
+ "learning_rate": 7.58205169628433e-06,
291142
+ "loss": 0.8056,
291143
+ "step": 150730
291144
+ },
291145
+ {
291146
+ "epoch": 1205.26,
291147
+ "learning_rate": 7.581970920840065e-06,
291148
+ "loss": 0.2615,
291149
+ "step": 150735
291150
+ },
291151
+ {
291152
+ "epoch": 1205.3,
291153
+ "learning_rate": 7.5818901453958e-06,
291154
+ "loss": 0.2697,
291155
+ "step": 150740
291156
+ },
291157
+ {
291158
+ "epoch": 1205.34,
291159
+ "learning_rate": 7.581809369951535e-06,
291160
+ "loss": 0.3135,
291161
+ "step": 150745
291162
+ },
291163
+ {
291164
+ "epoch": 1205.38,
291165
+ "learning_rate": 7.58172859450727e-06,
291166
+ "loss": 0.7361,
291167
+ "step": 150750
291168
+ },
291169
+ {
291170
+ "epoch": 1205.42,
291171
+ "learning_rate": 7.581647819063005e-06,
291172
+ "loss": 0.778,
291173
+ "step": 150755
291174
+ },
291175
+ {
291176
+ "epoch": 1205.46,
291177
+ "learning_rate": 7.58156704361874e-06,
291178
+ "loss": 0.2723,
291179
+ "step": 150760
291180
+ },
291181
+ {
291182
+ "epoch": 1205.5,
291183
+ "learning_rate": 7.581486268174476e-06,
291184
+ "loss": 0.3067,
291185
+ "step": 150765
291186
+ },
291187
+ {
291188
+ "epoch": 1205.54,
291189
+ "learning_rate": 7.58140549273021e-06,
291190
+ "loss": 0.2966,
291191
+ "step": 150770
291192
+ },
291193
+ {
291194
+ "epoch": 1205.58,
291195
+ "learning_rate": 7.581324717285946e-06,
291196
+ "loss": 0.7155,
291197
+ "step": 150775
291198
+ },
291199
+ {
291200
+ "epoch": 1205.62,
291201
+ "learning_rate": 7.581243941841681e-06,
291202
+ "loss": 0.9073,
291203
+ "step": 150780
291204
+ },
291205
+ {
291206
+ "epoch": 1205.66,
291207
+ "learning_rate": 7.581163166397416e-06,
291208
+ "loss": 0.2664,
291209
+ "step": 150785
291210
+ },
291211
+ {
291212
+ "epoch": 1205.7,
291213
+ "learning_rate": 7.581082390953151e-06,
291214
+ "loss": 0.2695,
291215
+ "step": 150790
291216
+ },
291217
+ {
291218
+ "epoch": 1205.74,
291219
+ "learning_rate": 7.581001615508886e-06,
291220
+ "loss": 0.3299,
291221
+ "step": 150795
291222
+ },
291223
+ {
291224
+ "epoch": 1205.78,
291225
+ "learning_rate": 7.580920840064621e-06,
291226
+ "loss": 0.7078,
291227
+ "step": 150800
291228
+ },
291229
+ {
291230
+ "epoch": 1205.82,
291231
+ "learning_rate": 7.580840064620356e-06,
291232
+ "loss": 0.9233,
291233
+ "step": 150805
291234
+ },
291235
+ {
291236
+ "epoch": 1205.86,
291237
+ "learning_rate": 7.580759289176091e-06,
291238
+ "loss": 0.2724,
291239
+ "step": 150810
291240
+ },
291241
+ {
291242
+ "epoch": 1205.9,
291243
+ "learning_rate": 7.580678513731826e-06,
291244
+ "loss": 0.3251,
291245
+ "step": 150815
291246
+ },
291247
+ {
291248
+ "epoch": 1205.94,
291249
+ "learning_rate": 7.580597738287562e-06,
291250
+ "loss": 0.3044,
291251
+ "step": 150820
291252
+ },
291253
+ {
291254
+ "epoch": 1205.98,
291255
+ "learning_rate": 7.580516962843296e-06,
291256
+ "loss": 0.614,
291257
+ "step": 150825
291258
+ },
291259
+ {
291260
+ "epoch": 1206.0,
291261
+ "eval_loss": 0.3523024320602417,
291262
+ "eval_runtime": 40.2575,
291263
+ "eval_samples_per_second": 20.766,
291264
+ "eval_steps_per_second": 0.671,
291265
+ "eval_wer": 0.182829413960433,
291266
+ "step": 150827
291267
  }
291268
  ],
291269
+ "max_steps": 625000,
291270
  "num_train_epochs": 5000,
291271
+ "total_flos": 4.2444315493114164e+20,
291272
  "trial_name": null,
291273
  "trial_params": null
291274
  }
model-bin/finetune/base/{checkpoint-150205 β†’ checkpoint-150827}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630244012.7530751/events.out.tfevents.1630244012.cc93b136ebf5.1086.179 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be26c9daead194841c91de81f386a51de6ca433a7e1908af359da8e170158788
3
+ size 4194
model-bin/finetune/base/log/1630244448.5792234/events.out.tfevents.1630244448.cc93b136ebf5.1086.181 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52d266b5f028fc38c9d1a6a414c1c393506f6832953cb7488df1efa39b8bf893
3
+ size 4194
model-bin/finetune/base/log/1630244993.356596/events.out.tfevents.1630244993.cc93b136ebf5.1086.183 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db7a02b0a48cfebf6119ec88b66e2fb707cbbd8d04c1f4c7556763fea904dbd8
3
+ size 4194
model-bin/finetune/base/log/1630245427.5586958/events.out.tfevents.1630245427.cc93b136ebf5.1086.185 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c397d04b557890d4c7955c897cc3fe2286c2fbf5b49d17e2f205594fa1ca5726
3
+ size 4194
model-bin/finetune/base/log/1630245858.8737807/events.out.tfevents.1630245858.cc93b136ebf5.1086.187 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:816366c9a07099ac79c7220c098da88f70aabc2b8cfd69ef728187b81a69a4f1
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630244012.cc93b136ebf5.1086.178 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aa3f0de0578fd1d28673ee4fdbb0ebfb66f41ea2387c6285d9df7e4d33ecad0
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630244448.cc93b136ebf5.1086.180 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e96fea5cd10e9f7da659de83310ebaee8c9de0e99e1f2c10547f556dcea96d12
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630244993.cc93b136ebf5.1086.182 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afc99b9f992b08112289a38952d17cabf09acb3de0e8da1ccd48ac875887989b
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630245427.cc93b136ebf5.1086.184 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d7955bc7c27f04e99fcd1c98830cccfa12ac8c2b0edbaff0d6a30315c0b3e1d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630245858.cc93b136ebf5.1086.186 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef4e6f78f20a1befa18e737a50ec4c4d90bf6a887d10552ee89424c85ea0395d
3
+ size 8622