Check commited on
Commit
ef6c207
Β·
1 Parent(s): c68fa3e

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/trainer_state.json +164 -5
  9. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/optimizer.pt +1 -1
  12. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/rng_state.pth +1 -1
  15. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/trainer_state.json +1902 -6
  18. model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1630230484.4148061/events.out.tfevents.1630230484.cc93b136ebf5.1086.119 +3 -0
  20. model-bin/finetune/base/log/1630230917.0829825/events.out.tfevents.1630230917.cc93b136ebf5.1086.121 +3 -0
  21. model-bin/finetune/base/log/1630231362.9249244/events.out.tfevents.1630231362.cc93b136ebf5.1086.123 +3 -0
  22. model-bin/finetune/base/log/1630231799.7950127/events.out.tfevents.1630231799.cc93b136ebf5.1086.125 +3 -0
  23. model-bin/finetune/base/log/1630232234.2438745/events.out.tfevents.1630232234.cc93b136ebf5.1086.127 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1630230484.cc93b136ebf5.1086.118 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1630230917.cc93b136ebf5.1086.120 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1630231362.cc93b136ebf5.1086.122 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1630231799.cc93b136ebf5.1086.124 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1630232234.cc93b136ebf5.1086.126 +3 -0
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:425263039b2e17958ef29df62ab9ab8878277ad6c4ead7176bda8210b2ed3fe6
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0ea9552f899d2d8f535fc434267063e77290df22b3ea0d489d5f3bcb93f46d0
3
  size 722165393
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a22ed2c49e0003da5e3bff5a28ee7e00afcf8ecb8920666d8355ae042273f4b
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dc25b2405ed70ef3243f1c3b26f7665ff79883dac75f72b56452985fad0a79b
3
  size 377909911
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49db60bb7be014c03e20f099f8b5b41640dd5b10966a8d382a375eaacf0dba1f
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9340a78af637a8c0ce5c4b0fa3f2905a42f8cd1ccd9b627b31325f5eaab00977
3
  size 14503
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e32e4b8d64538adb2e3a82fff0984d5e19137de747f31f2ccfe398a635069ee
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:998b6be1a485973ded56fe9124d5ea9cfa3bd59e3499608d8b1f8a68c1d7c92c
3
  size 559
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfefb4c24229027a9d0175af6026b93ceb96966446b698af9ac66780b0583d09
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70ffda1a735fe44bfbcdb5fdd2fcf7dbc1ddead2a83c5e723492366f2b5337c5
3
  size 623
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-146596}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.16716790141568252,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
4
- "epoch": 1180.995983935743,
5
- "global_step": 146472,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -285723,11 +285723,170 @@
285723
  "eval_steps_per_second": 0.645,
285724
  "eval_wer": 0.1758169934640523,
285725
  "step": 146472
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
285726
  }
285727
  ],
285728
  "max_steps": 620000,
285729
  "num_train_epochs": 5000,
285730
- "total_flos": 4.1218134700506854e+20,
285731
  "trial_name": null,
285732
  "trial_params": null
285733
  }
 
1
  {
2
+ "best_metric": 0.1668011442822563,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1181.995983935743,
5
+ "global_step": 146596,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
285723
  "eval_steps_per_second": 0.645,
285724
  "eval_wer": 0.1758169934640523,
285725
  "step": 146472
285726
+ },
285727
+ {
285728
+ "epoch": 1181.02,
285729
+ "learning_rate": 7.65075928917609e-06,
285730
+ "loss": 0.306,
285731
+ "step": 146475
285732
+ },
285733
+ {
285734
+ "epoch": 1181.06,
285735
+ "learning_rate": 7.650678513731826e-06,
285736
+ "loss": 0.3149,
285737
+ "step": 146480
285738
+ },
285739
+ {
285740
+ "epoch": 1181.1,
285741
+ "learning_rate": 7.65059773828756e-06,
285742
+ "loss": 0.2999,
285743
+ "step": 146485
285744
+ },
285745
+ {
285746
+ "epoch": 1181.14,
285747
+ "learning_rate": 7.650516962843296e-06,
285748
+ "loss": 0.3347,
285749
+ "step": 146490
285750
+ },
285751
+ {
285752
+ "epoch": 1181.18,
285753
+ "learning_rate": 7.650436187399032e-06,
285754
+ "loss": 0.6613,
285755
+ "step": 146495
285756
+ },
285757
+ {
285758
+ "epoch": 1181.22,
285759
+ "learning_rate": 7.650355411954766e-06,
285760
+ "loss": 0.8353,
285761
+ "step": 146500
285762
+ },
285763
+ {
285764
+ "epoch": 1181.27,
285765
+ "learning_rate": 7.650274636510502e-06,
285766
+ "loss": 0.3035,
285767
+ "step": 146505
285768
+ },
285769
+ {
285770
+ "epoch": 1181.31,
285771
+ "learning_rate": 7.650193861066236e-06,
285772
+ "loss": 0.3257,
285773
+ "step": 146510
285774
+ },
285775
+ {
285776
+ "epoch": 1181.35,
285777
+ "learning_rate": 7.650113085621972e-06,
285778
+ "loss": 0.4288,
285779
+ "step": 146515
285780
+ },
285781
+ {
285782
+ "epoch": 1181.39,
285783
+ "learning_rate": 7.650032310177706e-06,
285784
+ "loss": 0.6413,
285785
+ "step": 146520
285786
+ },
285787
+ {
285788
+ "epoch": 1181.43,
285789
+ "learning_rate": 7.649951534733442e-06,
285790
+ "loss": 0.8146,
285791
+ "step": 146525
285792
+ },
285793
+ {
285794
+ "epoch": 1181.47,
285795
+ "learning_rate": 7.649870759289176e-06,
285796
+ "loss": 0.2767,
285797
+ "step": 146530
285798
+ },
285799
+ {
285800
+ "epoch": 1181.51,
285801
+ "learning_rate": 7.649789983844912e-06,
285802
+ "loss": 0.2579,
285803
+ "step": 146535
285804
+ },
285805
+ {
285806
+ "epoch": 1181.55,
285807
+ "learning_rate": 7.649709208400646e-06,
285808
+ "loss": 0.3754,
285809
+ "step": 146540
285810
+ },
285811
+ {
285812
+ "epoch": 1181.59,
285813
+ "learning_rate": 7.649628432956382e-06,
285814
+ "loss": 0.7602,
285815
+ "step": 146545
285816
+ },
285817
+ {
285818
+ "epoch": 1181.63,
285819
+ "learning_rate": 7.649547657512118e-06,
285820
+ "loss": 0.879,
285821
+ "step": 146550
285822
+ },
285823
+ {
285824
+ "epoch": 1181.67,
285825
+ "learning_rate": 7.649466882067852e-06,
285826
+ "loss": 0.2737,
285827
+ "step": 146555
285828
+ },
285829
+ {
285830
+ "epoch": 1181.71,
285831
+ "learning_rate": 7.649386106623588e-06,
285832
+ "loss": 0.2831,
285833
+ "step": 146560
285834
+ },
285835
+ {
285836
+ "epoch": 1181.75,
285837
+ "learning_rate": 7.649305331179322e-06,
285838
+ "loss": 0.3641,
285839
+ "step": 146565
285840
+ },
285841
+ {
285842
+ "epoch": 1181.79,
285843
+ "learning_rate": 7.649224555735058e-06,
285844
+ "loss": 0.7179,
285845
+ "step": 146570
285846
+ },
285847
+ {
285848
+ "epoch": 1181.83,
285849
+ "learning_rate": 7.649143780290792e-06,
285850
+ "loss": 0.8389,
285851
+ "step": 146575
285852
+ },
285853
+ {
285854
+ "epoch": 1181.87,
285855
+ "learning_rate": 7.649063004846527e-06,
285856
+ "loss": 0.2527,
285857
+ "step": 146580
285858
+ },
285859
+ {
285860
+ "epoch": 1181.91,
285861
+ "learning_rate": 7.648982229402262e-06,
285862
+ "loss": 0.2547,
285863
+ "step": 146585
285864
+ },
285865
+ {
285866
+ "epoch": 1181.95,
285867
+ "learning_rate": 7.648901453957997e-06,
285868
+ "loss": 0.3927,
285869
+ "step": 146590
285870
+ },
285871
+ {
285872
+ "epoch": 1181.99,
285873
+ "learning_rate": 7.648820678513732e-06,
285874
+ "loss": 0.8367,
285875
+ "step": 146595
285876
+ },
285877
+ {
285878
+ "epoch": 1182.0,
285879
+ "eval_loss": 0.34015288949012756,
285880
+ "eval_runtime": 41.2224,
285881
+ "eval_samples_per_second": 20.304,
285882
+ "eval_steps_per_second": 0.655,
285883
+ "eval_wer": 0.1668011442822563,
285884
+ "step": 146596
285885
  }
285886
  ],
285887
  "max_steps": 620000,
285888
  "num_train_epochs": 5000,
285889
+ "total_flos": 4.125294303144466e+20,
285890
  "trial_name": null,
285891
  "trial_params": null
285892
  }
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-146596}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f7568a2befca944be90026427116813896320af24320ff0d6d6cce3345d3e75
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bd46150aabe97ab3943950ceac270eaf7ba5a905ee8fc857f765d37689b06b8
3
  size 722165393
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2376bae8ab9408163ddcedcab213b016840a625804da3cbf1526010b8abc141
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e642ff64d498eff54009c654ed787928c5cbb4d47b0eda3d26f1a951e1f4ec32
3
  size 377909911
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:80390ad562fb737219eaef200a69cb1cb5c9e3d1cb1722a1fdddeff6f6402b7d
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6232982667952bbc19502a4f0ef987e311dcdf6d628a958e5239d2cb0215ece
3
  size 14503
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d760abfe85cbbe513ba97c70e9515e7e485b75c78f8b50035843503942c2409
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c47fa02eeeae7ce0eb2caa75c3dd36cc9207f4a30770a12e539672a58d4fadb9
3
  size 559
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75a35ac11ff6fdc1ff80807c5620a77c06912969c11e90901d0e650569700174
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e7f4fe6d1acaa457416128487b7a022a2db612c7ffc4ebe77644069b686eb9a
3
  size 623
model-bin/finetune/base/{checkpoint-145602 β†’ checkpoint-147094}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.16716790141568252,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
4
- "epoch": 1164.0,
5
- "global_step": 145602,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -284616,11 +284616,1907 @@
284616
  "eval_steps_per_second": 0.687,
284617
  "eval_wer": 0.16716790141568252,
284618
  "step": 145602
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
284619
  }
284620
  ],
284621
- "max_steps": 625000,
284622
  "num_train_epochs": 5000,
284623
- "total_flos": 4.097272814922444e+20,
284624
  "trial_name": null,
284625
  "trial_params": null
284626
  }
 
1
  {
2
+ "best_metric": 0.1668011442822563,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1185.995983935743,
5
+ "global_step": 147094,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
284616
  "eval_steps_per_second": 0.687,
284617
  "eval_wer": 0.16716790141568252,
284618
  "step": 145602
284619
+ },
284620
+ {
284621
+ "epoch": 1164.02,
284622
+ "learning_rate": 7.66481421647819e-06,
284623
+ "loss": 0.3273,
284624
+ "step": 145605
284625
+ },
284626
+ {
284627
+ "epoch": 1164.06,
284628
+ "learning_rate": 7.664733441033926e-06,
284629
+ "loss": 0.2867,
284630
+ "step": 145610
284631
+ },
284632
+ {
284633
+ "epoch": 1164.1,
284634
+ "learning_rate": 7.664652665589662e-06,
284635
+ "loss": 0.3066,
284636
+ "step": 145615
284637
+ },
284638
+ {
284639
+ "epoch": 1164.14,
284640
+ "learning_rate": 7.664571890145396e-06,
284641
+ "loss": 0.3405,
284642
+ "step": 145620
284643
+ },
284644
+ {
284645
+ "epoch": 1164.18,
284646
+ "learning_rate": 7.664491114701132e-06,
284647
+ "loss": 0.7475,
284648
+ "step": 145625
284649
+ },
284650
+ {
284651
+ "epoch": 1164.22,
284652
+ "learning_rate": 7.664410339256866e-06,
284653
+ "loss": 0.9464,
284654
+ "step": 145630
284655
+ },
284656
+ {
284657
+ "epoch": 1164.26,
284658
+ "learning_rate": 7.664329563812602e-06,
284659
+ "loss": 0.3219,
284660
+ "step": 145635
284661
+ },
284662
+ {
284663
+ "epoch": 1164.3,
284664
+ "learning_rate": 7.664248788368336e-06,
284665
+ "loss": 0.3095,
284666
+ "step": 145640
284667
+ },
284668
+ {
284669
+ "epoch": 1164.34,
284670
+ "learning_rate": 7.664168012924072e-06,
284671
+ "loss": 0.3189,
284672
+ "step": 145645
284673
+ },
284674
+ {
284675
+ "epoch": 1164.38,
284676
+ "learning_rate": 7.664087237479806e-06,
284677
+ "loss": 0.7207,
284678
+ "step": 145650
284679
+ },
284680
+ {
284681
+ "epoch": 1164.42,
284682
+ "learning_rate": 7.664006462035542e-06,
284683
+ "loss": 0.8142,
284684
+ "step": 145655
284685
+ },
284686
+ {
284687
+ "epoch": 1164.46,
284688
+ "learning_rate": 7.663925686591276e-06,
284689
+ "loss": 0.2481,
284690
+ "step": 145660
284691
+ },
284692
+ {
284693
+ "epoch": 1164.5,
284694
+ "learning_rate": 7.663844911147012e-06,
284695
+ "loss": 0.2636,
284696
+ "step": 145665
284697
+ },
284698
+ {
284699
+ "epoch": 1164.54,
284700
+ "learning_rate": 7.663764135702746e-06,
284701
+ "loss": 0.3821,
284702
+ "step": 145670
284703
+ },
284704
+ {
284705
+ "epoch": 1164.58,
284706
+ "learning_rate": 7.663683360258482e-06,
284707
+ "loss": 0.7099,
284708
+ "step": 145675
284709
+ },
284710
+ {
284711
+ "epoch": 1164.62,
284712
+ "learning_rate": 7.663602584814218e-06,
284713
+ "loss": 0.7693,
284714
+ "step": 145680
284715
+ },
284716
+ {
284717
+ "epoch": 1164.66,
284718
+ "learning_rate": 7.663521809369952e-06,
284719
+ "loss": 0.3184,
284720
+ "step": 145685
284721
+ },
284722
+ {
284723
+ "epoch": 1164.7,
284724
+ "learning_rate": 7.663441033925688e-06,
284725
+ "loss": 0.2638,
284726
+ "step": 145690
284727
+ },
284728
+ {
284729
+ "epoch": 1164.74,
284730
+ "learning_rate": 7.663360258481422e-06,
284731
+ "loss": 0.4138,
284732
+ "step": 145695
284733
+ },
284734
+ {
284735
+ "epoch": 1164.78,
284736
+ "learning_rate": 7.663279483037158e-06,
284737
+ "loss": 0.7006,
284738
+ "step": 145700
284739
+ },
284740
+ {
284741
+ "epoch": 1164.82,
284742
+ "learning_rate": 7.663198707592892e-06,
284743
+ "loss": 0.9005,
284744
+ "step": 145705
284745
+ },
284746
+ {
284747
+ "epoch": 1164.86,
284748
+ "learning_rate": 7.663117932148628e-06,
284749
+ "loss": 0.2489,
284750
+ "step": 145710
284751
+ },
284752
+ {
284753
+ "epoch": 1164.9,
284754
+ "learning_rate": 7.663037156704362e-06,
284755
+ "loss": 0.2708,
284756
+ "step": 145715
284757
+ },
284758
+ {
284759
+ "epoch": 1164.94,
284760
+ "learning_rate": 7.662956381260098e-06,
284761
+ "loss": 0.3742,
284762
+ "step": 145720
284763
+ },
284764
+ {
284765
+ "epoch": 1164.98,
284766
+ "learning_rate": 7.662875605815832e-06,
284767
+ "loss": 0.7858,
284768
+ "step": 145725
284769
+ },
284770
+ {
284771
+ "epoch": 1165.0,
284772
+ "eval_loss": 0.38368138670921326,
284773
+ "eval_runtime": 40.6993,
284774
+ "eval_samples_per_second": 20.565,
284775
+ "eval_steps_per_second": 0.663,
284776
+ "eval_wer": 0.1798013485101138,
284777
+ "step": 145727
284778
+ },
284779
+ {
284780
+ "epoch": 1165.02,
284781
+ "learning_rate": 7.662794830371568e-06,
284782
+ "loss": 0.4956,
284783
+ "step": 145730
284784
+ },
284785
+ {
284786
+ "epoch": 1165.06,
284787
+ "learning_rate": 7.662714054927302e-06,
284788
+ "loss": 0.2444,
284789
+ "step": 145735
284790
+ },
284791
+ {
284792
+ "epoch": 1165.1,
284793
+ "learning_rate": 7.662633279483038e-06,
284794
+ "loss": 0.2835,
284795
+ "step": 145740
284796
+ },
284797
+ {
284798
+ "epoch": 1165.14,
284799
+ "learning_rate": 7.662552504038772e-06,
284800
+ "loss": 0.4172,
284801
+ "step": 145745
284802
+ },
284803
+ {
284804
+ "epoch": 1165.18,
284805
+ "learning_rate": 7.662471728594508e-06,
284806
+ "loss": 0.656,
284807
+ "step": 145750
284808
+ },
284809
+ {
284810
+ "epoch": 1165.22,
284811
+ "learning_rate": 7.662390953150244e-06,
284812
+ "loss": 0.8101,
284813
+ "step": 145755
284814
+ },
284815
+ {
284816
+ "epoch": 1165.26,
284817
+ "learning_rate": 7.662310177705978e-06,
284818
+ "loss": 0.2991,
284819
+ "step": 145760
284820
+ },
284821
+ {
284822
+ "epoch": 1165.3,
284823
+ "learning_rate": 7.662229402261714e-06,
284824
+ "loss": 0.2851,
284825
+ "step": 145765
284826
+ },
284827
+ {
284828
+ "epoch": 1165.34,
284829
+ "learning_rate": 7.662148626817448e-06,
284830
+ "loss": 0.3243,
284831
+ "step": 145770
284832
+ },
284833
+ {
284834
+ "epoch": 1165.38,
284835
+ "learning_rate": 7.662067851373183e-06,
284836
+ "loss": 0.7754,
284837
+ "step": 145775
284838
+ },
284839
+ {
284840
+ "epoch": 1165.42,
284841
+ "learning_rate": 7.661987075928918e-06,
284842
+ "loss": 0.9141,
284843
+ "step": 145780
284844
+ },
284845
+ {
284846
+ "epoch": 1165.46,
284847
+ "learning_rate": 7.661906300484653e-06,
284848
+ "loss": 0.2886,
284849
+ "step": 145785
284850
+ },
284851
+ {
284852
+ "epoch": 1165.5,
284853
+ "learning_rate": 7.661825525040388e-06,
284854
+ "loss": 0.305,
284855
+ "step": 145790
284856
+ },
284857
+ {
284858
+ "epoch": 1165.54,
284859
+ "learning_rate": 7.661744749596123e-06,
284860
+ "loss": 0.5382,
284861
+ "step": 145795
284862
+ },
284863
+ {
284864
+ "epoch": 1165.58,
284865
+ "learning_rate": 7.661663974151858e-06,
284866
+ "loss": 0.7651,
284867
+ "step": 145800
284868
+ },
284869
+ {
284870
+ "epoch": 1165.62,
284871
+ "learning_rate": 7.661583198707593e-06,
284872
+ "loss": 0.8388,
284873
+ "step": 145805
284874
+ },
284875
+ {
284876
+ "epoch": 1165.66,
284877
+ "learning_rate": 7.661502423263328e-06,
284878
+ "loss": 0.2597,
284879
+ "step": 145810
284880
+ },
284881
+ {
284882
+ "epoch": 1165.7,
284883
+ "learning_rate": 7.661421647819063e-06,
284884
+ "loss": 0.3229,
284885
+ "step": 145815
284886
+ },
284887
+ {
284888
+ "epoch": 1165.74,
284889
+ "learning_rate": 7.6613408723748e-06,
284890
+ "loss": 0.376,
284891
+ "step": 145820
284892
+ },
284893
+ {
284894
+ "epoch": 1165.78,
284895
+ "learning_rate": 7.661260096930533e-06,
284896
+ "loss": 0.6114,
284897
+ "step": 145825
284898
+ },
284899
+ {
284900
+ "epoch": 1165.82,
284901
+ "learning_rate": 7.66117932148627e-06,
284902
+ "loss": 0.8276,
284903
+ "step": 145830
284904
+ },
284905
+ {
284906
+ "epoch": 1165.86,
284907
+ "learning_rate": 7.661098546042003e-06,
284908
+ "loss": 0.2959,
284909
+ "step": 145835
284910
+ },
284911
+ {
284912
+ "epoch": 1165.9,
284913
+ "learning_rate": 7.66101777059774e-06,
284914
+ "loss": 0.2697,
284915
+ "step": 145840
284916
+ },
284917
+ {
284918
+ "epoch": 1165.94,
284919
+ "learning_rate": 7.660936995153473e-06,
284920
+ "loss": 0.3293,
284921
+ "step": 145845
284922
+ },
284923
+ {
284924
+ "epoch": 1165.98,
284925
+ "learning_rate": 7.660856219709209e-06,
284926
+ "loss": 0.7433,
284927
+ "step": 145850
284928
+ },
284929
+ {
284930
+ "epoch": 1166.0,
284931
+ "eval_loss": 0.33688074350357056,
284932
+ "eval_runtime": 41.2402,
284933
+ "eval_samples_per_second": 20.296,
284934
+ "eval_steps_per_second": 0.655,
284935
+ "eval_wer": 0.17012599228024178,
284936
+ "step": 145852
284937
+ },
284938
+ {
284939
+ "epoch": 1176.02,
284940
+ "learning_rate": 7.660775444264943e-06,
284941
+ "loss": 0.4127,
284942
+ "step": 145855
284943
+ },
284944
+ {
284945
+ "epoch": 1176.06,
284946
+ "learning_rate": 7.660694668820679e-06,
284947
+ "loss": 0.2818,
284948
+ "step": 145860
284949
+ },
284950
+ {
284951
+ "epoch": 1176.1,
284952
+ "learning_rate": 7.660613893376413e-06,
284953
+ "loss": 0.2966,
284954
+ "step": 145865
284955
+ },
284956
+ {
284957
+ "epoch": 1176.15,
284958
+ "learning_rate": 7.660533117932149e-06,
284959
+ "loss": 0.3262,
284960
+ "step": 145870
284961
+ },
284962
+ {
284963
+ "epoch": 1176.19,
284964
+ "learning_rate": 7.660452342487883e-06,
284965
+ "loss": 0.744,
284966
+ "step": 145875
284967
+ },
284968
+ {
284969
+ "epoch": 1176.23,
284970
+ "learning_rate": 7.660371567043619e-06,
284971
+ "loss": 0.8761,
284972
+ "step": 145880
284973
+ },
284974
+ {
284975
+ "epoch": 1176.27,
284976
+ "learning_rate": 7.660290791599355e-06,
284977
+ "loss": 0.2513,
284978
+ "step": 145885
284979
+ },
284980
+ {
284981
+ "epoch": 1176.31,
284982
+ "learning_rate": 7.660210016155089e-06,
284983
+ "loss": 0.2938,
284984
+ "step": 145890
284985
+ },
284986
+ {
284987
+ "epoch": 1176.35,
284988
+ "learning_rate": 7.660129240710825e-06,
284989
+ "loss": 0.3237,
284990
+ "step": 145895
284991
+ },
284992
+ {
284993
+ "epoch": 1176.39,
284994
+ "learning_rate": 7.660048465266559e-06,
284995
+ "loss": 0.7375,
284996
+ "step": 145900
284997
+ },
284998
+ {
284999
+ "epoch": 1176.43,
285000
+ "learning_rate": 7.659967689822295e-06,
285001
+ "loss": 0.8132,
285002
+ "step": 145905
285003
+ },
285004
+ {
285005
+ "epoch": 1176.47,
285006
+ "learning_rate": 7.659886914378029e-06,
285007
+ "loss": 0.3004,
285008
+ "step": 145910
285009
+ },
285010
+ {
285011
+ "epoch": 1176.51,
285012
+ "learning_rate": 7.659806138933765e-06,
285013
+ "loss": 0.2726,
285014
+ "step": 145915
285015
+ },
285016
+ {
285017
+ "epoch": 1176.55,
285018
+ "learning_rate": 7.659725363489499e-06,
285019
+ "loss": 0.3235,
285020
+ "step": 145920
285021
+ },
285022
+ {
285023
+ "epoch": 1176.59,
285024
+ "learning_rate": 7.659644588045235e-06,
285025
+ "loss": 0.7467,
285026
+ "step": 145925
285027
+ },
285028
+ {
285029
+ "epoch": 1176.63,
285030
+ "learning_rate": 7.659563812600969e-06,
285031
+ "loss": 0.7942,
285032
+ "step": 145930
285033
+ },
285034
+ {
285035
+ "epoch": 1176.67,
285036
+ "learning_rate": 7.659483037156705e-06,
285037
+ "loss": 0.2774,
285038
+ "step": 145935
285039
+ },
285040
+ {
285041
+ "epoch": 1176.71,
285042
+ "learning_rate": 7.659402261712439e-06,
285043
+ "loss": 0.2752,
285044
+ "step": 145940
285045
+ },
285046
+ {
285047
+ "epoch": 1176.75,
285048
+ "learning_rate": 7.659321486268175e-06,
285049
+ "loss": 0.3578,
285050
+ "step": 145945
285051
+ },
285052
+ {
285053
+ "epoch": 1176.79,
285054
+ "learning_rate": 7.659240710823909e-06,
285055
+ "loss": 0.6986,
285056
+ "step": 145950
285057
+ },
285058
+ {
285059
+ "epoch": 1176.83,
285060
+ "learning_rate": 7.659159935379645e-06,
285061
+ "loss": 0.7739,
285062
+ "step": 145955
285063
+ },
285064
+ {
285065
+ "epoch": 1176.87,
285066
+ "learning_rate": 7.65907915993538e-06,
285067
+ "loss": 0.2683,
285068
+ "step": 145960
285069
+ },
285070
+ {
285071
+ "epoch": 1176.91,
285072
+ "learning_rate": 7.658998384491115e-06,
285073
+ "loss": 0.3132,
285074
+ "step": 145965
285075
+ },
285076
+ {
285077
+ "epoch": 1176.95,
285078
+ "learning_rate": 7.65891760904685e-06,
285079
+ "loss": 0.373,
285080
+ "step": 145970
285081
+ },
285082
+ {
285083
+ "epoch": 1176.99,
285084
+ "learning_rate": 7.658836833602585e-06,
285085
+ "loss": 1.049,
285086
+ "step": 145975
285087
+ },
285088
+ {
285089
+ "epoch": 1177.0,
285090
+ "eval_loss": 0.3793235719203949,
285091
+ "eval_runtime": 40.3575,
285092
+ "eval_samples_per_second": 20.74,
285093
+ "eval_steps_per_second": 0.669,
285094
+ "eval_wer": 0.17174677608440797,
285095
+ "step": 145976
285096
+ },
285097
+ {
285098
+ "epoch": 1177.03,
285099
+ "learning_rate": 7.65875605815832e-06,
285100
+ "loss": 0.3211,
285101
+ "step": 145980
285102
+ },
285103
+ {
285104
+ "epoch": 1177.07,
285105
+ "learning_rate": 7.658675282714055e-06,
285106
+ "loss": 0.2587,
285107
+ "step": 145985
285108
+ },
285109
+ {
285110
+ "epoch": 1177.11,
285111
+ "learning_rate": 7.65859450726979e-06,
285112
+ "loss": 0.3043,
285113
+ "step": 145990
285114
+ },
285115
+ {
285116
+ "epoch": 1177.15,
285117
+ "learning_rate": 7.658513731825525e-06,
285118
+ "loss": 0.374,
285119
+ "step": 145995
285120
+ },
285121
+ {
285122
+ "epoch": 1177.19,
285123
+ "learning_rate": 7.65843295638126e-06,
285124
+ "loss": 0.9153,
285125
+ "step": 146000
285126
+ },
285127
+ {
285128
+ "epoch": 1177.23,
285129
+ "learning_rate": 7.658352180936995e-06,
285130
+ "loss": 0.7194,
285131
+ "step": 146005
285132
+ },
285133
+ {
285134
+ "epoch": 1177.27,
285135
+ "learning_rate": 7.65827140549273e-06,
285136
+ "loss": 0.2684,
285137
+ "step": 146010
285138
+ },
285139
+ {
285140
+ "epoch": 1177.31,
285141
+ "learning_rate": 7.658190630048465e-06,
285142
+ "loss": 0.2866,
285143
+ "step": 146015
285144
+ },
285145
+ {
285146
+ "epoch": 1177.35,
285147
+ "learning_rate": 7.6581098546042e-06,
285148
+ "loss": 0.3972,
285149
+ "step": 146020
285150
+ },
285151
+ {
285152
+ "epoch": 1177.39,
285153
+ "learning_rate": 7.658029079159936e-06,
285154
+ "loss": 1.0221,
285155
+ "step": 146025
285156
+ },
285157
+ {
285158
+ "epoch": 1177.43,
285159
+ "learning_rate": 7.65794830371567e-06,
285160
+ "loss": 0.5779,
285161
+ "step": 146030
285162
+ },
285163
+ {
285164
+ "epoch": 1177.47,
285165
+ "learning_rate": 7.657867528271406e-06,
285166
+ "loss": 0.3161,
285167
+ "step": 146035
285168
+ },
285169
+ {
285170
+ "epoch": 1177.51,
285171
+ "learning_rate": 7.65778675282714e-06,
285172
+ "loss": 0.2923,
285173
+ "step": 146040
285174
+ },
285175
+ {
285176
+ "epoch": 1177.55,
285177
+ "learning_rate": 7.657705977382876e-06,
285178
+ "loss": 0.4056,
285179
+ "step": 146045
285180
+ },
285181
+ {
285182
+ "epoch": 1177.59,
285183
+ "learning_rate": 7.65762520193861e-06,
285184
+ "loss": 0.8064,
285185
+ "step": 146050
285186
+ },
285187
+ {
285188
+ "epoch": 1177.63,
285189
+ "learning_rate": 7.657544426494346e-06,
285190
+ "loss": 0.6026,
285191
+ "step": 146055
285192
+ },
285193
+ {
285194
+ "epoch": 1177.67,
285195
+ "learning_rate": 7.65746365105008e-06,
285196
+ "loss": 0.2731,
285197
+ "step": 146060
285198
+ },
285199
+ {
285200
+ "epoch": 1177.71,
285201
+ "learning_rate": 7.657382875605816e-06,
285202
+ "loss": 0.2761,
285203
+ "step": 146065
285204
+ },
285205
+ {
285206
+ "epoch": 1177.76,
285207
+ "learning_rate": 7.65730210016155e-06,
285208
+ "loss": 0.3563,
285209
+ "step": 146070
285210
+ },
285211
+ {
285212
+ "epoch": 1177.8,
285213
+ "learning_rate": 7.657221324717286e-06,
285214
+ "loss": 0.8102,
285215
+ "step": 146075
285216
+ },
285217
+ {
285218
+ "epoch": 1177.84,
285219
+ "learning_rate": 7.65714054927302e-06,
285220
+ "loss": 0.5813,
285221
+ "step": 146080
285222
+ },
285223
+ {
285224
+ "epoch": 1177.88,
285225
+ "learning_rate": 7.657059773828756e-06,
285226
+ "loss": 0.3342,
285227
+ "step": 146085
285228
+ },
285229
+ {
285230
+ "epoch": 1177.92,
285231
+ "learning_rate": 7.656978998384492e-06,
285232
+ "loss": 0.3355,
285233
+ "step": 146090
285234
+ },
285235
+ {
285236
+ "epoch": 1177.96,
285237
+ "learning_rate": 7.656898222940226e-06,
285238
+ "loss": 0.4923,
285239
+ "step": 146095
285240
+ },
285241
+ {
285242
+ "epoch": 1178.0,
285243
+ "learning_rate": 7.656817447495962e-06,
285244
+ "loss": 1.2601,
285245
+ "step": 146100
285246
+ },
285247
+ {
285248
+ "epoch": 1178.0,
285249
+ "eval_loss": 0.4012053608894348,
285250
+ "eval_runtime": 40.577,
285251
+ "eval_samples_per_second": 20.627,
285252
+ "eval_steps_per_second": 0.665,
285253
+ "eval_wer": 0.1786023678891135,
285254
+ "step": 146100
285255
+ },
285256
+ {
285257
+ "epoch": 1178.04,
285258
+ "learning_rate": 7.656736672051696e-06,
285259
+ "loss": 0.2932,
285260
+ "step": 146105
285261
+ },
285262
+ {
285263
+ "epoch": 1178.08,
285264
+ "learning_rate": 7.656655896607432e-06,
285265
+ "loss": 0.2603,
285266
+ "step": 146110
285267
+ },
285268
+ {
285269
+ "epoch": 1178.12,
285270
+ "learning_rate": 7.656575121163166e-06,
285271
+ "loss": 0.2993,
285272
+ "step": 146115
285273
+ },
285274
+ {
285275
+ "epoch": 1178.16,
285276
+ "learning_rate": 7.656494345718902e-06,
285277
+ "loss": 0.4776,
285278
+ "step": 146120
285279
+ },
285280
+ {
285281
+ "epoch": 1178.2,
285282
+ "learning_rate": 7.656413570274636e-06,
285283
+ "loss": 1.1354,
285284
+ "step": 146125
285285
+ },
285286
+ {
285287
+ "epoch": 1178.24,
285288
+ "learning_rate": 7.656332794830372e-06,
285289
+ "loss": 0.2897,
285290
+ "step": 146130
285291
+ },
285292
+ {
285293
+ "epoch": 1178.28,
285294
+ "learning_rate": 7.656252019386106e-06,
285295
+ "loss": 0.2256,
285296
+ "step": 146135
285297
+ },
285298
+ {
285299
+ "epoch": 1178.32,
285300
+ "learning_rate": 7.656171243941842e-06,
285301
+ "loss": 0.2845,
285302
+ "step": 146140
285303
+ },
285304
+ {
285305
+ "epoch": 1178.36,
285306
+ "learning_rate": 7.656090468497578e-06,
285307
+ "loss": 0.4063,
285308
+ "step": 146145
285309
+ },
285310
+ {
285311
+ "epoch": 1178.4,
285312
+ "learning_rate": 7.656009693053312e-06,
285313
+ "loss": 1.1799,
285314
+ "step": 146150
285315
+ },
285316
+ {
285317
+ "epoch": 1178.44,
285318
+ "learning_rate": 7.655928917609048e-06,
285319
+ "loss": 0.3141,
285320
+ "step": 146155
285321
+ },
285322
+ {
285323
+ "epoch": 1178.48,
285324
+ "learning_rate": 7.655848142164782e-06,
285325
+ "loss": 0.2586,
285326
+ "step": 146160
285327
+ },
285328
+ {
285329
+ "epoch": 1178.52,
285330
+ "learning_rate": 7.655767366720518e-06,
285331
+ "loss": 0.3502,
285332
+ "step": 146165
285333
+ },
285334
+ {
285335
+ "epoch": 1178.56,
285336
+ "learning_rate": 7.655686591276252e-06,
285337
+ "loss": 0.4228,
285338
+ "step": 146170
285339
+ },
285340
+ {
285341
+ "epoch": 1178.6,
285342
+ "learning_rate": 7.655605815831988e-06,
285343
+ "loss": 1.2601,
285344
+ "step": 146175
285345
+ },
285346
+ {
285347
+ "epoch": 1178.64,
285348
+ "learning_rate": 7.655525040387722e-06,
285349
+ "loss": 0.2679,
285350
+ "step": 146180
285351
+ },
285352
+ {
285353
+ "epoch": 1178.68,
285354
+ "learning_rate": 7.655444264943458e-06,
285355
+ "loss": 0.2357,
285356
+ "step": 146185
285357
+ },
285358
+ {
285359
+ "epoch": 1178.72,
285360
+ "learning_rate": 7.655363489499192e-06,
285361
+ "loss": 0.3045,
285362
+ "step": 146190
285363
+ },
285364
+ {
285365
+ "epoch": 1178.76,
285366
+ "learning_rate": 7.655282714054928e-06,
285367
+ "loss": 0.5178,
285368
+ "step": 146195
285369
+ },
285370
+ {
285371
+ "epoch": 1178.8,
285372
+ "learning_rate": 7.655201938610663e-06,
285373
+ "loss": 1.2614,
285374
+ "step": 146200
285375
+ },
285376
+ {
285377
+ "epoch": 1178.84,
285378
+ "learning_rate": 7.655121163166397e-06,
285379
+ "loss": 0.3759,
285380
+ "step": 146205
285381
+ },
285382
+ {
285383
+ "epoch": 1178.88,
285384
+ "learning_rate": 7.655040387722133e-06,
285385
+ "loss": 0.2394,
285386
+ "step": 146210
285387
+ },
285388
+ {
285389
+ "epoch": 1178.92,
285390
+ "learning_rate": 7.654959612277867e-06,
285391
+ "loss": 0.275,
285392
+ "step": 146215
285393
+ },
285394
+ {
285395
+ "epoch": 1178.96,
285396
+ "learning_rate": 7.654878836833603e-06,
285397
+ "loss": 0.5791,
285398
+ "step": 146220
285399
+ },
285400
+ {
285401
+ "epoch": 1179.0,
285402
+ "eval_loss": 0.358010470867157,
285403
+ "eval_runtime": 40.5201,
285404
+ "eval_samples_per_second": 20.656,
285405
+ "eval_steps_per_second": 0.666,
285406
+ "eval_wer": 0.176530911221799,
285407
+ "step": 146224
285408
+ },
285409
+ {
285410
+ "epoch": 1179.01,
285411
+ "learning_rate": 7.654798061389337e-06,
285412
+ "loss": 0.5013,
285413
+ "step": 146225
285414
+ },
285415
+ {
285416
+ "epoch": 1179.05,
285417
+ "learning_rate": 7.654717285945073e-06,
285418
+ "loss": 0.2654,
285419
+ "step": 146230
285420
+ },
285421
+ {
285422
+ "epoch": 1179.09,
285423
+ "learning_rate": 7.654636510500807e-06,
285424
+ "loss": 0.2647,
285425
+ "step": 146235
285426
+ },
285427
+ {
285428
+ "epoch": 1179.13,
285429
+ "learning_rate": 7.654555735056543e-06,
285430
+ "loss": 0.3248,
285431
+ "step": 146240
285432
+ },
285433
+ {
285434
+ "epoch": 1179.17,
285435
+ "learning_rate": 7.654474959612277e-06,
285436
+ "loss": 0.4694,
285437
+ "step": 146245
285438
+ },
285439
+ {
285440
+ "epoch": 1179.21,
285441
+ "learning_rate": 7.654394184168013e-06,
285442
+ "loss": 1.0148,
285443
+ "step": 146250
285444
+ },
285445
+ {
285446
+ "epoch": 1179.25,
285447
+ "learning_rate": 7.654313408723747e-06,
285448
+ "loss": 0.2833,
285449
+ "step": 146255
285450
+ },
285451
+ {
285452
+ "epoch": 1179.29,
285453
+ "learning_rate": 7.654232633279483e-06,
285454
+ "loss": 0.2983,
285455
+ "step": 146260
285456
+ },
285457
+ {
285458
+ "epoch": 1179.33,
285459
+ "learning_rate": 7.654151857835219e-06,
285460
+ "loss": 0.3522,
285461
+ "step": 146265
285462
+ },
285463
+ {
285464
+ "epoch": 1179.37,
285465
+ "learning_rate": 7.654071082390953e-06,
285466
+ "loss": 0.5271,
285467
+ "step": 146270
285468
+ },
285469
+ {
285470
+ "epoch": 1179.41,
285471
+ "learning_rate": 7.653990306946689e-06,
285472
+ "loss": 1.1747,
285473
+ "step": 146275
285474
+ },
285475
+ {
285476
+ "epoch": 1179.45,
285477
+ "learning_rate": 7.653909531502423e-06,
285478
+ "loss": 0.2796,
285479
+ "step": 146280
285480
+ },
285481
+ {
285482
+ "epoch": 1179.49,
285483
+ "learning_rate": 7.653828756058159e-06,
285484
+ "loss": 0.2518,
285485
+ "step": 146285
285486
+ },
285487
+ {
285488
+ "epoch": 1179.53,
285489
+ "learning_rate": 7.653747980613893e-06,
285490
+ "loss": 0.307,
285491
+ "step": 146290
285492
+ },
285493
+ {
285494
+ "epoch": 1179.57,
285495
+ "learning_rate": 7.653667205169629e-06,
285496
+ "loss": 0.4902,
285497
+ "step": 146295
285498
+ },
285499
+ {
285500
+ "epoch": 1179.61,
285501
+ "learning_rate": 7.653586429725363e-06,
285502
+ "loss": 1.0019,
285503
+ "step": 146300
285504
+ },
285505
+ {
285506
+ "epoch": 1179.65,
285507
+ "learning_rate": 7.653505654281099e-06,
285508
+ "loss": 0.278,
285509
+ "step": 146305
285510
+ },
285511
+ {
285512
+ "epoch": 1179.69,
285513
+ "learning_rate": 7.653424878836833e-06,
285514
+ "loss": 0.3384,
285515
+ "step": 146310
285516
+ },
285517
+ {
285518
+ "epoch": 1179.73,
285519
+ "learning_rate": 7.653344103392569e-06,
285520
+ "loss": 0.3353,
285521
+ "step": 146315
285522
+ },
285523
+ {
285524
+ "epoch": 1179.77,
285525
+ "learning_rate": 7.653263327948305e-06,
285526
+ "loss": 0.5559,
285527
+ "step": 146320
285528
+ },
285529
+ {
285530
+ "epoch": 1179.81,
285531
+ "learning_rate": 7.653182552504039e-06,
285532
+ "loss": 1.4385,
285533
+ "step": 146325
285534
+ },
285535
+ {
285536
+ "epoch": 1179.85,
285537
+ "learning_rate": 7.653101777059775e-06,
285538
+ "loss": 0.2435,
285539
+ "step": 146330
285540
+ },
285541
+ {
285542
+ "epoch": 1179.9,
285543
+ "learning_rate": 7.653021001615509e-06,
285544
+ "loss": 0.2579,
285545
+ "step": 146335
285546
+ },
285547
+ {
285548
+ "epoch": 1179.94,
285549
+ "learning_rate": 7.652940226171245e-06,
285550
+ "loss": 0.345,
285551
+ "step": 146340
285552
+ },
285553
+ {
285554
+ "epoch": 1179.98,
285555
+ "learning_rate": 7.652859450726979e-06,
285556
+ "loss": 0.5621,
285557
+ "step": 146345
285558
+ },
285559
+ {
285560
+ "epoch": 1180.0,
285561
+ "eval_loss": 0.449453204870224,
285562
+ "eval_runtime": 40.0278,
285563
+ "eval_samples_per_second": 20.91,
285564
+ "eval_steps_per_second": 0.675,
285565
+ "eval_wer": 0.18728573557907288,
285566
+ "step": 146348
285567
+ },
285568
+ {
285569
+ "epoch": 1180.02,
285570
+ "learning_rate": 7.652778675282715e-06,
285571
+ "loss": 0.2882,
285572
+ "step": 146350
285573
+ },
285574
+ {
285575
+ "epoch": 1180.06,
285576
+ "learning_rate": 7.652697899838449e-06,
285577
+ "loss": 0.2908,
285578
+ "step": 146355
285579
+ },
285580
+ {
285581
+ "epoch": 1180.1,
285582
+ "learning_rate": 7.652617124394185e-06,
285583
+ "loss": 0.2317,
285584
+ "step": 146360
285585
+ },
285586
+ {
285587
+ "epoch": 1180.14,
285588
+ "learning_rate": 7.652536348949919e-06,
285589
+ "loss": 0.3416,
285590
+ "step": 146365
285591
+ },
285592
+ {
285593
+ "epoch": 1180.18,
285594
+ "learning_rate": 7.652455573505655e-06,
285595
+ "loss": 0.6812,
285596
+ "step": 146370
285597
+ },
285598
+ {
285599
+ "epoch": 1180.22,
285600
+ "learning_rate": 7.65237479806139e-06,
285601
+ "loss": 0.8339,
285602
+ "step": 146375
285603
+ },
285604
+ {
285605
+ "epoch": 1180.26,
285606
+ "learning_rate": 7.652294022617125e-06,
285607
+ "loss": 0.431,
285608
+ "step": 146380
285609
+ },
285610
+ {
285611
+ "epoch": 1180.3,
285612
+ "learning_rate": 7.65221324717286e-06,
285613
+ "loss": 0.3809,
285614
+ "step": 146385
285615
+ },
285616
+ {
285617
+ "epoch": 1180.34,
285618
+ "learning_rate": 7.652132471728595e-06,
285619
+ "loss": 0.3004,
285620
+ "step": 146390
285621
+ },
285622
+ {
285623
+ "epoch": 1180.38,
285624
+ "learning_rate": 7.65205169628433e-06,
285625
+ "loss": 0.578,
285626
+ "step": 146395
285627
+ },
285628
+ {
285629
+ "epoch": 1180.42,
285630
+ "learning_rate": 7.651970920840065e-06,
285631
+ "loss": 0.9797,
285632
+ "step": 146400
285633
+ },
285634
+ {
285635
+ "epoch": 1180.46,
285636
+ "learning_rate": 7.6518901453958e-06,
285637
+ "loss": 0.2545,
285638
+ "step": 146405
285639
+ },
285640
+ {
285641
+ "epoch": 1180.5,
285642
+ "learning_rate": 7.651809369951535e-06,
285643
+ "loss": 0.2621,
285644
+ "step": 146410
285645
+ },
285646
+ {
285647
+ "epoch": 1180.54,
285648
+ "learning_rate": 7.65172859450727e-06,
285649
+ "loss": 0.3193,
285650
+ "step": 146415
285651
+ },
285652
+ {
285653
+ "epoch": 1180.58,
285654
+ "learning_rate": 7.651647819063005e-06,
285655
+ "loss": 0.6539,
285656
+ "step": 146420
285657
+ },
285658
+ {
285659
+ "epoch": 1180.62,
285660
+ "learning_rate": 7.65156704361874e-06,
285661
+ "loss": 1.0436,
285662
+ "step": 146425
285663
+ },
285664
+ {
285665
+ "epoch": 1180.66,
285666
+ "learning_rate": 7.651486268174474e-06,
285667
+ "loss": 0.2563,
285668
+ "step": 146430
285669
+ },
285670
+ {
285671
+ "epoch": 1180.7,
285672
+ "learning_rate": 7.65140549273021e-06,
285673
+ "loss": 0.3142,
285674
+ "step": 146435
285675
+ },
285676
+ {
285677
+ "epoch": 1180.74,
285678
+ "learning_rate": 7.651324717285946e-06,
285679
+ "loss": 0.3735,
285680
+ "step": 146440
285681
+ },
285682
+ {
285683
+ "epoch": 1180.78,
285684
+ "learning_rate": 7.65124394184168e-06,
285685
+ "loss": 0.6211,
285686
+ "step": 146445
285687
+ },
285688
+ {
285689
+ "epoch": 1180.82,
285690
+ "learning_rate": 7.651163166397416e-06,
285691
+ "loss": 1.0659,
285692
+ "step": 146450
285693
+ },
285694
+ {
285695
+ "epoch": 1180.86,
285696
+ "learning_rate": 7.65108239095315e-06,
285697
+ "loss": 0.3007,
285698
+ "step": 146455
285699
+ },
285700
+ {
285701
+ "epoch": 1180.9,
285702
+ "learning_rate": 7.651001615508886e-06,
285703
+ "loss": 0.2761,
285704
+ "step": 146460
285705
+ },
285706
+ {
285707
+ "epoch": 1180.94,
285708
+ "learning_rate": 7.65092084006462e-06,
285709
+ "loss": 0.3487,
285710
+ "step": 146465
285711
+ },
285712
+ {
285713
+ "epoch": 1180.98,
285714
+ "learning_rate": 7.650840064620356e-06,
285715
+ "loss": 0.7013,
285716
+ "step": 146470
285717
+ },
285718
+ {
285719
+ "epoch": 1181.0,
285720
+ "eval_loss": 0.34985774755477905,
285721
+ "eval_runtime": 41.8541,
285722
+ "eval_samples_per_second": 19.998,
285723
+ "eval_steps_per_second": 0.645,
285724
+ "eval_wer": 0.1758169934640523,
285725
+ "step": 146472
285726
+ },
285727
+ {
285728
+ "epoch": 1181.02,
285729
+ "learning_rate": 7.65075928917609e-06,
285730
+ "loss": 0.306,
285731
+ "step": 146475
285732
+ },
285733
+ {
285734
+ "epoch": 1181.06,
285735
+ "learning_rate": 7.650678513731826e-06,
285736
+ "loss": 0.3149,
285737
+ "step": 146480
285738
+ },
285739
+ {
285740
+ "epoch": 1181.1,
285741
+ "learning_rate": 7.65059773828756e-06,
285742
+ "loss": 0.2999,
285743
+ "step": 146485
285744
+ },
285745
+ {
285746
+ "epoch": 1181.14,
285747
+ "learning_rate": 7.650516962843296e-06,
285748
+ "loss": 0.3347,
285749
+ "step": 146490
285750
+ },
285751
+ {
285752
+ "epoch": 1181.18,
285753
+ "learning_rate": 7.650436187399032e-06,
285754
+ "loss": 0.6613,
285755
+ "step": 146495
285756
+ },
285757
+ {
285758
+ "epoch": 1181.22,
285759
+ "learning_rate": 7.650355411954766e-06,
285760
+ "loss": 0.8353,
285761
+ "step": 146500
285762
+ },
285763
+ {
285764
+ "epoch": 1181.27,
285765
+ "learning_rate": 7.650274636510502e-06,
285766
+ "loss": 0.3035,
285767
+ "step": 146505
285768
+ },
285769
+ {
285770
+ "epoch": 1181.31,
285771
+ "learning_rate": 7.650193861066236e-06,
285772
+ "loss": 0.3257,
285773
+ "step": 146510
285774
+ },
285775
+ {
285776
+ "epoch": 1181.35,
285777
+ "learning_rate": 7.650113085621972e-06,
285778
+ "loss": 0.4288,
285779
+ "step": 146515
285780
+ },
285781
+ {
285782
+ "epoch": 1181.39,
285783
+ "learning_rate": 7.650032310177706e-06,
285784
+ "loss": 0.6413,
285785
+ "step": 146520
285786
+ },
285787
+ {
285788
+ "epoch": 1181.43,
285789
+ "learning_rate": 7.649951534733442e-06,
285790
+ "loss": 0.8146,
285791
+ "step": 146525
285792
+ },
285793
+ {
285794
+ "epoch": 1181.47,
285795
+ "learning_rate": 7.649870759289176e-06,
285796
+ "loss": 0.2767,
285797
+ "step": 146530
285798
+ },
285799
+ {
285800
+ "epoch": 1181.51,
285801
+ "learning_rate": 7.649789983844912e-06,
285802
+ "loss": 0.2579,
285803
+ "step": 146535
285804
+ },
285805
+ {
285806
+ "epoch": 1181.55,
285807
+ "learning_rate": 7.649709208400646e-06,
285808
+ "loss": 0.3754,
285809
+ "step": 146540
285810
+ },
285811
+ {
285812
+ "epoch": 1181.59,
285813
+ "learning_rate": 7.649628432956382e-06,
285814
+ "loss": 0.7602,
285815
+ "step": 146545
285816
+ },
285817
+ {
285818
+ "epoch": 1181.63,
285819
+ "learning_rate": 7.649547657512118e-06,
285820
+ "loss": 0.879,
285821
+ "step": 146550
285822
+ },
285823
+ {
285824
+ "epoch": 1181.67,
285825
+ "learning_rate": 7.649466882067852e-06,
285826
+ "loss": 0.2737,
285827
+ "step": 146555
285828
+ },
285829
+ {
285830
+ "epoch": 1181.71,
285831
+ "learning_rate": 7.649386106623588e-06,
285832
+ "loss": 0.2831,
285833
+ "step": 146560
285834
+ },
285835
+ {
285836
+ "epoch": 1181.75,
285837
+ "learning_rate": 7.649305331179322e-06,
285838
+ "loss": 0.3641,
285839
+ "step": 146565
285840
+ },
285841
+ {
285842
+ "epoch": 1181.79,
285843
+ "learning_rate": 7.649224555735058e-06,
285844
+ "loss": 0.7179,
285845
+ "step": 146570
285846
+ },
285847
+ {
285848
+ "epoch": 1181.83,
285849
+ "learning_rate": 7.649143780290792e-06,
285850
+ "loss": 0.8389,
285851
+ "step": 146575
285852
+ },
285853
+ {
285854
+ "epoch": 1181.87,
285855
+ "learning_rate": 7.649063004846527e-06,
285856
+ "loss": 0.2527,
285857
+ "step": 146580
285858
+ },
285859
+ {
285860
+ "epoch": 1181.91,
285861
+ "learning_rate": 7.648982229402262e-06,
285862
+ "loss": 0.2547,
285863
+ "step": 146585
285864
+ },
285865
+ {
285866
+ "epoch": 1181.95,
285867
+ "learning_rate": 7.648901453957997e-06,
285868
+ "loss": 0.3927,
285869
+ "step": 146590
285870
+ },
285871
+ {
285872
+ "epoch": 1181.99,
285873
+ "learning_rate": 7.648820678513732e-06,
285874
+ "loss": 0.8367,
285875
+ "step": 146595
285876
+ },
285877
+ {
285878
+ "epoch": 1182.0,
285879
+ "eval_loss": 0.34015288949012756,
285880
+ "eval_runtime": 41.2224,
285881
+ "eval_samples_per_second": 20.304,
285882
+ "eval_steps_per_second": 0.655,
285883
+ "eval_wer": 0.1668011442822563,
285884
+ "step": 146596
285885
+ },
285886
+ {
285887
+ "epoch": 1172.03,
285888
+ "learning_rate": 7.648739903069467e-06,
285889
+ "loss": 0.3097,
285890
+ "step": 146600
285891
+ },
285892
+ {
285893
+ "epoch": 1172.07,
285894
+ "learning_rate": 7.648659127625202e-06,
285895
+ "loss": 0.2985,
285896
+ "step": 146605
285897
+ },
285898
+ {
285899
+ "epoch": 1172.11,
285900
+ "learning_rate": 7.648578352180937e-06,
285901
+ "loss": 0.2698,
285902
+ "step": 146610
285903
+ },
285904
+ {
285905
+ "epoch": 1172.15,
285906
+ "learning_rate": 7.648497576736673e-06,
285907
+ "loss": 0.3778,
285908
+ "step": 146615
285909
+ },
285910
+ {
285911
+ "epoch": 1172.19,
285912
+ "learning_rate": 7.648416801292407e-06,
285913
+ "loss": 1.0694,
285914
+ "step": 146620
285915
+ },
285916
+ {
285917
+ "epoch": 1172.23,
285918
+ "learning_rate": 7.648336025848143e-06,
285919
+ "loss": 0.5919,
285920
+ "step": 146625
285921
+ },
285922
+ {
285923
+ "epoch": 1172.27,
285924
+ "learning_rate": 7.648255250403877e-06,
285925
+ "loss": 0.2701,
285926
+ "step": 146630
285927
+ },
285928
+ {
285929
+ "epoch": 1172.31,
285930
+ "learning_rate": 7.648174474959613e-06,
285931
+ "loss": 0.2998,
285932
+ "step": 146635
285933
+ },
285934
+ {
285935
+ "epoch": 1172.35,
285936
+ "learning_rate": 7.648093699515347e-06,
285937
+ "loss": 0.3737,
285938
+ "step": 146640
285939
+ },
285940
+ {
285941
+ "epoch": 1172.39,
285942
+ "learning_rate": 7.648012924071083e-06,
285943
+ "loss": 1.0077,
285944
+ "step": 146645
285945
+ },
285946
+ {
285947
+ "epoch": 1172.43,
285948
+ "learning_rate": 7.647932148626817e-06,
285949
+ "loss": 0.6127,
285950
+ "step": 146650
285951
+ },
285952
+ {
285953
+ "epoch": 1172.47,
285954
+ "learning_rate": 7.647851373182553e-06,
285955
+ "loss": 0.3308,
285956
+ "step": 146655
285957
+ },
285958
+ {
285959
+ "epoch": 1172.51,
285960
+ "learning_rate": 7.647770597738287e-06,
285961
+ "loss": 0.312,
285962
+ "step": 146660
285963
+ },
285964
+ {
285965
+ "epoch": 1172.55,
285966
+ "learning_rate": 7.647689822294023e-06,
285967
+ "loss": 0.4248,
285968
+ "step": 146665
285969
+ },
285970
+ {
285971
+ "epoch": 1172.59,
285972
+ "learning_rate": 7.647609046849759e-06,
285973
+ "loss": 0.8503,
285974
+ "step": 146670
285975
+ },
285976
+ {
285977
+ "epoch": 1172.63,
285978
+ "learning_rate": 7.647528271405493e-06,
285979
+ "loss": 0.7404,
285980
+ "step": 146675
285981
+ },
285982
+ {
285983
+ "epoch": 1172.67,
285984
+ "learning_rate": 7.647447495961229e-06,
285985
+ "loss": 0.3356,
285986
+ "step": 146680
285987
+ },
285988
+ {
285989
+ "epoch": 1172.71,
285990
+ "learning_rate": 7.647366720516963e-06,
285991
+ "loss": 0.2776,
285992
+ "step": 146685
285993
+ },
285994
+ {
285995
+ "epoch": 1172.75,
285996
+ "learning_rate": 7.647285945072699e-06,
285997
+ "loss": 0.4439,
285998
+ "step": 146690
285999
+ },
286000
+ {
286001
+ "epoch": 1172.79,
286002
+ "learning_rate": 7.647205169628433e-06,
286003
+ "loss": 0.957,
286004
+ "step": 146695
286005
+ },
286006
+ {
286007
+ "epoch": 1172.83,
286008
+ "learning_rate": 7.647124394184169e-06,
286009
+ "loss": 0.5475,
286010
+ "step": 146700
286011
+ },
286012
+ {
286013
+ "epoch": 1172.87,
286014
+ "learning_rate": 7.647043618739903e-06,
286015
+ "loss": 0.2893,
286016
+ "step": 146705
286017
+ },
286018
+ {
286019
+ "epoch": 1172.91,
286020
+ "learning_rate": 7.646962843295639e-06,
286021
+ "loss": 0.2689,
286022
+ "step": 146710
286023
+ },
286024
+ {
286025
+ "epoch": 1172.95,
286026
+ "learning_rate": 7.646882067851373e-06,
286027
+ "loss": 0.3002,
286028
+ "step": 146715
286029
+ },
286030
+ {
286031
+ "epoch": 1172.99,
286032
+ "learning_rate": 7.646801292407109e-06,
286033
+ "loss": 0.9223,
286034
+ "step": 146720
286035
+ },
286036
+ {
286037
+ "epoch": 1173.0,
286038
+ "eval_loss": 0.3900996744632721,
286039
+ "eval_runtime": 41.2507,
286040
+ "eval_samples_per_second": 20.291,
286041
+ "eval_steps_per_second": 0.655,
286042
+ "eval_wer": 0.1805263539476546,
286043
+ "step": 146721
286044
+ },
286045
+ {
286046
+ "epoch": 1183.03,
286047
+ "learning_rate": 7.646720516962845e-06,
286048
+ "loss": 0.3527,
286049
+ "step": 146725
286050
+ },
286051
+ {
286052
+ "epoch": 1183.07,
286053
+ "learning_rate": 7.646639741518579e-06,
286054
+ "loss": 0.2931,
286055
+ "step": 146730
286056
+ },
286057
+ {
286058
+ "epoch": 1183.11,
286059
+ "learning_rate": 7.646558966074315e-06,
286060
+ "loss": 0.2955,
286061
+ "step": 146735
286062
+ },
286063
+ {
286064
+ "epoch": 1183.15,
286065
+ "learning_rate": 7.646478190630049e-06,
286066
+ "loss": 0.3639,
286067
+ "step": 146740
286068
+ },
286069
+ {
286070
+ "epoch": 1183.19,
286071
+ "learning_rate": 7.646397415185785e-06,
286072
+ "loss": 0.8325,
286073
+ "step": 146745
286074
+ },
286075
+ {
286076
+ "epoch": 1183.23,
286077
+ "learning_rate": 7.646316639741519e-06,
286078
+ "loss": 0.5682,
286079
+ "step": 146750
286080
+ },
286081
+ {
286082
+ "epoch": 1183.27,
286083
+ "learning_rate": 7.646235864297255e-06,
286084
+ "loss": 0.3174,
286085
+ "step": 146755
286086
+ },
286087
+ {
286088
+ "epoch": 1183.31,
286089
+ "learning_rate": 7.646155088852989e-06,
286090
+ "loss": 0.3424,
286091
+ "step": 146760
286092
+ },
286093
+ {
286094
+ "epoch": 1183.35,
286095
+ "learning_rate": 7.646074313408725e-06,
286096
+ "loss": 0.3159,
286097
+ "step": 146765
286098
+ },
286099
+ {
286100
+ "epoch": 1183.39,
286101
+ "learning_rate": 7.645993537964459e-06,
286102
+ "loss": 0.7583,
286103
+ "step": 146770
286104
+ },
286105
+ {
286106
+ "epoch": 1183.43,
286107
+ "learning_rate": 7.645912762520195e-06,
286108
+ "loss": 0.5913,
286109
+ "step": 146775
286110
+ },
286111
+ {
286112
+ "epoch": 1183.47,
286113
+ "learning_rate": 7.645831987075929e-06,
286114
+ "loss": 0.2356,
286115
+ "step": 146780
286116
+ },
286117
+ {
286118
+ "epoch": 1183.51,
286119
+ "learning_rate": 7.645751211631665e-06,
286120
+ "loss": 0.3036,
286121
+ "step": 146785
286122
+ },
286123
+ {
286124
+ "epoch": 1183.55,
286125
+ "learning_rate": 7.6456704361874e-06,
286126
+ "loss": 0.4368,
286127
+ "step": 146790
286128
+ },
286129
+ {
286130
+ "epoch": 1183.59,
286131
+ "learning_rate": 7.645589660743134e-06,
286132
+ "loss": 0.8289,
286133
+ "step": 146795
286134
+ },
286135
+ {
286136
+ "epoch": 1183.63,
286137
+ "learning_rate": 7.64550888529887e-06,
286138
+ "loss": 0.6256,
286139
+ "step": 146800
286140
+ },
286141
+ {
286142
+ "epoch": 1183.67,
286143
+ "learning_rate": 7.645428109854604e-06,
286144
+ "loss": 0.2853,
286145
+ "step": 146805
286146
+ },
286147
+ {
286148
+ "epoch": 1183.71,
286149
+ "learning_rate": 7.64534733441034e-06,
286150
+ "loss": 0.3083,
286151
+ "step": 146810
286152
+ },
286153
+ {
286154
+ "epoch": 1183.76,
286155
+ "learning_rate": 7.645266558966074e-06,
286156
+ "loss": 0.4051,
286157
+ "step": 146815
286158
+ },
286159
+ {
286160
+ "epoch": 1183.8,
286161
+ "learning_rate": 7.64518578352181e-06,
286162
+ "loss": 1.006,
286163
+ "step": 146820
286164
+ },
286165
+ {
286166
+ "epoch": 1183.84,
286167
+ "learning_rate": 7.645105008077544e-06,
286168
+ "loss": 0.7374,
286169
+ "step": 146825
286170
+ },
286171
+ {
286172
+ "epoch": 1183.88,
286173
+ "learning_rate": 7.64502423263328e-06,
286174
+ "loss": 0.3358,
286175
+ "step": 146830
286176
+ },
286177
+ {
286178
+ "epoch": 1183.92,
286179
+ "learning_rate": 7.644943457189014e-06,
286180
+ "loss": 0.2254,
286181
+ "step": 146835
286182
+ },
286183
+ {
286184
+ "epoch": 1183.96,
286185
+ "learning_rate": 7.64486268174475e-06,
286186
+ "loss": 0.4519,
286187
+ "step": 146840
286188
+ },
286189
+ {
286190
+ "epoch": 1184.0,
286191
+ "learning_rate": 7.644781906300486e-06,
286192
+ "loss": 1.0456,
286193
+ "step": 146845
286194
+ },
286195
+ {
286196
+ "epoch": 1184.0,
286197
+ "eval_loss": 0.3610660433769226,
286198
+ "eval_runtime": 41.7394,
286199
+ "eval_samples_per_second": 20.053,
286200
+ "eval_steps_per_second": 0.647,
286201
+ "eval_wer": 0.16990750855727915,
286202
+ "step": 146845
286203
+ },
286204
+ {
286205
+ "epoch": 1174.04,
286206
+ "learning_rate": 7.64470113085622e-06,
286207
+ "loss": 0.2998,
286208
+ "step": 146850
286209
+ },
286210
+ {
286211
+ "epoch": 1174.08,
286212
+ "learning_rate": 7.644620355411956e-06,
286213
+ "loss": 0.369,
286214
+ "step": 146855
286215
+ },
286216
+ {
286217
+ "epoch": 1174.12,
286218
+ "learning_rate": 7.64453957996769e-06,
286219
+ "loss": 0.3144,
286220
+ "step": 146860
286221
+ },
286222
+ {
286223
+ "epoch": 1174.16,
286224
+ "learning_rate": 7.644458804523426e-06,
286225
+ "loss": 0.4513,
286226
+ "step": 146865
286227
+ },
286228
+ {
286229
+ "epoch": 1174.2,
286230
+ "learning_rate": 7.64437802907916e-06,
286231
+ "loss": 1.1218,
286232
+ "step": 146870
286233
+ },
286234
+ {
286235
+ "epoch": 1174.24,
286236
+ "learning_rate": 7.644297253634896e-06,
286237
+ "loss": 0.2873,
286238
+ "step": 146875
286239
+ },
286240
+ {
286241
+ "epoch": 1174.28,
286242
+ "learning_rate": 7.64421647819063e-06,
286243
+ "loss": 0.2713,
286244
+ "step": 146880
286245
+ },
286246
+ {
286247
+ "epoch": 1174.32,
286248
+ "learning_rate": 7.644135702746366e-06,
286249
+ "loss": 0.2787,
286250
+ "step": 146885
286251
+ },
286252
+ {
286253
+ "epoch": 1174.36,
286254
+ "learning_rate": 7.6440549273021e-06,
286255
+ "loss": 0.4724,
286256
+ "step": 146890
286257
+ },
286258
+ {
286259
+ "epoch": 1174.4,
286260
+ "learning_rate": 7.643974151857836e-06,
286261
+ "loss": 1.149,
286262
+ "step": 146895
286263
+ },
286264
+ {
286265
+ "epoch": 1174.44,
286266
+ "learning_rate": 7.643893376413572e-06,
286267
+ "loss": 0.3304,
286268
+ "step": 146900
286269
+ },
286270
+ {
286271
+ "epoch": 1174.48,
286272
+ "learning_rate": 7.643812600969306e-06,
286273
+ "loss": 0.2587,
286274
+ "step": 146905
286275
+ },
286276
+ {
286277
+ "epoch": 1174.52,
286278
+ "learning_rate": 7.643731825525042e-06,
286279
+ "loss": 0.2845,
286280
+ "step": 146910
286281
+ },
286282
+ {
286283
+ "epoch": 1174.56,
286284
+ "learning_rate": 7.643651050080776e-06,
286285
+ "loss": 0.4071,
286286
+ "step": 146915
286287
+ },
286288
+ {
286289
+ "epoch": 1174.6,
286290
+ "learning_rate": 7.643570274636512e-06,
286291
+ "loss": 1.1262,
286292
+ "step": 146920
286293
+ },
286294
+ {
286295
+ "epoch": 1174.64,
286296
+ "learning_rate": 7.643489499192246e-06,
286297
+ "loss": 0.3429,
286298
+ "step": 146925
286299
+ },
286300
+ {
286301
+ "epoch": 1174.68,
286302
+ "learning_rate": 7.643408723747982e-06,
286303
+ "loss": 0.3127,
286304
+ "step": 146930
286305
+ },
286306
+ {
286307
+ "epoch": 1174.72,
286308
+ "learning_rate": 7.643327948303716e-06,
286309
+ "loss": 0.2756,
286310
+ "step": 146935
286311
+ },
286312
+ {
286313
+ "epoch": 1174.76,
286314
+ "learning_rate": 7.643247172859452e-06,
286315
+ "loss": 0.5054,
286316
+ "step": 146940
286317
+ },
286318
+ {
286319
+ "epoch": 1174.8,
286320
+ "learning_rate": 7.643166397415186e-06,
286321
+ "loss": 1.0996,
286322
+ "step": 146945
286323
+ },
286324
+ {
286325
+ "epoch": 1174.84,
286326
+ "learning_rate": 7.643085621970922e-06,
286327
+ "loss": 0.3266,
286328
+ "step": 146950
286329
+ },
286330
+ {
286331
+ "epoch": 1174.88,
286332
+ "learning_rate": 7.643004846526656e-06,
286333
+ "loss": 0.2945,
286334
+ "step": 146955
286335
+ },
286336
+ {
286337
+ "epoch": 1174.92,
286338
+ "learning_rate": 7.642924071082392e-06,
286339
+ "loss": 0.3572,
286340
+ "step": 146960
286341
+ },
286342
+ {
286343
+ "epoch": 1174.96,
286344
+ "learning_rate": 7.642843295638127e-06,
286345
+ "loss": 0.4849,
286346
+ "step": 146965
286347
+ },
286348
+ {
286349
+ "epoch": 1175.0,
286350
+ "learning_rate": 7.642762520193862e-06,
286351
+ "loss": 1.3137,
286352
+ "step": 146970
286353
+ },
286354
+ {
286355
+ "epoch": 1175.0,
286356
+ "eval_loss": 0.3988247811794281,
286357
+ "eval_runtime": 41.5097,
286358
+ "eval_samples_per_second": 20.164,
286359
+ "eval_steps_per_second": 0.65,
286360
+ "eval_wer": 0.1729923798358734,
286361
+ "step": 146970
286362
+ },
286363
+ {
286364
+ "epoch": 1185.04,
286365
+ "learning_rate": 7.642681744749597e-06,
286366
+ "loss": 0.3428,
286367
+ "step": 146975
286368
+ },
286369
+ {
286370
+ "epoch": 1185.08,
286371
+ "learning_rate": 7.642600969305332e-06,
286372
+ "loss": 0.2385,
286373
+ "step": 146980
286374
+ },
286375
+ {
286376
+ "epoch": 1185.12,
286377
+ "learning_rate": 7.642520193861067e-06,
286378
+ "loss": 0.3427,
286379
+ "step": 146985
286380
+ },
286381
+ {
286382
+ "epoch": 1185.16,
286383
+ "learning_rate": 7.642439418416802e-06,
286384
+ "loss": 0.3936,
286385
+ "step": 146990
286386
+ },
286387
+ {
286388
+ "epoch": 1185.2,
286389
+ "learning_rate": 7.642358642972537e-06,
286390
+ "loss": 1.0638,
286391
+ "step": 146995
286392
+ },
286393
+ {
286394
+ "epoch": 1185.24,
286395
+ "learning_rate": 7.642277867528272e-06,
286396
+ "loss": 0.304,
286397
+ "step": 147000
286398
+ },
286399
+ {
286400
+ "epoch": 1185.28,
286401
+ "learning_rate": 7.642197092084007e-06,
286402
+ "loss": 0.2536,
286403
+ "step": 147005
286404
+ },
286405
+ {
286406
+ "epoch": 1185.32,
286407
+ "learning_rate": 7.642116316639741e-06,
286408
+ "loss": 0.3188,
286409
+ "step": 147010
286410
+ },
286411
+ {
286412
+ "epoch": 1185.36,
286413
+ "learning_rate": 7.642035541195477e-06,
286414
+ "loss": 0.5025,
286415
+ "step": 147015
286416
+ },
286417
+ {
286418
+ "epoch": 1185.4,
286419
+ "learning_rate": 7.641954765751213e-06,
286420
+ "loss": 1.2492,
286421
+ "step": 147020
286422
+ },
286423
+ {
286424
+ "epoch": 1185.44,
286425
+ "learning_rate": 7.641873990306947e-06,
286426
+ "loss": 0.2905,
286427
+ "step": 147025
286428
+ },
286429
+ {
286430
+ "epoch": 1185.48,
286431
+ "learning_rate": 7.641793214862683e-06,
286432
+ "loss": 0.3111,
286433
+ "step": 147030
286434
+ },
286435
+ {
286436
+ "epoch": 1185.52,
286437
+ "learning_rate": 7.641712439418417e-06,
286438
+ "loss": 0.3018,
286439
+ "step": 147035
286440
+ },
286441
+ {
286442
+ "epoch": 1185.56,
286443
+ "learning_rate": 7.641631663974153e-06,
286444
+ "loss": 0.4451,
286445
+ "step": 147040
286446
+ },
286447
+ {
286448
+ "epoch": 1185.6,
286449
+ "learning_rate": 7.641550888529887e-06,
286450
+ "loss": 1.0976,
286451
+ "step": 147045
286452
+ },
286453
+ {
286454
+ "epoch": 1185.64,
286455
+ "learning_rate": 7.641470113085623e-06,
286456
+ "loss": 0.3226,
286457
+ "step": 147050
286458
+ },
286459
+ {
286460
+ "epoch": 1185.68,
286461
+ "learning_rate": 7.641389337641357e-06,
286462
+ "loss": 0.2418,
286463
+ "step": 147055
286464
+ },
286465
+ {
286466
+ "epoch": 1185.72,
286467
+ "learning_rate": 7.641308562197093e-06,
286468
+ "loss": 0.3772,
286469
+ "step": 147060
286470
+ },
286471
+ {
286472
+ "epoch": 1185.76,
286473
+ "learning_rate": 7.641227786752827e-06,
286474
+ "loss": 0.4226,
286475
+ "step": 147065
286476
+ },
286477
+ {
286478
+ "epoch": 1185.8,
286479
+ "learning_rate": 7.641147011308563e-06,
286480
+ "loss": 1.1949,
286481
+ "step": 147070
286482
+ },
286483
+ {
286484
+ "epoch": 1185.84,
286485
+ "learning_rate": 7.641066235864299e-06,
286486
+ "loss": 0.3455,
286487
+ "step": 147075
286488
+ },
286489
+ {
286490
+ "epoch": 1185.88,
286491
+ "learning_rate": 7.640985460420033e-06,
286492
+ "loss": 0.2664,
286493
+ "step": 147080
286494
+ },
286495
+ {
286496
+ "epoch": 1185.92,
286497
+ "learning_rate": 7.640904684975769e-06,
286498
+ "loss": 0.3259,
286499
+ "step": 147085
286500
+ },
286501
+ {
286502
+ "epoch": 1185.96,
286503
+ "learning_rate": 7.640823909531503e-06,
286504
+ "loss": 0.4639,
286505
+ "step": 147090
286506
+ },
286507
+ {
286508
+ "epoch": 1186.0,
286509
+ "eval_loss": 0.41105982661247253,
286510
+ "eval_runtime": 41.11,
286511
+ "eval_samples_per_second": 20.36,
286512
+ "eval_steps_per_second": 0.657,
286513
+ "eval_wer": 0.1767975743574935,
286514
+ "step": 147094
286515
  }
286516
  ],
286517
+ "max_steps": 620000,
286518
  "num_train_epochs": 5000,
286519
+ "total_flos": 4.1393121623086616e+20,
286520
  "trial_name": null,
286521
  "trial_params": null
286522
  }
model-bin/finetune/base/{checkpoint-146472 β†’ checkpoint-147094}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630230484.4148061/events.out.tfevents.1630230484.cc93b136ebf5.1086.119 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:373e03a0065b2e921eefd867e202f84f548054ff66160ac42f034369030f3220
3
+ size 4194
model-bin/finetune/base/log/1630230917.0829825/events.out.tfevents.1630230917.cc93b136ebf5.1086.121 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4941891e313e1a07773f5f32179f173fb50469156bae41861b1c5fdb30df5d6
3
+ size 4194
model-bin/finetune/base/log/1630231362.9249244/events.out.tfevents.1630231362.cc93b136ebf5.1086.123 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a30f575e5d92c7236fc30cd24fb7cea6c129bed326a9687dcf47af1671d3698
3
+ size 4194
model-bin/finetune/base/log/1630231799.7950127/events.out.tfevents.1630231799.cc93b136ebf5.1086.125 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d581e2ac1cb2ca93023856f14c8f46437e5fbfd5ed33ec467f2dd73f4f8a4907
3
+ size 4194
model-bin/finetune/base/log/1630232234.2438745/events.out.tfevents.1630232234.cc93b136ebf5.1086.127 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b656bb8581cc3b0e8e81c57ed8bf041161d884c948fa9c079028a2323c3785ce
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630230484.cc93b136ebf5.1086.118 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21341d712d1a1d11a64108a5815e19580ec3c3b91f3adb0fcfa9f363754af853
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630230917.cc93b136ebf5.1086.120 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9207bff65ec5f514cf53a6f839848b0d88cffaba3a0673819f5b98e62a4475cd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630231362.cc93b136ebf5.1086.122 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e143394f3a4ff3d7ad1b06893d174d817d4c9b3c2fd9b14acd48ed85157171bd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630231799.cc93b136ebf5.1086.124 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13d1c476968ed905c4cb56413c983b39e1d176290a9c27dfd7a277e45fbe8511
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630232234.cc93b136ebf5.1086.126 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17df0783581cbdfcc60e0903bf27fdb281f33af1809e9e2e65cabef2ab06909c
3
+ size 8462