"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630113694.5899777/events.out.tfevents.1630113694.86bb0ddabf9b.1042.31 +3 -0
- model-bin/finetune/base/log/1630114108.4351602/events.out.tfevents.1630114108.86bb0ddabf9b.1042.33 +3 -0
- model-bin/finetune/base/log/1630114521.4023952/events.out.tfevents.1630114521.86bb0ddabf9b.1042.35 +3 -0
- model-bin/finetune/base/log/1630114935.8797095/events.out.tfevents.1630114935.86bb0ddabf9b.1042.37 +3 -0
- model-bin/finetune/base/log/1630115344.0735083/events.out.tfevents.1630115344.86bb0ddabf9b.1042.39 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630113694.86bb0ddabf9b.1042.30 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630114108.86bb0ddabf9b.1042.32 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630114521.86bb0ddabf9b.1042.34 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630114935.86bb0ddabf9b.1042.36 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630115344.86bb0ddabf9b.1042.38 +3 -0
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed6a4a1b60032720f1c3b30726a6595b0807d4670f546ca8940edf524a4e7cd6
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2bf7f96504e1c0a6174b9f2d5915d9873c10b1506cc00abe352873e28517867
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:796e263a25f4a0a7434c18d8024159eaa3656ad1fa20d7a25b7eae15f6439f16
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8ec6882b7592c0ac7b9cd7d7fe3da25ac3c6fd19100a89c73fb078e31a1ddc3
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1cd22da4f6bd733f8bb533d0e36af7533566eafa7ac974fe7b8b64dcc9e30a39
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -246771,11 +246771,800 @@
|
|
246771 |
"eval_steps_per_second": 0.696,
|
246772 |
"eval_wer": 0.18200635765506026,
|
246773 |
"step": 115856
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
246774 |
}
|
246775 |
],
|
246776 |
-
"max_steps":
|
246777 |
"num_train_epochs": 5000,
|
246778 |
-
"total_flos": 3.
|
246779 |
"trial_name": null,
|
246780 |
"trial_params": null
|
246781 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 931.0,
|
5 |
+
"global_step": 116478,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
246771 |
"eval_steps_per_second": 0.696,
|
246772 |
"eval_wer": 0.18200635765506026,
|
246773 |
"step": 115856
|
246774 |
+
},
|
246775 |
+
{
|
246776 |
+
"epoch": 934.03,
|
246777 |
+
"learning_rate": 8.160048076923078e-06,
|
246778 |
+
"loss": 0.3382,
|
246779 |
+
"step": 115860
|
246780 |
+
},
|
246781 |
+
{
|
246782 |
+
"epoch": 934.07,
|
246783 |
+
"learning_rate": 8.15996794871795e-06,
|
246784 |
+
"loss": 0.2679,
|
246785 |
+
"step": 115865
|
246786 |
+
},
|
246787 |
+
{
|
246788 |
+
"epoch": 934.11,
|
246789 |
+
"learning_rate": 8.15988782051282e-06,
|
246790 |
+
"loss": 0.277,
|
246791 |
+
"step": 115870
|
246792 |
+
},
|
246793 |
+
{
|
246794 |
+
"epoch": 934.15,
|
246795 |
+
"learning_rate": 8.159807692307694e-06,
|
246796 |
+
"loss": 0.4102,
|
246797 |
+
"step": 115875
|
246798 |
+
},
|
246799 |
+
{
|
246800 |
+
"epoch": 934.19,
|
246801 |
+
"learning_rate": 8.159727564102565e-06,
|
246802 |
+
"loss": 1.1364,
|
246803 |
+
"step": 115880
|
246804 |
+
},
|
246805 |
+
{
|
246806 |
+
"epoch": 934.23,
|
246807 |
+
"learning_rate": 8.159647435897436e-06,
|
246808 |
+
"loss": 0.627,
|
246809 |
+
"step": 115885
|
246810 |
+
},
|
246811 |
+
{
|
246812 |
+
"epoch": 934.27,
|
246813 |
+
"learning_rate": 8.15956730769231e-06,
|
246814 |
+
"loss": 0.2331,
|
246815 |
+
"step": 115890
|
246816 |
+
},
|
246817 |
+
{
|
246818 |
+
"epoch": 934.31,
|
246819 |
+
"learning_rate": 8.15948717948718e-06,
|
246820 |
+
"loss": 0.2426,
|
246821 |
+
"step": 115895
|
246822 |
+
},
|
246823 |
+
{
|
246824 |
+
"epoch": 934.35,
|
246825 |
+
"learning_rate": 8.159407051282052e-06,
|
246826 |
+
"loss": 0.3757,
|
246827 |
+
"step": 115900
|
246828 |
+
},
|
246829 |
+
{
|
246830 |
+
"epoch": 934.39,
|
246831 |
+
"learning_rate": 8.159326923076924e-06,
|
246832 |
+
"loss": 0.8871,
|
246833 |
+
"step": 115905
|
246834 |
+
},
|
246835 |
+
{
|
246836 |
+
"epoch": 934.43,
|
246837 |
+
"learning_rate": 8.159246794871795e-06,
|
246838 |
+
"loss": 0.6067,
|
246839 |
+
"step": 115910
|
246840 |
+
},
|
246841 |
+
{
|
246842 |
+
"epoch": 934.47,
|
246843 |
+
"learning_rate": 8.159166666666666e-06,
|
246844 |
+
"loss": 0.3254,
|
246845 |
+
"step": 115915
|
246846 |
+
},
|
246847 |
+
{
|
246848 |
+
"epoch": 934.51,
|
246849 |
+
"learning_rate": 8.15908653846154e-06,
|
246850 |
+
"loss": 0.3118,
|
246851 |
+
"step": 115920
|
246852 |
+
},
|
246853 |
+
{
|
246854 |
+
"epoch": 934.55,
|
246855 |
+
"learning_rate": 8.15900641025641e-06,
|
246856 |
+
"loss": 0.425,
|
246857 |
+
"step": 115925
|
246858 |
+
},
|
246859 |
+
{
|
246860 |
+
"epoch": 934.59,
|
246861 |
+
"learning_rate": 8.158926282051282e-06,
|
246862 |
+
"loss": 0.9399,
|
246863 |
+
"step": 115930
|
246864 |
+
},
|
246865 |
+
{
|
246866 |
+
"epoch": 934.63,
|
246867 |
+
"learning_rate": 8.158846153846155e-06,
|
246868 |
+
"loss": 0.7605,
|
246869 |
+
"step": 115935
|
246870 |
+
},
|
246871 |
+
{
|
246872 |
+
"epoch": 934.67,
|
246873 |
+
"learning_rate": 8.158766025641027e-06,
|
246874 |
+
"loss": 0.2649,
|
246875 |
+
"step": 115940
|
246876 |
+
},
|
246877 |
+
{
|
246878 |
+
"epoch": 934.71,
|
246879 |
+
"learning_rate": 8.158685897435898e-06,
|
246880 |
+
"loss": 0.3038,
|
246881 |
+
"step": 115945
|
246882 |
+
},
|
246883 |
+
{
|
246884 |
+
"epoch": 934.76,
|
246885 |
+
"learning_rate": 8.15860576923077e-06,
|
246886 |
+
"loss": 0.3894,
|
246887 |
+
"step": 115950
|
246888 |
+
},
|
246889 |
+
{
|
246890 |
+
"epoch": 934.8,
|
246891 |
+
"learning_rate": 8.158525641025642e-06,
|
246892 |
+
"loss": 0.8665,
|
246893 |
+
"step": 115955
|
246894 |
+
},
|
246895 |
+
{
|
246896 |
+
"epoch": 934.84,
|
246897 |
+
"learning_rate": 8.158445512820514e-06,
|
246898 |
+
"loss": 0.6454,
|
246899 |
+
"step": 115960
|
246900 |
+
},
|
246901 |
+
{
|
246902 |
+
"epoch": 934.88,
|
246903 |
+
"learning_rate": 8.158365384615385e-06,
|
246904 |
+
"loss": 0.2806,
|
246905 |
+
"step": 115965
|
246906 |
+
},
|
246907 |
+
{
|
246908 |
+
"epoch": 934.92,
|
246909 |
+
"learning_rate": 8.158285256410256e-06,
|
246910 |
+
"loss": 0.2804,
|
246911 |
+
"step": 115970
|
246912 |
+
},
|
246913 |
+
{
|
246914 |
+
"epoch": 934.96,
|
246915 |
+
"learning_rate": 8.15820512820513e-06,
|
246916 |
+
"loss": 0.4341,
|
246917 |
+
"step": 115975
|
246918 |
+
},
|
246919 |
+
{
|
246920 |
+
"epoch": 935.0,
|
246921 |
+
"learning_rate": 8.158125e-06,
|
246922 |
+
"loss": 0.9677,
|
246923 |
+
"step": 115980
|
246924 |
+
},
|
246925 |
+
{
|
246926 |
+
"epoch": 935.0,
|
246927 |
+
"eval_loss": 0.3683865964412689,
|
246928 |
+
"eval_runtime": 39.1683,
|
246929 |
+
"eval_samples_per_second": 21.523,
|
246930 |
+
"eval_steps_per_second": 0.689,
|
246931 |
+
"eval_wer": 0.18747801618009144,
|
246932 |
+
"step": 115980
|
246933 |
+
},
|
246934 |
+
{
|
246935 |
+
"epoch": 935.04,
|
246936 |
+
"learning_rate": 8.158044871794872e-06,
|
246937 |
+
"loss": 0.2787,
|
246938 |
+
"step": 115985
|
246939 |
+
},
|
246940 |
+
{
|
246941 |
+
"epoch": 935.08,
|
246942 |
+
"learning_rate": 8.157964743589745e-06,
|
246943 |
+
"loss": 0.292,
|
246944 |
+
"step": 115990
|
246945 |
+
},
|
246946 |
+
{
|
246947 |
+
"epoch": 935.12,
|
246948 |
+
"learning_rate": 8.157884615384617e-06,
|
246949 |
+
"loss": 0.3054,
|
246950 |
+
"step": 115995
|
246951 |
+
},
|
246952 |
+
{
|
246953 |
+
"epoch": 935.16,
|
246954 |
+
"learning_rate": 8.157804487179488e-06,
|
246955 |
+
"loss": 0.5644,
|
246956 |
+
"step": 116000
|
246957 |
+
},
|
246958 |
+
{
|
246959 |
+
"epoch": 935.2,
|
246960 |
+
"learning_rate": 8.15772435897436e-06,
|
246961 |
+
"loss": 1.306,
|
246962 |
+
"step": 116005
|
246963 |
+
},
|
246964 |
+
{
|
246965 |
+
"epoch": 935.24,
|
246966 |
+
"learning_rate": 8.157644230769232e-06,
|
246967 |
+
"loss": 0.3061,
|
246968 |
+
"step": 116010
|
246969 |
+
},
|
246970 |
+
{
|
246971 |
+
"epoch": 935.28,
|
246972 |
+
"learning_rate": 8.157564102564102e-06,
|
246973 |
+
"loss": 0.2666,
|
246974 |
+
"step": 116015
|
246975 |
+
},
|
246976 |
+
{
|
246977 |
+
"epoch": 935.32,
|
246978 |
+
"learning_rate": 8.157483974358975e-06,
|
246979 |
+
"loss": 0.3729,
|
246980 |
+
"step": 116020
|
246981 |
+
},
|
246982 |
+
{
|
246983 |
+
"epoch": 935.36,
|
246984 |
+
"learning_rate": 8.157403846153846e-06,
|
246985 |
+
"loss": 0.4313,
|
246986 |
+
"step": 116025
|
246987 |
+
},
|
246988 |
+
{
|
246989 |
+
"epoch": 935.4,
|
246990 |
+
"learning_rate": 8.157323717948718e-06,
|
246991 |
+
"loss": 1.1394,
|
246992 |
+
"step": 116030
|
246993 |
+
},
|
246994 |
+
{
|
246995 |
+
"epoch": 935.44,
|
246996 |
+
"learning_rate": 8.15724358974359e-06,
|
246997 |
+
"loss": 0.3986,
|
246998 |
+
"step": 116035
|
246999 |
+
},
|
247000 |
+
{
|
247001 |
+
"epoch": 935.48,
|
247002 |
+
"learning_rate": 8.157163461538462e-06,
|
247003 |
+
"loss": 0.2467,
|
247004 |
+
"step": 116040
|
247005 |
+
},
|
247006 |
+
{
|
247007 |
+
"epoch": 935.52,
|
247008 |
+
"learning_rate": 8.157083333333334e-06,
|
247009 |
+
"loss": 0.2822,
|
247010 |
+
"step": 116045
|
247011 |
+
},
|
247012 |
+
{
|
247013 |
+
"epoch": 935.56,
|
247014 |
+
"learning_rate": 8.157003205128205e-06,
|
247015 |
+
"loss": 0.4465,
|
247016 |
+
"step": 116050
|
247017 |
+
},
|
247018 |
+
{
|
247019 |
+
"epoch": 935.6,
|
247020 |
+
"learning_rate": 8.156923076923078e-06,
|
247021 |
+
"loss": 1.1917,
|
247022 |
+
"step": 116055
|
247023 |
+
},
|
247024 |
+
{
|
247025 |
+
"epoch": 935.64,
|
247026 |
+
"learning_rate": 8.15684294871795e-06,
|
247027 |
+
"loss": 0.3459,
|
247028 |
+
"step": 116060
|
247029 |
+
},
|
247030 |
+
{
|
247031 |
+
"epoch": 935.68,
|
247032 |
+
"learning_rate": 8.15676282051282e-06,
|
247033 |
+
"loss": 0.2845,
|
247034 |
+
"step": 116065
|
247035 |
+
},
|
247036 |
+
{
|
247037 |
+
"epoch": 935.72,
|
247038 |
+
"learning_rate": 8.156682692307692e-06,
|
247039 |
+
"loss": 0.2771,
|
247040 |
+
"step": 116070
|
247041 |
+
},
|
247042 |
+
{
|
247043 |
+
"epoch": 935.76,
|
247044 |
+
"learning_rate": 8.156602564102565e-06,
|
247045 |
+
"loss": 0.4098,
|
247046 |
+
"step": 116075
|
247047 |
+
},
|
247048 |
+
{
|
247049 |
+
"epoch": 935.8,
|
247050 |
+
"learning_rate": 8.156522435897436e-06,
|
247051 |
+
"loss": 1.3603,
|
247052 |
+
"step": 116080
|
247053 |
+
},
|
247054 |
+
{
|
247055 |
+
"epoch": 935.84,
|
247056 |
+
"learning_rate": 8.156442307692308e-06,
|
247057 |
+
"loss": 0.301,
|
247058 |
+
"step": 116085
|
247059 |
+
},
|
247060 |
+
{
|
247061 |
+
"epoch": 935.88,
|
247062 |
+
"learning_rate": 8.15636217948718e-06,
|
247063 |
+
"loss": 0.2724,
|
247064 |
+
"step": 116090
|
247065 |
+
},
|
247066 |
+
{
|
247067 |
+
"epoch": 935.92,
|
247068 |
+
"learning_rate": 8.156282051282052e-06,
|
247069 |
+
"loss": 0.3323,
|
247070 |
+
"step": 116095
|
247071 |
+
},
|
247072 |
+
{
|
247073 |
+
"epoch": 935.96,
|
247074 |
+
"learning_rate": 8.156201923076924e-06,
|
247075 |
+
"loss": 0.4345,
|
247076 |
+
"step": 116100
|
247077 |
+
},
|
247078 |
+
{
|
247079 |
+
"epoch": 936.0,
|
247080 |
+
"eval_loss": 0.3591098189353943,
|
247081 |
+
"eval_runtime": 40.0141,
|
247082 |
+
"eval_samples_per_second": 21.068,
|
247083 |
+
"eval_steps_per_second": 0.675,
|
247084 |
+
"eval_wer": 0.1825556353194544,
|
247085 |
+
"step": 116104
|
247086 |
+
},
|
247087 |
+
{
|
247088 |
+
"epoch": 936.01,
|
247089 |
+
"learning_rate": 8.156121794871795e-06,
|
247090 |
+
"loss": 0.3954,
|
247091 |
+
"step": 116105
|
247092 |
+
},
|
247093 |
+
{
|
247094 |
+
"epoch": 936.05,
|
247095 |
+
"learning_rate": 8.156041666666668e-06,
|
247096 |
+
"loss": 0.3009,
|
247097 |
+
"step": 116110
|
247098 |
+
},
|
247099 |
+
{
|
247100 |
+
"epoch": 936.09,
|
247101 |
+
"learning_rate": 8.15596153846154e-06,
|
247102 |
+
"loss": 0.3602,
|
247103 |
+
"step": 116115
|
247104 |
+
},
|
247105 |
+
{
|
247106 |
+
"epoch": 936.13,
|
247107 |
+
"learning_rate": 8.15588141025641e-06,
|
247108 |
+
"loss": 0.279,
|
247109 |
+
"step": 116120
|
247110 |
+
},
|
247111 |
+
{
|
247112 |
+
"epoch": 936.17,
|
247113 |
+
"learning_rate": 8.155801282051282e-06,
|
247114 |
+
"loss": 0.5056,
|
247115 |
+
"step": 116125
|
247116 |
+
},
|
247117 |
+
{
|
247118 |
+
"epoch": 936.21,
|
247119 |
+
"learning_rate": 8.155721153846155e-06,
|
247120 |
+
"loss": 1.0783,
|
247121 |
+
"step": 116130
|
247122 |
+
},
|
247123 |
+
{
|
247124 |
+
"epoch": 936.25,
|
247125 |
+
"learning_rate": 8.155641025641026e-06,
|
247126 |
+
"loss": 0.3115,
|
247127 |
+
"step": 116135
|
247128 |
+
},
|
247129 |
+
{
|
247130 |
+
"epoch": 936.29,
|
247131 |
+
"learning_rate": 8.155560897435898e-06,
|
247132 |
+
"loss": 0.2519,
|
247133 |
+
"step": 116140
|
247134 |
+
},
|
247135 |
+
{
|
247136 |
+
"epoch": 936.33,
|
247137 |
+
"learning_rate": 8.155480769230771e-06,
|
247138 |
+
"loss": 0.361,
|
247139 |
+
"step": 116145
|
247140 |
+
},
|
247141 |
+
{
|
247142 |
+
"epoch": 936.37,
|
247143 |
+
"learning_rate": 8.155400641025642e-06,
|
247144 |
+
"loss": 0.5359,
|
247145 |
+
"step": 116150
|
247146 |
+
},
|
247147 |
+
{
|
247148 |
+
"epoch": 936.41,
|
247149 |
+
"learning_rate": 8.155320512820514e-06,
|
247150 |
+
"loss": 1.0703,
|
247151 |
+
"step": 116155
|
247152 |
+
},
|
247153 |
+
{
|
247154 |
+
"epoch": 936.45,
|
247155 |
+
"learning_rate": 8.155240384615385e-06,
|
247156 |
+
"loss": 0.3052,
|
247157 |
+
"step": 116160
|
247158 |
+
},
|
247159 |
+
{
|
247160 |
+
"epoch": 936.49,
|
247161 |
+
"learning_rate": 8.155160256410258e-06,
|
247162 |
+
"loss": 0.2976,
|
247163 |
+
"step": 116165
|
247164 |
+
},
|
247165 |
+
{
|
247166 |
+
"epoch": 936.53,
|
247167 |
+
"learning_rate": 8.155080128205128e-06,
|
247168 |
+
"loss": 0.316,
|
247169 |
+
"step": 116170
|
247170 |
+
},
|
247171 |
+
{
|
247172 |
+
"epoch": 936.57,
|
247173 |
+
"learning_rate": 8.155e-06,
|
247174 |
+
"loss": 0.5374,
|
247175 |
+
"step": 116175
|
247176 |
+
},
|
247177 |
+
{
|
247178 |
+
"epoch": 936.61,
|
247179 |
+
"learning_rate": 8.154919871794874e-06,
|
247180 |
+
"loss": 0.9842,
|
247181 |
+
"step": 116180
|
247182 |
+
},
|
247183 |
+
{
|
247184 |
+
"epoch": 936.65,
|
247185 |
+
"learning_rate": 8.154839743589743e-06,
|
247186 |
+
"loss": 0.2637,
|
247187 |
+
"step": 116185
|
247188 |
+
},
|
247189 |
+
{
|
247190 |
+
"epoch": 936.69,
|
247191 |
+
"learning_rate": 8.154759615384616e-06,
|
247192 |
+
"loss": 0.2658,
|
247193 |
+
"step": 116190
|
247194 |
+
},
|
247195 |
+
{
|
247196 |
+
"epoch": 936.73,
|
247197 |
+
"learning_rate": 8.154679487179488e-06,
|
247198 |
+
"loss": 0.4132,
|
247199 |
+
"step": 116195
|
247200 |
+
},
|
247201 |
+
{
|
247202 |
+
"epoch": 936.77,
|
247203 |
+
"learning_rate": 8.15459935897436e-06,
|
247204 |
+
"loss": 0.4989,
|
247205 |
+
"step": 116200
|
247206 |
+
},
|
247207 |
+
{
|
247208 |
+
"epoch": 936.81,
|
247209 |
+
"learning_rate": 8.15451923076923e-06,
|
247210 |
+
"loss": 1.1565,
|
247211 |
+
"step": 116205
|
247212 |
+
},
|
247213 |
+
{
|
247214 |
+
"epoch": 936.85,
|
247215 |
+
"learning_rate": 8.154439102564104e-06,
|
247216 |
+
"loss": 0.37,
|
247217 |
+
"step": 116210
|
247218 |
+
},
|
247219 |
+
{
|
247220 |
+
"epoch": 936.89,
|
247221 |
+
"learning_rate": 8.154358974358975e-06,
|
247222 |
+
"loss": 0.4002,
|
247223 |
+
"step": 116215
|
247224 |
+
},
|
247225 |
+
{
|
247226 |
+
"epoch": 936.93,
|
247227 |
+
"learning_rate": 8.154278846153846e-06,
|
247228 |
+
"loss": 1.0027,
|
247229 |
+
"step": 116220
|
247230 |
+
},
|
247231 |
+
{
|
247232 |
+
"epoch": 936.97,
|
247233 |
+
"learning_rate": 8.154198717948718e-06,
|
247234 |
+
"loss": 0.5068,
|
247235 |
+
"step": 116225
|
247236 |
+
},
|
247237 |
+
{
|
247238 |
+
"epoch": 937.0,
|
247239 |
+
"eval_loss": 0.37429821491241455,
|
247240 |
+
"eval_runtime": 38.8986,
|
247241 |
+
"eval_samples_per_second": 21.697,
|
247242 |
+
"eval_steps_per_second": 0.694,
|
247243 |
+
"eval_wer": 0.18772589806004278,
|
247244 |
+
"step": 116228
|
247245 |
+
},
|
247246 |
+
{
|
247247 |
+
"epoch": 929.02,
|
247248 |
+
"learning_rate": 8.15411858974359e-06,
|
247249 |
+
"loss": 0.3236,
|
247250 |
+
"step": 116230
|
247251 |
+
},
|
247252 |
+
{
|
247253 |
+
"epoch": 929.06,
|
247254 |
+
"learning_rate": 8.154038461538462e-06,
|
247255 |
+
"loss": 0.3169,
|
247256 |
+
"step": 116235
|
247257 |
+
},
|
247258 |
+
{
|
247259 |
+
"epoch": 929.1,
|
247260 |
+
"learning_rate": 8.153958333333333e-06,
|
247261 |
+
"loss": 0.303,
|
247262 |
+
"step": 116240
|
247263 |
+
},
|
247264 |
+
{
|
247265 |
+
"epoch": 929.14,
|
247266 |
+
"learning_rate": 8.153878205128206e-06,
|
247267 |
+
"loss": 0.3253,
|
247268 |
+
"step": 116245
|
247269 |
+
},
|
247270 |
+
{
|
247271 |
+
"epoch": 929.18,
|
247272 |
+
"learning_rate": 8.153798076923078e-06,
|
247273 |
+
"loss": 0.6763,
|
247274 |
+
"step": 116250
|
247275 |
+
},
|
247276 |
+
{
|
247277 |
+
"epoch": 929.22,
|
247278 |
+
"learning_rate": 8.15371794871795e-06,
|
247279 |
+
"loss": 1.0281,
|
247280 |
+
"step": 116255
|
247281 |
+
},
|
247282 |
+
{
|
247283 |
+
"epoch": 929.26,
|
247284 |
+
"learning_rate": 8.15363782051282e-06,
|
247285 |
+
"loss": 0.2937,
|
247286 |
+
"step": 116260
|
247287 |
+
},
|
247288 |
+
{
|
247289 |
+
"epoch": 929.3,
|
247290 |
+
"learning_rate": 8.153557692307694e-06,
|
247291 |
+
"loss": 0.28,
|
247292 |
+
"step": 116265
|
247293 |
+
},
|
247294 |
+
{
|
247295 |
+
"epoch": 929.34,
|
247296 |
+
"learning_rate": 8.153477564102565e-06,
|
247297 |
+
"loss": 0.282,
|
247298 |
+
"step": 116270
|
247299 |
+
},
|
247300 |
+
{
|
247301 |
+
"epoch": 929.38,
|
247302 |
+
"learning_rate": 8.153397435897436e-06,
|
247303 |
+
"loss": 0.6766,
|
247304 |
+
"step": 116275
|
247305 |
+
},
|
247306 |
+
{
|
247307 |
+
"epoch": 929.42,
|
247308 |
+
"learning_rate": 8.15331730769231e-06,
|
247309 |
+
"loss": 0.8845,
|
247310 |
+
"step": 116280
|
247311 |
+
},
|
247312 |
+
{
|
247313 |
+
"epoch": 929.46,
|
247314 |
+
"learning_rate": 8.15323717948718e-06,
|
247315 |
+
"loss": 0.2686,
|
247316 |
+
"step": 116285
|
247317 |
+
},
|
247318 |
+
{
|
247319 |
+
"epoch": 929.5,
|
247320 |
+
"learning_rate": 8.153157051282052e-06,
|
247321 |
+
"loss": 0.2538,
|
247322 |
+
"step": 116290
|
247323 |
+
},
|
247324 |
+
{
|
247325 |
+
"epoch": 929.54,
|
247326 |
+
"learning_rate": 8.153076923076923e-06,
|
247327 |
+
"loss": 0.2814,
|
247328 |
+
"step": 116295
|
247329 |
+
},
|
247330 |
+
{
|
247331 |
+
"epoch": 929.58,
|
247332 |
+
"learning_rate": 8.152996794871797e-06,
|
247333 |
+
"loss": 0.5815,
|
247334 |
+
"step": 116300
|
247335 |
+
},
|
247336 |
+
{
|
247337 |
+
"epoch": 929.62,
|
247338 |
+
"learning_rate": 8.152916666666668e-06,
|
247339 |
+
"loss": 1.0262,
|
247340 |
+
"step": 116305
|
247341 |
+
},
|
247342 |
+
{
|
247343 |
+
"epoch": 929.66,
|
247344 |
+
"learning_rate": 8.15283653846154e-06,
|
247345 |
+
"loss": 0.2782,
|
247346 |
+
"step": 116310
|
247347 |
+
},
|
247348 |
+
{
|
247349 |
+
"epoch": 929.7,
|
247350 |
+
"learning_rate": 8.15275641025641e-06,
|
247351 |
+
"loss": 0.3347,
|
247352 |
+
"step": 116315
|
247353 |
+
},
|
247354 |
+
{
|
247355 |
+
"epoch": 929.74,
|
247356 |
+
"learning_rate": 8.152676282051284e-06,
|
247357 |
+
"loss": 0.3231,
|
247358 |
+
"step": 116320
|
247359 |
+
},
|
247360 |
+
{
|
247361 |
+
"epoch": 929.78,
|
247362 |
+
"learning_rate": 8.152596153846153e-06,
|
247363 |
+
"loss": 0.5842,
|
247364 |
+
"step": 116325
|
247365 |
+
},
|
247366 |
+
{
|
247367 |
+
"epoch": 929.82,
|
247368 |
+
"learning_rate": 8.152516025641026e-06,
|
247369 |
+
"loss": 1.0521,
|
247370 |
+
"step": 116330
|
247371 |
+
},
|
247372 |
+
{
|
247373 |
+
"epoch": 929.86,
|
247374 |
+
"learning_rate": 8.1524358974359e-06,
|
247375 |
+
"loss": 0.2646,
|
247376 |
+
"step": 116335
|
247377 |
+
},
|
247378 |
+
{
|
247379 |
+
"epoch": 929.9,
|
247380 |
+
"learning_rate": 8.152355769230769e-06,
|
247381 |
+
"loss": 0.3342,
|
247382 |
+
"step": 116340
|
247383 |
+
},
|
247384 |
+
{
|
247385 |
+
"epoch": 929.94,
|
247386 |
+
"learning_rate": 8.152275641025642e-06,
|
247387 |
+
"loss": 0.4099,
|
247388 |
+
"step": 116345
|
247389 |
+
},
|
247390 |
+
{
|
247391 |
+
"epoch": 929.98,
|
247392 |
+
"learning_rate": 8.152195512820513e-06,
|
247393 |
+
"loss": 0.6846,
|
247394 |
+
"step": 116350
|
247395 |
+
},
|
247396 |
+
{
|
247397 |
+
"epoch": 930.0,
|
247398 |
+
"eval_loss": 0.42502668499946594,
|
247399 |
+
"eval_runtime": 39.0001,
|
247400 |
+
"eval_samples_per_second": 21.615,
|
247401 |
+
"eval_steps_per_second": 0.692,
|
247402 |
+
"eval_wer": 0.1844075759800323,
|
247403 |
+
"step": 116353
|
247404 |
+
},
|
247405 |
+
{
|
247406 |
+
"epoch": 930.02,
|
247407 |
+
"learning_rate": 8.152115384615385e-06,
|
247408 |
+
"loss": 0.4367,
|
247409 |
+
"step": 116355
|
247410 |
+
},
|
247411 |
+
{
|
247412 |
+
"epoch": 930.06,
|
247413 |
+
"learning_rate": 8.152035256410256e-06,
|
247414 |
+
"loss": 0.2788,
|
247415 |
+
"step": 116360
|
247416 |
+
},
|
247417 |
+
{
|
247418 |
+
"epoch": 930.1,
|
247419 |
+
"learning_rate": 8.15195512820513e-06,
|
247420 |
+
"loss": 0.289,
|
247421 |
+
"step": 116365
|
247422 |
+
},
|
247423 |
+
{
|
247424 |
+
"epoch": 930.14,
|
247425 |
+
"learning_rate": 8.151875e-06,
|
247426 |
+
"loss": 0.364,
|
247427 |
+
"step": 116370
|
247428 |
+
},
|
247429 |
+
{
|
247430 |
+
"epoch": 930.18,
|
247431 |
+
"learning_rate": 8.151794871794872e-06,
|
247432 |
+
"loss": 0.5893,
|
247433 |
+
"step": 116375
|
247434 |
+
},
|
247435 |
+
{
|
247436 |
+
"epoch": 930.22,
|
247437 |
+
"learning_rate": 8.151714743589745e-06,
|
247438 |
+
"loss": 1.0691,
|
247439 |
+
"step": 116380
|
247440 |
+
},
|
247441 |
+
{
|
247442 |
+
"epoch": 930.26,
|
247443 |
+
"learning_rate": 8.151634615384616e-06,
|
247444 |
+
"loss": 0.2525,
|
247445 |
+
"step": 116385
|
247446 |
+
},
|
247447 |
+
{
|
247448 |
+
"epoch": 930.3,
|
247449 |
+
"learning_rate": 8.151554487179488e-06,
|
247450 |
+
"loss": 0.2818,
|
247451 |
+
"step": 116390
|
247452 |
+
},
|
247453 |
+
{
|
247454 |
+
"epoch": 930.34,
|
247455 |
+
"learning_rate": 8.151474358974359e-06,
|
247456 |
+
"loss": 0.3698,
|
247457 |
+
"step": 116395
|
247458 |
+
},
|
247459 |
+
{
|
247460 |
+
"epoch": 930.38,
|
247461 |
+
"learning_rate": 8.151394230769232e-06,
|
247462 |
+
"loss": 0.6021,
|
247463 |
+
"step": 116400
|
247464 |
+
},
|
247465 |
+
{
|
247466 |
+
"epoch": 930.42,
|
247467 |
+
"learning_rate": 8.151314102564104e-06,
|
247468 |
+
"loss": 1.0208,
|
247469 |
+
"step": 116405
|
247470 |
+
},
|
247471 |
+
{
|
247472 |
+
"epoch": 930.46,
|
247473 |
+
"learning_rate": 8.151233974358975e-06,
|
247474 |
+
"loss": 0.2736,
|
247475 |
+
"step": 116410
|
247476 |
+
},
|
247477 |
+
{
|
247478 |
+
"epoch": 930.5,
|
247479 |
+
"learning_rate": 8.151153846153846e-06,
|
247480 |
+
"loss": 0.3401,
|
247481 |
+
"step": 116415
|
247482 |
+
},
|
247483 |
+
{
|
247484 |
+
"epoch": 930.54,
|
247485 |
+
"learning_rate": 8.15107371794872e-06,
|
247486 |
+
"loss": 0.3492,
|
247487 |
+
"step": 116420
|
247488 |
+
},
|
247489 |
+
{
|
247490 |
+
"epoch": 930.58,
|
247491 |
+
"learning_rate": 8.15099358974359e-06,
|
247492 |
+
"loss": 0.7202,
|
247493 |
+
"step": 116425
|
247494 |
+
},
|
247495 |
+
{
|
247496 |
+
"epoch": 930.62,
|
247497 |
+
"learning_rate": 8.150913461538462e-06,
|
247498 |
+
"loss": 1.0942,
|
247499 |
+
"step": 116430
|
247500 |
+
},
|
247501 |
+
{
|
247502 |
+
"epoch": 930.66,
|
247503 |
+
"learning_rate": 8.150833333333335e-06,
|
247504 |
+
"loss": 0.2699,
|
247505 |
+
"step": 116435
|
247506 |
+
},
|
247507 |
+
{
|
247508 |
+
"epoch": 930.7,
|
247509 |
+
"learning_rate": 8.150753205128206e-06,
|
247510 |
+
"loss": 0.2856,
|
247511 |
+
"step": 116440
|
247512 |
+
},
|
247513 |
+
{
|
247514 |
+
"epoch": 930.74,
|
247515 |
+
"learning_rate": 8.150673076923078e-06,
|
247516 |
+
"loss": 0.3806,
|
247517 |
+
"step": 116445
|
247518 |
+
},
|
247519 |
+
{
|
247520 |
+
"epoch": 930.78,
|
247521 |
+
"learning_rate": 8.150592948717949e-06,
|
247522 |
+
"loss": 0.6902,
|
247523 |
+
"step": 116450
|
247524 |
+
},
|
247525 |
+
{
|
247526 |
+
"epoch": 930.82,
|
247527 |
+
"learning_rate": 8.150512820512822e-06,
|
247528 |
+
"loss": 1.0085,
|
247529 |
+
"step": 116455
|
247530 |
+
},
|
247531 |
+
{
|
247532 |
+
"epoch": 930.86,
|
247533 |
+
"learning_rate": 8.150432692307692e-06,
|
247534 |
+
"loss": 0.2629,
|
247535 |
+
"step": 116460
|
247536 |
+
},
|
247537 |
+
{
|
247538 |
+
"epoch": 930.9,
|
247539 |
+
"learning_rate": 8.150352564102565e-06,
|
247540 |
+
"loss": 0.2615,
|
247541 |
+
"step": 116465
|
247542 |
+
},
|
247543 |
+
{
|
247544 |
+
"epoch": 930.94,
|
247545 |
+
"learning_rate": 8.150272435897436e-06,
|
247546 |
+
"loss": 0.3081,
|
247547 |
+
"step": 116470
|
247548 |
+
},
|
247549 |
+
{
|
247550 |
+
"epoch": 930.98,
|
247551 |
+
"learning_rate": 8.150192307692308e-06,
|
247552 |
+
"loss": 0.5409,
|
247553 |
+
"step": 116475
|
247554 |
+
},
|
247555 |
+
{
|
247556 |
+
"epoch": 931.0,
|
247557 |
+
"eval_loss": 0.427347868680954,
|
247558 |
+
"eval_runtime": 39.3615,
|
247559 |
+
"eval_samples_per_second": 21.417,
|
247560 |
+
"eval_steps_per_second": 0.686,
|
247561 |
+
"eval_wer": 0.18822154026011984,
|
247562 |
+
"step": 116478
|
247563 |
}
|
247564 |
],
|
247565 |
+
"max_steps": 625000,
|
247566 |
"num_train_epochs": 5000,
|
247567 |
+
"total_flos": 3.2779879386977056e+20,
|
247568 |
"trial_name": null,
|
247569 |
"trial_params": null
|
247570 |
}
|
model-bin/finetune/base/{checkpoint-115856 β checkpoint-116478}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630113694.5899777/events.out.tfevents.1630113694.86bb0ddabf9b.1042.31
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2bafd466c630fd1f651cfc6bbaac47320578e8cbe124912388e189cdd02f369
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630114108.4351602/events.out.tfevents.1630114108.86bb0ddabf9b.1042.33
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07a0a4953561629052e2d0bc1801579b24c34391dc125a4716a5f56a2a3045fa
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630114521.4023952/events.out.tfevents.1630114521.86bb0ddabf9b.1042.35
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f29c80132da81c5a6dd7313800425f0fb049072d5f15e5425c2f62b5e3e9619b
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630114935.8797095/events.out.tfevents.1630114935.86bb0ddabf9b.1042.37
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:840305c14ee3814d1fdf3db5aca2eb7cc8afa2665aa999949f65e2e0cb1dee27
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630115344.0735083/events.out.tfevents.1630115344.86bb0ddabf9b.1042.39
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:755d20a6b48427f9c481163078b8315b0f5e80a84035f6c92eaf75a877a3c1bf
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630113694.86bb0ddabf9b.1042.30
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c0bce1d69ec0c9a328677eb09f0ad0770a749ad9c3fea932b09d28be9fbd1a70
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630114108.86bb0ddabf9b.1042.32
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba95ea0c440eb81e806d98f8334fc2cca8bd634a1c8b8264996852ee87ee21d0
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630114521.86bb0ddabf9b.1042.34
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3cebd1f50853e7c5b4a05ca51952ae5a5a5f3c76be45e7ddf63b91f9739d919c
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630114935.86bb0ddabf9b.1042.36
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:683c8af391b89ef17e13a05e99e3522534eda937abf38257a90c3d8c661bdd3e
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630115344.86bb0ddabf9b.1042.38
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:830c5154d8a6e363532f21d0a69517dce85a978e3fa9e0cb2b35a98412204cb1
|
3 |
+
size 8622
|