"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629869061.2637658/events.out.tfevents.1629869061.7e498afd5545.905.53 +3 -0
- model-bin/finetune/base/log/1629869731.455241/events.out.tfevents.1629869731.7e498afd5545.905.55 +3 -0
- model-bin/finetune/base/log/1629870382.7916713/events.out.tfevents.1629870382.7e498afd5545.905.57 +3 -0
- model-bin/finetune/base/log/1629871038.1608121/events.out.tfevents.1629871038.7e498afd5545.905.59 +3 -0
- model-bin/finetune/base/log/1629871680.9885855/events.out.tfevents.1629871680.7e498afd5545.905.61 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629869061.7e498afd5545.905.52 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629869731.7e498afd5545.905.54 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629870382.7e498afd5545.905.56 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629871038.7e498afd5545.905.58 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629871680.7e498afd5545.905.60 +3 -0
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:474f662d85a0ed7f5bc206a95d86114fc46bc01fa819c5ec2777457c77a00761
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2515fce69a86617d9a349a139a5baf90c4eb5c34710ae7f1938eb4041c8573b3
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9e7b3e60ec7f2a8985b5a92be7a019bc55dd61ea1a80f5e0727362168793b07
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0db8edbac0eb024da8220cc21e3842882d8da7392af65962fe8a5134ce0473dd
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6bd477d4820649f1a048e2fc31a66af8ecccd0bbb65e45ea5df7cdf6444fcd34
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.18412114350410416,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -191826,11 +191826,800 @@
|
|
191826 |
"eval_steps_per_second": 0.669,
|
191827 |
"eval_wer": 0.19356979573777744,
|
191828 |
"step": 72674
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
191829 |
}
|
191830 |
],
|
191831 |
"max_steps": 625000,
|
191832 |
"num_train_epochs": 5000,
|
191833 |
-
"total_flos": 2.
|
191834 |
"trial_name": null,
|
191835 |
"trial_params": null
|
191836 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.18412114350410416,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
4 |
+
"epoch": 586.0,
|
5 |
+
"global_step": 73294,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
191826 |
"eval_steps_per_second": 0.669,
|
191827 |
"eval_wer": 0.19356979573777744,
|
191828 |
"step": 72674
|
191829 |
+
},
|
191830 |
+
{
|
191831 |
+
"epoch": 586.01,
|
191832 |
+
"learning_rate": 8.851794871794872e-06,
|
191833 |
+
"loss": 0.4982,
|
191834 |
+
"step": 72675
|
191835 |
+
},
|
191836 |
+
{
|
191837 |
+
"epoch": 586.05,
|
191838 |
+
"learning_rate": 8.851714743589745e-06,
|
191839 |
+
"loss": 0.3086,
|
191840 |
+
"step": 72680
|
191841 |
+
},
|
191842 |
+
{
|
191843 |
+
"epoch": 586.09,
|
191844 |
+
"learning_rate": 8.851634615384616e-06,
|
191845 |
+
"loss": 0.3143,
|
191846 |
+
"step": 72685
|
191847 |
+
},
|
191848 |
+
{
|
191849 |
+
"epoch": 586.13,
|
191850 |
+
"learning_rate": 8.851554487179488e-06,
|
191851 |
+
"loss": 0.3705,
|
191852 |
+
"step": 72690
|
191853 |
+
},
|
191854 |
+
{
|
191855 |
+
"epoch": 586.17,
|
191856 |
+
"learning_rate": 8.851474358974361e-06,
|
191857 |
+
"loss": 0.5783,
|
191858 |
+
"step": 72695
|
191859 |
+
},
|
191860 |
+
{
|
191861 |
+
"epoch": 586.21,
|
191862 |
+
"learning_rate": 8.85139423076923e-06,
|
191863 |
+
"loss": 1.174,
|
191864 |
+
"step": 72700
|
191865 |
+
},
|
191866 |
+
{
|
191867 |
+
"epoch": 586.25,
|
191868 |
+
"learning_rate": 8.851314102564104e-06,
|
191869 |
+
"loss": 0.4436,
|
191870 |
+
"step": 72705
|
191871 |
+
},
|
191872 |
+
{
|
191873 |
+
"epoch": 586.29,
|
191874 |
+
"learning_rate": 8.851233974358975e-06,
|
191875 |
+
"loss": 0.2785,
|
191876 |
+
"step": 72710
|
191877 |
+
},
|
191878 |
+
{
|
191879 |
+
"epoch": 586.33,
|
191880 |
+
"learning_rate": 8.851153846153846e-06,
|
191881 |
+
"loss": 0.4354,
|
191882 |
+
"step": 72715
|
191883 |
+
},
|
191884 |
+
{
|
191885 |
+
"epoch": 586.37,
|
191886 |
+
"learning_rate": 8.851073717948718e-06,
|
191887 |
+
"loss": 0.674,
|
191888 |
+
"step": 72720
|
191889 |
+
},
|
191890 |
+
{
|
191891 |
+
"epoch": 586.41,
|
191892 |
+
"learning_rate": 8.85099358974359e-06,
|
191893 |
+
"loss": 1.2375,
|
191894 |
+
"step": 72725
|
191895 |
+
},
|
191896 |
+
{
|
191897 |
+
"epoch": 586.45,
|
191898 |
+
"learning_rate": 8.850913461538462e-06,
|
191899 |
+
"loss": 0.4444,
|
191900 |
+
"step": 72730
|
191901 |
+
},
|
191902 |
+
{
|
191903 |
+
"epoch": 586.49,
|
191904 |
+
"learning_rate": 8.850833333333333e-06,
|
191905 |
+
"loss": 0.3199,
|
191906 |
+
"step": 72735
|
191907 |
+
},
|
191908 |
+
{
|
191909 |
+
"epoch": 586.53,
|
191910 |
+
"learning_rate": 8.850753205128206e-06,
|
191911 |
+
"loss": 0.3793,
|
191912 |
+
"step": 72740
|
191913 |
+
},
|
191914 |
+
{
|
191915 |
+
"epoch": 586.57,
|
191916 |
+
"learning_rate": 8.850673076923078e-06,
|
191917 |
+
"loss": 0.5258,
|
191918 |
+
"step": 72745
|
191919 |
+
},
|
191920 |
+
{
|
191921 |
+
"epoch": 586.61,
|
191922 |
+
"learning_rate": 8.85059294871795e-06,
|
191923 |
+
"loss": 1.1779,
|
191924 |
+
"step": 72750
|
191925 |
+
},
|
191926 |
+
{
|
191927 |
+
"epoch": 586.65,
|
191928 |
+
"learning_rate": 8.85051282051282e-06,
|
191929 |
+
"loss": 0.308,
|
191930 |
+
"step": 72755
|
191931 |
+
},
|
191932 |
+
{
|
191933 |
+
"epoch": 586.69,
|
191934 |
+
"learning_rate": 8.850432692307694e-06,
|
191935 |
+
"loss": 0.3619,
|
191936 |
+
"step": 72760
|
191937 |
+
},
|
191938 |
+
{
|
191939 |
+
"epoch": 586.73,
|
191940 |
+
"learning_rate": 8.850352564102565e-06,
|
191941 |
+
"loss": 0.3743,
|
191942 |
+
"step": 72765
|
191943 |
+
},
|
191944 |
+
{
|
191945 |
+
"epoch": 586.77,
|
191946 |
+
"learning_rate": 8.850272435897436e-06,
|
191947 |
+
"loss": 0.6303,
|
191948 |
+
"step": 72770
|
191949 |
+
},
|
191950 |
+
{
|
191951 |
+
"epoch": 586.81,
|
191952 |
+
"learning_rate": 8.850192307692308e-06,
|
191953 |
+
"loss": 1.0891,
|
191954 |
+
"step": 72775
|
191955 |
+
},
|
191956 |
+
{
|
191957 |
+
"epoch": 586.85,
|
191958 |
+
"learning_rate": 8.85011217948718e-06,
|
191959 |
+
"loss": 0.3039,
|
191960 |
+
"step": 72780
|
191961 |
+
},
|
191962 |
+
{
|
191963 |
+
"epoch": 586.89,
|
191964 |
+
"learning_rate": 8.850032051282052e-06,
|
191965 |
+
"loss": 0.2967,
|
191966 |
+
"step": 72785
|
191967 |
+
},
|
191968 |
+
{
|
191969 |
+
"epoch": 586.93,
|
191970 |
+
"learning_rate": 8.849951923076923e-06,
|
191971 |
+
"loss": 0.3696,
|
191972 |
+
"step": 72790
|
191973 |
+
},
|
191974 |
+
{
|
191975 |
+
"epoch": 586.97,
|
191976 |
+
"learning_rate": 8.849871794871797e-06,
|
191977 |
+
"loss": 0.5783,
|
191978 |
+
"step": 72795
|
191979 |
+
},
|
191980 |
+
{
|
191981 |
+
"epoch": 587.0,
|
191982 |
+
"eval_loss": 0.37081632018089294,
|
191983 |
+
"eval_runtime": 40.9633,
|
191984 |
+
"eval_samples_per_second": 20.506,
|
191985 |
+
"eval_steps_per_second": 0.659,
|
191986 |
+
"eval_wer": 0.1872572990076226,
|
191987 |
+
"step": 72798
|
191988 |
+
},
|
191989 |
+
{
|
191990 |
+
"epoch": 591.02,
|
191991 |
+
"learning_rate": 8.849791666666668e-06,
|
191992 |
+
"loss": 0.4085,
|
191993 |
+
"step": 72800
|
191994 |
+
},
|
191995 |
+
{
|
191996 |
+
"epoch": 591.06,
|
191997 |
+
"learning_rate": 8.84971153846154e-06,
|
191998 |
+
"loss": 0.35,
|
191999 |
+
"step": 72805
|
192000 |
+
},
|
192001 |
+
{
|
192002 |
+
"epoch": 591.1,
|
192003 |
+
"learning_rate": 8.84963141025641e-06,
|
192004 |
+
"loss": 0.3162,
|
192005 |
+
"step": 72810
|
192006 |
+
},
|
192007 |
+
{
|
192008 |
+
"epoch": 591.14,
|
192009 |
+
"learning_rate": 8.849551282051284e-06,
|
192010 |
+
"loss": 0.3517,
|
192011 |
+
"step": 72815
|
192012 |
+
},
|
192013 |
+
{
|
192014 |
+
"epoch": 591.18,
|
192015 |
+
"learning_rate": 8.849471153846155e-06,
|
192016 |
+
"loss": 0.8103,
|
192017 |
+
"step": 72820
|
192018 |
+
},
|
192019 |
+
{
|
192020 |
+
"epoch": 591.22,
|
192021 |
+
"learning_rate": 8.849391025641026e-06,
|
192022 |
+
"loss": 1.0763,
|
192023 |
+
"step": 72825
|
192024 |
+
},
|
192025 |
+
{
|
192026 |
+
"epoch": 591.26,
|
192027 |
+
"learning_rate": 8.849310897435898e-06,
|
192028 |
+
"loss": 0.3132,
|
192029 |
+
"step": 72830
|
192030 |
+
},
|
192031 |
+
{
|
192032 |
+
"epoch": 591.3,
|
192033 |
+
"learning_rate": 8.84923076923077e-06,
|
192034 |
+
"loss": 0.2745,
|
192035 |
+
"step": 72835
|
192036 |
+
},
|
192037 |
+
{
|
192038 |
+
"epoch": 591.34,
|
192039 |
+
"learning_rate": 8.849150641025642e-06,
|
192040 |
+
"loss": 0.3993,
|
192041 |
+
"step": 72840
|
192042 |
+
},
|
192043 |
+
{
|
192044 |
+
"epoch": 591.38,
|
192045 |
+
"learning_rate": 8.849070512820513e-06,
|
192046 |
+
"loss": 0.6536,
|
192047 |
+
"step": 72845
|
192048 |
+
},
|
192049 |
+
{
|
192050 |
+
"epoch": 591.42,
|
192051 |
+
"learning_rate": 8.848990384615387e-06,
|
192052 |
+
"loss": 1.1869,
|
192053 |
+
"step": 72850
|
192054 |
+
},
|
192055 |
+
{
|
192056 |
+
"epoch": 591.46,
|
192057 |
+
"learning_rate": 8.848910256410256e-06,
|
192058 |
+
"loss": 0.2996,
|
192059 |
+
"step": 72855
|
192060 |
+
},
|
192061 |
+
{
|
192062 |
+
"epoch": 591.5,
|
192063 |
+
"learning_rate": 8.84883012820513e-06,
|
192064 |
+
"loss": 0.3045,
|
192065 |
+
"step": 72860
|
192066 |
+
},
|
192067 |
+
{
|
192068 |
+
"epoch": 591.54,
|
192069 |
+
"learning_rate": 8.84875e-06,
|
192070 |
+
"loss": 0.3851,
|
192071 |
+
"step": 72865
|
192072 |
+
},
|
192073 |
+
{
|
192074 |
+
"epoch": 591.58,
|
192075 |
+
"learning_rate": 8.848669871794872e-06,
|
192076 |
+
"loss": 0.6637,
|
192077 |
+
"step": 72870
|
192078 |
+
},
|
192079 |
+
{
|
192080 |
+
"epoch": 591.62,
|
192081 |
+
"learning_rate": 8.848589743589743e-06,
|
192082 |
+
"loss": 1.0274,
|
192083 |
+
"step": 72875
|
192084 |
+
},
|
192085 |
+
{
|
192086 |
+
"epoch": 591.66,
|
192087 |
+
"learning_rate": 8.848509615384616e-06,
|
192088 |
+
"loss": 0.3225,
|
192089 |
+
"step": 72880
|
192090 |
+
},
|
192091 |
+
{
|
192092 |
+
"epoch": 591.7,
|
192093 |
+
"learning_rate": 8.848429487179488e-06,
|
192094 |
+
"loss": 0.3141,
|
192095 |
+
"step": 72885
|
192096 |
+
},
|
192097 |
+
{
|
192098 |
+
"epoch": 591.74,
|
192099 |
+
"learning_rate": 8.848349358974359e-06,
|
192100 |
+
"loss": 0.373,
|
192101 |
+
"step": 72890
|
192102 |
+
},
|
192103 |
+
{
|
192104 |
+
"epoch": 591.79,
|
192105 |
+
"learning_rate": 8.848269230769232e-06,
|
192106 |
+
"loss": 0.9046,
|
192107 |
+
"step": 72895
|
192108 |
+
},
|
192109 |
+
{
|
192110 |
+
"epoch": 591.83,
|
192111 |
+
"learning_rate": 8.848189102564104e-06,
|
192112 |
+
"loss": 1.149,
|
192113 |
+
"step": 72900
|
192114 |
+
},
|
192115 |
+
{
|
192116 |
+
"epoch": 591.87,
|
192117 |
+
"learning_rate": 8.848108974358975e-06,
|
192118 |
+
"loss": 0.3039,
|
192119 |
+
"step": 72905
|
192120 |
+
},
|
192121 |
+
{
|
192122 |
+
"epoch": 591.91,
|
192123 |
+
"learning_rate": 8.848028846153846e-06,
|
192124 |
+
"loss": 0.562,
|
192125 |
+
"step": 72910
|
192126 |
+
},
|
192127 |
+
{
|
192128 |
+
"epoch": 591.95,
|
192129 |
+
"learning_rate": 8.84794871794872e-06,
|
192130 |
+
"loss": 0.4599,
|
192131 |
+
"step": 72915
|
192132 |
+
},
|
192133 |
+
{
|
192134 |
+
"epoch": 591.99,
|
192135 |
+
"learning_rate": 8.84786858974359e-06,
|
192136 |
+
"loss": 0.9062,
|
192137 |
+
"step": 72920
|
192138 |
+
},
|
192139 |
+
{
|
192140 |
+
"epoch": 592.0,
|
192141 |
+
"eval_loss": 0.39401471614837646,
|
192142 |
+
"eval_runtime": 38.9267,
|
192143 |
+
"eval_samples_per_second": 21.579,
|
192144 |
+
"eval_steps_per_second": 0.694,
|
192145 |
+
"eval_wer": 0.18718740806119447,
|
192146 |
+
"step": 72921
|
192147 |
+
},
|
192148 |
+
{
|
192149 |
+
"epoch": 588.03,
|
192150 |
+
"learning_rate": 8.847788461538462e-06,
|
192151 |
+
"loss": 0.3517,
|
192152 |
+
"step": 72925
|
192153 |
+
},
|
192154 |
+
{
|
192155 |
+
"epoch": 588.07,
|
192156 |
+
"learning_rate": 8.847708333333333e-06,
|
192157 |
+
"loss": 0.3238,
|
192158 |
+
"step": 72930
|
192159 |
+
},
|
192160 |
+
{
|
192161 |
+
"epoch": 588.11,
|
192162 |
+
"learning_rate": 8.847628205128206e-06,
|
192163 |
+
"loss": 0.3437,
|
192164 |
+
"step": 72935
|
192165 |
+
},
|
192166 |
+
{
|
192167 |
+
"epoch": 588.15,
|
192168 |
+
"learning_rate": 8.847548076923078e-06,
|
192169 |
+
"loss": 0.4455,
|
192170 |
+
"step": 72940
|
192171 |
+
},
|
192172 |
+
{
|
192173 |
+
"epoch": 588.19,
|
192174 |
+
"learning_rate": 8.847467948717949e-06,
|
192175 |
+
"loss": 0.9492,
|
192176 |
+
"step": 72945
|
192177 |
+
},
|
192178 |
+
{
|
192179 |
+
"epoch": 588.23,
|
192180 |
+
"learning_rate": 8.847387820512822e-06,
|
192181 |
+
"loss": 0.6762,
|
192182 |
+
"step": 72950
|
192183 |
+
},
|
192184 |
+
{
|
192185 |
+
"epoch": 588.27,
|
192186 |
+
"learning_rate": 8.847307692307694e-06,
|
192187 |
+
"loss": 0.3302,
|
192188 |
+
"step": 72955
|
192189 |
+
},
|
192190 |
+
{
|
192191 |
+
"epoch": 588.31,
|
192192 |
+
"learning_rate": 8.847227564102565e-06,
|
192193 |
+
"loss": 0.3051,
|
192194 |
+
"step": 72960
|
192195 |
+
},
|
192196 |
+
{
|
192197 |
+
"epoch": 588.35,
|
192198 |
+
"learning_rate": 8.847147435897436e-06,
|
192199 |
+
"loss": 0.3876,
|
192200 |
+
"step": 72965
|
192201 |
+
},
|
192202 |
+
{
|
192203 |
+
"epoch": 588.39,
|
192204 |
+
"learning_rate": 8.84706730769231e-06,
|
192205 |
+
"loss": 0.921,
|
192206 |
+
"step": 72970
|
192207 |
+
},
|
192208 |
+
{
|
192209 |
+
"epoch": 588.43,
|
192210 |
+
"learning_rate": 8.846987179487179e-06,
|
192211 |
+
"loss": 0.6965,
|
192212 |
+
"step": 72975
|
192213 |
+
},
|
192214 |
+
{
|
192215 |
+
"epoch": 588.47,
|
192216 |
+
"learning_rate": 8.846907051282052e-06,
|
192217 |
+
"loss": 0.2984,
|
192218 |
+
"step": 72980
|
192219 |
+
},
|
192220 |
+
{
|
192221 |
+
"epoch": 588.51,
|
192222 |
+
"learning_rate": 8.846826923076923e-06,
|
192223 |
+
"loss": 0.3759,
|
192224 |
+
"step": 72985
|
192225 |
+
},
|
192226 |
+
{
|
192227 |
+
"epoch": 588.55,
|
192228 |
+
"learning_rate": 8.846746794871795e-06,
|
192229 |
+
"loss": 0.4898,
|
192230 |
+
"step": 72990
|
192231 |
+
},
|
192232 |
+
{
|
192233 |
+
"epoch": 588.59,
|
192234 |
+
"learning_rate": 8.846666666666668e-06,
|
192235 |
+
"loss": 1.0942,
|
192236 |
+
"step": 72995
|
192237 |
+
},
|
192238 |
+
{
|
192239 |
+
"epoch": 588.63,
|
192240 |
+
"learning_rate": 8.84658653846154e-06,
|
192241 |
+
"loss": 0.6618,
|
192242 |
+
"step": 73000
|
192243 |
+
},
|
192244 |
+
{
|
192245 |
+
"epoch": 588.67,
|
192246 |
+
"learning_rate": 8.84650641025641e-06,
|
192247 |
+
"loss": 0.355,
|
192248 |
+
"step": 73005
|
192249 |
+
},
|
192250 |
+
{
|
192251 |
+
"epoch": 588.71,
|
192252 |
+
"learning_rate": 8.846426282051282e-06,
|
192253 |
+
"loss": 0.3374,
|
192254 |
+
"step": 73010
|
192255 |
+
},
|
192256 |
+
{
|
192257 |
+
"epoch": 588.76,
|
192258 |
+
"learning_rate": 8.846346153846155e-06,
|
192259 |
+
"loss": 0.3638,
|
192260 |
+
"step": 73015
|
192261 |
+
},
|
192262 |
+
{
|
192263 |
+
"epoch": 588.8,
|
192264 |
+
"learning_rate": 8.846266025641026e-06,
|
192265 |
+
"loss": 0.9145,
|
192266 |
+
"step": 73020
|
192267 |
+
},
|
192268 |
+
{
|
192269 |
+
"epoch": 588.84,
|
192270 |
+
"learning_rate": 8.846185897435898e-06,
|
192271 |
+
"loss": 0.6553,
|
192272 |
+
"step": 73025
|
192273 |
+
},
|
192274 |
+
{
|
192275 |
+
"epoch": 588.88,
|
192276 |
+
"learning_rate": 8.846105769230769e-06,
|
192277 |
+
"loss": 0.3028,
|
192278 |
+
"step": 73030
|
192279 |
+
},
|
192280 |
+
{
|
192281 |
+
"epoch": 588.92,
|
192282 |
+
"learning_rate": 8.846025641025642e-06,
|
192283 |
+
"loss": 0.3749,
|
192284 |
+
"step": 73035
|
192285 |
+
},
|
192286 |
+
{
|
192287 |
+
"epoch": 588.96,
|
192288 |
+
"learning_rate": 8.845945512820513e-06,
|
192289 |
+
"loss": 0.4571,
|
192290 |
+
"step": 73040
|
192291 |
+
},
|
192292 |
+
{
|
192293 |
+
"epoch": 589.0,
|
192294 |
+
"learning_rate": 8.845865384615385e-06,
|
192295 |
+
"loss": 1.2306,
|
192296 |
+
"step": 73045
|
192297 |
+
},
|
192298 |
+
{
|
192299 |
+
"epoch": 589.0,
|
192300 |
+
"eval_loss": 0.4083371162414551,
|
192301 |
+
"eval_runtime": 39.5334,
|
192302 |
+
"eval_samples_per_second": 21.248,
|
192303 |
+
"eval_steps_per_second": 0.683,
|
192304 |
+
"eval_wer": 0.19103616596405593,
|
192305 |
+
"step": 73045
|
192306 |
+
},
|
192307 |
+
{
|
192308 |
+
"epoch": 589.04,
|
192309 |
+
"learning_rate": 8.845785256410258e-06,
|
192310 |
+
"loss": 0.3375,
|
192311 |
+
"step": 73050
|
192312 |
+
},
|
192313 |
+
{
|
192314 |
+
"epoch": 589.08,
|
192315 |
+
"learning_rate": 8.84570512820513e-06,
|
192316 |
+
"loss": 0.2793,
|
192317 |
+
"step": 73055
|
192318 |
+
},
|
192319 |
+
{
|
192320 |
+
"epoch": 589.12,
|
192321 |
+
"learning_rate": 8.845625e-06,
|
192322 |
+
"loss": 0.3388,
|
192323 |
+
"step": 73060
|
192324 |
+
},
|
192325 |
+
{
|
192326 |
+
"epoch": 589.16,
|
192327 |
+
"learning_rate": 8.845544871794872e-06,
|
192328 |
+
"loss": 0.4248,
|
192329 |
+
"step": 73065
|
192330 |
+
},
|
192331 |
+
{
|
192332 |
+
"epoch": 589.2,
|
192333 |
+
"learning_rate": 8.845464743589745e-06,
|
192334 |
+
"loss": 1.1882,
|
192335 |
+
"step": 73070
|
192336 |
+
},
|
192337 |
+
{
|
192338 |
+
"epoch": 589.24,
|
192339 |
+
"learning_rate": 8.845384615384616e-06,
|
192340 |
+
"loss": 0.2749,
|
192341 |
+
"step": 73075
|
192342 |
+
},
|
192343 |
+
{
|
192344 |
+
"epoch": 589.28,
|
192345 |
+
"learning_rate": 8.845304487179488e-06,
|
192346 |
+
"loss": 0.4075,
|
192347 |
+
"step": 73080
|
192348 |
+
},
|
192349 |
+
{
|
192350 |
+
"epoch": 589.32,
|
192351 |
+
"learning_rate": 8.845224358974359e-06,
|
192352 |
+
"loss": 0.3066,
|
192353 |
+
"step": 73085
|
192354 |
+
},
|
192355 |
+
{
|
192356 |
+
"epoch": 589.36,
|
192357 |
+
"learning_rate": 8.845144230769232e-06,
|
192358 |
+
"loss": 0.5385,
|
192359 |
+
"step": 73090
|
192360 |
+
},
|
192361 |
+
{
|
192362 |
+
"epoch": 589.4,
|
192363 |
+
"learning_rate": 8.845064102564103e-06,
|
192364 |
+
"loss": 1.1468,
|
192365 |
+
"step": 73095
|
192366 |
+
},
|
192367 |
+
{
|
192368 |
+
"epoch": 589.44,
|
192369 |
+
"learning_rate": 8.844983974358975e-06,
|
192370 |
+
"loss": 0.3261,
|
192371 |
+
"step": 73100
|
192372 |
+
},
|
192373 |
+
{
|
192374 |
+
"epoch": 589.48,
|
192375 |
+
"learning_rate": 8.844903846153848e-06,
|
192376 |
+
"loss": 0.2908,
|
192377 |
+
"step": 73105
|
192378 |
+
},
|
192379 |
+
{
|
192380 |
+
"epoch": 589.52,
|
192381 |
+
"learning_rate": 8.84482371794872e-06,
|
192382 |
+
"loss": 0.418,
|
192383 |
+
"step": 73110
|
192384 |
+
},
|
192385 |
+
{
|
192386 |
+
"epoch": 589.56,
|
192387 |
+
"learning_rate": 8.84474358974359e-06,
|
192388 |
+
"loss": 0.4959,
|
192389 |
+
"step": 73115
|
192390 |
+
},
|
192391 |
+
{
|
192392 |
+
"epoch": 589.6,
|
192393 |
+
"learning_rate": 8.844663461538462e-06,
|
192394 |
+
"loss": 1.0528,
|
192395 |
+
"step": 73120
|
192396 |
+
},
|
192397 |
+
{
|
192398 |
+
"epoch": 589.64,
|
192399 |
+
"learning_rate": 8.844583333333335e-06,
|
192400 |
+
"loss": 0.3883,
|
192401 |
+
"step": 73125
|
192402 |
+
},
|
192403 |
+
{
|
192404 |
+
"epoch": 589.68,
|
192405 |
+
"learning_rate": 8.844503205128205e-06,
|
192406 |
+
"loss": 0.3486,
|
192407 |
+
"step": 73130
|
192408 |
+
},
|
192409 |
+
{
|
192410 |
+
"epoch": 589.72,
|
192411 |
+
"learning_rate": 8.844423076923078e-06,
|
192412 |
+
"loss": 0.2976,
|
192413 |
+
"step": 73135
|
192414 |
+
},
|
192415 |
+
{
|
192416 |
+
"epoch": 589.76,
|
192417 |
+
"learning_rate": 8.84434294871795e-06,
|
192418 |
+
"loss": 0.4688,
|
192419 |
+
"step": 73140
|
192420 |
+
},
|
192421 |
+
{
|
192422 |
+
"epoch": 589.8,
|
192423 |
+
"learning_rate": 8.84426282051282e-06,
|
192424 |
+
"loss": 1.2549,
|
192425 |
+
"step": 73145
|
192426 |
+
},
|
192427 |
+
{
|
192428 |
+
"epoch": 589.84,
|
192429 |
+
"learning_rate": 8.844182692307693e-06,
|
192430 |
+
"loss": 0.3377,
|
192431 |
+
"step": 73150
|
192432 |
+
},
|
192433 |
+
{
|
192434 |
+
"epoch": 589.88,
|
192435 |
+
"learning_rate": 8.844102564102565e-06,
|
192436 |
+
"loss": 0.2784,
|
192437 |
+
"step": 73155
|
192438 |
+
},
|
192439 |
+
{
|
192440 |
+
"epoch": 589.92,
|
192441 |
+
"learning_rate": 8.844022435897436e-06,
|
192442 |
+
"loss": 0.3856,
|
192443 |
+
"step": 73160
|
192444 |
+
},
|
192445 |
+
{
|
192446 |
+
"epoch": 589.96,
|
192447 |
+
"learning_rate": 8.843942307692308e-06,
|
192448 |
+
"loss": 0.5648,
|
192449 |
+
"step": 73165
|
192450 |
+
},
|
192451 |
+
{
|
192452 |
+
"epoch": 590.0,
|
192453 |
+
"eval_loss": 0.3960420787334442,
|
192454 |
+
"eval_runtime": 39.2969,
|
192455 |
+
"eval_samples_per_second": 21.376,
|
192456 |
+
"eval_steps_per_second": 0.687,
|
192457 |
+
"eval_wer": 0.19401051040241885,
|
192458 |
+
"step": 73169
|
192459 |
+
},
|
192460 |
+
{
|
192461 |
+
"epoch": 585.01,
|
192462 |
+
"learning_rate": 8.84386217948718e-06,
|
192463 |
+
"loss": 0.4584,
|
192464 |
+
"step": 73170
|
192465 |
+
},
|
192466 |
+
{
|
192467 |
+
"epoch": 585.05,
|
192468 |
+
"learning_rate": 8.843782051282052e-06,
|
192469 |
+
"loss": 0.3261,
|
192470 |
+
"step": 73175
|
192471 |
+
},
|
192472 |
+
{
|
192473 |
+
"epoch": 585.09,
|
192474 |
+
"learning_rate": 8.843701923076923e-06,
|
192475 |
+
"loss": 0.3012,
|
192476 |
+
"step": 73180
|
192477 |
+
},
|
192478 |
+
{
|
192479 |
+
"epoch": 585.13,
|
192480 |
+
"learning_rate": 8.843621794871795e-06,
|
192481 |
+
"loss": 0.3411,
|
192482 |
+
"step": 73185
|
192483 |
+
},
|
192484 |
+
{
|
192485 |
+
"epoch": 585.17,
|
192486 |
+
"learning_rate": 8.843541666666668e-06,
|
192487 |
+
"loss": 0.5135,
|
192488 |
+
"step": 73190
|
192489 |
+
},
|
192490 |
+
{
|
192491 |
+
"epoch": 585.21,
|
192492 |
+
"learning_rate": 8.843461538461539e-06,
|
192493 |
+
"loss": 1.1694,
|
192494 |
+
"step": 73195
|
192495 |
+
},
|
192496 |
+
{
|
192497 |
+
"epoch": 585.25,
|
192498 |
+
"learning_rate": 8.84338141025641e-06,
|
192499 |
+
"loss": 0.3068,
|
192500 |
+
"step": 73200
|
192501 |
+
},
|
192502 |
+
{
|
192503 |
+
"epoch": 585.29,
|
192504 |
+
"learning_rate": 8.843301282051283e-06,
|
192505 |
+
"loss": 0.3249,
|
192506 |
+
"step": 73205
|
192507 |
+
},
|
192508 |
+
{
|
192509 |
+
"epoch": 585.33,
|
192510 |
+
"learning_rate": 8.843221153846155e-06,
|
192511 |
+
"loss": 0.3034,
|
192512 |
+
"step": 73210
|
192513 |
+
},
|
192514 |
+
{
|
192515 |
+
"epoch": 585.37,
|
192516 |
+
"learning_rate": 8.843141025641026e-06,
|
192517 |
+
"loss": 0.5366,
|
192518 |
+
"step": 73215
|
192519 |
+
},
|
192520 |
+
{
|
192521 |
+
"epoch": 585.41,
|
192522 |
+
"learning_rate": 8.843060897435898e-06,
|
192523 |
+
"loss": 1.1144,
|
192524 |
+
"step": 73220
|
192525 |
+
},
|
192526 |
+
{
|
192527 |
+
"epoch": 585.45,
|
192528 |
+
"learning_rate": 8.84298076923077e-06,
|
192529 |
+
"loss": 0.3196,
|
192530 |
+
"step": 73225
|
192531 |
+
},
|
192532 |
+
{
|
192533 |
+
"epoch": 585.49,
|
192534 |
+
"learning_rate": 8.842900641025642e-06,
|
192535 |
+
"loss": 0.2776,
|
192536 |
+
"step": 73230
|
192537 |
+
},
|
192538 |
+
{
|
192539 |
+
"epoch": 585.53,
|
192540 |
+
"learning_rate": 8.842820512820513e-06,
|
192541 |
+
"loss": 0.3681,
|
192542 |
+
"step": 73235
|
192543 |
+
},
|
192544 |
+
{
|
192545 |
+
"epoch": 585.57,
|
192546 |
+
"learning_rate": 8.842740384615386e-06,
|
192547 |
+
"loss": 0.5304,
|
192548 |
+
"step": 73240
|
192549 |
+
},
|
192550 |
+
{
|
192551 |
+
"epoch": 585.61,
|
192552 |
+
"learning_rate": 8.842660256410258e-06,
|
192553 |
+
"loss": 1.0968,
|
192554 |
+
"step": 73245
|
192555 |
+
},
|
192556 |
+
{
|
192557 |
+
"epoch": 585.65,
|
192558 |
+
"learning_rate": 8.842580128205129e-06,
|
192559 |
+
"loss": 0.3225,
|
192560 |
+
"step": 73250
|
192561 |
+
},
|
192562 |
+
{
|
192563 |
+
"epoch": 585.69,
|
192564 |
+
"learning_rate": 8.8425e-06,
|
192565 |
+
"loss": 0.3974,
|
192566 |
+
"step": 73255
|
192567 |
+
},
|
192568 |
+
{
|
192569 |
+
"epoch": 585.73,
|
192570 |
+
"learning_rate": 8.842419871794874e-06,
|
192571 |
+
"loss": 0.3627,
|
192572 |
+
"step": 73260
|
192573 |
+
},
|
192574 |
+
{
|
192575 |
+
"epoch": 585.77,
|
192576 |
+
"learning_rate": 8.842339743589743e-06,
|
192577 |
+
"loss": 0.6047,
|
192578 |
+
"step": 73265
|
192579 |
+
},
|
192580 |
+
{
|
192581 |
+
"epoch": 585.81,
|
192582 |
+
"learning_rate": 8.842259615384616e-06,
|
192583 |
+
"loss": 1.4464,
|
192584 |
+
"step": 73270
|
192585 |
+
},
|
192586 |
+
{
|
192587 |
+
"epoch": 585.85,
|
192588 |
+
"learning_rate": 8.842179487179488e-06,
|
192589 |
+
"loss": 0.3017,
|
192590 |
+
"step": 73275
|
192591 |
+
},
|
192592 |
+
{
|
192593 |
+
"epoch": 585.89,
|
192594 |
+
"learning_rate": 8.842099358974359e-06,
|
192595 |
+
"loss": 0.3094,
|
192596 |
+
"step": 73280
|
192597 |
+
},
|
192598 |
+
{
|
192599 |
+
"epoch": 585.93,
|
192600 |
+
"learning_rate": 8.84201923076923e-06,
|
192601 |
+
"loss": 0.3192,
|
192602 |
+
"step": 73285
|
192603 |
+
},
|
192604 |
+
{
|
192605 |
+
"epoch": 585.97,
|
192606 |
+
"learning_rate": 8.841939102564103e-06,
|
192607 |
+
"loss": 0.592,
|
192608 |
+
"step": 73290
|
192609 |
+
},
|
192610 |
+
{
|
192611 |
+
"epoch": 586.0,
|
192612 |
+
"eval_loss": 0.44540178775787354,
|
192613 |
+
"eval_runtime": 41.2543,
|
192614 |
+
"eval_samples_per_second": 20.362,
|
192615 |
+
"eval_steps_per_second": 0.654,
|
192616 |
+
"eval_wer": 0.19331724793448857,
|
192617 |
+
"step": 73294
|
192618 |
}
|
192619 |
],
|
192620 |
"max_steps": 625000,
|
192621 |
"num_train_epochs": 5000,
|
192622 |
+
"total_flos": 2.0626163435744595e+20,
|
192623 |
"trial_name": null,
|
192624 |
"trial_params": null
|
192625 |
}
|
model-bin/finetune/base/{checkpoint-72674 β checkpoint-73294}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1629869061.2637658/events.out.tfevents.1629869061.7e498afd5545.905.53
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ae8b52d1cd0d21233e15daa0c3fd1f10410692d55db06adb0bad90d9bd14f04
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629869731.455241/events.out.tfevents.1629869731.7e498afd5545.905.55
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7b7bba2db6b9aa26631c81409a7a206463883d621c7b84c66643a8fa0a16ac4
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629870382.7916713/events.out.tfevents.1629870382.7e498afd5545.905.57
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d6e6a23d33d554930a0d80fc0746a3000a6f3f14c8bb3842ebee55a1d8b1ac80
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629871038.1608121/events.out.tfevents.1629871038.7e498afd5545.905.59
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97f08523b6dc5e662dc16ebc30317ef82b0be6dfd390188c36e6c2fbe39ccb62
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629871680.9885855/events.out.tfevents.1629871680.7e498afd5545.905.61
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb95ec1ce9f723a2f319980d8803678ca13d76e0d9bf34a758a9cf0fbff7f4c8
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629869061.7e498afd5545.905.52
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db1917f0e53b299d476406dc2c4e430220decc4c355ed43929b974e548f41a5c
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629869731.7e498afd5545.905.54
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:142e02e858d8dafb0636532ea92a864b0da56c0bfe2dbd84abfd6af74ec9493c
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629870382.7e498afd5545.905.56
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e64d8a2f979a06e5f77a62fafba02e43e3c683e6fd3b4277f7583f5ff75d66b3
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629871038.7e498afd5545.905.58
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d783b37f9b96d3ce9ed3ff18f0a35d766c1a1e41750eca9b79e9ccda15e23af7
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1629871680.7e498afd5545.905.60
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1d523aa3e926f22f482c94d89c4d63a07b628e2e1859a7970fd874cfb8e35232
|
3 |
+
size 8622
|