"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630250753.668995/events.out.tfevents.1630250753.cc93b136ebf5.1086.209 +3 -0
- model-bin/finetune/base/log/1630251188.88167/events.out.tfevents.1630251188.cc93b136ebf5.1086.211 +3 -0
- model-bin/finetune/base/log/1630251628.1797204/events.out.tfevents.1630251628.cc93b136ebf5.1086.213 +3 -0
- model-bin/finetune/base/log/1630252065.7514985/events.out.tfevents.1630252065.cc93b136ebf5.1086.215 +3 -0
- model-bin/finetune/base/log/1630252500.5741549/events.out.tfevents.1630252500.cc93b136ebf5.1086.217 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630250753.cc93b136ebf5.1086.208 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630251188.cc93b136ebf5.1086.210 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630251628.cc93b136ebf5.1086.212 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630252065.cc93b136ebf5.1086.214 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630252500.cc93b136ebf5.1086.216 +3 -0
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:013e7596c9833c0f02cb6d858875c8b42145994a60410ff7acc22c55ed3a3f78
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:358bbb0a31f3eada3dcf3518c389f01e7260b174069b7a039e533c23adcbc695
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c64726b931c7ed69c0aa3b500fe65f0f56bb20ca55efd46d255f359a9d38051f
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e38d467b638b9bb9004c2051e9035f041fc23da780dcb2db9c2dbc1bb4f1e23d
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1740c53bf7ef796ec727899064cf449f45e9e8026922bf22f64ed1a0b45bd32d
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -292842,11 +292842,806 @@
|
|
292842 |
"eval_steps_per_second": 0.667,
|
292843 |
"eval_wer": 0.1764970733903647,
|
292844 |
"step": 152068
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
292845 |
}
|
292846 |
],
|
292847 |
-
"max_steps":
|
292848 |
"num_train_epochs": 5000,
|
292849 |
-
"total_flos": 4.
|
292850 |
"trial_name": null,
|
292851 |
"trial_params": null
|
292852 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1221.0,
|
5 |
+
"global_step": 152690,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
292842 |
"eval_steps_per_second": 0.667,
|
292843 |
"eval_wer": 0.1764970733903647,
|
292844 |
"step": 152068
|
292845 |
+
},
|
292846 |
+
{
|
292847 |
+
"epoch": 1226.02,
|
292848 |
+
"learning_rate": 7.560403877221325e-06,
|
292849 |
+
"loss": 0.3566,
|
292850 |
+
"step": 152070
|
292851 |
+
},
|
292852 |
+
{
|
292853 |
+
"epoch": 1226.06,
|
292854 |
+
"learning_rate": 7.5603231017770605e-06,
|
292855 |
+
"loss": 0.2873,
|
292856 |
+
"step": 152075
|
292857 |
+
},
|
292858 |
+
{
|
292859 |
+
"epoch": 1226.1,
|
292860 |
+
"learning_rate": 7.5602423263327955e-06,
|
292861 |
+
"loss": 0.2621,
|
292862 |
+
"step": 152080
|
292863 |
+
},
|
292864 |
+
{
|
292865 |
+
"epoch": 1226.14,
|
292866 |
+
"learning_rate": 7.5601615508885305e-06,
|
292867 |
+
"loss": 0.3207,
|
292868 |
+
"step": 152085
|
292869 |
+
},
|
292870 |
+
{
|
292871 |
+
"epoch": 1226.18,
|
292872 |
+
"learning_rate": 7.5600807754442655e-06,
|
292873 |
+
"loss": 0.7188,
|
292874 |
+
"step": 152090
|
292875 |
+
},
|
292876 |
+
{
|
292877 |
+
"epoch": 1226.22,
|
292878 |
+
"learning_rate": 7.5600000000000005e-06,
|
292879 |
+
"loss": 1.0223,
|
292880 |
+
"step": 152095
|
292881 |
+
},
|
292882 |
+
{
|
292883 |
+
"epoch": 1226.26,
|
292884 |
+
"learning_rate": 7.5599192245557355e-06,
|
292885 |
+
"loss": 0.2622,
|
292886 |
+
"step": 152100
|
292887 |
+
},
|
292888 |
+
{
|
292889 |
+
"epoch": 1226.3,
|
292890 |
+
"learning_rate": 7.5598384491114704e-06,
|
292891 |
+
"loss": 0.2759,
|
292892 |
+
"step": 152105
|
292893 |
+
},
|
292894 |
+
{
|
292895 |
+
"epoch": 1226.34,
|
292896 |
+
"learning_rate": 7.559757673667206e-06,
|
292897 |
+
"loss": 0.3349,
|
292898 |
+
"step": 152110
|
292899 |
+
},
|
292900 |
+
{
|
292901 |
+
"epoch": 1226.38,
|
292902 |
+
"learning_rate": 7.55967689822294e-06,
|
292903 |
+
"loss": 0.5856,
|
292904 |
+
"step": 152115
|
292905 |
+
},
|
292906 |
+
{
|
292907 |
+
"epoch": 1226.42,
|
292908 |
+
"learning_rate": 7.559596122778676e-06,
|
292909 |
+
"loss": 0.9452,
|
292910 |
+
"step": 152120
|
292911 |
+
},
|
292912 |
+
{
|
292913 |
+
"epoch": 1226.46,
|
292914 |
+
"learning_rate": 7.55951534733441e-06,
|
292915 |
+
"loss": 0.2844,
|
292916 |
+
"step": 152125
|
292917 |
+
},
|
292918 |
+
{
|
292919 |
+
"epoch": 1226.5,
|
292920 |
+
"learning_rate": 7.559434571890146e-06,
|
292921 |
+
"loss": 0.2637,
|
292922 |
+
"step": 152130
|
292923 |
+
},
|
292924 |
+
{
|
292925 |
+
"epoch": 1226.54,
|
292926 |
+
"learning_rate": 7.55935379644588e-06,
|
292927 |
+
"loss": 0.2986,
|
292928 |
+
"step": 152135
|
292929 |
+
},
|
292930 |
+
{
|
292931 |
+
"epoch": 1226.58,
|
292932 |
+
"learning_rate": 7.559273021001616e-06,
|
292933 |
+
"loss": 0.6334,
|
292934 |
+
"step": 152140
|
292935 |
+
},
|
292936 |
+
{
|
292937 |
+
"epoch": 1226.62,
|
292938 |
+
"learning_rate": 7.55919224555735e-06,
|
292939 |
+
"loss": 0.9182,
|
292940 |
+
"step": 152145
|
292941 |
+
},
|
292942 |
+
{
|
292943 |
+
"epoch": 1226.66,
|
292944 |
+
"learning_rate": 7.559111470113086e-06,
|
292945 |
+
"loss": 0.2254,
|
292946 |
+
"step": 152150
|
292947 |
+
},
|
292948 |
+
{
|
292949 |
+
"epoch": 1226.7,
|
292950 |
+
"learning_rate": 7.559030694668821e-06,
|
292951 |
+
"loss": 0.2388,
|
292952 |
+
"step": 152155
|
292953 |
+
},
|
292954 |
+
{
|
292955 |
+
"epoch": 1226.74,
|
292956 |
+
"learning_rate": 7.558949919224556e-06,
|
292957 |
+
"loss": 0.3118,
|
292958 |
+
"step": 152160
|
292959 |
+
},
|
292960 |
+
{
|
292961 |
+
"epoch": 1226.78,
|
292962 |
+
"learning_rate": 7.558869143780292e-06,
|
292963 |
+
"loss": 0.6253,
|
292964 |
+
"step": 152165
|
292965 |
+
},
|
292966 |
+
{
|
292967 |
+
"epoch": 1226.82,
|
292968 |
+
"learning_rate": 7.558788368336026e-06,
|
292969 |
+
"loss": 0.9901,
|
292970 |
+
"step": 152170
|
292971 |
+
},
|
292972 |
+
{
|
292973 |
+
"epoch": 1226.86,
|
292974 |
+
"learning_rate": 7.558707592891762e-06,
|
292975 |
+
"loss": 0.2865,
|
292976 |
+
"step": 152175
|
292977 |
+
},
|
292978 |
+
{
|
292979 |
+
"epoch": 1226.9,
|
292980 |
+
"learning_rate": 7.558626817447496e-06,
|
292981 |
+
"loss": 0.3002,
|
292982 |
+
"step": 152180
|
292983 |
+
},
|
292984 |
+
{
|
292985 |
+
"epoch": 1226.94,
|
292986 |
+
"learning_rate": 7.558546042003232e-06,
|
292987 |
+
"loss": 0.3956,
|
292988 |
+
"step": 152185
|
292989 |
+
},
|
292990 |
+
{
|
292991 |
+
"epoch": 1226.98,
|
292992 |
+
"learning_rate": 7.558465266558966e-06,
|
292993 |
+
"loss": 0.687,
|
292994 |
+
"step": 152190
|
292995 |
+
},
|
292996 |
+
{
|
292997 |
+
"epoch": 1227.0,
|
292998 |
+
"eval_loss": 0.39228349924087524,
|
292999 |
+
"eval_runtime": 41.5829,
|
293000 |
+
"eval_samples_per_second": 20.104,
|
293001 |
+
"eval_steps_per_second": 0.649,
|
293002 |
+
"eval_wer": 0.18111669214530102,
|
293003 |
+
"step": 152192
|
293004 |
+
},
|
293005 |
+
{
|
293006 |
+
"epoch": 1227.02,
|
293007 |
+
"learning_rate": 7.558384491114702e-06,
|
293008 |
+
"loss": 0.2944,
|
293009 |
+
"step": 152195
|
293010 |
+
},
|
293011 |
+
{
|
293012 |
+
"epoch": 1227.06,
|
293013 |
+
"learning_rate": 7.558303715670436e-06,
|
293014 |
+
"loss": 0.2883,
|
293015 |
+
"step": 152200
|
293016 |
+
},
|
293017 |
+
{
|
293018 |
+
"epoch": 1227.1,
|
293019 |
+
"learning_rate": 7.558222940226172e-06,
|
293020 |
+
"loss": 0.299,
|
293021 |
+
"step": 152205
|
293022 |
+
},
|
293023 |
+
{
|
293024 |
+
"epoch": 1227.14,
|
293025 |
+
"learning_rate": 7.558142164781906e-06,
|
293026 |
+
"loss": 0.2715,
|
293027 |
+
"step": 152210
|
293028 |
+
},
|
293029 |
+
{
|
293030 |
+
"epoch": 1227.18,
|
293031 |
+
"learning_rate": 7.558061389337642e-06,
|
293032 |
+
"loss": 0.7032,
|
293033 |
+
"step": 152215
|
293034 |
+
},
|
293035 |
+
{
|
293036 |
+
"epoch": 1227.22,
|
293037 |
+
"learning_rate": 7.557980613893378e-06,
|
293038 |
+
"loss": 0.889,
|
293039 |
+
"step": 152220
|
293040 |
+
},
|
293041 |
+
{
|
293042 |
+
"epoch": 1227.27,
|
293043 |
+
"learning_rate": 7.557899838449112e-06,
|
293044 |
+
"loss": 0.2777,
|
293045 |
+
"step": 152225
|
293046 |
+
},
|
293047 |
+
{
|
293048 |
+
"epoch": 1227.31,
|
293049 |
+
"learning_rate": 7.557819063004848e-06,
|
293050 |
+
"loss": 0.3006,
|
293051 |
+
"step": 152230
|
293052 |
+
},
|
293053 |
+
{
|
293054 |
+
"epoch": 1227.35,
|
293055 |
+
"learning_rate": 7.557738287560582e-06,
|
293056 |
+
"loss": 0.3308,
|
293057 |
+
"step": 152235
|
293058 |
+
},
|
293059 |
+
{
|
293060 |
+
"epoch": 1227.39,
|
293061 |
+
"learning_rate": 7.557657512116318e-06,
|
293062 |
+
"loss": 0.7281,
|
293063 |
+
"step": 152240
|
293064 |
+
},
|
293065 |
+
{
|
293066 |
+
"epoch": 1227.43,
|
293067 |
+
"learning_rate": 7.557576736672052e-06,
|
293068 |
+
"loss": 0.8714,
|
293069 |
+
"step": 152245
|
293070 |
+
},
|
293071 |
+
{
|
293072 |
+
"epoch": 1227.47,
|
293073 |
+
"learning_rate": 7.557495961227788e-06,
|
293074 |
+
"loss": 0.2962,
|
293075 |
+
"step": 152250
|
293076 |
+
},
|
293077 |
+
{
|
293078 |
+
"epoch": 1227.51,
|
293079 |
+
"learning_rate": 7.557415185783522e-06,
|
293080 |
+
"loss": 0.2503,
|
293081 |
+
"step": 152255
|
293082 |
+
},
|
293083 |
+
{
|
293084 |
+
"epoch": 1227.55,
|
293085 |
+
"learning_rate": 7.557334410339258e-06,
|
293086 |
+
"loss": 0.3374,
|
293087 |
+
"step": 152260
|
293088 |
+
},
|
293089 |
+
{
|
293090 |
+
"epoch": 1227.59,
|
293091 |
+
"learning_rate": 7.557253634894992e-06,
|
293092 |
+
"loss": 0.6683,
|
293093 |
+
"step": 152265
|
293094 |
+
},
|
293095 |
+
{
|
293096 |
+
"epoch": 1227.63,
|
293097 |
+
"learning_rate": 7.5571728594507276e-06,
|
293098 |
+
"loss": 0.7008,
|
293099 |
+
"step": 152270
|
293100 |
+
},
|
293101 |
+
{
|
293102 |
+
"epoch": 1227.67,
|
293103 |
+
"learning_rate": 7.557092084006462e-06,
|
293104 |
+
"loss": 0.3346,
|
293105 |
+
"step": 152275
|
293106 |
+
},
|
293107 |
+
{
|
293108 |
+
"epoch": 1227.71,
|
293109 |
+
"learning_rate": 7.5570113085621975e-06,
|
293110 |
+
"loss": 0.2314,
|
293111 |
+
"step": 152280
|
293112 |
+
},
|
293113 |
+
{
|
293114 |
+
"epoch": 1227.75,
|
293115 |
+
"learning_rate": 7.556930533117933e-06,
|
293116 |
+
"loss": 0.4819,
|
293117 |
+
"step": 152285
|
293118 |
+
},
|
293119 |
+
{
|
293120 |
+
"epoch": 1227.79,
|
293121 |
+
"learning_rate": 7.5568497576736675e-06,
|
293122 |
+
"loss": 0.5945,
|
293123 |
+
"step": 152290
|
293124 |
+
},
|
293125 |
+
{
|
293126 |
+
"epoch": 1227.83,
|
293127 |
+
"learning_rate": 7.556768982229403e-06,
|
293128 |
+
"loss": 0.8223,
|
293129 |
+
"step": 152295
|
293130 |
+
},
|
293131 |
+
{
|
293132 |
+
"epoch": 1227.87,
|
293133 |
+
"learning_rate": 7.5566882067851375e-06,
|
293134 |
+
"loss": 0.2747,
|
293135 |
+
"step": 152300
|
293136 |
+
},
|
293137 |
+
{
|
293138 |
+
"epoch": 1227.91,
|
293139 |
+
"learning_rate": 7.556607431340873e-06,
|
293140 |
+
"loss": 0.3093,
|
293141 |
+
"step": 152305
|
293142 |
+
},
|
293143 |
+
{
|
293144 |
+
"epoch": 1227.95,
|
293145 |
+
"learning_rate": 7.5565266558966075e-06,
|
293146 |
+
"loss": 0.4192,
|
293147 |
+
"step": 152310
|
293148 |
+
},
|
293149 |
+
{
|
293150 |
+
"epoch": 1227.99,
|
293151 |
+
"learning_rate": 7.556445880452343e-06,
|
293152 |
+
"loss": 0.7994,
|
293153 |
+
"step": 152315
|
293154 |
+
},
|
293155 |
+
{
|
293156 |
+
"epoch": 1228.0,
|
293157 |
+
"eval_loss": 0.4167693257331848,
|
293158 |
+
"eval_runtime": 41.5016,
|
293159 |
+
"eval_samples_per_second": 20.144,
|
293160 |
+
"eval_steps_per_second": 0.651,
|
293161 |
+
"eval_wer": 0.18405892219062203,
|
293162 |
+
"step": 152316
|
293163 |
+
},
|
293164 |
+
{
|
293165 |
+
"epoch": 1218.03,
|
293166 |
+
"learning_rate": 7.5563651050080774e-06,
|
293167 |
+
"loss": 0.2815,
|
293168 |
+
"step": 152320
|
293169 |
+
},
|
293170 |
+
{
|
293171 |
+
"epoch": 1218.07,
|
293172 |
+
"learning_rate": 7.556284329563813e-06,
|
293173 |
+
"loss": 0.2471,
|
293174 |
+
"step": 152325
|
293175 |
+
},
|
293176 |
+
{
|
293177 |
+
"epoch": 1218.11,
|
293178 |
+
"learning_rate": 7.556203554119547e-06,
|
293179 |
+
"loss": 0.2546,
|
293180 |
+
"step": 152330
|
293181 |
+
},
|
293182 |
+
{
|
293183 |
+
"epoch": 1218.15,
|
293184 |
+
"learning_rate": 7.556122778675283e-06,
|
293185 |
+
"loss": 0.339,
|
293186 |
+
"step": 152335
|
293187 |
+
},
|
293188 |
+
{
|
293189 |
+
"epoch": 1218.19,
|
293190 |
+
"learning_rate": 7.556042003231019e-06,
|
293191 |
+
"loss": 0.7478,
|
293192 |
+
"step": 152340
|
293193 |
+
},
|
293194 |
+
{
|
293195 |
+
"epoch": 1218.23,
|
293196 |
+
"learning_rate": 7.555961227786753e-06,
|
293197 |
+
"loss": 0.6217,
|
293198 |
+
"step": 152345
|
293199 |
+
},
|
293200 |
+
{
|
293201 |
+
"epoch": 1218.27,
|
293202 |
+
"learning_rate": 7.555880452342489e-06,
|
293203 |
+
"loss": 0.2577,
|
293204 |
+
"step": 152350
|
293205 |
+
},
|
293206 |
+
{
|
293207 |
+
"epoch": 1218.31,
|
293208 |
+
"learning_rate": 7.555799676898223e-06,
|
293209 |
+
"loss": 0.2436,
|
293210 |
+
"step": 152355
|
293211 |
+
},
|
293212 |
+
{
|
293213 |
+
"epoch": 1218.35,
|
293214 |
+
"learning_rate": 7.555718901453959e-06,
|
293215 |
+
"loss": 0.437,
|
293216 |
+
"step": 152360
|
293217 |
+
},
|
293218 |
+
{
|
293219 |
+
"epoch": 1218.39,
|
293220 |
+
"learning_rate": 7.555638126009693e-06,
|
293221 |
+
"loss": 0.829,
|
293222 |
+
"step": 152365
|
293223 |
+
},
|
293224 |
+
{
|
293225 |
+
"epoch": 1218.43,
|
293226 |
+
"learning_rate": 7.555557350565429e-06,
|
293227 |
+
"loss": 0.5973,
|
293228 |
+
"step": 152370
|
293229 |
+
},
|
293230 |
+
{
|
293231 |
+
"epoch": 1218.47,
|
293232 |
+
"learning_rate": 7.555476575121163e-06,
|
293233 |
+
"loss": 0.2641,
|
293234 |
+
"step": 152375
|
293235 |
+
},
|
293236 |
+
{
|
293237 |
+
"epoch": 1218.51,
|
293238 |
+
"learning_rate": 7.555395799676899e-06,
|
293239 |
+
"loss": 0.2739,
|
293240 |
+
"step": 152380
|
293241 |
+
},
|
293242 |
+
{
|
293243 |
+
"epoch": 1218.55,
|
293244 |
+
"learning_rate": 7.555315024232633e-06,
|
293245 |
+
"loss": 0.4165,
|
293246 |
+
"step": 152385
|
293247 |
+
},
|
293248 |
+
{
|
293249 |
+
"epoch": 1218.59,
|
293250 |
+
"learning_rate": 7.555234248788369e-06,
|
293251 |
+
"loss": 0.826,
|
293252 |
+
"step": 152390
|
293253 |
+
},
|
293254 |
+
{
|
293255 |
+
"epoch": 1218.63,
|
293256 |
+
"learning_rate": 7.555153473344105e-06,
|
293257 |
+
"loss": 0.6447,
|
293258 |
+
"step": 152395
|
293259 |
+
},
|
293260 |
+
{
|
293261 |
+
"epoch": 1218.67,
|
293262 |
+
"learning_rate": 7.555072697899839e-06,
|
293263 |
+
"loss": 0.2597,
|
293264 |
+
"step": 152400
|
293265 |
+
},
|
293266 |
+
{
|
293267 |
+
"epoch": 1218.71,
|
293268 |
+
"learning_rate": 7.554991922455575e-06,
|
293269 |
+
"loss": 0.2887,
|
293270 |
+
"step": 152405
|
293271 |
+
},
|
293272 |
+
{
|
293273 |
+
"epoch": 1218.75,
|
293274 |
+
"learning_rate": 7.554911147011309e-06,
|
293275 |
+
"loss": 0.4304,
|
293276 |
+
"step": 152410
|
293277 |
+
},
|
293278 |
+
{
|
293279 |
+
"epoch": 1218.79,
|
293280 |
+
"learning_rate": 7.554830371567045e-06,
|
293281 |
+
"loss": 0.9143,
|
293282 |
+
"step": 152415
|
293283 |
+
},
|
293284 |
+
{
|
293285 |
+
"epoch": 1218.83,
|
293286 |
+
"learning_rate": 7.554749596122779e-06,
|
293287 |
+
"loss": 0.6371,
|
293288 |
+
"step": 152420
|
293289 |
+
},
|
293290 |
+
{
|
293291 |
+
"epoch": 1218.87,
|
293292 |
+
"learning_rate": 7.554668820678515e-06,
|
293293 |
+
"loss": 0.2591,
|
293294 |
+
"step": 152425
|
293295 |
+
},
|
293296 |
+
{
|
293297 |
+
"epoch": 1218.91,
|
293298 |
+
"learning_rate": 7.554588045234249e-06,
|
293299 |
+
"loss": 0.3687,
|
293300 |
+
"step": 152430
|
293301 |
+
},
|
293302 |
+
{
|
293303 |
+
"epoch": 1218.95,
|
293304 |
+
"learning_rate": 7.554507269789985e-06,
|
293305 |
+
"loss": 0.4046,
|
293306 |
+
"step": 152435
|
293307 |
+
},
|
293308 |
+
{
|
293309 |
+
"epoch": 1218.99,
|
293310 |
+
"learning_rate": 7.554426494345719e-06,
|
293311 |
+
"loss": 0.902,
|
293312 |
+
"step": 152440
|
293313 |
+
},
|
293314 |
+
{
|
293315 |
+
"epoch": 1219.0,
|
293316 |
+
"eval_loss": 0.35546669363975525,
|
293317 |
+
"eval_runtime": 42.7299,
|
293318 |
+
"eval_samples_per_second": 19.565,
|
293319 |
+
"eval_steps_per_second": 0.632,
|
293320 |
+
"eval_wer": 0.17738750176329524,
|
293321 |
+
"step": 152441
|
293322 |
+
},
|
293323 |
+
{
|
293324 |
+
"epoch": 1229.03,
|
293325 |
+
"learning_rate": 7.554345718901455e-06,
|
293326 |
+
"loss": 0.2988,
|
293327 |
+
"step": 152445
|
293328 |
+
},
|
293329 |
+
{
|
293330 |
+
"epoch": 1229.07,
|
293331 |
+
"learning_rate": 7.554264943457189e-06,
|
293332 |
+
"loss": 0.2575,
|
293333 |
+
"step": 152450
|
293334 |
+
},
|
293335 |
+
{
|
293336 |
+
"epoch": 1229.11,
|
293337 |
+
"learning_rate": 7.554184168012925e-06,
|
293338 |
+
"loss": 0.3322,
|
293339 |
+
"step": 152455
|
293340 |
+
},
|
293341 |
+
{
|
293342 |
+
"epoch": 1229.15,
|
293343 |
+
"learning_rate": 7.5541033925686605e-06,
|
293344 |
+
"loss": 0.4137,
|
293345 |
+
"step": 152460
|
293346 |
+
},
|
293347 |
+
{
|
293348 |
+
"epoch": 1229.19,
|
293349 |
+
"learning_rate": 7.554022617124395e-06,
|
293350 |
+
"loss": 0.8767,
|
293351 |
+
"step": 152465
|
293352 |
+
},
|
293353 |
+
{
|
293354 |
+
"epoch": 1229.23,
|
293355 |
+
"learning_rate": 7.5539418416801305e-06,
|
293356 |
+
"loss": 0.625,
|
293357 |
+
"step": 152470
|
293358 |
+
},
|
293359 |
+
{
|
293360 |
+
"epoch": 1229.27,
|
293361 |
+
"learning_rate": 7.553861066235865e-06,
|
293362 |
+
"loss": 0.272,
|
293363 |
+
"step": 152475
|
293364 |
+
},
|
293365 |
+
{
|
293366 |
+
"epoch": 1229.31,
|
293367 |
+
"learning_rate": 7.5537802907916004e-06,
|
293368 |
+
"loss": 0.3351,
|
293369 |
+
"step": 152480
|
293370 |
+
},
|
293371 |
+
{
|
293372 |
+
"epoch": 1229.35,
|
293373 |
+
"learning_rate": 7.5536995153473346e-06,
|
293374 |
+
"loss": 0.356,
|
293375 |
+
"step": 152485
|
293376 |
+
},
|
293377 |
+
{
|
293378 |
+
"epoch": 1229.39,
|
293379 |
+
"learning_rate": 7.55361873990307e-06,
|
293380 |
+
"loss": 0.8533,
|
293381 |
+
"step": 152490
|
293382 |
+
},
|
293383 |
+
{
|
293384 |
+
"epoch": 1229.43,
|
293385 |
+
"learning_rate": 7.5535379644588045e-06,
|
293386 |
+
"loss": 0.6315,
|
293387 |
+
"step": 152495
|
293388 |
+
},
|
293389 |
+
{
|
293390 |
+
"epoch": 1229.47,
|
293391 |
+
"learning_rate": 7.55345718901454e-06,
|
293392 |
+
"loss": 0.2547,
|
293393 |
+
"step": 152500
|
293394 |
+
},
|
293395 |
+
{
|
293396 |
+
"epoch": 1229.51,
|
293397 |
+
"learning_rate": 7.5533764135702745e-06,
|
293398 |
+
"loss": 0.2782,
|
293399 |
+
"step": 152505
|
293400 |
+
},
|
293401 |
+
{
|
293402 |
+
"epoch": 1229.55,
|
293403 |
+
"learning_rate": 7.55329563812601e-06,
|
293404 |
+
"loss": 0.3844,
|
293405 |
+
"step": 152510
|
293406 |
+
},
|
293407 |
+
{
|
293408 |
+
"epoch": 1229.59,
|
293409 |
+
"learning_rate": 7.553214862681746e-06,
|
293410 |
+
"loss": 1.0582,
|
293411 |
+
"step": 152515
|
293412 |
+
},
|
293413 |
+
{
|
293414 |
+
"epoch": 1229.63,
|
293415 |
+
"learning_rate": 7.55313408723748e-06,
|
293416 |
+
"loss": 0.6986,
|
293417 |
+
"step": 152520
|
293418 |
+
},
|
293419 |
+
{
|
293420 |
+
"epoch": 1229.67,
|
293421 |
+
"learning_rate": 7.553053311793216e-06,
|
293422 |
+
"loss": 0.2519,
|
293423 |
+
"step": 152525
|
293424 |
+
},
|
293425 |
+
{
|
293426 |
+
"epoch": 1229.71,
|
293427 |
+
"learning_rate": 7.55297253634895e-06,
|
293428 |
+
"loss": 0.3056,
|
293429 |
+
"step": 152530
|
293430 |
+
},
|
293431 |
+
{
|
293432 |
+
"epoch": 1229.76,
|
293433 |
+
"learning_rate": 7.552891760904686e-06,
|
293434 |
+
"loss": 0.3928,
|
293435 |
+
"step": 152535
|
293436 |
+
},
|
293437 |
+
{
|
293438 |
+
"epoch": 1229.8,
|
293439 |
+
"learning_rate": 7.55281098546042e-06,
|
293440 |
+
"loss": 0.8044,
|
293441 |
+
"step": 152540
|
293442 |
+
},
|
293443 |
+
{
|
293444 |
+
"epoch": 1229.84,
|
293445 |
+
"learning_rate": 7.552730210016156e-06,
|
293446 |
+
"loss": 0.651,
|
293447 |
+
"step": 152545
|
293448 |
+
},
|
293449 |
+
{
|
293450 |
+
"epoch": 1229.88,
|
293451 |
+
"learning_rate": 7.55264943457189e-06,
|
293452 |
+
"loss": 0.2421,
|
293453 |
+
"step": 152550
|
293454 |
+
},
|
293455 |
+
{
|
293456 |
+
"epoch": 1229.92,
|
293457 |
+
"learning_rate": 7.552568659127626e-06,
|
293458 |
+
"loss": 0.3304,
|
293459 |
+
"step": 152555
|
293460 |
+
},
|
293461 |
+
{
|
293462 |
+
"epoch": 1229.96,
|
293463 |
+
"learning_rate": 7.55248788368336e-06,
|
293464 |
+
"loss": 0.4315,
|
293465 |
+
"step": 152560
|
293466 |
+
},
|
293467 |
+
{
|
293468 |
+
"epoch": 1230.0,
|
293469 |
+
"learning_rate": 7.552407108239096e-06,
|
293470 |
+
"loss": 1.0926,
|
293471 |
+
"step": 152565
|
293472 |
+
},
|
293473 |
+
{
|
293474 |
+
"epoch": 1230.0,
|
293475 |
+
"eval_loss": 0.5449385643005371,
|
293476 |
+
"eval_runtime": 40.681,
|
293477 |
+
"eval_samples_per_second": 20.55,
|
293478 |
+
"eval_steps_per_second": 0.664,
|
293479 |
+
"eval_wer": 0.18418858485560524,
|
293480 |
+
"step": 152565
|
293481 |
+
},
|
293482 |
+
{
|
293483 |
+
"epoch": 1220.04,
|
293484 |
+
"learning_rate": 7.552326332794832e-06,
|
293485 |
+
"loss": 0.2789,
|
293486 |
+
"step": 152570
|
293487 |
+
},
|
293488 |
+
{
|
293489 |
+
"epoch": 1220.08,
|
293490 |
+
"learning_rate": 7.552245557350566e-06,
|
293491 |
+
"loss": 0.271,
|
293492 |
+
"step": 152575
|
293493 |
+
},
|
293494 |
+
{
|
293495 |
+
"epoch": 1220.12,
|
293496 |
+
"learning_rate": 7.552164781906302e-06,
|
293497 |
+
"loss": 0.243,
|
293498 |
+
"step": 152580
|
293499 |
+
},
|
293500 |
+
{
|
293501 |
+
"epoch": 1220.16,
|
293502 |
+
"learning_rate": 7.552084006462036e-06,
|
293503 |
+
"loss": 0.4346,
|
293504 |
+
"step": 152585
|
293505 |
+
},
|
293506 |
+
{
|
293507 |
+
"epoch": 1220.2,
|
293508 |
+
"learning_rate": 7.552003231017772e-06,
|
293509 |
+
"loss": 1.2121,
|
293510 |
+
"step": 152590
|
293511 |
+
},
|
293512 |
+
{
|
293513 |
+
"epoch": 1220.24,
|
293514 |
+
"learning_rate": 7.551922455573506e-06,
|
293515 |
+
"loss": 0.373,
|
293516 |
+
"step": 152595
|
293517 |
+
},
|
293518 |
+
{
|
293519 |
+
"epoch": 1220.28,
|
293520 |
+
"learning_rate": 7.551841680129242e-06,
|
293521 |
+
"loss": 0.2537,
|
293522 |
+
"step": 152600
|
293523 |
+
},
|
293524 |
+
{
|
293525 |
+
"epoch": 1220.32,
|
293526 |
+
"learning_rate": 7.551760904684976e-06,
|
293527 |
+
"loss": 0.2496,
|
293528 |
+
"step": 152605
|
293529 |
+
},
|
293530 |
+
{
|
293531 |
+
"epoch": 1220.36,
|
293532 |
+
"learning_rate": 7.551680129240712e-06,
|
293533 |
+
"loss": 0.5073,
|
293534 |
+
"step": 152610
|
293535 |
+
},
|
293536 |
+
{
|
293537 |
+
"epoch": 1220.4,
|
293538 |
+
"learning_rate": 7.551599353796446e-06,
|
293539 |
+
"loss": 1.2567,
|
293540 |
+
"step": 152615
|
293541 |
+
},
|
293542 |
+
{
|
293543 |
+
"epoch": 1220.44,
|
293544 |
+
"learning_rate": 7.551518578352182e-06,
|
293545 |
+
"loss": 0.3609,
|
293546 |
+
"step": 152620
|
293547 |
+
},
|
293548 |
+
{
|
293549 |
+
"epoch": 1220.48,
|
293550 |
+
"learning_rate": 7.551437802907916e-06,
|
293551 |
+
"loss": 0.2748,
|
293552 |
+
"step": 152625
|
293553 |
+
},
|
293554 |
+
{
|
293555 |
+
"epoch": 1220.52,
|
293556 |
+
"learning_rate": 7.551357027463652e-06,
|
293557 |
+
"loss": 0.3177,
|
293558 |
+
"step": 152630
|
293559 |
+
},
|
293560 |
+
{
|
293561 |
+
"epoch": 1220.56,
|
293562 |
+
"learning_rate": 7.551276252019388e-06,
|
293563 |
+
"loss": 0.437,
|
293564 |
+
"step": 152635
|
293565 |
+
},
|
293566 |
+
{
|
293567 |
+
"epoch": 1220.6,
|
293568 |
+
"learning_rate": 7.551195476575122e-06,
|
293569 |
+
"loss": 1.3232,
|
293570 |
+
"step": 152640
|
293571 |
+
},
|
293572 |
+
{
|
293573 |
+
"epoch": 1220.64,
|
293574 |
+
"learning_rate": 7.5511147011308576e-06,
|
293575 |
+
"loss": 0.2868,
|
293576 |
+
"step": 152645
|
293577 |
+
},
|
293578 |
+
{
|
293579 |
+
"epoch": 1220.68,
|
293580 |
+
"learning_rate": 7.551033925686592e-06,
|
293581 |
+
"loss": 0.3602,
|
293582 |
+
"step": 152650
|
293583 |
+
},
|
293584 |
+
{
|
293585 |
+
"epoch": 1220.72,
|
293586 |
+
"learning_rate": 7.5509531502423275e-06,
|
293587 |
+
"loss": 0.2706,
|
293588 |
+
"step": 152655
|
293589 |
+
},
|
293590 |
+
{
|
293591 |
+
"epoch": 1220.76,
|
293592 |
+
"learning_rate": 7.550872374798062e-06,
|
293593 |
+
"loss": 0.4301,
|
293594 |
+
"step": 152660
|
293595 |
+
},
|
293596 |
+
{
|
293597 |
+
"epoch": 1220.8,
|
293598 |
+
"learning_rate": 7.5507915993537975e-06,
|
293599 |
+
"loss": 1.0362,
|
293600 |
+
"step": 152665
|
293601 |
+
},
|
293602 |
+
{
|
293603 |
+
"epoch": 1220.84,
|
293604 |
+
"learning_rate": 7.550710823909532e-06,
|
293605 |
+
"loss": 0.2926,
|
293606 |
+
"step": 152670
|
293607 |
+
},
|
293608 |
+
{
|
293609 |
+
"epoch": 1220.88,
|
293610 |
+
"learning_rate": 7.5506300484652675e-06,
|
293611 |
+
"loss": 0.2753,
|
293612 |
+
"step": 152675
|
293613 |
+
},
|
293614 |
+
{
|
293615 |
+
"epoch": 1220.92,
|
293616 |
+
"learning_rate": 7.550549273021002e-06,
|
293617 |
+
"loss": 0.3452,
|
293618 |
+
"step": 152680
|
293619 |
+
},
|
293620 |
+
{
|
293621 |
+
"epoch": 1220.96,
|
293622 |
+
"learning_rate": 7.5504684975767375e-06,
|
293623 |
+
"loss": 0.4678,
|
293624 |
+
"step": 152685
|
293625 |
+
},
|
293626 |
+
{
|
293627 |
+
"epoch": 1221.0,
|
293628 |
+
"learning_rate": 7.5503877221324724e-06,
|
293629 |
+
"loss": 1.3258,
|
293630 |
+
"step": 152690
|
293631 |
+
},
|
293632 |
+
{
|
293633 |
+
"epoch": 1221.0,
|
293634 |
+
"eval_loss": 0.3559703230857849,
|
293635 |
+
"eval_runtime": 43.0859,
|
293636 |
+
"eval_samples_per_second": 19.403,
|
293637 |
+
"eval_steps_per_second": 0.627,
|
293638 |
+
"eval_wer": 0.17048071387449626,
|
293639 |
+
"step": 152690
|
293640 |
}
|
293641 |
],
|
293642 |
+
"max_steps": 625000,
|
293643 |
"num_train_epochs": 5000,
|
293644 |
+
"total_flos": 4.2968328943638807e+20,
|
293645 |
"trial_name": null,
|
293646 |
"trial_params": null
|
293647 |
}
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630250753.668995/events.out.tfevents.1630250753.cc93b136ebf5.1086.209
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d9c0aee1ce7801ac75359895fcb1668e7c7233a293e1e85b17034353ff0003ee
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630251188.88167/events.out.tfevents.1630251188.cc93b136ebf5.1086.211
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc10d5878cb756554085f94d5e325de11d0130822fcaaad5d044e2996b03f139
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630251628.1797204/events.out.tfevents.1630251628.cc93b136ebf5.1086.213
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a75e79f95a32dbd01c2a774a76e9bb2824f0d5b1551051b1f8085c6c884e1d3
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630252065.7514985/events.out.tfevents.1630252065.cc93b136ebf5.1086.215
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e323b0936deda44bf91875918182857d58ba5428f8afdebac01580255cb0e269
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630252500.5741549/events.out.tfevents.1630252500.cc93b136ebf5.1086.217
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb9ebf7e8137afa957d4d2026869e20b8bd962c5bce171c25618696ce1316d6e
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630250753.cc93b136ebf5.1086.208
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c6b329615b997a29dc2441fdc952ec848b8f5de5513d43042f86ea38b90097d
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630251188.cc93b136ebf5.1086.210
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28f4b9a76a7866c02a50de0f72d113f2770ef0fb3e160da1f4087186466043da
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630251628.cc93b136ebf5.1086.212
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c24d8523270a58fc6c0a96845b48c37ed90191d876506302e6a8a7e1c204bf0
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630252065.cc93b136ebf5.1086.214
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ad901e5a01c00724eaa3e2c6d1064f0725f8468b8af7d9f4449a542c3c3cd99
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630252500.cc93b136ebf5.1086.216
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1c9e6fc6b8beb6d7bda1c768e2664d3a7732f3851f11540d7cf4be9cc49d64f
|
3 |
+
size 8622
|