"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630248561.4202452/events.out.tfevents.1630248561.cc93b136ebf5.1086.199 +3 -0
- model-bin/finetune/base/log/1630248992.3105023/events.out.tfevents.1630248992.cc93b136ebf5.1086.201 +3 -0
- model-bin/finetune/base/log/1630249423.7833743/events.out.tfevents.1630249423.cc93b136ebf5.1086.203 +3 -0
- model-bin/finetune/base/log/1630249857.8968918/events.out.tfevents.1630249857.cc93b136ebf5.1086.205 +3 -0
- model-bin/finetune/base/log/1630250287.4046617/events.out.tfevents.1630250287.cc93b136ebf5.1086.207 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630248561.cc93b136ebf5.1086.198 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630248992.cc93b136ebf5.1086.200 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630249423.cc93b136ebf5.1086.202 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630249857.cc93b136ebf5.1086.204 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630250287.cc93b136ebf5.1086.206 +3 -0
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88106383b07c091dad16598681c368eb1f7d5dd6754f75dcdede18a251a76193
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e90c5a61e1081391cac20c2749dbcf1cb5ad34446b8a2074676f8de81385b325
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e5c8178a08780b7e94358c9559057df556e07dd698f9fa5dc769d640159e51e
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5fe336fce1f24fb6ba716a3ec77454f1b6f380bbf0069997841f401a2fb77c2a
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1c8fd19316f91f2cc4123c75c86425f3af495dca75999776dc0193fa6dd50cb
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -292053,11 +292053,800 @@
|
|
292053 |
"eval_steps_per_second": 0.639,
|
292054 |
"eval_wer": 0.17625899280575538,
|
292055 |
"step": 151448
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
292056 |
}
|
292057 |
],
|
292058 |
"max_steps": 620000,
|
292059 |
"num_train_epochs": 5000,
|
292060 |
-
"total_flos": 4.
|
292061 |
"trial_name": null,
|
292062 |
"trial_params": null
|
292063 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1225.995983935743,
|
5 |
+
"global_step": 152068,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
292053 |
"eval_steps_per_second": 0.639,
|
292054 |
"eval_wer": 0.17625899280575538,
|
292055 |
"step": 151448
|
292056 |
+
},
|
292057 |
+
{
|
292058 |
+
"epoch": 1221.02,
|
292059 |
+
"learning_rate": 7.570420032310179e-06,
|
292060 |
+
"loss": 0.3857,
|
292061 |
+
"step": 151450
|
292062 |
+
},
|
292063 |
+
{
|
292064 |
+
"epoch": 1221.06,
|
292065 |
+
"learning_rate": 7.570339256865913e-06,
|
292066 |
+
"loss": 0.27,
|
292067 |
+
"step": 151455
|
292068 |
+
},
|
292069 |
+
{
|
292070 |
+
"epoch": 1221.1,
|
292071 |
+
"learning_rate": 7.5702584814216486e-06,
|
292072 |
+
"loss": 0.2653,
|
292073 |
+
"step": 151460
|
292074 |
+
},
|
292075 |
+
{
|
292076 |
+
"epoch": 1221.14,
|
292077 |
+
"learning_rate": 7.570177705977384e-06,
|
292078 |
+
"loss": 0.34,
|
292079 |
+
"step": 151465
|
292080 |
+
},
|
292081 |
+
{
|
292082 |
+
"epoch": 1221.18,
|
292083 |
+
"learning_rate": 7.5700969305331186e-06,
|
292084 |
+
"loss": 0.5222,
|
292085 |
+
"step": 151470
|
292086 |
+
},
|
292087 |
+
{
|
292088 |
+
"epoch": 1221.22,
|
292089 |
+
"learning_rate": 7.570016155088854e-06,
|
292090 |
+
"loss": 0.8796,
|
292091 |
+
"step": 151475
|
292092 |
+
},
|
292093 |
+
{
|
292094 |
+
"epoch": 1221.26,
|
292095 |
+
"learning_rate": 7.5699353796445885e-06,
|
292096 |
+
"loss": 0.2732,
|
292097 |
+
"step": 151480
|
292098 |
+
},
|
292099 |
+
{
|
292100 |
+
"epoch": 1221.3,
|
292101 |
+
"learning_rate": 7.569854604200324e-06,
|
292102 |
+
"loss": 0.2637,
|
292103 |
+
"step": 151485
|
292104 |
+
},
|
292105 |
+
{
|
292106 |
+
"epoch": 1221.34,
|
292107 |
+
"learning_rate": 7.5697738287560585e-06,
|
292108 |
+
"loss": 0.3104,
|
292109 |
+
"step": 151490
|
292110 |
+
},
|
292111 |
+
{
|
292112 |
+
"epoch": 1221.38,
|
292113 |
+
"learning_rate": 7.569693053311794e-06,
|
292114 |
+
"loss": 0.5735,
|
292115 |
+
"step": 151495
|
292116 |
+
},
|
292117 |
+
{
|
292118 |
+
"epoch": 1221.42,
|
292119 |
+
"learning_rate": 7.5696122778675285e-06,
|
292120 |
+
"loss": 0.9314,
|
292121 |
+
"step": 151500
|
292122 |
+
},
|
292123 |
+
{
|
292124 |
+
"epoch": 1221.46,
|
292125 |
+
"learning_rate": 7.569531502423264e-06,
|
292126 |
+
"loss": 0.2415,
|
292127 |
+
"step": 151505
|
292128 |
+
},
|
292129 |
+
{
|
292130 |
+
"epoch": 1221.5,
|
292131 |
+
"learning_rate": 7.5694507269789985e-06,
|
292132 |
+
"loss": 0.2433,
|
292133 |
+
"step": 151510
|
292134 |
+
},
|
292135 |
+
{
|
292136 |
+
"epoch": 1221.54,
|
292137 |
+
"learning_rate": 7.569369951534734e-06,
|
292138 |
+
"loss": 0.3259,
|
292139 |
+
"step": 151515
|
292140 |
+
},
|
292141 |
+
{
|
292142 |
+
"epoch": 1221.58,
|
292143 |
+
"learning_rate": 7.569289176090469e-06,
|
292144 |
+
"loss": 0.5686,
|
292145 |
+
"step": 151520
|
292146 |
+
},
|
292147 |
+
{
|
292148 |
+
"epoch": 1221.62,
|
292149 |
+
"learning_rate": 7.569208400646204e-06,
|
292150 |
+
"loss": 0.9219,
|
292151 |
+
"step": 151525
|
292152 |
+
},
|
292153 |
+
{
|
292154 |
+
"epoch": 1221.66,
|
292155 |
+
"learning_rate": 7.569127625201939e-06,
|
292156 |
+
"loss": 0.2692,
|
292157 |
+
"step": 151530
|
292158 |
+
},
|
292159 |
+
{
|
292160 |
+
"epoch": 1221.7,
|
292161 |
+
"learning_rate": 7.569046849757674e-06,
|
292162 |
+
"loss": 0.29,
|
292163 |
+
"step": 151535
|
292164 |
+
},
|
292165 |
+
{
|
292166 |
+
"epoch": 1221.74,
|
292167 |
+
"learning_rate": 7.56896607431341e-06,
|
292168 |
+
"loss": 0.2842,
|
292169 |
+
"step": 151540
|
292170 |
+
},
|
292171 |
+
{
|
292172 |
+
"epoch": 1221.78,
|
292173 |
+
"learning_rate": 7.568885298869144e-06,
|
292174 |
+
"loss": 0.5224,
|
292175 |
+
"step": 151545
|
292176 |
+
},
|
292177 |
+
{
|
292178 |
+
"epoch": 1221.82,
|
292179 |
+
"learning_rate": 7.56880452342488e-06,
|
292180 |
+
"loss": 0.9927,
|
292181 |
+
"step": 151550
|
292182 |
+
},
|
292183 |
+
{
|
292184 |
+
"epoch": 1221.86,
|
292185 |
+
"learning_rate": 7.568723747980614e-06,
|
292186 |
+
"loss": 0.2828,
|
292187 |
+
"step": 151555
|
292188 |
+
},
|
292189 |
+
{
|
292190 |
+
"epoch": 1221.9,
|
292191 |
+
"learning_rate": 7.56864297253635e-06,
|
292192 |
+
"loss": 0.2777,
|
292193 |
+
"step": 151560
|
292194 |
+
},
|
292195 |
+
{
|
292196 |
+
"epoch": 1221.94,
|
292197 |
+
"learning_rate": 7.568562197092084e-06,
|
292198 |
+
"loss": 0.3088,
|
292199 |
+
"step": 151565
|
292200 |
+
},
|
292201 |
+
{
|
292202 |
+
"epoch": 1221.98,
|
292203 |
+
"learning_rate": 7.56848142164782e-06,
|
292204 |
+
"loss": 0.6365,
|
292205 |
+
"step": 151570
|
292206 |
+
},
|
292207 |
+
{
|
292208 |
+
"epoch": 1222.0,
|
292209 |
+
"eval_loss": 0.5525439977645874,
|
292210 |
+
"eval_runtime": 39.5867,
|
292211 |
+
"eval_samples_per_second": 21.118,
|
292212 |
+
"eval_steps_per_second": 0.682,
|
292213 |
+
"eval_wer": 0.1825968316531494,
|
292214 |
+
"step": 151572
|
292215 |
+
},
|
292216 |
+
{
|
292217 |
+
"epoch": 1222.02,
|
292218 |
+
"learning_rate": 7.568400646203554e-06,
|
292219 |
+
"loss": 0.2871,
|
292220 |
+
"step": 151575
|
292221 |
+
},
|
292222 |
+
{
|
292223 |
+
"epoch": 1222.06,
|
292224 |
+
"learning_rate": 7.56831987075929e-06,
|
292225 |
+
"loss": 0.2949,
|
292226 |
+
"step": 151580
|
292227 |
+
},
|
292228 |
+
{
|
292229 |
+
"epoch": 1222.1,
|
292230 |
+
"learning_rate": 7.568239095315025e-06,
|
292231 |
+
"loss": 0.2781,
|
292232 |
+
"step": 151585
|
292233 |
+
},
|
292234 |
+
{
|
292235 |
+
"epoch": 1222.14,
|
292236 |
+
"learning_rate": 7.56815831987076e-06,
|
292237 |
+
"loss": 0.4072,
|
292238 |
+
"step": 151590
|
292239 |
+
},
|
292240 |
+
{
|
292241 |
+
"epoch": 1222.18,
|
292242 |
+
"learning_rate": 7.568077544426495e-06,
|
292243 |
+
"loss": 0.69,
|
292244 |
+
"step": 151595
|
292245 |
+
},
|
292246 |
+
{
|
292247 |
+
"epoch": 1222.22,
|
292248 |
+
"learning_rate": 7.56799676898223e-06,
|
292249 |
+
"loss": 0.7913,
|
292250 |
+
"step": 151600
|
292251 |
+
},
|
292252 |
+
{
|
292253 |
+
"epoch": 1222.27,
|
292254 |
+
"learning_rate": 7.567915993537966e-06,
|
292255 |
+
"loss": 0.3416,
|
292256 |
+
"step": 151605
|
292257 |
+
},
|
292258 |
+
{
|
292259 |
+
"epoch": 1222.31,
|
292260 |
+
"learning_rate": 7.5678352180937e-06,
|
292261 |
+
"loss": 0.2677,
|
292262 |
+
"step": 151610
|
292263 |
+
},
|
292264 |
+
{
|
292265 |
+
"epoch": 1222.35,
|
292266 |
+
"learning_rate": 7.567754442649436e-06,
|
292267 |
+
"loss": 0.3702,
|
292268 |
+
"step": 151615
|
292269 |
+
},
|
292270 |
+
{
|
292271 |
+
"epoch": 1222.39,
|
292272 |
+
"learning_rate": 7.56767366720517e-06,
|
292273 |
+
"loss": 0.7914,
|
292274 |
+
"step": 151620
|
292275 |
+
},
|
292276 |
+
{
|
292277 |
+
"epoch": 1222.43,
|
292278 |
+
"learning_rate": 7.567592891760906e-06,
|
292279 |
+
"loss": 0.7734,
|
292280 |
+
"step": 151625
|
292281 |
+
},
|
292282 |
+
{
|
292283 |
+
"epoch": 1222.47,
|
292284 |
+
"learning_rate": 7.56751211631664e-06,
|
292285 |
+
"loss": 0.4037,
|
292286 |
+
"step": 151630
|
292287 |
+
},
|
292288 |
+
{
|
292289 |
+
"epoch": 1222.51,
|
292290 |
+
"learning_rate": 7.567431340872376e-06,
|
292291 |
+
"loss": 0.2796,
|
292292 |
+
"step": 151635
|
292293 |
+
},
|
292294 |
+
{
|
292295 |
+
"epoch": 1222.55,
|
292296 |
+
"learning_rate": 7.567350565428111e-06,
|
292297 |
+
"loss": 0.4157,
|
292298 |
+
"step": 151640
|
292299 |
+
},
|
292300 |
+
{
|
292301 |
+
"epoch": 1222.59,
|
292302 |
+
"learning_rate": 7.567269789983846e-06,
|
292303 |
+
"loss": 0.7692,
|
292304 |
+
"step": 151645
|
292305 |
+
},
|
292306 |
+
{
|
292307 |
+
"epoch": 1222.63,
|
292308 |
+
"learning_rate": 7.567189014539581e-06,
|
292309 |
+
"loss": 0.8732,
|
292310 |
+
"step": 151650
|
292311 |
+
},
|
292312 |
+
{
|
292313 |
+
"epoch": 1222.67,
|
292314 |
+
"learning_rate": 7.567108239095316e-06,
|
292315 |
+
"loss": 0.256,
|
292316 |
+
"step": 151655
|
292317 |
+
},
|
292318 |
+
{
|
292319 |
+
"epoch": 1222.71,
|
292320 |
+
"learning_rate": 7.567027463651051e-06,
|
292321 |
+
"loss": 0.2665,
|
292322 |
+
"step": 151660
|
292323 |
+
},
|
292324 |
+
{
|
292325 |
+
"epoch": 1222.75,
|
292326 |
+
"learning_rate": 7.566946688206786e-06,
|
292327 |
+
"loss": 0.4612,
|
292328 |
+
"step": 151665
|
292329 |
+
},
|
292330 |
+
{
|
292331 |
+
"epoch": 1222.79,
|
292332 |
+
"learning_rate": 7.5668659127625214e-06,
|
292333 |
+
"loss": 0.7535,
|
292334 |
+
"step": 151670
|
292335 |
+
},
|
292336 |
+
{
|
292337 |
+
"epoch": 1222.83,
|
292338 |
+
"learning_rate": 7.566785137318256e-06,
|
292339 |
+
"loss": 0.9049,
|
292340 |
+
"step": 151675
|
292341 |
+
},
|
292342 |
+
{
|
292343 |
+
"epoch": 1222.87,
|
292344 |
+
"learning_rate": 7.566704361873991e-06,
|
292345 |
+
"loss": 0.3755,
|
292346 |
+
"step": 151680
|
292347 |
+
},
|
292348 |
+
{
|
292349 |
+
"epoch": 1222.91,
|
292350 |
+
"learning_rate": 7.5666235864297256e-06,
|
292351 |
+
"loss": 0.2588,
|
292352 |
+
"step": 151685
|
292353 |
+
},
|
292354 |
+
{
|
292355 |
+
"epoch": 1222.95,
|
292356 |
+
"learning_rate": 7.566542810985461e-06,
|
292357 |
+
"loss": 0.365,
|
292358 |
+
"step": 151690
|
292359 |
+
},
|
292360 |
+
{
|
292361 |
+
"epoch": 1222.99,
|
292362 |
+
"learning_rate": 7.566462035541196e-06,
|
292363 |
+
"loss": 0.7382,
|
292364 |
+
"step": 151695
|
292365 |
+
},
|
292366 |
+
{
|
292367 |
+
"epoch": 1223.0,
|
292368 |
+
"eval_loss": 0.3656209409236908,
|
292369 |
+
"eval_runtime": 41.975,
|
292370 |
+
"eval_samples_per_second": 19.917,
|
292371 |
+
"eval_steps_per_second": 0.643,
|
292372 |
+
"eval_wer": 0.16889752446747267,
|
292373 |
+
"step": 151696
|
292374 |
+
},
|
292375 |
+
{
|
292376 |
+
"epoch": 1223.03,
|
292377 |
+
"learning_rate": 7.566381260096931e-06,
|
292378 |
+
"loss": 0.3057,
|
292379 |
+
"step": 151700
|
292380 |
+
},
|
292381 |
+
{
|
292382 |
+
"epoch": 1223.07,
|
292383 |
+
"learning_rate": 7.566300484652666e-06,
|
292384 |
+
"loss": 0.3804,
|
292385 |
+
"step": 151705
|
292386 |
+
},
|
292387 |
+
{
|
292388 |
+
"epoch": 1223.11,
|
292389 |
+
"learning_rate": 7.566219709208401e-06,
|
292390 |
+
"loss": 0.3112,
|
292391 |
+
"step": 151710
|
292392 |
+
},
|
292393 |
+
{
|
292394 |
+
"epoch": 1223.15,
|
292395 |
+
"learning_rate": 7.566138933764136e-06,
|
292396 |
+
"loss": 0.3271,
|
292397 |
+
"step": 151715
|
292398 |
+
},
|
292399 |
+
{
|
292400 |
+
"epoch": 1223.19,
|
292401 |
+
"learning_rate": 7.566058158319871e-06,
|
292402 |
+
"loss": 0.8103,
|
292403 |
+
"step": 151720
|
292404 |
+
},
|
292405 |
+
{
|
292406 |
+
"epoch": 1223.23,
|
292407 |
+
"learning_rate": 7.565977382875606e-06,
|
292408 |
+
"loss": 0.6019,
|
292409 |
+
"step": 151725
|
292410 |
+
},
|
292411 |
+
{
|
292412 |
+
"epoch": 1223.27,
|
292413 |
+
"learning_rate": 7.565896607431341e-06,
|
292414 |
+
"loss": 0.2601,
|
292415 |
+
"step": 151730
|
292416 |
+
},
|
292417 |
+
{
|
292418 |
+
"epoch": 1223.31,
|
292419 |
+
"learning_rate": 7.565815831987076e-06,
|
292420 |
+
"loss": 0.2529,
|
292421 |
+
"step": 151735
|
292422 |
+
},
|
292423 |
+
{
|
292424 |
+
"epoch": 1223.35,
|
292425 |
+
"learning_rate": 7.565735056542811e-06,
|
292426 |
+
"loss": 0.3854,
|
292427 |
+
"step": 151740
|
292428 |
+
},
|
292429 |
+
{
|
292430 |
+
"epoch": 1223.4,
|
292431 |
+
"learning_rate": 7.565654281098547e-06,
|
292432 |
+
"loss": 0.8185,
|
292433 |
+
"step": 151745
|
292434 |
+
},
|
292435 |
+
{
|
292436 |
+
"epoch": 1223.44,
|
292437 |
+
"learning_rate": 7.565573505654281e-06,
|
292438 |
+
"loss": 0.5565,
|
292439 |
+
"step": 151750
|
292440 |
+
},
|
292441 |
+
{
|
292442 |
+
"epoch": 1223.48,
|
292443 |
+
"learning_rate": 7.565492730210017e-06,
|
292444 |
+
"loss": 0.2581,
|
292445 |
+
"step": 151755
|
292446 |
+
},
|
292447 |
+
{
|
292448 |
+
"epoch": 1223.52,
|
292449 |
+
"learning_rate": 7.565411954765752e-06,
|
292450 |
+
"loss": 0.2536,
|
292451 |
+
"step": 151760
|
292452 |
+
},
|
292453 |
+
{
|
292454 |
+
"epoch": 1223.56,
|
292455 |
+
"learning_rate": 7.565331179321487e-06,
|
292456 |
+
"loss": 0.3955,
|
292457 |
+
"step": 151765
|
292458 |
+
},
|
292459 |
+
{
|
292460 |
+
"epoch": 1223.6,
|
292461 |
+
"learning_rate": 7.565250403877222e-06,
|
292462 |
+
"loss": 0.8402,
|
292463 |
+
"step": 151770
|
292464 |
+
},
|
292465 |
+
{
|
292466 |
+
"epoch": 1223.64,
|
292467 |
+
"learning_rate": 7.565169628432957e-06,
|
292468 |
+
"loss": 0.6756,
|
292469 |
+
"step": 151775
|
292470 |
+
},
|
292471 |
+
{
|
292472 |
+
"epoch": 1223.68,
|
292473 |
+
"learning_rate": 7.565088852988692e-06,
|
292474 |
+
"loss": 0.2822,
|
292475 |
+
"step": 151780
|
292476 |
+
},
|
292477 |
+
{
|
292478 |
+
"epoch": 1223.72,
|
292479 |
+
"learning_rate": 7.565008077544427e-06,
|
292480 |
+
"loss": 0.3014,
|
292481 |
+
"step": 151785
|
292482 |
+
},
|
292483 |
+
{
|
292484 |
+
"epoch": 1223.76,
|
292485 |
+
"learning_rate": 7.564927302100162e-06,
|
292486 |
+
"loss": 0.4023,
|
292487 |
+
"step": 151790
|
292488 |
+
},
|
292489 |
+
{
|
292490 |
+
"epoch": 1223.8,
|
292491 |
+
"learning_rate": 7.564846526655897e-06,
|
292492 |
+
"loss": 0.8358,
|
292493 |
+
"step": 151795
|
292494 |
+
},
|
292495 |
+
{
|
292496 |
+
"epoch": 1223.84,
|
292497 |
+
"learning_rate": 7.564765751211632e-06,
|
292498 |
+
"loss": 0.7451,
|
292499 |
+
"step": 151800
|
292500 |
+
},
|
292501 |
+
{
|
292502 |
+
"epoch": 1223.88,
|
292503 |
+
"learning_rate": 7.564684975767367e-06,
|
292504 |
+
"loss": 0.2822,
|
292505 |
+
"step": 151805
|
292506 |
+
},
|
292507 |
+
{
|
292508 |
+
"epoch": 1223.92,
|
292509 |
+
"learning_rate": 7.564604200323103e-06,
|
292510 |
+
"loss": 0.2669,
|
292511 |
+
"step": 151810
|
292512 |
+
},
|
292513 |
+
{
|
292514 |
+
"epoch": 1223.96,
|
292515 |
+
"learning_rate": 7.564523424878838e-06,
|
292516 |
+
"loss": 0.4182,
|
292517 |
+
"step": 151815
|
292518 |
+
},
|
292519 |
+
{
|
292520 |
+
"epoch": 1224.0,
|
292521 |
+
"learning_rate": 7.564442649434573e-06,
|
292522 |
+
"loss": 1.1147,
|
292523 |
+
"step": 151820
|
292524 |
+
},
|
292525 |
+
{
|
292526 |
+
"epoch": 1224.0,
|
292527 |
+
"eval_loss": 0.34463566541671753,
|
292528 |
+
"eval_runtime": 39.9445,
|
292529 |
+
"eval_samples_per_second": 20.929,
|
292530 |
+
"eval_steps_per_second": 0.676,
|
292531 |
+
"eval_wer": 0.1835759611795446,
|
292532 |
+
"step": 151820
|
292533 |
+
},
|
292534 |
+
{
|
292535 |
+
"epoch": 1224.04,
|
292536 |
+
"learning_rate": 7.564361873990308e-06,
|
292537 |
+
"loss": 0.277,
|
292538 |
+
"step": 151825
|
292539 |
+
},
|
292540 |
+
{
|
292541 |
+
"epoch": 1224.08,
|
292542 |
+
"learning_rate": 7.564281098546043e-06,
|
292543 |
+
"loss": 0.2902,
|
292544 |
+
"step": 151830
|
292545 |
+
},
|
292546 |
+
{
|
292547 |
+
"epoch": 1224.12,
|
292548 |
+
"learning_rate": 7.564200323101778e-06,
|
292549 |
+
"loss": 0.2973,
|
292550 |
+
"step": 151835
|
292551 |
+
},
|
292552 |
+
{
|
292553 |
+
"epoch": 1224.16,
|
292554 |
+
"learning_rate": 7.564119547657513e-06,
|
292555 |
+
"loss": 0.4297,
|
292556 |
+
"step": 151840
|
292557 |
+
},
|
292558 |
+
{
|
292559 |
+
"epoch": 1224.2,
|
292560 |
+
"learning_rate": 7.564038772213248e-06,
|
292561 |
+
"loss": 1.2619,
|
292562 |
+
"step": 151845
|
292563 |
+
},
|
292564 |
+
{
|
292565 |
+
"epoch": 1224.24,
|
292566 |
+
"learning_rate": 7.563957996768983e-06,
|
292567 |
+
"loss": 0.3364,
|
292568 |
+
"step": 151850
|
292569 |
+
},
|
292570 |
+
{
|
292571 |
+
"epoch": 1224.28,
|
292572 |
+
"learning_rate": 7.563877221324718e-06,
|
292573 |
+
"loss": 0.2814,
|
292574 |
+
"step": 151855
|
292575 |
+
},
|
292576 |
+
{
|
292577 |
+
"epoch": 1224.32,
|
292578 |
+
"learning_rate": 7.563796445880453e-06,
|
292579 |
+
"loss": 0.3734,
|
292580 |
+
"step": 151860
|
292581 |
+
},
|
292582 |
+
{
|
292583 |
+
"epoch": 1224.36,
|
292584 |
+
"learning_rate": 7.563715670436188e-06,
|
292585 |
+
"loss": 0.4837,
|
292586 |
+
"step": 151865
|
292587 |
+
},
|
292588 |
+
{
|
292589 |
+
"epoch": 1224.4,
|
292590 |
+
"learning_rate": 7.5636348949919235e-06,
|
292591 |
+
"loss": 1.1229,
|
292592 |
+
"step": 151870
|
292593 |
+
},
|
292594 |
+
{
|
292595 |
+
"epoch": 1224.44,
|
292596 |
+
"learning_rate": 7.5635541195476585e-06,
|
292597 |
+
"loss": 0.3491,
|
292598 |
+
"step": 151875
|
292599 |
+
},
|
292600 |
+
{
|
292601 |
+
"epoch": 1224.48,
|
292602 |
+
"learning_rate": 7.5634733441033935e-06,
|
292603 |
+
"loss": 0.2809,
|
292604 |
+
"step": 151880
|
292605 |
+
},
|
292606 |
+
{
|
292607 |
+
"epoch": 1224.52,
|
292608 |
+
"learning_rate": 7.5633925686591284e-06,
|
292609 |
+
"loss": 0.2803,
|
292610 |
+
"step": 151885
|
292611 |
+
},
|
292612 |
+
{
|
292613 |
+
"epoch": 1224.56,
|
292614 |
+
"learning_rate": 7.5633117932148634e-06,
|
292615 |
+
"loss": 0.4462,
|
292616 |
+
"step": 151890
|
292617 |
+
},
|
292618 |
+
{
|
292619 |
+
"epoch": 1224.6,
|
292620 |
+
"learning_rate": 7.563231017770598e-06,
|
292621 |
+
"loss": 1.3136,
|
292622 |
+
"step": 151895
|
292623 |
+
},
|
292624 |
+
{
|
292625 |
+
"epoch": 1224.65,
|
292626 |
+
"learning_rate": 7.563150242326333e-06,
|
292627 |
+
"loss": 0.2907,
|
292628 |
+
"step": 151900
|
292629 |
+
},
|
292630 |
+
{
|
292631 |
+
"epoch": 1224.69,
|
292632 |
+
"learning_rate": 7.563069466882068e-06,
|
292633 |
+
"loss": 0.2524,
|
292634 |
+
"step": 151905
|
292635 |
+
},
|
292636 |
+
{
|
292637 |
+
"epoch": 1224.73,
|
292638 |
+
"learning_rate": 7.562988691437803e-06,
|
292639 |
+
"loss": 0.3055,
|
292640 |
+
"step": 151910
|
292641 |
+
},
|
292642 |
+
{
|
292643 |
+
"epoch": 1224.77,
|
292644 |
+
"learning_rate": 7.562907915993538e-06,
|
292645 |
+
"loss": 0.4168,
|
292646 |
+
"step": 151915
|
292647 |
+
},
|
292648 |
+
{
|
292649 |
+
"epoch": 1224.81,
|
292650 |
+
"learning_rate": 7.562827140549273e-06,
|
292651 |
+
"loss": 0.9734,
|
292652 |
+
"step": 151920
|
292653 |
+
},
|
292654 |
+
{
|
292655 |
+
"epoch": 1224.85,
|
292656 |
+
"learning_rate": 7.562746365105008e-06,
|
292657 |
+
"loss": 0.2457,
|
292658 |
+
"step": 151925
|
292659 |
+
},
|
292660 |
+
{
|
292661 |
+
"epoch": 1224.89,
|
292662 |
+
"learning_rate": 7.562665589660743e-06,
|
292663 |
+
"loss": 0.2464,
|
292664 |
+
"step": 151930
|
292665 |
+
},
|
292666 |
+
{
|
292667 |
+
"epoch": 1224.93,
|
292668 |
+
"learning_rate": 7.562584814216479e-06,
|
292669 |
+
"loss": 0.2614,
|
292670 |
+
"step": 151935
|
292671 |
+
},
|
292672 |
+
{
|
292673 |
+
"epoch": 1224.97,
|
292674 |
+
"learning_rate": 7.562504038772213e-06,
|
292675 |
+
"loss": 0.4991,
|
292676 |
+
"step": 151940
|
292677 |
+
},
|
292678 |
+
{
|
292679 |
+
"epoch": 1225.0,
|
292680 |
+
"eval_loss": 0.344937801361084,
|
292681 |
+
"eval_runtime": 40.5017,
|
292682 |
+
"eval_samples_per_second": 20.666,
|
292683 |
+
"eval_steps_per_second": 0.667,
|
292684 |
+
"eval_wer": 0.1767148683722655,
|
292685 |
+
"step": 151944
|
292686 |
+
},
|
292687 |
+
{
|
292688 |
+
"epoch": 1225.01,
|
292689 |
+
"learning_rate": 7.562423263327949e-06,
|
292690 |
+
"loss": 0.2553,
|
292691 |
+
"step": 151945
|
292692 |
+
},
|
292693 |
+
{
|
292694 |
+
"epoch": 1225.05,
|
292695 |
+
"learning_rate": 7.562342487883684e-06,
|
292696 |
+
"loss": 0.2594,
|
292697 |
+
"step": 151950
|
292698 |
+
},
|
292699 |
+
{
|
292700 |
+
"epoch": 1225.09,
|
292701 |
+
"learning_rate": 7.562261712439419e-06,
|
292702 |
+
"loss": 0.2837,
|
292703 |
+
"step": 151955
|
292704 |
+
},
|
292705 |
+
{
|
292706 |
+
"epoch": 1225.13,
|
292707 |
+
"learning_rate": 7.562180936995154e-06,
|
292708 |
+
"loss": 0.2949,
|
292709 |
+
"step": 151960
|
292710 |
+
},
|
292711 |
+
{
|
292712 |
+
"epoch": 1225.17,
|
292713 |
+
"learning_rate": 7.562100161550889e-06,
|
292714 |
+
"loss": 0.4855,
|
292715 |
+
"step": 151965
|
292716 |
+
},
|
292717 |
+
{
|
292718 |
+
"epoch": 1225.21,
|
292719 |
+
"learning_rate": 7.562019386106624e-06,
|
292720 |
+
"loss": 1.1717,
|
292721 |
+
"step": 151970
|
292722 |
+
},
|
292723 |
+
{
|
292724 |
+
"epoch": 1225.25,
|
292725 |
+
"learning_rate": 7.561938610662359e-06,
|
292726 |
+
"loss": 0.3164,
|
292727 |
+
"step": 151975
|
292728 |
+
},
|
292729 |
+
{
|
292730 |
+
"epoch": 1225.29,
|
292731 |
+
"learning_rate": 7.561857835218094e-06,
|
292732 |
+
"loss": 0.2502,
|
292733 |
+
"step": 151980
|
292734 |
+
},
|
292735 |
+
{
|
292736 |
+
"epoch": 1225.33,
|
292737 |
+
"learning_rate": 7.561777059773829e-06,
|
292738 |
+
"loss": 0.2958,
|
292739 |
+
"step": 151985
|
292740 |
+
},
|
292741 |
+
{
|
292742 |
+
"epoch": 1225.37,
|
292743 |
+
"learning_rate": 7.561696284329565e-06,
|
292744 |
+
"loss": 0.4835,
|
292745 |
+
"step": 151990
|
292746 |
+
},
|
292747 |
+
{
|
292748 |
+
"epoch": 1225.41,
|
292749 |
+
"learning_rate": 7.561615508885299e-06,
|
292750 |
+
"loss": 1.2292,
|
292751 |
+
"step": 151995
|
292752 |
+
},
|
292753 |
+
{
|
292754 |
+
"epoch": 1225.45,
|
292755 |
+
"learning_rate": 7.561534733441035e-06,
|
292756 |
+
"loss": 0.3822,
|
292757 |
+
"step": 152000
|
292758 |
+
},
|
292759 |
+
{
|
292760 |
+
"epoch": 1225.49,
|
292761 |
+
"learning_rate": 7.561453957996769e-06,
|
292762 |
+
"loss": 0.326,
|
292763 |
+
"step": 152005
|
292764 |
+
},
|
292765 |
+
{
|
292766 |
+
"epoch": 1225.53,
|
292767 |
+
"learning_rate": 7.561373182552505e-06,
|
292768 |
+
"loss": 0.3352,
|
292769 |
+
"step": 152010
|
292770 |
+
},
|
292771 |
+
{
|
292772 |
+
"epoch": 1225.57,
|
292773 |
+
"learning_rate": 7.56129240710824e-06,
|
292774 |
+
"loss": 0.5148,
|
292775 |
+
"step": 152015
|
292776 |
+
},
|
292777 |
+
{
|
292778 |
+
"epoch": 1225.61,
|
292779 |
+
"learning_rate": 7.561211631663975e-06,
|
292780 |
+
"loss": 1.0644,
|
292781 |
+
"step": 152020
|
292782 |
+
},
|
292783 |
+
{
|
292784 |
+
"epoch": 1225.65,
|
292785 |
+
"learning_rate": 7.56113085621971e-06,
|
292786 |
+
"loss": 0.2705,
|
292787 |
+
"step": 152025
|
292788 |
+
},
|
292789 |
+
{
|
292790 |
+
"epoch": 1225.69,
|
292791 |
+
"learning_rate": 7.561050080775445e-06,
|
292792 |
+
"loss": 0.2654,
|
292793 |
+
"step": 152030
|
292794 |
+
},
|
292795 |
+
{
|
292796 |
+
"epoch": 1225.73,
|
292797 |
+
"learning_rate": 7.56096930533118e-06,
|
292798 |
+
"loss": 0.2655,
|
292799 |
+
"step": 152035
|
292800 |
+
},
|
292801 |
+
{
|
292802 |
+
"epoch": 1225.77,
|
292803 |
+
"learning_rate": 7.560888529886915e-06,
|
292804 |
+
"loss": 0.5798,
|
292805 |
+
"step": 152040
|
292806 |
+
},
|
292807 |
+
{
|
292808 |
+
"epoch": 1225.81,
|
292809 |
+
"learning_rate": 7.560807754442651e-06,
|
292810 |
+
"loss": 1.0464,
|
292811 |
+
"step": 152045
|
292812 |
+
},
|
292813 |
+
{
|
292814 |
+
"epoch": 1225.85,
|
292815 |
+
"learning_rate": 7.560726978998385e-06,
|
292816 |
+
"loss": 0.264,
|
292817 |
+
"step": 152050
|
292818 |
+
},
|
292819 |
+
{
|
292820 |
+
"epoch": 1225.89,
|
292821 |
+
"learning_rate": 7.5606462035541206e-06,
|
292822 |
+
"loss": 0.2871,
|
292823 |
+
"step": 152055
|
292824 |
+
},
|
292825 |
+
{
|
292826 |
+
"epoch": 1225.93,
|
292827 |
+
"learning_rate": 7.560565428109855e-06,
|
292828 |
+
"loss": 0.3312,
|
292829 |
+
"step": 152060
|
292830 |
+
},
|
292831 |
+
{
|
292832 |
+
"epoch": 1225.97,
|
292833 |
+
"learning_rate": 7.5604846526655905e-06,
|
292834 |
+
"loss": 0.5482,
|
292835 |
+
"step": 152065
|
292836 |
+
},
|
292837 |
+
{
|
292838 |
+
"epoch": 1226.0,
|
292839 |
+
"eval_loss": 0.3725050091743469,
|
292840 |
+
"eval_runtime": 40.5085,
|
292841 |
+
"eval_samples_per_second": 20.662,
|
292842 |
+
"eval_steps_per_second": 0.667,
|
292843 |
+
"eval_wer": 0.1764970733903647,
|
292844 |
+
"step": 152068
|
292845 |
}
|
292846 |
],
|
292847 |
"max_steps": 620000,
|
292848 |
"num_train_epochs": 5000,
|
292849 |
+
"total_flos": 4.2793273628046655e+20,
|
292850 |
"trial_name": null,
|
292851 |
"trial_params": null
|
292852 |
}
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630248561.4202452/events.out.tfevents.1630248561.cc93b136ebf5.1086.199
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e69c53824a99a78f62d60a3191904b4929d958e37a96aed121cea82ef6f44c6
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630248992.3105023/events.out.tfevents.1630248992.cc93b136ebf5.1086.201
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:18c11de576210df8c1bcd2e338686d0380f1a420b31373ffd06eb3aed8c1f453
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630249423.7833743/events.out.tfevents.1630249423.cc93b136ebf5.1086.203
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:568ece09adb0211ba3f15921d31806d68262b8cfa9857606486c11d4bae9edd6
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630249857.8968918/events.out.tfevents.1630249857.cc93b136ebf5.1086.205
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca6c23b7959ed4eff92db23ba2ba96ded88c6e6bd51c3726d065275e78ced855
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630250287.4046617/events.out.tfevents.1630250287.cc93b136ebf5.1086.207
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d4f64e2892b1172fc9ed28d3b5df11b5c1e87ce978ca5299337293b8153e478
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630248561.cc93b136ebf5.1086.198
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4839731a799003ae2c63f39a2b9d8d658d064e4e4d52528b686ea21a72782e95
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630248992.cc93b136ebf5.1086.200
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d0ef6e85b448f1e9daa36e511987fdd78a3da2fe85a0475a40bc89341de9b3cb
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630249423.cc93b136ebf5.1086.202
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7ad462f18f3ef5e7674be715c076f3a6912a2bdaa53ab1bde8f0f57357adc4f4
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630249857.cc93b136ebf5.1086.204
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8beff9185670015d1b0453cddc98932f4cf7524e092ff8056b49425d192625e8
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630250287.cc93b136ebf5.1086.206
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4c75d446f14703e0a0c02a6515f68440f0868da9b0a01d30d83ccc76744caace
|
3 |
+
size 8622
|