"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/trainer_state.json +639 -3
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630080962.0011778/events.out.tfevents.1630080962.dea8124df033.1100.1 +3 -0
- model-bin/finetune/base/log/1630081546.0007536/events.out.tfevents.1630081546.dea8124df033.1391.1 +3 -0
- model-bin/finetune/base/log/1630081975.5740092/events.out.tfevents.1630081975.dea8124df033.1391.3 +3 -0
- model-bin/finetune/base/log/1630082400.0108204/events.out.tfevents.1630082400.dea8124df033.1391.5 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630080961.dea8124df033.1100.0 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630081545.dea8124df033.1391.0 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630081975.dea8124df033.1391.2 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630082400.dea8124df033.1391.4 +3 -0
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8f3ce461a41ebb478a3033a35cb4191908dec25dc3b09f575609fc6aa146dd60
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed4c4bc6da00d916aba5b2523194212a682bc5e8e07fee75fc5d3fcb170a87ce
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8135104c63b5165310558e47d8ba777edf4c3b459df9a523c789c28dcc1e549
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:531a84460dc36ad58f317625a8b2df924c497e322bd18a71440cc0d51bac941b
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:57fccb27f6be770c60514fd2f134fe1b6e0a8965992c964c9bd1421740a9ae29
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -242184,11 +242184,647 @@
|
|
242184 |
"eval_steps_per_second": 0.643,
|
242185 |
"eval_wer": 0.18449336594766882,
|
242186 |
"step": 112251
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
242187 |
}
|
242188 |
],
|
242189 |
"max_steps": 625000,
|
242190 |
"num_train_epochs": 5000,
|
242191 |
-
"total_flos": 3.
|
242192 |
"trial_name": null,
|
242193 |
"trial_params": null
|
242194 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 902.0,
|
5 |
+
"global_step": 112750,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
242184 |
"eval_steps_per_second": 0.643,
|
242185 |
"eval_wer": 0.18449336594766882,
|
242186 |
"step": 112251
|
242187 |
+
},
|
242188 |
+
{
|
242189 |
+
"epoch": 898.03,
|
242190 |
+
"learning_rate": 8.217772435897437e-06,
|
242191 |
+
"loss": 0.302,
|
242192 |
+
"step": 112255
|
242193 |
+
},
|
242194 |
+
{
|
242195 |
+
"epoch": 898.07,
|
242196 |
+
"learning_rate": 8.217692307692308e-06,
|
242197 |
+
"loss": 0.2726,
|
242198 |
+
"step": 112260
|
242199 |
+
},
|
242200 |
+
{
|
242201 |
+
"epoch": 898.11,
|
242202 |
+
"learning_rate": 8.21761217948718e-06,
|
242203 |
+
"loss": 0.3286,
|
242204 |
+
"step": 112265
|
242205 |
+
},
|
242206 |
+
{
|
242207 |
+
"epoch": 898.15,
|
242208 |
+
"learning_rate": 8.217532051282053e-06,
|
242209 |
+
"loss": 0.4197,
|
242210 |
+
"step": 112270
|
242211 |
+
},
|
242212 |
+
{
|
242213 |
+
"epoch": 898.19,
|
242214 |
+
"learning_rate": 8.217451923076924e-06,
|
242215 |
+
"loss": 0.8389,
|
242216 |
+
"step": 112275
|
242217 |
+
},
|
242218 |
+
{
|
242219 |
+
"epoch": 898.23,
|
242220 |
+
"learning_rate": 8.217371794871795e-06,
|
242221 |
+
"loss": 0.6037,
|
242222 |
+
"step": 112280
|
242223 |
+
},
|
242224 |
+
{
|
242225 |
+
"epoch": 898.27,
|
242226 |
+
"learning_rate": 8.217291666666667e-06,
|
242227 |
+
"loss": 0.2566,
|
242228 |
+
"step": 112285
|
242229 |
+
},
|
242230 |
+
{
|
242231 |
+
"epoch": 898.31,
|
242232 |
+
"learning_rate": 8.21721153846154e-06,
|
242233 |
+
"loss": 0.3006,
|
242234 |
+
"step": 112290
|
242235 |
+
},
|
242236 |
+
{
|
242237 |
+
"epoch": 898.35,
|
242238 |
+
"learning_rate": 8.217131410256411e-06,
|
242239 |
+
"loss": 0.3725,
|
242240 |
+
"step": 112295
|
242241 |
+
},
|
242242 |
+
{
|
242243 |
+
"epoch": 898.39,
|
242244 |
+
"learning_rate": 8.217051282051282e-06,
|
242245 |
+
"loss": 0.8954,
|
242246 |
+
"step": 112300
|
242247 |
+
},
|
242248 |
+
{
|
242249 |
+
"epoch": 898.43,
|
242250 |
+
"learning_rate": 8.216971153846156e-06,
|
242251 |
+
"loss": 0.7274,
|
242252 |
+
"step": 112305
|
242253 |
+
},
|
242254 |
+
{
|
242255 |
+
"epoch": 898.47,
|
242256 |
+
"learning_rate": 8.216891025641027e-06,
|
242257 |
+
"loss": 0.293,
|
242258 |
+
"step": 112310
|
242259 |
+
},
|
242260 |
+
{
|
242261 |
+
"epoch": 898.51,
|
242262 |
+
"learning_rate": 8.216810897435898e-06,
|
242263 |
+
"loss": 0.2938,
|
242264 |
+
"step": 112315
|
242265 |
+
},
|
242266 |
+
{
|
242267 |
+
"epoch": 898.55,
|
242268 |
+
"learning_rate": 8.21673076923077e-06,
|
242269 |
+
"loss": 0.3959,
|
242270 |
+
"step": 112320
|
242271 |
+
},
|
242272 |
+
{
|
242273 |
+
"epoch": 898.59,
|
242274 |
+
"learning_rate": 8.216650641025643e-06,
|
242275 |
+
"loss": 0.8866,
|
242276 |
+
"step": 112325
|
242277 |
+
},
|
242278 |
+
{
|
242279 |
+
"epoch": 898.63,
|
242280 |
+
"learning_rate": 8.216570512820512e-06,
|
242281 |
+
"loss": 0.8329,
|
242282 |
+
"step": 112330
|
242283 |
+
},
|
242284 |
+
{
|
242285 |
+
"epoch": 898.67,
|
242286 |
+
"learning_rate": 8.216490384615385e-06,
|
242287 |
+
"loss": 0.3571,
|
242288 |
+
"step": 112335
|
242289 |
+
},
|
242290 |
+
{
|
242291 |
+
"epoch": 898.71,
|
242292 |
+
"learning_rate": 8.216410256410257e-06,
|
242293 |
+
"loss": 0.3254,
|
242294 |
+
"step": 112340
|
242295 |
+
},
|
242296 |
+
{
|
242297 |
+
"epoch": 898.75,
|
242298 |
+
"learning_rate": 8.216330128205128e-06,
|
242299 |
+
"loss": 0.3964,
|
242300 |
+
"step": 112345
|
242301 |
+
},
|
242302 |
+
{
|
242303 |
+
"epoch": 898.79,
|
242304 |
+
"learning_rate": 8.216250000000001e-06,
|
242305 |
+
"loss": 1.0306,
|
242306 |
+
"step": 112350
|
242307 |
+
},
|
242308 |
+
{
|
242309 |
+
"epoch": 898.83,
|
242310 |
+
"learning_rate": 8.216169871794873e-06,
|
242311 |
+
"loss": 0.6421,
|
242312 |
+
"step": 112355
|
242313 |
+
},
|
242314 |
+
{
|
242315 |
+
"epoch": 898.87,
|
242316 |
+
"learning_rate": 8.216089743589744e-06,
|
242317 |
+
"loss": 0.3203,
|
242318 |
+
"step": 112360
|
242319 |
+
},
|
242320 |
+
{
|
242321 |
+
"epoch": 898.91,
|
242322 |
+
"learning_rate": 8.216009615384615e-06,
|
242323 |
+
"loss": 0.3004,
|
242324 |
+
"step": 112365
|
242325 |
+
},
|
242326 |
+
{
|
242327 |
+
"epoch": 898.95,
|
242328 |
+
"learning_rate": 8.215929487179488e-06,
|
242329 |
+
"loss": 0.3823,
|
242330 |
+
"step": 112370
|
242331 |
+
},
|
242332 |
+
{
|
242333 |
+
"epoch": 898.99,
|
242334 |
+
"learning_rate": 8.21584935897436e-06,
|
242335 |
+
"loss": 0.8118,
|
242336 |
+
"step": 112375
|
242337 |
+
},
|
242338 |
+
{
|
242339 |
+
"epoch": 899.0,
|
242340 |
+
"eval_loss": 0.40806370973587036,
|
242341 |
+
"eval_runtime": 41.149,
|
242342 |
+
"eval_samples_per_second": 20.341,
|
242343 |
+
"eval_steps_per_second": 0.656,
|
242344 |
+
"eval_wer": 0.18508635386553998,
|
242345 |
+
"step": 112376
|
242346 |
+
},
|
242347 |
+
{
|
242348 |
+
"epoch": 906.03,
|
242349 |
+
"learning_rate": 8.201357027463652e-06,
|
242350 |
+
"loss": 0.3272,
|
242351 |
+
"step": 112380
|
242352 |
+
},
|
242353 |
+
{
|
242354 |
+
"epoch": 906.07,
|
242355 |
+
"learning_rate": 8.201276252019386e-06,
|
242356 |
+
"loss": 0.2657,
|
242357 |
+
"step": 112385
|
242358 |
+
},
|
242359 |
+
{
|
242360 |
+
"epoch": 906.11,
|
242361 |
+
"learning_rate": 8.201195476575122e-06,
|
242362 |
+
"loss": 0.304,
|
242363 |
+
"step": 112390
|
242364 |
+
},
|
242365 |
+
{
|
242366 |
+
"epoch": 906.15,
|
242367 |
+
"learning_rate": 8.201114701130856e-06,
|
242368 |
+
"loss": 0.4485,
|
242369 |
+
"step": 112395
|
242370 |
+
},
|
242371 |
+
{
|
242372 |
+
"epoch": 906.19,
|
242373 |
+
"learning_rate": 8.201033925686592e-06,
|
242374 |
+
"loss": 0.8494,
|
242375 |
+
"step": 112400
|
242376 |
+
},
|
242377 |
+
{
|
242378 |
+
"epoch": 906.23,
|
242379 |
+
"learning_rate": 8.200953150242326e-06,
|
242380 |
+
"loss": 0.5495,
|
242381 |
+
"step": 112405
|
242382 |
+
},
|
242383 |
+
{
|
242384 |
+
"epoch": 906.27,
|
242385 |
+
"learning_rate": 8.200872374798062e-06,
|
242386 |
+
"loss": 0.2863,
|
242387 |
+
"step": 112410
|
242388 |
+
},
|
242389 |
+
{
|
242390 |
+
"epoch": 906.31,
|
242391 |
+
"learning_rate": 8.200791599353796e-06,
|
242392 |
+
"loss": 0.306,
|
242393 |
+
"step": 112415
|
242394 |
+
},
|
242395 |
+
{
|
242396 |
+
"epoch": 906.35,
|
242397 |
+
"learning_rate": 8.200710823909532e-06,
|
242398 |
+
"loss": 0.369,
|
242399 |
+
"step": 112420
|
242400 |
+
},
|
242401 |
+
{
|
242402 |
+
"epoch": 906.4,
|
242403 |
+
"learning_rate": 8.200630048465266e-06,
|
242404 |
+
"loss": 0.912,
|
242405 |
+
"step": 112425
|
242406 |
+
},
|
242407 |
+
{
|
242408 |
+
"epoch": 906.44,
|
242409 |
+
"learning_rate": 8.200549273021002e-06,
|
242410 |
+
"loss": 0.7257,
|
242411 |
+
"step": 112430
|
242412 |
+
},
|
242413 |
+
{
|
242414 |
+
"epoch": 906.48,
|
242415 |
+
"learning_rate": 8.200468497576736e-06,
|
242416 |
+
"loss": 0.2877,
|
242417 |
+
"step": 112435
|
242418 |
+
},
|
242419 |
+
{
|
242420 |
+
"epoch": 906.52,
|
242421 |
+
"learning_rate": 8.200387722132472e-06,
|
242422 |
+
"loss": 0.3152,
|
242423 |
+
"step": 112440
|
242424 |
+
},
|
242425 |
+
{
|
242426 |
+
"epoch": 906.56,
|
242427 |
+
"learning_rate": 8.200306946688208e-06,
|
242428 |
+
"loss": 0.3952,
|
242429 |
+
"step": 112445
|
242430 |
+
},
|
242431 |
+
{
|
242432 |
+
"epoch": 906.6,
|
242433 |
+
"learning_rate": 8.200226171243942e-06,
|
242434 |
+
"loss": 0.8972,
|
242435 |
+
"step": 112450
|
242436 |
+
},
|
242437 |
+
{
|
242438 |
+
"epoch": 906.64,
|
242439 |
+
"learning_rate": 8.200145395799678e-06,
|
242440 |
+
"loss": 0.8771,
|
242441 |
+
"step": 112455
|
242442 |
+
},
|
242443 |
+
{
|
242444 |
+
"epoch": 906.68,
|
242445 |
+
"learning_rate": 8.200064620355412e-06,
|
242446 |
+
"loss": 0.3608,
|
242447 |
+
"step": 112460
|
242448 |
+
},
|
242449 |
+
{
|
242450 |
+
"epoch": 906.72,
|
242451 |
+
"learning_rate": 8.199983844911148e-06,
|
242452 |
+
"loss": 0.3328,
|
242453 |
+
"step": 112465
|
242454 |
+
},
|
242455 |
+
{
|
242456 |
+
"epoch": 906.76,
|
242457 |
+
"learning_rate": 8.199903069466882e-06,
|
242458 |
+
"loss": 0.3953,
|
242459 |
+
"step": 112470
|
242460 |
+
},
|
242461 |
+
{
|
242462 |
+
"epoch": 906.8,
|
242463 |
+
"learning_rate": 8.199822294022618e-06,
|
242464 |
+
"loss": 0.9336,
|
242465 |
+
"step": 112475
|
242466 |
+
},
|
242467 |
+
{
|
242468 |
+
"epoch": 906.84,
|
242469 |
+
"learning_rate": 8.199741518578352e-06,
|
242470 |
+
"loss": 0.7011,
|
242471 |
+
"step": 112480
|
242472 |
+
},
|
242473 |
+
{
|
242474 |
+
"epoch": 906.88,
|
242475 |
+
"learning_rate": 8.199660743134088e-06,
|
242476 |
+
"loss": 0.3199,
|
242477 |
+
"step": 112485
|
242478 |
+
},
|
242479 |
+
{
|
242480 |
+
"epoch": 906.92,
|
242481 |
+
"learning_rate": 8.199579967689822e-06,
|
242482 |
+
"loss": 0.2874,
|
242483 |
+
"step": 112490
|
242484 |
+
},
|
242485 |
+
{
|
242486 |
+
"epoch": 906.96,
|
242487 |
+
"learning_rate": 8.199499192245558e-06,
|
242488 |
+
"loss": 0.378,
|
242489 |
+
"step": 112495
|
242490 |
+
},
|
242491 |
+
{
|
242492 |
+
"epoch": 907.0,
|
242493 |
+
"learning_rate": 8.199418416801294e-06,
|
242494 |
+
"loss": 1.1028,
|
242495 |
+
"step": 112500
|
242496 |
+
},
|
242497 |
+
{
|
242498 |
+
"epoch": 907.0,
|
242499 |
+
"eval_loss": 0.3665727376937866,
|
242500 |
+
"eval_runtime": 40.8442,
|
242501 |
+
"eval_samples_per_second": 20.492,
|
242502 |
+
"eval_steps_per_second": 0.661,
|
242503 |
+
"eval_wer": 0.1838938822847202,
|
242504 |
+
"step": 112500
|
242505 |
+
},
|
242506 |
+
{
|
242507 |
+
"epoch": 900.04,
|
242508 |
+
"learning_rate": 8.199337641357028e-06,
|
242509 |
+
"loss": 0.3578,
|
242510 |
+
"step": 112505
|
242511 |
+
},
|
242512 |
+
{
|
242513 |
+
"epoch": 900.08,
|
242514 |
+
"learning_rate": 8.199256865912764e-06,
|
242515 |
+
"loss": 0.2938,
|
242516 |
+
"step": 112510
|
242517 |
+
},
|
242518 |
+
{
|
242519 |
+
"epoch": 900.12,
|
242520 |
+
"learning_rate": 8.199176090468498e-06,
|
242521 |
+
"loss": 0.3098,
|
242522 |
+
"step": 112515
|
242523 |
+
},
|
242524 |
+
{
|
242525 |
+
"epoch": 900.16,
|
242526 |
+
"learning_rate": 8.199095315024234e-06,
|
242527 |
+
"loss": 0.4641,
|
242528 |
+
"step": 112520
|
242529 |
+
},
|
242530 |
+
{
|
242531 |
+
"epoch": 900.2,
|
242532 |
+
"learning_rate": 8.199014539579968e-06,
|
242533 |
+
"loss": 1.0493,
|
242534 |
+
"step": 112525
|
242535 |
+
},
|
242536 |
+
{
|
242537 |
+
"epoch": 900.24,
|
242538 |
+
"learning_rate": 8.198933764135704e-06,
|
242539 |
+
"loss": 0.3391,
|
242540 |
+
"step": 112530
|
242541 |
+
},
|
242542 |
+
{
|
242543 |
+
"epoch": 900.28,
|
242544 |
+
"learning_rate": 8.198852988691438e-06,
|
242545 |
+
"loss": 0.3322,
|
242546 |
+
"step": 112535
|
242547 |
+
},
|
242548 |
+
{
|
242549 |
+
"epoch": 900.32,
|
242550 |
+
"learning_rate": 8.198772213247174e-06,
|
242551 |
+
"loss": 0.3384,
|
242552 |
+
"step": 112540
|
242553 |
+
},
|
242554 |
+
{
|
242555 |
+
"epoch": 900.36,
|
242556 |
+
"learning_rate": 8.198691437802908e-06,
|
242557 |
+
"loss": 0.4261,
|
242558 |
+
"step": 112545
|
242559 |
+
},
|
242560 |
+
{
|
242561 |
+
"epoch": 900.4,
|
242562 |
+
"learning_rate": 8.198610662358644e-06,
|
242563 |
+
"loss": 1.2445,
|
242564 |
+
"step": 112550
|
242565 |
+
},
|
242566 |
+
{
|
242567 |
+
"epoch": 900.44,
|
242568 |
+
"learning_rate": 8.19852988691438e-06,
|
242569 |
+
"loss": 0.2923,
|
242570 |
+
"step": 112555
|
242571 |
+
},
|
242572 |
+
{
|
242573 |
+
"epoch": 900.48,
|
242574 |
+
"learning_rate": 8.198449111470114e-06,
|
242575 |
+
"loss": 0.2442,
|
242576 |
+
"step": 112560
|
242577 |
+
},
|
242578 |
+
{
|
242579 |
+
"epoch": 900.52,
|
242580 |
+
"learning_rate": 8.19836833602585e-06,
|
242581 |
+
"loss": 0.3361,
|
242582 |
+
"step": 112565
|
242583 |
+
},
|
242584 |
+
{
|
242585 |
+
"epoch": 900.56,
|
242586 |
+
"learning_rate": 8.198287560581584e-06,
|
242587 |
+
"loss": 0.4356,
|
242588 |
+
"step": 112570
|
242589 |
+
},
|
242590 |
+
{
|
242591 |
+
"epoch": 900.6,
|
242592 |
+
"learning_rate": 8.19820678513732e-06,
|
242593 |
+
"loss": 1.1667,
|
242594 |
+
"step": 112575
|
242595 |
+
},
|
242596 |
+
{
|
242597 |
+
"epoch": 900.64,
|
242598 |
+
"learning_rate": 8.198126009693053e-06,
|
242599 |
+
"loss": 0.318,
|
242600 |
+
"step": 112580
|
242601 |
+
},
|
242602 |
+
{
|
242603 |
+
"epoch": 900.68,
|
242604 |
+
"learning_rate": 8.19804523424879e-06,
|
242605 |
+
"loss": 0.3111,
|
242606 |
+
"step": 112585
|
242607 |
+
},
|
242608 |
+
{
|
242609 |
+
"epoch": 900.72,
|
242610 |
+
"learning_rate": 8.197964458804523e-06,
|
242611 |
+
"loss": 0.3473,
|
242612 |
+
"step": 112590
|
242613 |
+
},
|
242614 |
+
{
|
242615 |
+
"epoch": 900.76,
|
242616 |
+
"learning_rate": 8.19788368336026e-06,
|
242617 |
+
"loss": 0.4569,
|
242618 |
+
"step": 112595
|
242619 |
+
},
|
242620 |
+
{
|
242621 |
+
"epoch": 900.8,
|
242622 |
+
"learning_rate": 8.197802907915993e-06,
|
242623 |
+
"loss": 1.2133,
|
242624 |
+
"step": 112600
|
242625 |
+
},
|
242626 |
+
{
|
242627 |
+
"epoch": 900.84,
|
242628 |
+
"learning_rate": 8.19772213247173e-06,
|
242629 |
+
"loss": 0.324,
|
242630 |
+
"step": 112605
|
242631 |
+
},
|
242632 |
+
{
|
242633 |
+
"epoch": 900.88,
|
242634 |
+
"learning_rate": 8.197641357027463e-06,
|
242635 |
+
"loss": 0.2444,
|
242636 |
+
"step": 112610
|
242637 |
+
},
|
242638 |
+
{
|
242639 |
+
"epoch": 900.92,
|
242640 |
+
"learning_rate": 8.1975605815832e-06,
|
242641 |
+
"loss": 0.4005,
|
242642 |
+
"step": 112615
|
242643 |
+
},
|
242644 |
+
{
|
242645 |
+
"epoch": 900.96,
|
242646 |
+
"learning_rate": 8.197479806138935e-06,
|
242647 |
+
"loss": 0.528,
|
242648 |
+
"step": 112620
|
242649 |
+
},
|
242650 |
+
{
|
242651 |
+
"epoch": 901.0,
|
242652 |
+
"learning_rate": 8.19739903069467e-06,
|
242653 |
+
"loss": 1.3845,
|
242654 |
+
"step": 112625
|
242655 |
+
},
|
242656 |
+
{
|
242657 |
+
"epoch": 901.0,
|
242658 |
+
"eval_loss": 0.4503331482410431,
|
242659 |
+
"eval_runtime": 40.4509,
|
242660 |
+
"eval_samples_per_second": 20.667,
|
242661 |
+
"eval_steps_per_second": 0.667,
|
242662 |
+
"eval_wer": 0.18393001543777107,
|
242663 |
+
"step": 112625
|
242664 |
+
},
|
242665 |
+
{
|
242666 |
+
"epoch": 901.04,
|
242667 |
+
"learning_rate": 8.197318255250405e-06,
|
242668 |
+
"loss": 0.3084,
|
242669 |
+
"step": 112630
|
242670 |
+
},
|
242671 |
+
{
|
242672 |
+
"epoch": 901.08,
|
242673 |
+
"learning_rate": 8.19723747980614e-06,
|
242674 |
+
"loss": 0.3249,
|
242675 |
+
"step": 112635
|
242676 |
+
},
|
242677 |
+
{
|
242678 |
+
"epoch": 901.12,
|
242679 |
+
"learning_rate": 8.197156704361875e-06,
|
242680 |
+
"loss": 0.3247,
|
242681 |
+
"step": 112640
|
242682 |
+
},
|
242683 |
+
{
|
242684 |
+
"epoch": 901.16,
|
242685 |
+
"learning_rate": 8.19707592891761e-06,
|
242686 |
+
"loss": 0.4088,
|
242687 |
+
"step": 112645
|
242688 |
+
},
|
242689 |
+
{
|
242690 |
+
"epoch": 901.2,
|
242691 |
+
"learning_rate": 8.196995153473345e-06,
|
242692 |
+
"loss": 1.0528,
|
242693 |
+
"step": 112650
|
242694 |
+
},
|
242695 |
+
{
|
242696 |
+
"epoch": 901.24,
|
242697 |
+
"learning_rate": 8.19691437802908e-06,
|
242698 |
+
"loss": 0.3751,
|
242699 |
+
"step": 112655
|
242700 |
+
},
|
242701 |
+
{
|
242702 |
+
"epoch": 901.28,
|
242703 |
+
"learning_rate": 8.196833602584815e-06,
|
242704 |
+
"loss": 0.3149,
|
242705 |
+
"step": 112660
|
242706 |
+
},
|
242707 |
+
{
|
242708 |
+
"epoch": 901.32,
|
242709 |
+
"learning_rate": 8.196752827140549e-06,
|
242710 |
+
"loss": 0.3781,
|
242711 |
+
"step": 112665
|
242712 |
+
},
|
242713 |
+
{
|
242714 |
+
"epoch": 901.36,
|
242715 |
+
"learning_rate": 8.196672051696285e-06,
|
242716 |
+
"loss": 0.4784,
|
242717 |
+
"step": 112670
|
242718 |
+
},
|
242719 |
+
{
|
242720 |
+
"epoch": 901.4,
|
242721 |
+
"learning_rate": 8.19659127625202e-06,
|
242722 |
+
"loss": 1.2451,
|
242723 |
+
"step": 112675
|
242724 |
+
},
|
242725 |
+
{
|
242726 |
+
"epoch": 901.44,
|
242727 |
+
"learning_rate": 8.196510500807755e-06,
|
242728 |
+
"loss": 0.3353,
|
242729 |
+
"step": 112680
|
242730 |
+
},
|
242731 |
+
{
|
242732 |
+
"epoch": 901.48,
|
242733 |
+
"learning_rate": 8.19642972536349e-06,
|
242734 |
+
"loss": 0.2894,
|
242735 |
+
"step": 112685
|
242736 |
+
},
|
242737 |
+
{
|
242738 |
+
"epoch": 901.52,
|
242739 |
+
"learning_rate": 8.196348949919225e-06,
|
242740 |
+
"loss": 0.3487,
|
242741 |
+
"step": 112690
|
242742 |
+
},
|
242743 |
+
{
|
242744 |
+
"epoch": 901.56,
|
242745 |
+
"learning_rate": 8.19626817447496e-06,
|
242746 |
+
"loss": 0.4494,
|
242747 |
+
"step": 112695
|
242748 |
+
},
|
242749 |
+
{
|
242750 |
+
"epoch": 901.6,
|
242751 |
+
"learning_rate": 8.196187399030695e-06,
|
242752 |
+
"loss": 1.2488,
|
242753 |
+
"step": 112700
|
242754 |
+
},
|
242755 |
+
{
|
242756 |
+
"epoch": 901.64,
|
242757 |
+
"learning_rate": 8.19610662358643e-06,
|
242758 |
+
"loss": 0.3036,
|
242759 |
+
"step": 112705
|
242760 |
+
},
|
242761 |
+
{
|
242762 |
+
"epoch": 901.68,
|
242763 |
+
"learning_rate": 8.196025848142165e-06,
|
242764 |
+
"loss": 0.2959,
|
242765 |
+
"step": 112710
|
242766 |
+
},
|
242767 |
+
{
|
242768 |
+
"epoch": 901.72,
|
242769 |
+
"learning_rate": 8.1959450726979e-06,
|
242770 |
+
"loss": 0.2841,
|
242771 |
+
"step": 112715
|
242772 |
+
},
|
242773 |
+
{
|
242774 |
+
"epoch": 901.76,
|
242775 |
+
"learning_rate": 8.195864297253635e-06,
|
242776 |
+
"loss": 0.4297,
|
242777 |
+
"step": 112720
|
242778 |
+
},
|
242779 |
+
{
|
242780 |
+
"epoch": 901.8,
|
242781 |
+
"learning_rate": 8.19578352180937e-06,
|
242782 |
+
"loss": 1.2188,
|
242783 |
+
"step": 112725
|
242784 |
+
},
|
242785 |
+
{
|
242786 |
+
"epoch": 901.84,
|
242787 |
+
"learning_rate": 8.195702746365107e-06,
|
242788 |
+
"loss": 0.2619,
|
242789 |
+
"step": 112730
|
242790 |
+
},
|
242791 |
+
{
|
242792 |
+
"epoch": 901.88,
|
242793 |
+
"learning_rate": 8.19562197092084e-06,
|
242794 |
+
"loss": 0.2855,
|
242795 |
+
"step": 112735
|
242796 |
+
},
|
242797 |
+
{
|
242798 |
+
"epoch": 901.92,
|
242799 |
+
"learning_rate": 8.195541195476576e-06,
|
242800 |
+
"loss": 0.3261,
|
242801 |
+
"step": 112740
|
242802 |
+
},
|
242803 |
+
{
|
242804 |
+
"epoch": 901.96,
|
242805 |
+
"learning_rate": 8.19546042003231e-06,
|
242806 |
+
"loss": 0.4363,
|
242807 |
+
"step": 112745
|
242808 |
+
},
|
242809 |
+
{
|
242810 |
+
"epoch": 902.0,
|
242811 |
+
"learning_rate": 8.195379644588046e-06,
|
242812 |
+
"loss": 1.1429,
|
242813 |
+
"step": 112750
|
242814 |
+
},
|
242815 |
+
{
|
242816 |
+
"epoch": 902.0,
|
242817 |
+
"eval_loss": 0.5105034708976746,
|
242818 |
+
"eval_runtime": 40.0763,
|
242819 |
+
"eval_samples_per_second": 20.86,
|
242820 |
+
"eval_steps_per_second": 0.674,
|
242821 |
+
"eval_wer": 0.19159501927067893,
|
242822 |
+
"step": 112750
|
242823 |
}
|
242824 |
],
|
242825 |
"max_steps": 625000,
|
242826 |
"num_train_epochs": 5000,
|
242827 |
+
"total_flos": 3.1730977775534506e+20,
|
242828 |
"trial_name": null,
|
242829 |
"trial_params": null
|
242830 |
}
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630080962.0011778/events.out.tfevents.1630080962.dea8124df033.1100.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65769844150994ff1555f45c9fd77bcdc22d6e17d7ec210515f7dcc46d700e00
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630081546.0007536/events.out.tfevents.1630081546.dea8124df033.1391.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ca7f4f4e86a2d00bb90bf95a40178cbfacb9d1728beb7d620e9f81b02c3ec64
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630081975.5740092/events.out.tfevents.1630081975.dea8124df033.1391.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a09512d50d1ef26d2f297f63456ce7b2241a6d47e49e31d1709f1a540a19d283
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630082400.0108204/events.out.tfevents.1630082400.dea8124df033.1391.5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:591d0fec865524e1bd554d8008efd614fcfcc6a63c2fd4f85f2741bdf87ecfe2
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630080961.dea8124df033.1100.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6e5f1d6e1f7f53e3ece49153d4e77fcb0baf6ef7d1ad168e857905f19d2fb26
|
3 |
+
size 8630
|
model-bin/finetune/base/log/events.out.tfevents.1630081545.dea8124df033.1391.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ac932a20bfe1b4c19ed8e11b0eed7b25a3e961e714f7aba090ff646d0b4daee
|
3 |
+
size 8630
|
model-bin/finetune/base/log/events.out.tfevents.1630081975.dea8124df033.1391.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3855248414824359cfde883f6e12381b6483dc45b106e824cc4e0814703b94cb
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630082400.dea8124df033.1391.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65473aae46d56c2a52378f6db4915b2bbf7ee5b710e3bb6d6b4cf153cab8aad5
|
3 |
+
size 8622
|