"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630117908.6872168/events.out.tfevents.1630117908.86bb0ddabf9b.1042.51 +3 -0
- model-bin/finetune/base/log/1630118323.7473285/events.out.tfevents.1630118323.86bb0ddabf9b.1042.53 +3 -0
- model-bin/finetune/base/log/1630118841.0029836/events.out.tfevents.1630118841.86bb0ddabf9b.1042.55 +3 -0
- model-bin/finetune/base/log/1630119249.7913108/events.out.tfevents.1630119249.86bb0ddabf9b.1042.57 +3 -0
- model-bin/finetune/base/log/1630119657.5811403/events.out.tfevents.1630119657.86bb0ddabf9b.1042.59 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630117908.86bb0ddabf9b.1042.50 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630118323.86bb0ddabf9b.1042.52 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630118840.86bb0ddabf9b.1042.54 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630119249.86bb0ddabf9b.1042.56 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630119657.86bb0ddabf9b.1042.58 +3 -0
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f8f9d9e29f9c4357eccefce34b1788641cbf79fdb30f31166c13d2a3991cec7
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d6fd2acf7aa2ca9a155077ce484b63264f0447cd4f800916e207dcb087cf3dc
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3883b24ab055ced9e5920bf2f787c0bf5fe045f5bc46d3e0a656d2fb11a7e73e
|
3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:85667687b163c74870e550a93ae8876dd8fc5752af02aeaab176e91efede68c0
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee5116ffc5a6492b87e5b79cfa6a431988c523d33002a0376caef480f1a04f73
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -248355,11 +248355,800 @@
|
|
248355 |
"eval_steps_per_second": 0.679,
|
248356 |
"eval_wer": 0.1851399856424982,
|
248357 |
"step": 117100
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
248358 |
}
|
248359 |
],
|
248360 |
"max_steps": 625000,
|
248361 |
"num_train_epochs": 5000,
|
248362 |
-
"total_flos": 3.
|
248363 |
"trial_name": null,
|
248364 |
"trial_params": null
|
248365 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 941.0,
|
5 |
+
"global_step": 117722,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
248355 |
"eval_steps_per_second": 0.679,
|
248356 |
"eval_wer": 0.1851399856424982,
|
248357 |
"step": 117100
|
248358 |
+
},
|
248359 |
+
{
|
248360 |
+
"epoch": 936.04,
|
248361 |
+
"learning_rate": 8.140096153846155e-06,
|
248362 |
+
"loss": 0.36,
|
248363 |
+
"step": 117105
|
248364 |
+
},
|
248365 |
+
{
|
248366 |
+
"epoch": 936.08,
|
248367 |
+
"learning_rate": 8.140016025641026e-06,
|
248368 |
+
"loss": 0.2816,
|
248369 |
+
"step": 117110
|
248370 |
+
},
|
248371 |
+
{
|
248372 |
+
"epoch": 936.12,
|
248373 |
+
"learning_rate": 8.139935897435897e-06,
|
248374 |
+
"loss": 0.302,
|
248375 |
+
"step": 117115
|
248376 |
+
},
|
248377 |
+
{
|
248378 |
+
"epoch": 936.16,
|
248379 |
+
"learning_rate": 8.13985576923077e-06,
|
248380 |
+
"loss": 0.4257,
|
248381 |
+
"step": 117120
|
248382 |
+
},
|
248383 |
+
{
|
248384 |
+
"epoch": 936.2,
|
248385 |
+
"learning_rate": 8.139775641025642e-06,
|
248386 |
+
"loss": 1.0676,
|
248387 |
+
"step": 117125
|
248388 |
+
},
|
248389 |
+
{
|
248390 |
+
"epoch": 936.24,
|
248391 |
+
"learning_rate": 8.139695512820513e-06,
|
248392 |
+
"loss": 0.3769,
|
248393 |
+
"step": 117130
|
248394 |
+
},
|
248395 |
+
{
|
248396 |
+
"epoch": 936.28,
|
248397 |
+
"learning_rate": 8.139615384615386e-06,
|
248398 |
+
"loss": 0.3151,
|
248399 |
+
"step": 117135
|
248400 |
+
},
|
248401 |
+
{
|
248402 |
+
"epoch": 936.32,
|
248403 |
+
"learning_rate": 8.139535256410258e-06,
|
248404 |
+
"loss": 0.2631,
|
248405 |
+
"step": 117140
|
248406 |
+
},
|
248407 |
+
{
|
248408 |
+
"epoch": 936.36,
|
248409 |
+
"learning_rate": 8.139455128205129e-06,
|
248410 |
+
"loss": 0.397,
|
248411 |
+
"step": 117145
|
248412 |
+
},
|
248413 |
+
{
|
248414 |
+
"epoch": 936.4,
|
248415 |
+
"learning_rate": 8.139375e-06,
|
248416 |
+
"loss": 1.1359,
|
248417 |
+
"step": 117150
|
248418 |
+
},
|
248419 |
+
{
|
248420 |
+
"epoch": 936.44,
|
248421 |
+
"learning_rate": 8.139294871794873e-06,
|
248422 |
+
"loss": 0.3287,
|
248423 |
+
"step": 117155
|
248424 |
+
},
|
248425 |
+
{
|
248426 |
+
"epoch": 936.48,
|
248427 |
+
"learning_rate": 8.139214743589745e-06,
|
248428 |
+
"loss": 0.2817,
|
248429 |
+
"step": 117160
|
248430 |
+
},
|
248431 |
+
{
|
248432 |
+
"epoch": 936.52,
|
248433 |
+
"learning_rate": 8.139134615384616e-06,
|
248434 |
+
"loss": 0.3736,
|
248435 |
+
"step": 117165
|
248436 |
+
},
|
248437 |
+
{
|
248438 |
+
"epoch": 936.56,
|
248439 |
+
"learning_rate": 8.139054487179487e-06,
|
248440 |
+
"loss": 0.4126,
|
248441 |
+
"step": 117170
|
248442 |
+
},
|
248443 |
+
{
|
248444 |
+
"epoch": 936.6,
|
248445 |
+
"learning_rate": 8.13897435897436e-06,
|
248446 |
+
"loss": 1.0384,
|
248447 |
+
"step": 117175
|
248448 |
+
},
|
248449 |
+
{
|
248450 |
+
"epoch": 936.64,
|
248451 |
+
"learning_rate": 8.13889423076923e-06,
|
248452 |
+
"loss": 0.3022,
|
248453 |
+
"step": 117180
|
248454 |
+
},
|
248455 |
+
{
|
248456 |
+
"epoch": 936.68,
|
248457 |
+
"learning_rate": 8.138814102564103e-06,
|
248458 |
+
"loss": 0.2988,
|
248459 |
+
"step": 117185
|
248460 |
+
},
|
248461 |
+
{
|
248462 |
+
"epoch": 936.72,
|
248463 |
+
"learning_rate": 8.138733974358976e-06,
|
248464 |
+
"loss": 0.3346,
|
248465 |
+
"step": 117190
|
248466 |
+
},
|
248467 |
+
{
|
248468 |
+
"epoch": 936.76,
|
248469 |
+
"learning_rate": 8.138653846153846e-06,
|
248470 |
+
"loss": 0.4354,
|
248471 |
+
"step": 117195
|
248472 |
+
},
|
248473 |
+
{
|
248474 |
+
"epoch": 936.8,
|
248475 |
+
"learning_rate": 8.138573717948719e-06,
|
248476 |
+
"loss": 1.2176,
|
248477 |
+
"step": 117200
|
248478 |
+
},
|
248479 |
+
{
|
248480 |
+
"epoch": 936.84,
|
248481 |
+
"learning_rate": 8.13849358974359e-06,
|
248482 |
+
"loss": 0.367,
|
248483 |
+
"step": 117205
|
248484 |
+
},
|
248485 |
+
{
|
248486 |
+
"epoch": 936.88,
|
248487 |
+
"learning_rate": 8.138413461538462e-06,
|
248488 |
+
"loss": 0.2743,
|
248489 |
+
"step": 117210
|
248490 |
+
},
|
248491 |
+
{
|
248492 |
+
"epoch": 936.92,
|
248493 |
+
"learning_rate": 8.138333333333333e-06,
|
248494 |
+
"loss": 0.3529,
|
248495 |
+
"step": 117215
|
248496 |
+
},
|
248497 |
+
{
|
248498 |
+
"epoch": 936.96,
|
248499 |
+
"learning_rate": 8.138253205128206e-06,
|
248500 |
+
"loss": 0.4974,
|
248501 |
+
"step": 117220
|
248502 |
+
},
|
248503 |
+
{
|
248504 |
+
"epoch": 937.0,
|
248505 |
+
"learning_rate": 8.138173076923077e-06,
|
248506 |
+
"loss": 1.2262,
|
248507 |
+
"step": 117225
|
248508 |
+
},
|
248509 |
+
{
|
248510 |
+
"epoch": 937.0,
|
248511 |
+
"eval_loss": 0.382039874792099,
|
248512 |
+
"eval_runtime": 39.12,
|
248513 |
+
"eval_samples_per_second": 21.575,
|
248514 |
+
"eval_steps_per_second": 0.69,
|
248515 |
+
"eval_wer": 0.18934867596075827,
|
248516 |
+
"step": 117225
|
248517 |
+
},
|
248518 |
+
{
|
248519 |
+
"epoch": 945.04,
|
248520 |
+
"learning_rate": 8.138092948717949e-06,
|
248521 |
+
"loss": 0.3032,
|
248522 |
+
"step": 117230
|
248523 |
+
},
|
248524 |
+
{
|
248525 |
+
"epoch": 945.08,
|
248526 |
+
"learning_rate": 8.138012820512822e-06,
|
248527 |
+
"loss": 0.273,
|
248528 |
+
"step": 117235
|
248529 |
+
},
|
248530 |
+
{
|
248531 |
+
"epoch": 945.12,
|
248532 |
+
"learning_rate": 8.137932692307693e-06,
|
248533 |
+
"loss": 0.3273,
|
248534 |
+
"step": 117240
|
248535 |
+
},
|
248536 |
+
{
|
248537 |
+
"epoch": 945.16,
|
248538 |
+
"learning_rate": 8.137852564102565e-06,
|
248539 |
+
"loss": 0.4947,
|
248540 |
+
"step": 117245
|
248541 |
+
},
|
248542 |
+
{
|
248543 |
+
"epoch": 945.2,
|
248544 |
+
"learning_rate": 8.137772435897436e-06,
|
248545 |
+
"loss": 1.2386,
|
248546 |
+
"step": 117250
|
248547 |
+
},
|
248548 |
+
{
|
248549 |
+
"epoch": 945.24,
|
248550 |
+
"learning_rate": 8.137692307692309e-06,
|
248551 |
+
"loss": 0.4844,
|
248552 |
+
"step": 117255
|
248553 |
+
},
|
248554 |
+
{
|
248555 |
+
"epoch": 945.28,
|
248556 |
+
"learning_rate": 8.13761217948718e-06,
|
248557 |
+
"loss": 0.2666,
|
248558 |
+
"step": 117260
|
248559 |
+
},
|
248560 |
+
{
|
248561 |
+
"epoch": 945.32,
|
248562 |
+
"learning_rate": 8.137532051282052e-06,
|
248563 |
+
"loss": 0.3561,
|
248564 |
+
"step": 117265
|
248565 |
+
},
|
248566 |
+
{
|
248567 |
+
"epoch": 945.36,
|
248568 |
+
"learning_rate": 8.137451923076923e-06,
|
248569 |
+
"loss": 0.5483,
|
248570 |
+
"step": 117270
|
248571 |
+
},
|
248572 |
+
{
|
248573 |
+
"epoch": 945.4,
|
248574 |
+
"learning_rate": 8.137371794871796e-06,
|
248575 |
+
"loss": 1.1604,
|
248576 |
+
"step": 117275
|
248577 |
+
},
|
248578 |
+
{
|
248579 |
+
"epoch": 945.44,
|
248580 |
+
"learning_rate": 8.137291666666667e-06,
|
248581 |
+
"loss": 0.4208,
|
248582 |
+
"step": 117280
|
248583 |
+
},
|
248584 |
+
{
|
248585 |
+
"epoch": 945.48,
|
248586 |
+
"learning_rate": 8.137211538461539e-06,
|
248587 |
+
"loss": 0.3461,
|
248588 |
+
"step": 117285
|
248589 |
+
},
|
248590 |
+
{
|
248591 |
+
"epoch": 945.52,
|
248592 |
+
"learning_rate": 8.137131410256412e-06,
|
248593 |
+
"loss": 0.3296,
|
248594 |
+
"step": 117290
|
248595 |
+
},
|
248596 |
+
{
|
248597 |
+
"epoch": 945.56,
|
248598 |
+
"learning_rate": 8.137051282051283e-06,
|
248599 |
+
"loss": 0.4776,
|
248600 |
+
"step": 117295
|
248601 |
+
},
|
248602 |
+
{
|
248603 |
+
"epoch": 945.6,
|
248604 |
+
"learning_rate": 8.136971153846155e-06,
|
248605 |
+
"loss": 1.2078,
|
248606 |
+
"step": 117300
|
248607 |
+
},
|
248608 |
+
{
|
248609 |
+
"epoch": 945.64,
|
248610 |
+
"learning_rate": 8.136891025641026e-06,
|
248611 |
+
"loss": 0.2939,
|
248612 |
+
"step": 117305
|
248613 |
+
},
|
248614 |
+
{
|
248615 |
+
"epoch": 945.68,
|
248616 |
+
"learning_rate": 8.136810897435899e-06,
|
248617 |
+
"loss": 0.3025,
|
248618 |
+
"step": 117310
|
248619 |
+
},
|
248620 |
+
{
|
248621 |
+
"epoch": 945.72,
|
248622 |
+
"learning_rate": 8.136730769230769e-06,
|
248623 |
+
"loss": 0.2963,
|
248624 |
+
"step": 117315
|
248625 |
+
},
|
248626 |
+
{
|
248627 |
+
"epoch": 945.76,
|
248628 |
+
"learning_rate": 8.136650641025642e-06,
|
248629 |
+
"loss": 0.474,
|
248630 |
+
"step": 117320
|
248631 |
+
},
|
248632 |
+
{
|
248633 |
+
"epoch": 945.8,
|
248634 |
+
"learning_rate": 8.136570512820513e-06,
|
248635 |
+
"loss": 1.1468,
|
248636 |
+
"step": 117325
|
248637 |
+
},
|
248638 |
+
{
|
248639 |
+
"epoch": 945.84,
|
248640 |
+
"learning_rate": 8.136490384615386e-06,
|
248641 |
+
"loss": 0.3025,
|
248642 |
+
"step": 117330
|
248643 |
+
},
|
248644 |
+
{
|
248645 |
+
"epoch": 945.88,
|
248646 |
+
"learning_rate": 8.136410256410258e-06,
|
248647 |
+
"loss": 0.2942,
|
248648 |
+
"step": 117335
|
248649 |
+
},
|
248650 |
+
{
|
248651 |
+
"epoch": 945.92,
|
248652 |
+
"learning_rate": 8.136330128205129e-06,
|
248653 |
+
"loss": 0.3408,
|
248654 |
+
"step": 117340
|
248655 |
+
},
|
248656 |
+
{
|
248657 |
+
"epoch": 945.96,
|
248658 |
+
"learning_rate": 8.136250000000002e-06,
|
248659 |
+
"loss": 0.5317,
|
248660 |
+
"step": 117345
|
248661 |
+
},
|
248662 |
+
{
|
248663 |
+
"epoch": 946.0,
|
248664 |
+
"eval_loss": 0.4116620719432831,
|
248665 |
+
"eval_runtime": 49.1871,
|
248666 |
+
"eval_samples_per_second": 17.139,
|
248667 |
+
"eval_steps_per_second": 0.549,
|
248668 |
+
"eval_wer": 0.1847012186169432,
|
248669 |
+
"step": 117349
|
248670 |
+
},
|
248671 |
+
{
|
248672 |
+
"epoch": 946.01,
|
248673 |
+
"learning_rate": 8.136169871794872e-06,
|
248674 |
+
"loss": 0.4503,
|
248675 |
+
"step": 117350
|
248676 |
+
},
|
248677 |
+
{
|
248678 |
+
"epoch": 946.05,
|
248679 |
+
"learning_rate": 8.136089743589745e-06,
|
248680 |
+
"loss": 0.3793,
|
248681 |
+
"step": 117355
|
248682 |
+
},
|
248683 |
+
{
|
248684 |
+
"epoch": 946.09,
|
248685 |
+
"learning_rate": 8.136009615384616e-06,
|
248686 |
+
"loss": 0.2987,
|
248687 |
+
"step": 117360
|
248688 |
+
},
|
248689 |
+
{
|
248690 |
+
"epoch": 946.13,
|
248691 |
+
"learning_rate": 8.135929487179487e-06,
|
248692 |
+
"loss": 0.3598,
|
248693 |
+
"step": 117365
|
248694 |
+
},
|
248695 |
+
{
|
248696 |
+
"epoch": 946.17,
|
248697 |
+
"learning_rate": 8.135849358974359e-06,
|
248698 |
+
"loss": 0.6339,
|
248699 |
+
"step": 117370
|
248700 |
+
},
|
248701 |
+
{
|
248702 |
+
"epoch": 946.21,
|
248703 |
+
"learning_rate": 8.135769230769232e-06,
|
248704 |
+
"loss": 1.1798,
|
248705 |
+
"step": 117375
|
248706 |
+
},
|
248707 |
+
{
|
248708 |
+
"epoch": 946.25,
|
248709 |
+
"learning_rate": 8.135689102564103e-06,
|
248710 |
+
"loss": 0.3248,
|
248711 |
+
"step": 117380
|
248712 |
+
},
|
248713 |
+
{
|
248714 |
+
"epoch": 946.29,
|
248715 |
+
"learning_rate": 8.135608974358974e-06,
|
248716 |
+
"loss": 0.4143,
|
248717 |
+
"step": 117385
|
248718 |
+
},
|
248719 |
+
{
|
248720 |
+
"epoch": 946.33,
|
248721 |
+
"learning_rate": 8.135528846153848e-06,
|
248722 |
+
"loss": 0.3344,
|
248723 |
+
"step": 117390
|
248724 |
+
},
|
248725 |
+
{
|
248726 |
+
"epoch": 946.37,
|
248727 |
+
"learning_rate": 8.135448717948719e-06,
|
248728 |
+
"loss": 0.515,
|
248729 |
+
"step": 117395
|
248730 |
+
},
|
248731 |
+
{
|
248732 |
+
"epoch": 946.41,
|
248733 |
+
"learning_rate": 8.13536858974359e-06,
|
248734 |
+
"loss": 1.1973,
|
248735 |
+
"step": 117400
|
248736 |
+
},
|
248737 |
+
{
|
248738 |
+
"epoch": 946.45,
|
248739 |
+
"learning_rate": 8.135288461538462e-06,
|
248740 |
+
"loss": 0.2734,
|
248741 |
+
"step": 117405
|
248742 |
+
},
|
248743 |
+
{
|
248744 |
+
"epoch": 946.49,
|
248745 |
+
"learning_rate": 8.135208333333335e-06,
|
248746 |
+
"loss": 0.2632,
|
248747 |
+
"step": 117410
|
248748 |
+
},
|
248749 |
+
{
|
248750 |
+
"epoch": 946.53,
|
248751 |
+
"learning_rate": 8.135128205128206e-06,
|
248752 |
+
"loss": 0.3458,
|
248753 |
+
"step": 117415
|
248754 |
+
},
|
248755 |
+
{
|
248756 |
+
"epoch": 946.57,
|
248757 |
+
"learning_rate": 8.135048076923077e-06,
|
248758 |
+
"loss": 0.5121,
|
248759 |
+
"step": 117420
|
248760 |
+
},
|
248761 |
+
{
|
248762 |
+
"epoch": 946.61,
|
248763 |
+
"learning_rate": 8.134967948717949e-06,
|
248764 |
+
"loss": 1.1818,
|
248765 |
+
"step": 117425
|
248766 |
+
},
|
248767 |
+
{
|
248768 |
+
"epoch": 946.65,
|
248769 |
+
"learning_rate": 8.134887820512822e-06,
|
248770 |
+
"loss": 0.3381,
|
248771 |
+
"step": 117430
|
248772 |
+
},
|
248773 |
+
{
|
248774 |
+
"epoch": 946.69,
|
248775 |
+
"learning_rate": 8.134807692307693e-06,
|
248776 |
+
"loss": 0.2735,
|
248777 |
+
"step": 117435
|
248778 |
+
},
|
248779 |
+
{
|
248780 |
+
"epoch": 946.73,
|
248781 |
+
"learning_rate": 8.134727564102565e-06,
|
248782 |
+
"loss": 0.3295,
|
248783 |
+
"step": 117440
|
248784 |
+
},
|
248785 |
+
{
|
248786 |
+
"epoch": 946.77,
|
248787 |
+
"learning_rate": 8.134647435897438e-06,
|
248788 |
+
"loss": 0.6012,
|
248789 |
+
"step": 117445
|
248790 |
+
},
|
248791 |
+
{
|
248792 |
+
"epoch": 946.81,
|
248793 |
+
"learning_rate": 8.134567307692309e-06,
|
248794 |
+
"loss": 1.0467,
|
248795 |
+
"step": 117450
|
248796 |
+
},
|
248797 |
+
{
|
248798 |
+
"epoch": 946.85,
|
248799 |
+
"learning_rate": 8.13448717948718e-06,
|
248800 |
+
"loss": 0.2941,
|
248801 |
+
"step": 117455
|
248802 |
+
},
|
248803 |
+
{
|
248804 |
+
"epoch": 946.89,
|
248805 |
+
"learning_rate": 8.134407051282052e-06,
|
248806 |
+
"loss": 0.488,
|
248807 |
+
"step": 117460
|
248808 |
+
},
|
248809 |
+
{
|
248810 |
+
"epoch": 946.93,
|
248811 |
+
"learning_rate": 8.134326923076925e-06,
|
248812 |
+
"loss": 0.3663,
|
248813 |
+
"step": 117465
|
248814 |
+
},
|
248815 |
+
{
|
248816 |
+
"epoch": 946.97,
|
248817 |
+
"learning_rate": 8.134246794871794e-06,
|
248818 |
+
"loss": 0.5698,
|
248819 |
+
"step": 117470
|
248820 |
+
},
|
248821 |
+
{
|
248822 |
+
"epoch": 947.0,
|
248823 |
+
"eval_loss": 0.3862200081348419,
|
248824 |
+
"eval_runtime": 39.1121,
|
248825 |
+
"eval_samples_per_second": 21.477,
|
248826 |
+
"eval_steps_per_second": 0.69,
|
248827 |
+
"eval_wer": 0.1748963711730056,
|
248828 |
+
"step": 117473
|
248829 |
+
},
|
248830 |
+
{
|
248831 |
+
"epoch": 947.02,
|
248832 |
+
"learning_rate": 8.134166666666667e-06,
|
248833 |
+
"loss": 0.3545,
|
248834 |
+
"step": 117475
|
248835 |
+
},
|
248836 |
+
{
|
248837 |
+
"epoch": 947.06,
|
248838 |
+
"learning_rate": 8.134086538461539e-06,
|
248839 |
+
"loss": 0.2982,
|
248840 |
+
"step": 117480
|
248841 |
+
},
|
248842 |
+
{
|
248843 |
+
"epoch": 947.1,
|
248844 |
+
"learning_rate": 8.13400641025641e-06,
|
248845 |
+
"loss": 0.2839,
|
248846 |
+
"step": 117485
|
248847 |
+
},
|
248848 |
+
{
|
248849 |
+
"epoch": 947.14,
|
248850 |
+
"learning_rate": 8.133926282051283e-06,
|
248851 |
+
"loss": 0.382,
|
248852 |
+
"step": 117490
|
248853 |
+
},
|
248854 |
+
{
|
248855 |
+
"epoch": 947.18,
|
248856 |
+
"learning_rate": 8.133846153846155e-06,
|
248857 |
+
"loss": 0.9134,
|
248858 |
+
"step": 117495
|
248859 |
+
},
|
248860 |
+
{
|
248861 |
+
"epoch": 947.22,
|
248862 |
+
"learning_rate": 8.133766025641026e-06,
|
248863 |
+
"loss": 0.9888,
|
248864 |
+
"step": 117500
|
248865 |
+
},
|
248866 |
+
{
|
248867 |
+
"epoch": 947.26,
|
248868 |
+
"learning_rate": 8.133685897435897e-06,
|
248869 |
+
"loss": 0.3261,
|
248870 |
+
"step": 117505
|
248871 |
+
},
|
248872 |
+
{
|
248873 |
+
"epoch": 947.3,
|
248874 |
+
"learning_rate": 8.13360576923077e-06,
|
248875 |
+
"loss": 0.2684,
|
248876 |
+
"step": 117510
|
248877 |
+
},
|
248878 |
+
{
|
248879 |
+
"epoch": 947.34,
|
248880 |
+
"learning_rate": 8.133525641025642e-06,
|
248881 |
+
"loss": 0.3223,
|
248882 |
+
"step": 117515
|
248883 |
+
},
|
248884 |
+
{
|
248885 |
+
"epoch": 947.38,
|
248886 |
+
"learning_rate": 8.133445512820513e-06,
|
248887 |
+
"loss": 0.534,
|
248888 |
+
"step": 117520
|
248889 |
+
},
|
248890 |
+
{
|
248891 |
+
"epoch": 947.42,
|
248892 |
+
"learning_rate": 8.133365384615384e-06,
|
248893 |
+
"loss": 1.0773,
|
248894 |
+
"step": 117525
|
248895 |
+
},
|
248896 |
+
{
|
248897 |
+
"epoch": 947.46,
|
248898 |
+
"learning_rate": 8.133285256410257e-06,
|
248899 |
+
"loss": 0.2496,
|
248900 |
+
"step": 117530
|
248901 |
+
},
|
248902 |
+
{
|
248903 |
+
"epoch": 947.5,
|
248904 |
+
"learning_rate": 8.133205128205129e-06,
|
248905 |
+
"loss": 0.3469,
|
248906 |
+
"step": 117535
|
248907 |
+
},
|
248908 |
+
{
|
248909 |
+
"epoch": 947.54,
|
248910 |
+
"learning_rate": 8.133125e-06,
|
248911 |
+
"loss": 0.3361,
|
248912 |
+
"step": 117540
|
248913 |
+
},
|
248914 |
+
{
|
248915 |
+
"epoch": 947.58,
|
248916 |
+
"learning_rate": 8.133044871794873e-06,
|
248917 |
+
"loss": 0.7251,
|
248918 |
+
"step": 117545
|
248919 |
+
},
|
248920 |
+
{
|
248921 |
+
"epoch": 947.62,
|
248922 |
+
"learning_rate": 8.132964743589745e-06,
|
248923 |
+
"loss": 1.0967,
|
248924 |
+
"step": 117550
|
248925 |
+
},
|
248926 |
+
{
|
248927 |
+
"epoch": 947.66,
|
248928 |
+
"learning_rate": 8.132884615384616e-06,
|
248929 |
+
"loss": 0.3342,
|
248930 |
+
"step": 117555
|
248931 |
+
},
|
248932 |
+
{
|
248933 |
+
"epoch": 947.7,
|
248934 |
+
"learning_rate": 8.132804487179487e-06,
|
248935 |
+
"loss": 0.2923,
|
248936 |
+
"step": 117560
|
248937 |
+
},
|
248938 |
+
{
|
248939 |
+
"epoch": 947.74,
|
248940 |
+
"learning_rate": 8.13272435897436e-06,
|
248941 |
+
"loss": 0.3418,
|
248942 |
+
"step": 117565
|
248943 |
+
},
|
248944 |
+
{
|
248945 |
+
"epoch": 947.78,
|
248946 |
+
"learning_rate": 8.132644230769232e-06,
|
248947 |
+
"loss": 0.6134,
|
248948 |
+
"step": 117570
|
248949 |
+
},
|
248950 |
+
{
|
248951 |
+
"epoch": 947.82,
|
248952 |
+
"learning_rate": 8.132564102564103e-06,
|
248953 |
+
"loss": 0.9037,
|
248954 |
+
"step": 117575
|
248955 |
+
},
|
248956 |
+
{
|
248957 |
+
"epoch": 947.86,
|
248958 |
+
"learning_rate": 8.132483974358974e-06,
|
248959 |
+
"loss": 0.3372,
|
248960 |
+
"step": 117580
|
248961 |
+
},
|
248962 |
+
{
|
248963 |
+
"epoch": 947.9,
|
248964 |
+
"learning_rate": 8.132403846153847e-06,
|
248965 |
+
"loss": 0.3168,
|
248966 |
+
"step": 117585
|
248967 |
+
},
|
248968 |
+
{
|
248969 |
+
"epoch": 947.94,
|
248970 |
+
"learning_rate": 8.132323717948719e-06,
|
248971 |
+
"loss": 0.3129,
|
248972 |
+
"step": 117590
|
248973 |
+
},
|
248974 |
+
{
|
248975 |
+
"epoch": 947.98,
|
248976 |
+
"learning_rate": 8.13224358974359e-06,
|
248977 |
+
"loss": 0.827,
|
248978 |
+
"step": 117595
|
248979 |
+
},
|
248980 |
+
{
|
248981 |
+
"epoch": 948.0,
|
248982 |
+
"eval_loss": 0.43474265933036804,
|
248983 |
+
"eval_runtime": 38.7119,
|
248984 |
+
"eval_samples_per_second": 21.673,
|
248985 |
+
"eval_steps_per_second": 0.697,
|
248986 |
+
"eval_wer": 0.19487402889711755,
|
248987 |
+
"step": 117597
|
248988 |
+
},
|
248989 |
+
{
|
248990 |
+
"epoch": 940.02,
|
248991 |
+
"learning_rate": 8.132163461538463e-06,
|
248992 |
+
"loss": 0.3282,
|
248993 |
+
"step": 117600
|
248994 |
+
},
|
248995 |
+
{
|
248996 |
+
"epoch": 940.06,
|
248997 |
+
"learning_rate": 8.132083333333335e-06,
|
248998 |
+
"loss": 0.3094,
|
248999 |
+
"step": 117605
|
249000 |
+
},
|
249001 |
+
{
|
249002 |
+
"epoch": 940.1,
|
249003 |
+
"learning_rate": 8.132003205128206e-06,
|
249004 |
+
"loss": 0.2889,
|
249005 |
+
"step": 117610
|
249006 |
+
},
|
249007 |
+
{
|
249008 |
+
"epoch": 940.14,
|
249009 |
+
"learning_rate": 8.131923076923077e-06,
|
249010 |
+
"loss": 0.3158,
|
249011 |
+
"step": 117615
|
249012 |
+
},
|
249013 |
+
{
|
249014 |
+
"epoch": 940.18,
|
249015 |
+
"learning_rate": 8.13184294871795e-06,
|
249016 |
+
"loss": 0.7108,
|
249017 |
+
"step": 117620
|
249018 |
+
},
|
249019 |
+
{
|
249020 |
+
"epoch": 940.22,
|
249021 |
+
"learning_rate": 8.13176282051282e-06,
|
249022 |
+
"loss": 0.8738,
|
249023 |
+
"step": 117625
|
249024 |
+
},
|
249025 |
+
{
|
249026 |
+
"epoch": 940.26,
|
249027 |
+
"learning_rate": 8.131682692307693e-06,
|
249028 |
+
"loss": 0.3388,
|
249029 |
+
"step": 117630
|
249030 |
+
},
|
249031 |
+
{
|
249032 |
+
"epoch": 940.3,
|
249033 |
+
"learning_rate": 8.131602564102566e-06,
|
249034 |
+
"loss": 0.2985,
|
249035 |
+
"step": 117635
|
249036 |
+
},
|
249037 |
+
{
|
249038 |
+
"epoch": 940.34,
|
249039 |
+
"learning_rate": 8.131522435897436e-06,
|
249040 |
+
"loss": 0.4881,
|
249041 |
+
"step": 117640
|
249042 |
+
},
|
249043 |
+
{
|
249044 |
+
"epoch": 940.38,
|
249045 |
+
"learning_rate": 8.131442307692309e-06,
|
249046 |
+
"loss": 0.7521,
|
249047 |
+
"step": 117645
|
249048 |
+
},
|
249049 |
+
{
|
249050 |
+
"epoch": 940.42,
|
249051 |
+
"learning_rate": 8.13136217948718e-06,
|
249052 |
+
"loss": 1.0756,
|
249053 |
+
"step": 117650
|
249054 |
+
},
|
249055 |
+
{
|
249056 |
+
"epoch": 940.46,
|
249057 |
+
"learning_rate": 8.131282051282052e-06,
|
249058 |
+
"loss": 0.2887,
|
249059 |
+
"step": 117655
|
249060 |
+
},
|
249061 |
+
{
|
249062 |
+
"epoch": 940.5,
|
249063 |
+
"learning_rate": 8.131201923076923e-06,
|
249064 |
+
"loss": 0.3438,
|
249065 |
+
"step": 117660
|
249066 |
+
},
|
249067 |
+
{
|
249068 |
+
"epoch": 940.54,
|
249069 |
+
"learning_rate": 8.131121794871796e-06,
|
249070 |
+
"loss": 0.3598,
|
249071 |
+
"step": 117665
|
249072 |
+
},
|
249073 |
+
{
|
249074 |
+
"epoch": 940.58,
|
249075 |
+
"learning_rate": 8.131041666666667e-06,
|
249076 |
+
"loss": 0.7274,
|
249077 |
+
"step": 117670
|
249078 |
+
},
|
249079 |
+
{
|
249080 |
+
"epoch": 940.62,
|
249081 |
+
"learning_rate": 8.130961538461539e-06,
|
249082 |
+
"loss": 0.876,
|
249083 |
+
"step": 117675
|
249084 |
+
},
|
249085 |
+
{
|
249086 |
+
"epoch": 940.66,
|
249087 |
+
"learning_rate": 8.13088141025641e-06,
|
249088 |
+
"loss": 0.2849,
|
249089 |
+
"step": 117680
|
249090 |
+
},
|
249091 |
+
{
|
249092 |
+
"epoch": 940.7,
|
249093 |
+
"learning_rate": 8.130801282051283e-06,
|
249094 |
+
"loss": 0.2736,
|
249095 |
+
"step": 117685
|
249096 |
+
},
|
249097 |
+
{
|
249098 |
+
"epoch": 940.74,
|
249099 |
+
"learning_rate": 8.130721153846154e-06,
|
249100 |
+
"loss": 0.3746,
|
249101 |
+
"step": 117690
|
249102 |
+
},
|
249103 |
+
{
|
249104 |
+
"epoch": 940.78,
|
249105 |
+
"learning_rate": 8.130641025641026e-06,
|
249106 |
+
"loss": 0.737,
|
249107 |
+
"step": 117695
|
249108 |
+
},
|
249109 |
+
{
|
249110 |
+
"epoch": 940.82,
|
249111 |
+
"learning_rate": 8.130560897435899e-06,
|
249112 |
+
"loss": 0.8436,
|
249113 |
+
"step": 117700
|
249114 |
+
},
|
249115 |
+
{
|
249116 |
+
"epoch": 940.86,
|
249117 |
+
"learning_rate": 8.13048076923077e-06,
|
249118 |
+
"loss": 0.2284,
|
249119 |
+
"step": 117705
|
249120 |
+
},
|
249121 |
+
{
|
249122 |
+
"epoch": 940.9,
|
249123 |
+
"learning_rate": 8.130400641025642e-06,
|
249124 |
+
"loss": 0.3026,
|
249125 |
+
"step": 117710
|
249126 |
+
},
|
249127 |
+
{
|
249128 |
+
"epoch": 940.94,
|
249129 |
+
"learning_rate": 8.130320512820513e-06,
|
249130 |
+
"loss": 0.426,
|
249131 |
+
"step": 117715
|
249132 |
+
},
|
249133 |
+
{
|
249134 |
+
"epoch": 940.98,
|
249135 |
+
"learning_rate": 8.130240384615386e-06,
|
249136 |
+
"loss": 0.7146,
|
249137 |
+
"step": 117720
|
249138 |
+
},
|
249139 |
+
{
|
249140 |
+
"epoch": 941.0,
|
249141 |
+
"eval_loss": 0.34532302618026733,
|
249142 |
+
"eval_runtime": 38.5318,
|
249143 |
+
"eval_samples_per_second": 21.774,
|
249144 |
+
"eval_steps_per_second": 0.701,
|
249145 |
+
"eval_wer": 0.18467364669191358,
|
249146 |
+
"step": 117722
|
249147 |
}
|
249148 |
],
|
249149 |
"max_steps": 625000,
|
249150 |
"num_train_epochs": 5000,
|
249151 |
+
"total_flos": 3.313008475394062e+20,
|
249152 |
"trial_name": null,
|
249153 |
"trial_params": null
|
249154 |
}
|
model-bin/finetune/base/{checkpoint-117100 β checkpoint-117722}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630117908.6872168/events.out.tfevents.1630117908.86bb0ddabf9b.1042.51
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a1a623804e4583586cde6273c613a8e5ca8fe0b77e9947bcc881c674bdfb457
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630118323.7473285/events.out.tfevents.1630118323.86bb0ddabf9b.1042.53
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f3bfba1097373b7feebc5b84cad11f506237fe60f000442bd3df78e16a4d2bc
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630118841.0029836/events.out.tfevents.1630118841.86bb0ddabf9b.1042.55
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74fe57cdf6563d18fe657f6b941c44795dcb541056449b18ebd5b5b124eb2665
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630119249.7913108/events.out.tfevents.1630119249.86bb0ddabf9b.1042.57
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4984439e0ef86c761adbbf1f111486bf5d33a618505376836a6244a0788d262e
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630119657.5811403/events.out.tfevents.1630119657.86bb0ddabf9b.1042.59
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f47351033db2b4e29e30eae5441bbc8da255b1f2602bd8952e9547a59c46ff48
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630117908.86bb0ddabf9b.1042.50
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5637e932748456917dc2a740db74ad845315b0eee9eacfb686a35cd15bbc1ccf
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630118323.86bb0ddabf9b.1042.52
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:48434c5e9458dce1d2b1b175ad63f08018d916ddd57dfcfe72f568f9b60528f2
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630118840.86bb0ddabf9b.1042.54
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c4728b4ef5f812daf8e8b9549659f7eb9b5cf1aae46a768dd0ba123bdd6956ee
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630119249.86bb0ddabf9b.1042.56
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4bb9f8eb0abfa4eca6066b388a060ff81f509bc71bc76654c434e2abfe87d079
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630119657.86bb0ddabf9b.1042.58
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3afeafde1841e315916058445ba3d2199ce216a28a5e292d148f36f9b76d7f48
|
3 |
+
size 8622
|