"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630255201.5100508/events.out.tfevents.1630255201.cc93b136ebf5.1086.229 +3 -0
- model-bin/finetune/base/log/1630255633.58935/events.out.tfevents.1630255633.cc93b136ebf5.1086.231 +3 -0
- model-bin/finetune/base/log/1630256177.9503975/events.out.tfevents.1630256177.cc93b136ebf5.1086.233 +3 -0
- model-bin/finetune/base/log/1630256616.2655816/events.out.tfevents.1630256616.cc93b136ebf5.1086.235 +3 -0
- model-bin/finetune/base/log/1630257046.964922/events.out.tfevents.1630257046.cc93b136ebf5.1086.237 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630255201.cc93b136ebf5.1086.228 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630255633.cc93b136ebf5.1086.230 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630256177.cc93b136ebf5.1086.232 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630256616.cc93b136ebf5.1086.234 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630257046.cc93b136ebf5.1086.236 +3 -0
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6c4e2f164fe612c9447b7fdd6a75bd151c34385c5157ec3091d0d894af2e00d
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd584b4168381271c31720a7885b9ee125cb30ba21f1597f757e00db3049a185
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f6ff2f68c3c8c25a6448582f63dfac8cbe057a7cbc620566ec94b3b6c1eb228
|
3 |
+
size 14439
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39306f1d255c1140e0b97c10ee337a2e56b425eb5149e6f8db137bc9f9496e80
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2facb014c0362e2448c4cd83ad0a2f81ed3eca6c82dcc98beab730981e6e8e16
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -294426,11 +294426,800 @@
|
|
294426 |
"eval_steps_per_second": 0.649,
|
294427 |
"eval_wer": 0.18537154524903376,
|
294428 |
"step": 153312
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
294429 |
}
|
294430 |
],
|
294431 |
-
"max_steps":
|
294432 |
"num_train_epochs": 5000,
|
294433 |
-
"total_flos": 4.
|
294434 |
"trial_name": null,
|
294435 |
"trial_params": null
|
294436 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1231.0,
|
5 |
+
"global_step": 153934,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
294426 |
"eval_steps_per_second": 0.649,
|
294427 |
"eval_wer": 0.18537154524903376,
|
294428 |
"step": 153312
|
294429 |
+
},
|
294430 |
+
{
|
294431 |
+
"epoch": 1236.02,
|
294432 |
+
"learning_rate": 7.5403069466882075e-06,
|
294433 |
+
"loss": 0.3773,
|
294434 |
+
"step": 153315
|
294435 |
+
},
|
294436 |
+
{
|
294437 |
+
"epoch": 1236.06,
|
294438 |
+
"learning_rate": 7.5402261712439425e-06,
|
294439 |
+
"loss": 0.3049,
|
294440 |
+
"step": 153320
|
294441 |
+
},
|
294442 |
+
{
|
294443 |
+
"epoch": 1236.1,
|
294444 |
+
"learning_rate": 7.5401453957996775e-06,
|
294445 |
+
"loss": 0.2696,
|
294446 |
+
"step": 153325
|
294447 |
+
},
|
294448 |
+
{
|
294449 |
+
"epoch": 1236.14,
|
294450 |
+
"learning_rate": 7.5400646203554125e-06,
|
294451 |
+
"loss": 0.3384,
|
294452 |
+
"step": 153330
|
294453 |
+
},
|
294454 |
+
{
|
294455 |
+
"epoch": 1236.18,
|
294456 |
+
"learning_rate": 7.5399838449111475e-06,
|
294457 |
+
"loss": 0.7549,
|
294458 |
+
"step": 153335
|
294459 |
+
},
|
294460 |
+
{
|
294461 |
+
"epoch": 1236.22,
|
294462 |
+
"learning_rate": 7.5399030694668825e-06,
|
294463 |
+
"loss": 0.9564,
|
294464 |
+
"step": 153340
|
294465 |
+
},
|
294466 |
+
{
|
294467 |
+
"epoch": 1236.27,
|
294468 |
+
"learning_rate": 7.5398222940226175e-06,
|
294469 |
+
"loss": 0.29,
|
294470 |
+
"step": 153345
|
294471 |
+
},
|
294472 |
+
{
|
294473 |
+
"epoch": 1236.31,
|
294474 |
+
"learning_rate": 7.539741518578353e-06,
|
294475 |
+
"loss": 0.2812,
|
294476 |
+
"step": 153350
|
294477 |
+
},
|
294478 |
+
{
|
294479 |
+
"epoch": 1236.35,
|
294480 |
+
"learning_rate": 7.5396607431340874e-06,
|
294481 |
+
"loss": 0.3392,
|
294482 |
+
"step": 153355
|
294483 |
+
},
|
294484 |
+
{
|
294485 |
+
"epoch": 1236.39,
|
294486 |
+
"learning_rate": 7.539579967689823e-06,
|
294487 |
+
"loss": 0.7306,
|
294488 |
+
"step": 153360
|
294489 |
+
},
|
294490 |
+
{
|
294491 |
+
"epoch": 1236.43,
|
294492 |
+
"learning_rate": 7.539499192245558e-06,
|
294493 |
+
"loss": 0.9617,
|
294494 |
+
"step": 153365
|
294495 |
+
},
|
294496 |
+
{
|
294497 |
+
"epoch": 1236.47,
|
294498 |
+
"learning_rate": 7.539418416801293e-06,
|
294499 |
+
"loss": 0.2778,
|
294500 |
+
"step": 153370
|
294501 |
+
},
|
294502 |
+
{
|
294503 |
+
"epoch": 1236.51,
|
294504 |
+
"learning_rate": 7.539337641357028e-06,
|
294505 |
+
"loss": 0.2961,
|
294506 |
+
"step": 153375
|
294507 |
+
},
|
294508 |
+
{
|
294509 |
+
"epoch": 1236.55,
|
294510 |
+
"learning_rate": 7.539256865912763e-06,
|
294511 |
+
"loss": 0.3302,
|
294512 |
+
"step": 153380
|
294513 |
+
},
|
294514 |
+
{
|
294515 |
+
"epoch": 1236.59,
|
294516 |
+
"learning_rate": 7.539176090468498e-06,
|
294517 |
+
"loss": 0.683,
|
294518 |
+
"step": 153385
|
294519 |
+
},
|
294520 |
+
{
|
294521 |
+
"epoch": 1236.63,
|
294522 |
+
"learning_rate": 7.539095315024233e-06,
|
294523 |
+
"loss": 0.9545,
|
294524 |
+
"step": 153390
|
294525 |
+
},
|
294526 |
+
{
|
294527 |
+
"epoch": 1236.67,
|
294528 |
+
"learning_rate": 7.539014539579968e-06,
|
294529 |
+
"loss": 0.3683,
|
294530 |
+
"step": 153395
|
294531 |
+
},
|
294532 |
+
{
|
294533 |
+
"epoch": 1236.71,
|
294534 |
+
"learning_rate": 7.538933764135703e-06,
|
294535 |
+
"loss": 0.2904,
|
294536 |
+
"step": 153400
|
294537 |
+
},
|
294538 |
+
{
|
294539 |
+
"epoch": 1236.75,
|
294540 |
+
"learning_rate": 7.538852988691438e-06,
|
294541 |
+
"loss": 0.29,
|
294542 |
+
"step": 153405
|
294543 |
+
},
|
294544 |
+
{
|
294545 |
+
"epoch": 1236.79,
|
294546 |
+
"learning_rate": 7.538772213247173e-06,
|
294547 |
+
"loss": 0.745,
|
294548 |
+
"step": 153410
|
294549 |
+
},
|
294550 |
+
{
|
294551 |
+
"epoch": 1236.83,
|
294552 |
+
"learning_rate": 7.538691437802909e-06,
|
294553 |
+
"loss": 0.9307,
|
294554 |
+
"step": 153415
|
294555 |
+
},
|
294556 |
+
{
|
294557 |
+
"epoch": 1236.87,
|
294558 |
+
"learning_rate": 7.538610662358644e-06,
|
294559 |
+
"loss": 0.3326,
|
294560 |
+
"step": 153420
|
294561 |
+
},
|
294562 |
+
{
|
294563 |
+
"epoch": 1236.91,
|
294564 |
+
"learning_rate": 7.538529886914379e-06,
|
294565 |
+
"loss": 0.268,
|
294566 |
+
"step": 153425
|
294567 |
+
},
|
294568 |
+
{
|
294569 |
+
"epoch": 1236.95,
|
294570 |
+
"learning_rate": 7.538449111470114e-06,
|
294571 |
+
"loss": 0.4247,
|
294572 |
+
"step": 153430
|
294573 |
+
},
|
294574 |
+
{
|
294575 |
+
"epoch": 1236.99,
|
294576 |
+
"learning_rate": 7.538368336025849e-06,
|
294577 |
+
"loss": 0.8271,
|
294578 |
+
"step": 153435
|
294579 |
+
},
|
294580 |
+
{
|
294581 |
+
"epoch": 1237.0,
|
294582 |
+
"eval_loss": 0.3603706955909729,
|
294583 |
+
"eval_runtime": 42.9904,
|
294584 |
+
"eval_samples_per_second": 19.446,
|
294585 |
+
"eval_steps_per_second": 0.628,
|
294586 |
+
"eval_wer": 0.1769643109042178,
|
294587 |
+
"step": 153436
|
294588 |
+
},
|
294589 |
+
{
|
294590 |
+
"epoch": 1237.03,
|
294591 |
+
"learning_rate": 7.538287560581584e-06,
|
294592 |
+
"loss": 0.3607,
|
294593 |
+
"step": 153440
|
294594 |
+
},
|
294595 |
+
{
|
294596 |
+
"epoch": 1237.07,
|
294597 |
+
"learning_rate": 7.538206785137319e-06,
|
294598 |
+
"loss": 0.2871,
|
294599 |
+
"step": 153445
|
294600 |
+
},
|
294601 |
+
{
|
294602 |
+
"epoch": 1237.11,
|
294603 |
+
"learning_rate": 7.538126009693054e-06,
|
294604 |
+
"loss": 0.266,
|
294605 |
+
"step": 153450
|
294606 |
+
},
|
294607 |
+
{
|
294608 |
+
"epoch": 1237.15,
|
294609 |
+
"learning_rate": 7.538045234248789e-06,
|
294610 |
+
"loss": 0.375,
|
294611 |
+
"step": 153455
|
294612 |
+
},
|
294613 |
+
{
|
294614 |
+
"epoch": 1237.19,
|
294615 |
+
"learning_rate": 7.537964458804524e-06,
|
294616 |
+
"loss": 0.835,
|
294617 |
+
"step": 153460
|
294618 |
+
},
|
294619 |
+
{
|
294620 |
+
"epoch": 1237.23,
|
294621 |
+
"learning_rate": 7.537883683360259e-06,
|
294622 |
+
"loss": 0.6919,
|
294623 |
+
"step": 153465
|
294624 |
+
},
|
294625 |
+
{
|
294626 |
+
"epoch": 1237.27,
|
294627 |
+
"learning_rate": 7.537802907915994e-06,
|
294628 |
+
"loss": 0.29,
|
294629 |
+
"step": 153470
|
294630 |
+
},
|
294631 |
+
{
|
294632 |
+
"epoch": 1237.31,
|
294633 |
+
"learning_rate": 7.537722132471729e-06,
|
294634 |
+
"loss": 0.2965,
|
294635 |
+
"step": 153475
|
294636 |
+
},
|
294637 |
+
{
|
294638 |
+
"epoch": 1237.35,
|
294639 |
+
"learning_rate": 7.537641357027465e-06,
|
294640 |
+
"loss": 0.3667,
|
294641 |
+
"step": 153480
|
294642 |
+
},
|
294643 |
+
{
|
294644 |
+
"epoch": 1237.39,
|
294645 |
+
"learning_rate": 7.5375605815832e-06,
|
294646 |
+
"loss": 0.8043,
|
294647 |
+
"step": 153485
|
294648 |
+
},
|
294649 |
+
{
|
294650 |
+
"epoch": 1237.43,
|
294651 |
+
"learning_rate": 7.537479806138935e-06,
|
294652 |
+
"loss": 0.6527,
|
294653 |
+
"step": 153490
|
294654 |
+
},
|
294655 |
+
{
|
294656 |
+
"epoch": 1237.47,
|
294657 |
+
"learning_rate": 7.53739903069467e-06,
|
294658 |
+
"loss": 0.2526,
|
294659 |
+
"step": 153495
|
294660 |
+
},
|
294661 |
+
{
|
294662 |
+
"epoch": 1237.51,
|
294663 |
+
"learning_rate": 7.537318255250405e-06,
|
294664 |
+
"loss": 0.3293,
|
294665 |
+
"step": 153500
|
294666 |
+
},
|
294667 |
+
{
|
294668 |
+
"epoch": 1237.55,
|
294669 |
+
"learning_rate": 7.53723747980614e-06,
|
294670 |
+
"loss": 0.4376,
|
294671 |
+
"step": 153505
|
294672 |
+
},
|
294673 |
+
{
|
294674 |
+
"epoch": 1237.59,
|
294675 |
+
"learning_rate": 7.537156704361875e-06,
|
294676 |
+
"loss": 0.7937,
|
294677 |
+
"step": 153510
|
294678 |
+
},
|
294679 |
+
{
|
294680 |
+
"epoch": 1237.63,
|
294681 |
+
"learning_rate": 7.5370759289176096e-06,
|
294682 |
+
"loss": 0.593,
|
294683 |
+
"step": 153515
|
294684 |
+
},
|
294685 |
+
{
|
294686 |
+
"epoch": 1237.67,
|
294687 |
+
"learning_rate": 7.5369951534733446e-06,
|
294688 |
+
"loss": 0.2965,
|
294689 |
+
"step": 153520
|
294690 |
+
},
|
294691 |
+
{
|
294692 |
+
"epoch": 1237.71,
|
294693 |
+
"learning_rate": 7.5369143780290795e-06,
|
294694 |
+
"loss": 0.3655,
|
294695 |
+
"step": 153525
|
294696 |
+
},
|
294697 |
+
{
|
294698 |
+
"epoch": 1237.76,
|
294699 |
+
"learning_rate": 7.5368336025848145e-06,
|
294700 |
+
"loss": 0.3352,
|
294701 |
+
"step": 153530
|
294702 |
+
},
|
294703 |
+
{
|
294704 |
+
"epoch": 1237.8,
|
294705 |
+
"learning_rate": 7.5367528271405495e-06,
|
294706 |
+
"loss": 1.0342,
|
294707 |
+
"step": 153535
|
294708 |
+
},
|
294709 |
+
{
|
294710 |
+
"epoch": 1237.84,
|
294711 |
+
"learning_rate": 7.536672051696285e-06,
|
294712 |
+
"loss": 0.7089,
|
294713 |
+
"step": 153540
|
294714 |
+
},
|
294715 |
+
{
|
294716 |
+
"epoch": 1237.88,
|
294717 |
+
"learning_rate": 7.5365912762520195e-06,
|
294718 |
+
"loss": 0.2906,
|
294719 |
+
"step": 153545
|
294720 |
+
},
|
294721 |
+
{
|
294722 |
+
"epoch": 1237.92,
|
294723 |
+
"learning_rate": 7.536510500807755e-06,
|
294724 |
+
"loss": 0.3161,
|
294725 |
+
"step": 153550
|
294726 |
+
},
|
294727 |
+
{
|
294728 |
+
"epoch": 1237.96,
|
294729 |
+
"learning_rate": 7.53642972536349e-06,
|
294730 |
+
"loss": 0.4502,
|
294731 |
+
"step": 153555
|
294732 |
+
},
|
294733 |
+
{
|
294734 |
+
"epoch": 1238.0,
|
294735 |
+
"learning_rate": 7.536348949919225e-06,
|
294736 |
+
"loss": 1.0289,
|
294737 |
+
"step": 153560
|
294738 |
+
},
|
294739 |
+
{
|
294740 |
+
"epoch": 1238.0,
|
294741 |
+
"eval_loss": 0.5478711724281311,
|
294742 |
+
"eval_runtime": 48.9827,
|
294743 |
+
"eval_samples_per_second": 17.067,
|
294744 |
+
"eval_steps_per_second": 0.551,
|
294745 |
+
"eval_wer": 0.1848707647995149,
|
294746 |
+
"step": 153560
|
294747 |
+
},
|
294748 |
+
{
|
294749 |
+
"epoch": 1238.04,
|
294750 |
+
"learning_rate": 7.53626817447496e-06,
|
294751 |
+
"loss": 0.3428,
|
294752 |
+
"step": 153565
|
294753 |
+
},
|
294754 |
+
{
|
294755 |
+
"epoch": 1238.08,
|
294756 |
+
"learning_rate": 7.536187399030695e-06,
|
294757 |
+
"loss": 0.3268,
|
294758 |
+
"step": 153570
|
294759 |
+
},
|
294760 |
+
{
|
294761 |
+
"epoch": 1238.12,
|
294762 |
+
"learning_rate": 7.53610662358643e-06,
|
294763 |
+
"loss": 0.2791,
|
294764 |
+
"step": 153575
|
294765 |
+
},
|
294766 |
+
{
|
294767 |
+
"epoch": 1238.16,
|
294768 |
+
"learning_rate": 7.536025848142165e-06,
|
294769 |
+
"loss": 0.4305,
|
294770 |
+
"step": 153580
|
294771 |
+
},
|
294772 |
+
{
|
294773 |
+
"epoch": 1238.2,
|
294774 |
+
"learning_rate": 7.5359450726979e-06,
|
294775 |
+
"loss": 0.9733,
|
294776 |
+
"step": 153585
|
294777 |
+
},
|
294778 |
+
{
|
294779 |
+
"epoch": 1238.24,
|
294780 |
+
"learning_rate": 7.535864297253635e-06,
|
294781 |
+
"loss": 0.3815,
|
294782 |
+
"step": 153590
|
294783 |
+
},
|
294784 |
+
{
|
294785 |
+
"epoch": 1238.28,
|
294786 |
+
"learning_rate": 7.535783521809371e-06,
|
294787 |
+
"loss": 0.2843,
|
294788 |
+
"step": 153595
|
294789 |
+
},
|
294790 |
+
{
|
294791 |
+
"epoch": 1238.32,
|
294792 |
+
"learning_rate": 7.535702746365105e-06,
|
294793 |
+
"loss": 0.2388,
|
294794 |
+
"step": 153600
|
294795 |
+
},
|
294796 |
+
{
|
294797 |
+
"epoch": 1238.36,
|
294798 |
+
"learning_rate": 7.535621970920841e-06,
|
294799 |
+
"loss": 0.4122,
|
294800 |
+
"step": 153605
|
294801 |
+
},
|
294802 |
+
{
|
294803 |
+
"epoch": 1238.4,
|
294804 |
+
"learning_rate": 7.535557350565429e-06,
|
294805 |
+
"loss": 1.0435,
|
294806 |
+
"step": 153610
|
294807 |
+
},
|
294808 |
+
{
|
294809 |
+
"epoch": 1238.44,
|
294810 |
+
"learning_rate": 7.535476575121163e-06,
|
294811 |
+
"loss": 0.295,
|
294812 |
+
"step": 153615
|
294813 |
+
},
|
294814 |
+
{
|
294815 |
+
"epoch": 1238.48,
|
294816 |
+
"learning_rate": 7.535395799676899e-06,
|
294817 |
+
"loss": 0.2713,
|
294818 |
+
"step": 153620
|
294819 |
+
},
|
294820 |
+
{
|
294821 |
+
"epoch": 1238.52,
|
294822 |
+
"learning_rate": 7.535315024232633e-06,
|
294823 |
+
"loss": 0.2524,
|
294824 |
+
"step": 153625
|
294825 |
+
},
|
294826 |
+
{
|
294827 |
+
"epoch": 1238.56,
|
294828 |
+
"learning_rate": 7.535234248788369e-06,
|
294829 |
+
"loss": 0.4259,
|
294830 |
+
"step": 153630
|
294831 |
+
},
|
294832 |
+
{
|
294833 |
+
"epoch": 1238.6,
|
294834 |
+
"learning_rate": 7.535153473344104e-06,
|
294835 |
+
"loss": 1.1058,
|
294836 |
+
"step": 153635
|
294837 |
+
},
|
294838 |
+
{
|
294839 |
+
"epoch": 1238.65,
|
294840 |
+
"learning_rate": 7.535072697899839e-06,
|
294841 |
+
"loss": 0.2864,
|
294842 |
+
"step": 153640
|
294843 |
+
},
|
294844 |
+
{
|
294845 |
+
"epoch": 1238.69,
|
294846 |
+
"learning_rate": 7.534991922455574e-06,
|
294847 |
+
"loss": 0.2977,
|
294848 |
+
"step": 153645
|
294849 |
+
},
|
294850 |
+
{
|
294851 |
+
"epoch": 1238.73,
|
294852 |
+
"learning_rate": 7.534911147011309e-06,
|
294853 |
+
"loss": 0.3155,
|
294854 |
+
"step": 153650
|
294855 |
+
},
|
294856 |
+
{
|
294857 |
+
"epoch": 1238.77,
|
294858 |
+
"learning_rate": 7.534830371567045e-06,
|
294859 |
+
"loss": 0.3997,
|
294860 |
+
"step": 153655
|
294861 |
+
},
|
294862 |
+
{
|
294863 |
+
"epoch": 1238.81,
|
294864 |
+
"learning_rate": 7.534749596122779e-06,
|
294865 |
+
"loss": 1.1922,
|
294866 |
+
"step": 153660
|
294867 |
+
},
|
294868 |
+
{
|
294869 |
+
"epoch": 1238.85,
|
294870 |
+
"learning_rate": 7.534668820678515e-06,
|
294871 |
+
"loss": 0.3735,
|
294872 |
+
"step": 153665
|
294873 |
+
},
|
294874 |
+
{
|
294875 |
+
"epoch": 1238.89,
|
294876 |
+
"learning_rate": 7.534588045234249e-06,
|
294877 |
+
"loss": 0.2743,
|
294878 |
+
"step": 153670
|
294879 |
+
},
|
294880 |
+
{
|
294881 |
+
"epoch": 1238.93,
|
294882 |
+
"learning_rate": 7.534507269789985e-06,
|
294883 |
+
"loss": 0.2918,
|
294884 |
+
"step": 153675
|
294885 |
+
},
|
294886 |
+
{
|
294887 |
+
"epoch": 1238.97,
|
294888 |
+
"learning_rate": 7.534426494345719e-06,
|
294889 |
+
"loss": 0.444,
|
294890 |
+
"step": 153680
|
294891 |
+
},
|
294892 |
+
{
|
294893 |
+
"epoch": 1239.0,
|
294894 |
+
"eval_loss": 0.41387853026390076,
|
294895 |
+
"eval_runtime": 42.5902,
|
294896 |
+
"eval_samples_per_second": 19.84,
|
294897 |
+
"eval_steps_per_second": 0.634,
|
294898 |
+
"eval_wer": 0.18675968316256086,
|
294899 |
+
"step": 153684
|
294900 |
+
},
|
294901 |
+
{
|
294902 |
+
"epoch": 1229.01,
|
294903 |
+
"learning_rate": 7.534345718901455e-06,
|
294904 |
+
"loss": 0.2827,
|
294905 |
+
"step": 153685
|
294906 |
+
},
|
294907 |
+
{
|
294908 |
+
"epoch": 1229.05,
|
294909 |
+
"learning_rate": 7.53426494345719e-06,
|
294910 |
+
"loss": 0.3284,
|
294911 |
+
"step": 153690
|
294912 |
+
},
|
294913 |
+
{
|
294914 |
+
"epoch": 1229.09,
|
294915 |
+
"learning_rate": 7.534184168012925e-06,
|
294916 |
+
"loss": 0.2527,
|
294917 |
+
"step": 153695
|
294918 |
+
},
|
294919 |
+
{
|
294920 |
+
"epoch": 1229.13,
|
294921 |
+
"learning_rate": 7.53410339256866e-06,
|
294922 |
+
"loss": 0.2886,
|
294923 |
+
"step": 153700
|
294924 |
+
},
|
294925 |
+
{
|
294926 |
+
"epoch": 1229.17,
|
294927 |
+
"learning_rate": 7.534022617124395e-06,
|
294928 |
+
"loss": 0.4841,
|
294929 |
+
"step": 153705
|
294930 |
+
},
|
294931 |
+
{
|
294932 |
+
"epoch": 1229.21,
|
294933 |
+
"learning_rate": 7.53394184168013e-06,
|
294934 |
+
"loss": 1.2124,
|
294935 |
+
"step": 153710
|
294936 |
+
},
|
294937 |
+
{
|
294938 |
+
"epoch": 1229.25,
|
294939 |
+
"learning_rate": 7.533861066235865e-06,
|
294940 |
+
"loss": 0.2612,
|
294941 |
+
"step": 153715
|
294942 |
+
},
|
294943 |
+
{
|
294944 |
+
"epoch": 1229.29,
|
294945 |
+
"learning_rate": 7.533780290791601e-06,
|
294946 |
+
"loss": 0.2457,
|
294947 |
+
"step": 153720
|
294948 |
+
},
|
294949 |
+
{
|
294950 |
+
"epoch": 1229.33,
|
294951 |
+
"learning_rate": 7.533699515347335e-06,
|
294952 |
+
"loss": 0.2777,
|
294953 |
+
"step": 153725
|
294954 |
+
},
|
294955 |
+
{
|
294956 |
+
"epoch": 1229.37,
|
294957 |
+
"learning_rate": 7.533618739903071e-06,
|
294958 |
+
"loss": 0.4545,
|
294959 |
+
"step": 153730
|
294960 |
+
},
|
294961 |
+
{
|
294962 |
+
"epoch": 1229.41,
|
294963 |
+
"learning_rate": 7.533537964458805e-06,
|
294964 |
+
"loss": 1.1277,
|
294965 |
+
"step": 153735
|
294966 |
+
},
|
294967 |
+
{
|
294968 |
+
"epoch": 1229.45,
|
294969 |
+
"learning_rate": 7.5334571890145406e-06,
|
294970 |
+
"loss": 0.2915,
|
294971 |
+
"step": 153740
|
294972 |
+
},
|
294973 |
+
{
|
294974 |
+
"epoch": 1229.49,
|
294975 |
+
"learning_rate": 7.533376413570275e-06,
|
294976 |
+
"loss": 0.2849,
|
294977 |
+
"step": 153745
|
294978 |
+
},
|
294979 |
+
{
|
294980 |
+
"epoch": 1229.53,
|
294981 |
+
"learning_rate": 7.5332956381260105e-06,
|
294982 |
+
"loss": 0.2834,
|
294983 |
+
"step": 153750
|
294984 |
+
},
|
294985 |
+
{
|
294986 |
+
"epoch": 1229.57,
|
294987 |
+
"learning_rate": 7.5332148626817455e-06,
|
294988 |
+
"loss": 0.4269,
|
294989 |
+
"step": 153755
|
294990 |
+
},
|
294991 |
+
{
|
294992 |
+
"epoch": 1229.61,
|
294993 |
+
"learning_rate": 7.5331340872374805e-06,
|
294994 |
+
"loss": 0.9413,
|
294995 |
+
"step": 153760
|
294996 |
+
},
|
294997 |
+
{
|
294998 |
+
"epoch": 1229.65,
|
294999 |
+
"learning_rate": 7.5330533117932155e-06,
|
295000 |
+
"loss": 0.3305,
|
295001 |
+
"step": 153765
|
295002 |
+
},
|
295003 |
+
{
|
295004 |
+
"epoch": 1229.69,
|
295005 |
+
"learning_rate": 7.5329725363489505e-06,
|
295006 |
+
"loss": 0.2752,
|
295007 |
+
"step": 153770
|
295008 |
+
},
|
295009 |
+
{
|
295010 |
+
"epoch": 1229.73,
|
295011 |
+
"learning_rate": 7.5328917609046855e-06,
|
295012 |
+
"loss": 0.3041,
|
295013 |
+
"step": 153775
|
295014 |
+
},
|
295015 |
+
{
|
295016 |
+
"epoch": 1229.77,
|
295017 |
+
"learning_rate": 7.5328109854604205e-06,
|
295018 |
+
"loss": 0.4273,
|
295019 |
+
"step": 153780
|
295020 |
+
},
|
295021 |
+
{
|
295022 |
+
"epoch": 1229.81,
|
295023 |
+
"learning_rate": 7.532730210016156e-06,
|
295024 |
+
"loss": 1.0783,
|
295025 |
+
"step": 153785
|
295026 |
+
},
|
295027 |
+
{
|
295028 |
+
"epoch": 1229.85,
|
295029 |
+
"learning_rate": 7.5326494345718904e-06,
|
295030 |
+
"loss": 0.2982,
|
295031 |
+
"step": 153790
|
295032 |
+
},
|
295033 |
+
{
|
295034 |
+
"epoch": 1229.89,
|
295035 |
+
"learning_rate": 7.532568659127626e-06,
|
295036 |
+
"loss": 0.2782,
|
295037 |
+
"step": 153795
|
295038 |
+
},
|
295039 |
+
{
|
295040 |
+
"epoch": 1229.93,
|
295041 |
+
"learning_rate": 7.53248788368336e-06,
|
295042 |
+
"loss": 0.3229,
|
295043 |
+
"step": 153800
|
295044 |
+
},
|
295045 |
+
{
|
295046 |
+
"epoch": 1229.97,
|
295047 |
+
"learning_rate": 7.532407108239096e-06,
|
295048 |
+
"loss": 0.4878,
|
295049 |
+
"step": 153805
|
295050 |
+
},
|
295051 |
+
{
|
295052 |
+
"epoch": 1230.0,
|
295053 |
+
"eval_loss": 0.426218718290329,
|
295054 |
+
"eval_runtime": 40.8122,
|
295055 |
+
"eval_samples_per_second": 20.68,
|
295056 |
+
"eval_steps_per_second": 0.662,
|
295057 |
+
"eval_wer": 0.17735651131425145,
|
295058 |
+
"step": 153809
|
295059 |
+
},
|
295060 |
+
{
|
295061 |
+
"epoch": 1230.01,
|
295062 |
+
"learning_rate": 7.532326332794831e-06,
|
295063 |
+
"loss": 0.3619,
|
295064 |
+
"step": 153810
|
295065 |
+
},
|
295066 |
+
{
|
295067 |
+
"epoch": 1230.05,
|
295068 |
+
"learning_rate": 7.532245557350566e-06,
|
295069 |
+
"loss": 0.2662,
|
295070 |
+
"step": 153815
|
295071 |
+
},
|
295072 |
+
{
|
295073 |
+
"epoch": 1230.09,
|
295074 |
+
"learning_rate": 7.532164781906301e-06,
|
295075 |
+
"loss": 0.2751,
|
295076 |
+
"step": 153820
|
295077 |
+
},
|
295078 |
+
{
|
295079 |
+
"epoch": 1230.13,
|
295080 |
+
"learning_rate": 7.532084006462036e-06,
|
295081 |
+
"loss": 0.3694,
|
295082 |
+
"step": 153825
|
295083 |
+
},
|
295084 |
+
{
|
295085 |
+
"epoch": 1230.17,
|
295086 |
+
"learning_rate": 7.532003231017771e-06,
|
295087 |
+
"loss": 0.5156,
|
295088 |
+
"step": 153830
|
295089 |
+
},
|
295090 |
+
{
|
295091 |
+
"epoch": 1230.21,
|
295092 |
+
"learning_rate": 7.531922455573506e-06,
|
295093 |
+
"loss": 1.0791,
|
295094 |
+
"step": 153835
|
295095 |
+
},
|
295096 |
+
{
|
295097 |
+
"epoch": 1230.25,
|
295098 |
+
"learning_rate": 7.531841680129241e-06,
|
295099 |
+
"loss": 0.2508,
|
295100 |
+
"step": 153840
|
295101 |
+
},
|
295102 |
+
{
|
295103 |
+
"epoch": 1230.29,
|
295104 |
+
"learning_rate": 7.531760904684976e-06,
|
295105 |
+
"loss": 0.2831,
|
295106 |
+
"step": 153845
|
295107 |
+
},
|
295108 |
+
{
|
295109 |
+
"epoch": 1230.33,
|
295110 |
+
"learning_rate": 7.531680129240711e-06,
|
295111 |
+
"loss": 0.3186,
|
295112 |
+
"step": 153850
|
295113 |
+
},
|
295114 |
+
{
|
295115 |
+
"epoch": 1230.37,
|
295116 |
+
"learning_rate": 7.531599353796446e-06,
|
295117 |
+
"loss": 0.4961,
|
295118 |
+
"step": 153855
|
295119 |
+
},
|
295120 |
+
{
|
295121 |
+
"epoch": 1230.41,
|
295122 |
+
"learning_rate": 7.531518578352182e-06,
|
295123 |
+
"loss": 1.0887,
|
295124 |
+
"step": 153860
|
295125 |
+
},
|
295126 |
+
{
|
295127 |
+
"epoch": 1230.45,
|
295128 |
+
"learning_rate": 7.531437802907917e-06,
|
295129 |
+
"loss": 0.2721,
|
295130 |
+
"step": 153865
|
295131 |
+
},
|
295132 |
+
{
|
295133 |
+
"epoch": 1230.49,
|
295134 |
+
"learning_rate": 7.531357027463652e-06,
|
295135 |
+
"loss": 0.2654,
|
295136 |
+
"step": 153870
|
295137 |
+
},
|
295138 |
+
{
|
295139 |
+
"epoch": 1230.53,
|
295140 |
+
"learning_rate": 7.531276252019387e-06,
|
295141 |
+
"loss": 0.2596,
|
295142 |
+
"step": 153875
|
295143 |
+
},
|
295144 |
+
{
|
295145 |
+
"epoch": 1230.57,
|
295146 |
+
"learning_rate": 7.531195476575122e-06,
|
295147 |
+
"loss": 0.5072,
|
295148 |
+
"step": 153880
|
295149 |
+
},
|
295150 |
+
{
|
295151 |
+
"epoch": 1230.61,
|
295152 |
+
"learning_rate": 7.531114701130857e-06,
|
295153 |
+
"loss": 1.0273,
|
295154 |
+
"step": 153885
|
295155 |
+
},
|
295156 |
+
{
|
295157 |
+
"epoch": 1230.65,
|
295158 |
+
"learning_rate": 7.531033925686592e-06,
|
295159 |
+
"loss": 0.3047,
|
295160 |
+
"step": 153890
|
295161 |
+
},
|
295162 |
+
{
|
295163 |
+
"epoch": 1230.69,
|
295164 |
+
"learning_rate": 7.530953150242327e-06,
|
295165 |
+
"loss": 0.2919,
|
295166 |
+
"step": 153895
|
295167 |
+
},
|
295168 |
+
{
|
295169 |
+
"epoch": 1230.73,
|
295170 |
+
"learning_rate": 7.530872374798062e-06,
|
295171 |
+
"loss": 0.3372,
|
295172 |
+
"step": 153900
|
295173 |
+
},
|
295174 |
+
{
|
295175 |
+
"epoch": 1230.77,
|
295176 |
+
"learning_rate": 7.530791599353797e-06,
|
295177 |
+
"loss": 0.4598,
|
295178 |
+
"step": 153905
|
295179 |
+
},
|
295180 |
+
{
|
295181 |
+
"epoch": 1230.81,
|
295182 |
+
"learning_rate": 7.530710823909532e-06,
|
295183 |
+
"loss": 1.2106,
|
295184 |
+
"step": 153910
|
295185 |
+
},
|
295186 |
+
{
|
295187 |
+
"epoch": 1230.85,
|
295188 |
+
"learning_rate": 7.530630048465267e-06,
|
295189 |
+
"loss": 0.3627,
|
295190 |
+
"step": 153915
|
295191 |
+
},
|
295192 |
+
{
|
295193 |
+
"epoch": 1230.89,
|
295194 |
+
"learning_rate": 7.530549273021002e-06,
|
295195 |
+
"loss": 0.2599,
|
295196 |
+
"step": 153920
|
295197 |
+
},
|
295198 |
+
{
|
295199 |
+
"epoch": 1230.93,
|
295200 |
+
"learning_rate": 7.530468497576738e-06,
|
295201 |
+
"loss": 0.2899,
|
295202 |
+
"step": 153925
|
295203 |
+
},
|
295204 |
+
{
|
295205 |
+
"epoch": 1230.97,
|
295206 |
+
"learning_rate": 7.530387722132473e-06,
|
295207 |
+
"loss": 0.4721,
|
295208 |
+
"step": 153930
|
295209 |
+
},
|
295210 |
+
{
|
295211 |
+
"epoch": 1231.0,
|
295212 |
+
"eval_loss": 0.3832179605960846,
|
295213 |
+
"eval_runtime": 42.084,
|
295214 |
+
"eval_samples_per_second": 20.079,
|
295215 |
+
"eval_steps_per_second": 0.642,
|
295216 |
+
"eval_wer": 0.17948529411764705,
|
295217 |
+
"step": 153934
|
295218 |
}
|
295219 |
],
|
295220 |
+
"max_steps": 625000,
|
295221 |
"num_train_epochs": 5000,
|
295222 |
+
"total_flos": 4.331791340637063e+20,
|
295223 |
"trial_name": null,
|
295224 |
"trial_params": null
|
295225 |
}
|
model-bin/finetune/base/{checkpoint-153312 β checkpoint-153934}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630255201.5100508/events.out.tfevents.1630255201.cc93b136ebf5.1086.229
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3c705e91bc7083cacdbb6e3ea2d47851baed071b6f5c8e4577088faed2362e2
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630255633.58935/events.out.tfevents.1630255633.cc93b136ebf5.1086.231
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b63fccc5857491b99b5739495ed2be5087930f24a10b687fbf40415f783d523b
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630256177.9503975/events.out.tfevents.1630256177.cc93b136ebf5.1086.233
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a630c947a6f333b8473ee3f23138bc64d6785f9d00986278b06a30f8020b218
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630256616.2655816/events.out.tfevents.1630256616.cc93b136ebf5.1086.235
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f50a9bac3409292271e60f77ab2e8fd4ac623bb8b7a62fc7b0b2b25a90957a55
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630257046.964922/events.out.tfevents.1630257046.cc93b136ebf5.1086.237
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd08c75684eb37b19f78400983c8e53fa797ee0de9523400f2e5ef29575155e4
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630255201.cc93b136ebf5.1086.228
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:782bdf8f3cfb090f13eab643e64ba25bab0b9d06fcc958f9f11eae9c08acc88c
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630255633.cc93b136ebf5.1086.230
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:abbb3352bfa1783fb894e721926ff0819978e7ef7567481e0f4e8ca1934056bf
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630256177.cc93b136ebf5.1086.232
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42cb727f7eb722f684c8f8c3d603d4a89ef672d7fa8e4dfd675ffd8227a59de7
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630256616.cc93b136ebf5.1086.234
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e2b66ef6a795218838308e9f89ad071f07bd164a9141a9f4ea4bf195f96b68e
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630257046.cc93b136ebf5.1086.236
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7d4dc274a35f05f7eb0c9f9e0ec2c3f3b907459fdfc1f66f38050fc947f4a3b
|
3 |
+
size 8622
|