"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630244012.7530751/events.out.tfevents.1630244012.cc93b136ebf5.1086.179 +3 -0
- model-bin/finetune/base/log/1630244448.5792234/events.out.tfevents.1630244448.cc93b136ebf5.1086.181 +3 -0
- model-bin/finetune/base/log/1630244993.356596/events.out.tfevents.1630244993.cc93b136ebf5.1086.183 +3 -0
- model-bin/finetune/base/log/1630245427.5586958/events.out.tfevents.1630245427.cc93b136ebf5.1086.185 +3 -0
- model-bin/finetune/base/log/1630245858.8737807/events.out.tfevents.1630245858.cc93b136ebf5.1086.187 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630244012.cc93b136ebf5.1086.178 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630244448.cc93b136ebf5.1086.180 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630244993.cc93b136ebf5.1086.182 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630245427.cc93b136ebf5.1086.184 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630245858.cc93b136ebf5.1086.186 +3 -0
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9665c6e1c6f2987ac982b4f02510d642ccf49cb18e475af98729715643f2744
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83712bb0bb146694098ef77fead19cec258379ae2b0f648e4edb18652a55565a
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f44a922c3d45f1ec6984b47ad3b841b6c8eca8288bc65f7fa757f180153fd40
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:adf29ee28b173423b576c771747dd93d436f5c3f2a3af9a31f91c508ff5dd1b2
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee134784171ae4f4005cf971653fad3f3e37c1a6c7b565c91f04852ab760d0e4
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -290475,11 +290475,800 @@
|
|
290475 |
"eval_steps_per_second": 0.644,
|
290476 |
"eval_wer": 0.17267648552564754,
|
290477 |
"step": 150205
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
290478 |
}
|
290479 |
],
|
290480 |
-
"max_steps":
|
290481 |
"num_train_epochs": 5000,
|
290482 |
-
"total_flos": 4.
|
290483 |
"trial_name": null,
|
290484 |
"trial_params": null
|
290485 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1206.0,
|
5 |
+
"global_step": 150827,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
290475 |
"eval_steps_per_second": 0.644,
|
290476 |
"eval_wer": 0.17267648552564754,
|
290477 |
"step": 150205
|
290478 |
+
},
|
290479 |
+
{
|
290480 |
+
"epoch": 1211.04,
|
290481 |
+
"learning_rate": 7.590452342487884e-06,
|
290482 |
+
"loss": 0.3201,
|
290483 |
+
"step": 150210
|
290484 |
+
},
|
290485 |
+
{
|
290486 |
+
"epoch": 1211.08,
|
290487 |
+
"learning_rate": 7.59037156704362e-06,
|
290488 |
+
"loss": 0.2736,
|
290489 |
+
"step": 150215
|
290490 |
+
},
|
290491 |
+
{
|
290492 |
+
"epoch": 1211.12,
|
290493 |
+
"learning_rate": 7.590290791599354e-06,
|
290494 |
+
"loss": 0.2423,
|
290495 |
+
"step": 150220
|
290496 |
+
},
|
290497 |
+
{
|
290498 |
+
"epoch": 1211.16,
|
290499 |
+
"learning_rate": 7.59021001615509e-06,
|
290500 |
+
"loss": 0.4379,
|
290501 |
+
"step": 150225
|
290502 |
+
},
|
290503 |
+
{
|
290504 |
+
"epoch": 1211.2,
|
290505 |
+
"learning_rate": 7.590129240710824e-06,
|
290506 |
+
"loss": 0.9755,
|
290507 |
+
"step": 150230
|
290508 |
+
},
|
290509 |
+
{
|
290510 |
+
"epoch": 1211.24,
|
290511 |
+
"learning_rate": 7.59004846526656e-06,
|
290512 |
+
"loss": 0.3417,
|
290513 |
+
"step": 150235
|
290514 |
+
},
|
290515 |
+
{
|
290516 |
+
"epoch": 1211.28,
|
290517 |
+
"learning_rate": 7.5899676898222955e-06,
|
290518 |
+
"loss": 0.2895,
|
290519 |
+
"step": 150240
|
290520 |
+
},
|
290521 |
+
{
|
290522 |
+
"epoch": 1211.32,
|
290523 |
+
"learning_rate": 7.58988691437803e-06,
|
290524 |
+
"loss": 0.2984,
|
290525 |
+
"step": 150245
|
290526 |
+
},
|
290527 |
+
{
|
290528 |
+
"epoch": 1211.36,
|
290529 |
+
"learning_rate": 7.5898061389337655e-06,
|
290530 |
+
"loss": 0.3888,
|
290531 |
+
"step": 150250
|
290532 |
+
},
|
290533 |
+
{
|
290534 |
+
"epoch": 1211.4,
|
290535 |
+
"learning_rate": 7.5897253634895e-06,
|
290536 |
+
"loss": 1.1877,
|
290537 |
+
"step": 150255
|
290538 |
+
},
|
290539 |
+
{
|
290540 |
+
"epoch": 1211.44,
|
290541 |
+
"learning_rate": 7.5896445880452355e-06,
|
290542 |
+
"loss": 0.2986,
|
290543 |
+
"step": 150260
|
290544 |
+
},
|
290545 |
+
{
|
290546 |
+
"epoch": 1211.48,
|
290547 |
+
"learning_rate": 7.58956381260097e-06,
|
290548 |
+
"loss": 0.2978,
|
290549 |
+
"step": 150265
|
290550 |
+
},
|
290551 |
+
{
|
290552 |
+
"epoch": 1211.52,
|
290553 |
+
"learning_rate": 7.5894830371567055e-06,
|
290554 |
+
"loss": 0.3136,
|
290555 |
+
"step": 150270
|
290556 |
+
},
|
290557 |
+
{
|
290558 |
+
"epoch": 1211.56,
|
290559 |
+
"learning_rate": 7.58940226171244e-06,
|
290560 |
+
"loss": 0.4057,
|
290561 |
+
"step": 150275
|
290562 |
+
},
|
290563 |
+
{
|
290564 |
+
"epoch": 1211.6,
|
290565 |
+
"learning_rate": 7.5893214862681754e-06,
|
290566 |
+
"loss": 1.0684,
|
290567 |
+
"step": 150280
|
290568 |
+
},
|
290569 |
+
{
|
290570 |
+
"epoch": 1211.64,
|
290571 |
+
"learning_rate": 7.5892407108239096e-06,
|
290572 |
+
"loss": 0.3111,
|
290573 |
+
"step": 150285
|
290574 |
+
},
|
290575 |
+
{
|
290576 |
+
"epoch": 1211.68,
|
290577 |
+
"learning_rate": 7.589159935379645e-06,
|
290578 |
+
"loss": 0.3245,
|
290579 |
+
"step": 150290
|
290580 |
+
},
|
290581 |
+
{
|
290582 |
+
"epoch": 1211.72,
|
290583 |
+
"learning_rate": 7.58907915993538e-06,
|
290584 |
+
"loss": 0.3207,
|
290585 |
+
"step": 150295
|
290586 |
+
},
|
290587 |
+
{
|
290588 |
+
"epoch": 1211.76,
|
290589 |
+
"learning_rate": 7.588998384491115e-06,
|
290590 |
+
"loss": 0.3932,
|
290591 |
+
"step": 150300
|
290592 |
+
},
|
290593 |
+
{
|
290594 |
+
"epoch": 1211.8,
|
290595 |
+
"learning_rate": 7.588917609046851e-06,
|
290596 |
+
"loss": 1.1473,
|
290597 |
+
"step": 150305
|
290598 |
+
},
|
290599 |
+
{
|
290600 |
+
"epoch": 1211.84,
|
290601 |
+
"learning_rate": 7.588836833602585e-06,
|
290602 |
+
"loss": 0.2849,
|
290603 |
+
"step": 150310
|
290604 |
+
},
|
290605 |
+
{
|
290606 |
+
"epoch": 1211.88,
|
290607 |
+
"learning_rate": 7.588756058158321e-06,
|
290608 |
+
"loss": 0.3008,
|
290609 |
+
"step": 150315
|
290610 |
+
},
|
290611 |
+
{
|
290612 |
+
"epoch": 1211.92,
|
290613 |
+
"learning_rate": 7.588675282714055e-06,
|
290614 |
+
"loss": 0.3498,
|
290615 |
+
"step": 150320
|
290616 |
+
},
|
290617 |
+
{
|
290618 |
+
"epoch": 1211.96,
|
290619 |
+
"learning_rate": 7.588594507269791e-06,
|
290620 |
+
"loss": 0.528,
|
290621 |
+
"step": 150325
|
290622 |
+
},
|
290623 |
+
{
|
290624 |
+
"epoch": 1212.0,
|
290625 |
+
"eval_loss": 0.38405805826187134,
|
290626 |
+
"eval_runtime": 42.6893,
|
290627 |
+
"eval_samples_per_second": 19.7,
|
290628 |
+
"eval_steps_per_second": 0.632,
|
290629 |
+
"eval_wer": 0.17788736420811893,
|
290630 |
+
"step": 150329
|
290631 |
+
},
|
290632 |
+
{
|
290633 |
+
"epoch": 1212.01,
|
290634 |
+
"learning_rate": 7.588513731825525e-06,
|
290635 |
+
"loss": 0.4414,
|
290636 |
+
"step": 150330
|
290637 |
+
},
|
290638 |
+
{
|
290639 |
+
"epoch": 1212.05,
|
290640 |
+
"learning_rate": 7.588432956381261e-06,
|
290641 |
+
"loss": 0.3031,
|
290642 |
+
"step": 150335
|
290643 |
+
},
|
290644 |
+
{
|
290645 |
+
"epoch": 1212.09,
|
290646 |
+
"learning_rate": 7.588352180936995e-06,
|
290647 |
+
"loss": 0.2945,
|
290648 |
+
"step": 150340
|
290649 |
+
},
|
290650 |
+
{
|
290651 |
+
"epoch": 1212.13,
|
290652 |
+
"learning_rate": 7.588271405492731e-06,
|
290653 |
+
"loss": 0.3029,
|
290654 |
+
"step": 150345
|
290655 |
+
},
|
290656 |
+
{
|
290657 |
+
"epoch": 1212.17,
|
290658 |
+
"learning_rate": 7.588190630048465e-06,
|
290659 |
+
"loss": 0.4425,
|
290660 |
+
"step": 150350
|
290661 |
+
},
|
290662 |
+
{
|
290663 |
+
"epoch": 1212.21,
|
290664 |
+
"learning_rate": 7.588109854604201e-06,
|
290665 |
+
"loss": 1.0489,
|
290666 |
+
"step": 150355
|
290667 |
+
},
|
290668 |
+
{
|
290669 |
+
"epoch": 1212.25,
|
290670 |
+
"learning_rate": 7.588029079159936e-06,
|
290671 |
+
"loss": 0.2773,
|
290672 |
+
"step": 150360
|
290673 |
+
},
|
290674 |
+
{
|
290675 |
+
"epoch": 1212.29,
|
290676 |
+
"learning_rate": 7.587948303715671e-06,
|
290677 |
+
"loss": 0.267,
|
290678 |
+
"step": 150365
|
290679 |
+
},
|
290680 |
+
{
|
290681 |
+
"epoch": 1212.33,
|
290682 |
+
"learning_rate": 7.587867528271407e-06,
|
290683 |
+
"loss": 0.3697,
|
290684 |
+
"step": 150370
|
290685 |
+
},
|
290686 |
+
{
|
290687 |
+
"epoch": 1212.37,
|
290688 |
+
"learning_rate": 7.587786752827141e-06,
|
290689 |
+
"loss": 0.4697,
|
290690 |
+
"step": 150375
|
290691 |
+
},
|
290692 |
+
{
|
290693 |
+
"epoch": 1212.41,
|
290694 |
+
"learning_rate": 7.587705977382877e-06,
|
290695 |
+
"loss": 1.0624,
|
290696 |
+
"step": 150380
|
290697 |
+
},
|
290698 |
+
{
|
290699 |
+
"epoch": 1212.45,
|
290700 |
+
"learning_rate": 7.587625201938611e-06,
|
290701 |
+
"loss": 0.2572,
|
290702 |
+
"step": 150385
|
290703 |
+
},
|
290704 |
+
{
|
290705 |
+
"epoch": 1212.49,
|
290706 |
+
"learning_rate": 7.587544426494347e-06,
|
290707 |
+
"loss": 0.3007,
|
290708 |
+
"step": 150390
|
290709 |
+
},
|
290710 |
+
{
|
290711 |
+
"epoch": 1212.53,
|
290712 |
+
"learning_rate": 7.587463651050081e-06,
|
290713 |
+
"loss": 0.2912,
|
290714 |
+
"step": 150395
|
290715 |
+
},
|
290716 |
+
{
|
290717 |
+
"epoch": 1212.57,
|
290718 |
+
"learning_rate": 7.587382875605817e-06,
|
290719 |
+
"loss": 0.5251,
|
290720 |
+
"step": 150400
|
290721 |
+
},
|
290722 |
+
{
|
290723 |
+
"epoch": 1212.61,
|
290724 |
+
"learning_rate": 7.587302100161551e-06,
|
290725 |
+
"loss": 1.0472,
|
290726 |
+
"step": 150405
|
290727 |
+
},
|
290728 |
+
{
|
290729 |
+
"epoch": 1212.65,
|
290730 |
+
"learning_rate": 7.587221324717287e-06,
|
290731 |
+
"loss": 0.3116,
|
290732 |
+
"step": 150410
|
290733 |
+
},
|
290734 |
+
{
|
290735 |
+
"epoch": 1212.69,
|
290736 |
+
"learning_rate": 7.587140549273022e-06,
|
290737 |
+
"loss": 0.2427,
|
290738 |
+
"step": 150415
|
290739 |
+
},
|
290740 |
+
{
|
290741 |
+
"epoch": 1212.73,
|
290742 |
+
"learning_rate": 7.587059773828757e-06,
|
290743 |
+
"loss": 0.3504,
|
290744 |
+
"step": 150420
|
290745 |
+
},
|
290746 |
+
{
|
290747 |
+
"epoch": 1212.77,
|
290748 |
+
"learning_rate": 7.586978998384492e-06,
|
290749 |
+
"loss": 0.4429,
|
290750 |
+
"step": 150425
|
290751 |
+
},
|
290752 |
+
{
|
290753 |
+
"epoch": 1212.81,
|
290754 |
+
"learning_rate": 7.586898222940227e-06,
|
290755 |
+
"loss": 0.945,
|
290756 |
+
"step": 150430
|
290757 |
+
},
|
290758 |
+
{
|
290759 |
+
"epoch": 1212.85,
|
290760 |
+
"learning_rate": 7.586817447495963e-06,
|
290761 |
+
"loss": 0.3042,
|
290762 |
+
"step": 150435
|
290763 |
+
},
|
290764 |
+
{
|
290765 |
+
"epoch": 1212.9,
|
290766 |
+
"learning_rate": 7.586736672051697e-06,
|
290767 |
+
"loss": 0.2662,
|
290768 |
+
"step": 150440
|
290769 |
+
},
|
290770 |
+
{
|
290771 |
+
"epoch": 1212.94,
|
290772 |
+
"learning_rate": 7.5866558966074326e-06,
|
290773 |
+
"loss": 0.3849,
|
290774 |
+
"step": 150445
|
290775 |
+
},
|
290776 |
+
{
|
290777 |
+
"epoch": 1212.98,
|
290778 |
+
"learning_rate": 7.586575121163167e-06,
|
290779 |
+
"loss": 0.6568,
|
290780 |
+
"step": 150450
|
290781 |
+
},
|
290782 |
+
{
|
290783 |
+
"epoch": 1213.0,
|
290784 |
+
"eval_loss": 0.3243919909000397,
|
290785 |
+
"eval_runtime": 45.4766,
|
290786 |
+
"eval_samples_per_second": 18.493,
|
290787 |
+
"eval_steps_per_second": 0.594,
|
290788 |
+
"eval_wer": 0.17089605734767024,
|
290789 |
+
"step": 150453
|
290790 |
+
},
|
290791 |
+
{
|
290792 |
+
"epoch": 1203.02,
|
290793 |
+
"learning_rate": 7.5864943457189025e-06,
|
290794 |
+
"loss": 0.2572,
|
290795 |
+
"step": 150455
|
290796 |
+
},
|
290797 |
+
{
|
290798 |
+
"epoch": 1203.06,
|
290799 |
+
"learning_rate": 7.586413570274637e-06,
|
290800 |
+
"loss": 0.2823,
|
290801 |
+
"step": 150460
|
290802 |
+
},
|
290803 |
+
{
|
290804 |
+
"epoch": 1203.1,
|
290805 |
+
"learning_rate": 7.5863327948303725e-06,
|
290806 |
+
"loss": 0.2366,
|
290807 |
+
"step": 150465
|
290808 |
+
},
|
290809 |
+
{
|
290810 |
+
"epoch": 1203.14,
|
290811 |
+
"learning_rate": 7.5862520193861075e-06,
|
290812 |
+
"loss": 0.3375,
|
290813 |
+
"step": 150470
|
290814 |
+
},
|
290815 |
+
{
|
290816 |
+
"epoch": 1203.18,
|
290817 |
+
"learning_rate": 7.5861712439418425e-06,
|
290818 |
+
"loss": 0.6214,
|
290819 |
+
"step": 150475
|
290820 |
+
},
|
290821 |
+
{
|
290822 |
+
"epoch": 1203.22,
|
290823 |
+
"learning_rate": 7.5860904684975775e-06,
|
290824 |
+
"loss": 0.9851,
|
290825 |
+
"step": 150480
|
290826 |
+
},
|
290827 |
+
{
|
290828 |
+
"epoch": 1203.26,
|
290829 |
+
"learning_rate": 7.5860096930533125e-06,
|
290830 |
+
"loss": 0.2788,
|
290831 |
+
"step": 150485
|
290832 |
+
},
|
290833 |
+
{
|
290834 |
+
"epoch": 1203.3,
|
290835 |
+
"learning_rate": 7.5859289176090474e-06,
|
290836 |
+
"loss": 0.2879,
|
290837 |
+
"step": 150490
|
290838 |
+
},
|
290839 |
+
{
|
290840 |
+
"epoch": 1203.34,
|
290841 |
+
"learning_rate": 7.5858481421647824e-06,
|
290842 |
+
"loss": 0.3532,
|
290843 |
+
"step": 150495
|
290844 |
+
},
|
290845 |
+
{
|
290846 |
+
"epoch": 1203.38,
|
290847 |
+
"learning_rate": 7.585767366720517e-06,
|
290848 |
+
"loss": 0.5532,
|
290849 |
+
"step": 150500
|
290850 |
+
},
|
290851 |
+
{
|
290852 |
+
"epoch": 1203.42,
|
290853 |
+
"learning_rate": 7.585686591276252e-06,
|
290854 |
+
"loss": 1.016,
|
290855 |
+
"step": 150505
|
290856 |
+
},
|
290857 |
+
{
|
290858 |
+
"epoch": 1203.46,
|
290859 |
+
"learning_rate": 7.585605815831988e-06,
|
290860 |
+
"loss": 0.303,
|
290861 |
+
"step": 150510
|
290862 |
+
},
|
290863 |
+
{
|
290864 |
+
"epoch": 1203.5,
|
290865 |
+
"learning_rate": 7.585525040387722e-06,
|
290866 |
+
"loss": 0.2581,
|
290867 |
+
"step": 150515
|
290868 |
+
},
|
290869 |
+
{
|
290870 |
+
"epoch": 1203.54,
|
290871 |
+
"learning_rate": 7.585444264943458e-06,
|
290872 |
+
"loss": 0.3192,
|
290873 |
+
"step": 150520
|
290874 |
+
},
|
290875 |
+
{
|
290876 |
+
"epoch": 1203.58,
|
290877 |
+
"learning_rate": 7.585363489499192e-06,
|
290878 |
+
"loss": 0.5658,
|
290879 |
+
"step": 150525
|
290880 |
+
},
|
290881 |
+
{
|
290882 |
+
"epoch": 1203.62,
|
290883 |
+
"learning_rate": 7.585282714054928e-06,
|
290884 |
+
"loss": 1.1043,
|
290885 |
+
"step": 150530
|
290886 |
+
},
|
290887 |
+
{
|
290888 |
+
"epoch": 1203.66,
|
290889 |
+
"learning_rate": 7.585201938610663e-06,
|
290890 |
+
"loss": 0.3109,
|
290891 |
+
"step": 150535
|
290892 |
+
},
|
290893 |
+
{
|
290894 |
+
"epoch": 1203.7,
|
290895 |
+
"learning_rate": 7.585121163166398e-06,
|
290896 |
+
"loss": 0.3193,
|
290897 |
+
"step": 150540
|
290898 |
+
},
|
290899 |
+
{
|
290900 |
+
"epoch": 1203.74,
|
290901 |
+
"learning_rate": 7.585040387722133e-06,
|
290902 |
+
"loss": 0.2706,
|
290903 |
+
"step": 150545
|
290904 |
+
},
|
290905 |
+
{
|
290906 |
+
"epoch": 1203.78,
|
290907 |
+
"learning_rate": 7.584959612277868e-06,
|
290908 |
+
"loss": 0.6651,
|
290909 |
+
"step": 150550
|
290910 |
+
},
|
290911 |
+
{
|
290912 |
+
"epoch": 1203.82,
|
290913 |
+
"learning_rate": 7.584878836833603e-06,
|
290914 |
+
"loss": 0.9646,
|
290915 |
+
"step": 150555
|
290916 |
+
},
|
290917 |
+
{
|
290918 |
+
"epoch": 1203.86,
|
290919 |
+
"learning_rate": 7.584798061389338e-06,
|
290920 |
+
"loss": 0.284,
|
290921 |
+
"step": 150560
|
290922 |
+
},
|
290923 |
+
{
|
290924 |
+
"epoch": 1203.9,
|
290925 |
+
"learning_rate": 7.584717285945073e-06,
|
290926 |
+
"loss": 0.2926,
|
290927 |
+
"step": 150565
|
290928 |
+
},
|
290929 |
+
{
|
290930 |
+
"epoch": 1203.94,
|
290931 |
+
"learning_rate": 7.584636510500808e-06,
|
290932 |
+
"loss": 0.427,
|
290933 |
+
"step": 150570
|
290934 |
+
},
|
290935 |
+
{
|
290936 |
+
"epoch": 1203.98,
|
290937 |
+
"learning_rate": 7.584555735056544e-06,
|
290938 |
+
"loss": 0.6464,
|
290939 |
+
"step": 150575
|
290940 |
+
},
|
290941 |
+
{
|
290942 |
+
"epoch": 1204.0,
|
290943 |
+
"eval_loss": 0.5391775965690613,
|
290944 |
+
"eval_runtime": 39.7033,
|
290945 |
+
"eval_samples_per_second": 21.056,
|
290946 |
+
"eval_steps_per_second": 0.68,
|
290947 |
+
"eval_wer": 0.1843401803986963,
|
290948 |
+
"step": 150578
|
290949 |
+
},
|
290950 |
+
{
|
290951 |
+
"epoch": 1214.02,
|
290952 |
+
"learning_rate": 7.584474959612278e-06,
|
290953 |
+
"loss": 0.3288,
|
290954 |
+
"step": 150580
|
290955 |
+
},
|
290956 |
+
{
|
290957 |
+
"epoch": 1214.06,
|
290958 |
+
"learning_rate": 7.584394184168014e-06,
|
290959 |
+
"loss": 0.9793,
|
290960 |
+
"step": 150585
|
290961 |
+
},
|
290962 |
+
{
|
290963 |
+
"epoch": 1214.1,
|
290964 |
+
"learning_rate": 7.584313408723749e-06,
|
290965 |
+
"loss": 0.2777,
|
290966 |
+
"step": 150590
|
290967 |
+
},
|
290968 |
+
{
|
290969 |
+
"epoch": 1214.14,
|
290970 |
+
"learning_rate": 7.584232633279484e-06,
|
290971 |
+
"loss": 0.3273,
|
290972 |
+
"step": 150595
|
290973 |
+
},
|
290974 |
+
{
|
290975 |
+
"epoch": 1214.18,
|
290976 |
+
"learning_rate": 7.584151857835219e-06,
|
290977 |
+
"loss": 0.5917,
|
290978 |
+
"step": 150600
|
290979 |
+
},
|
290980 |
+
{
|
290981 |
+
"epoch": 1214.22,
|
290982 |
+
"learning_rate": 7.584071082390954e-06,
|
290983 |
+
"loss": 0.8979,
|
290984 |
+
"step": 150605
|
290985 |
+
},
|
290986 |
+
{
|
290987 |
+
"epoch": 1214.26,
|
290988 |
+
"learning_rate": 7.583990306946689e-06,
|
290989 |
+
"loss": 0.3145,
|
290990 |
+
"step": 150610
|
290991 |
+
},
|
290992 |
+
{
|
290993 |
+
"epoch": 1214.3,
|
290994 |
+
"learning_rate": 7.583909531502424e-06,
|
290995 |
+
"loss": 0.2725,
|
290996 |
+
"step": 150615
|
290997 |
+
},
|
290998 |
+
{
|
290999 |
+
"epoch": 1214.34,
|
291000 |
+
"learning_rate": 7.583828756058159e-06,
|
291001 |
+
"loss": 0.3622,
|
291002 |
+
"step": 150620
|
291003 |
+
},
|
291004 |
+
{
|
291005 |
+
"epoch": 1214.38,
|
291006 |
+
"learning_rate": 7.583747980613894e-06,
|
291007 |
+
"loss": 0.5381,
|
291008 |
+
"step": 150625
|
291009 |
+
},
|
291010 |
+
{
|
291011 |
+
"epoch": 1214.42,
|
291012 |
+
"learning_rate": 7.583667205169629e-06,
|
291013 |
+
"loss": 1.0317,
|
291014 |
+
"step": 150630
|
291015 |
+
},
|
291016 |
+
{
|
291017 |
+
"epoch": 1214.46,
|
291018 |
+
"learning_rate": 7.583586429725364e-06,
|
291019 |
+
"loss": 0.3174,
|
291020 |
+
"step": 150635
|
291021 |
+
},
|
291022 |
+
{
|
291023 |
+
"epoch": 1214.5,
|
291024 |
+
"learning_rate": 7.5835056542811e-06,
|
291025 |
+
"loss": 1.059,
|
291026 |
+
"step": 150640
|
291027 |
+
},
|
291028 |
+
{
|
291029 |
+
"epoch": 1214.54,
|
291030 |
+
"learning_rate": 7.583424878836835e-06,
|
291031 |
+
"loss": 0.355,
|
291032 |
+
"step": 150645
|
291033 |
+
},
|
291034 |
+
{
|
291035 |
+
"epoch": 1214.58,
|
291036 |
+
"learning_rate": 7.58334410339257e-06,
|
291037 |
+
"loss": 0.6424,
|
291038 |
+
"step": 150650
|
291039 |
+
},
|
291040 |
+
{
|
291041 |
+
"epoch": 1214.62,
|
291042 |
+
"learning_rate": 7.5832633279483046e-06,
|
291043 |
+
"loss": 1.0002,
|
291044 |
+
"step": 150655
|
291045 |
+
},
|
291046 |
+
{
|
291047 |
+
"epoch": 1214.66,
|
291048 |
+
"learning_rate": 7.5831825525040396e-06,
|
291049 |
+
"loss": 0.2473,
|
291050 |
+
"step": 150660
|
291051 |
+
},
|
291052 |
+
{
|
291053 |
+
"epoch": 1214.7,
|
291054 |
+
"learning_rate": 7.5831017770597745e-06,
|
291055 |
+
"loss": 0.2639,
|
291056 |
+
"step": 150665
|
291057 |
+
},
|
291058 |
+
{
|
291059 |
+
"epoch": 1214.74,
|
291060 |
+
"learning_rate": 7.5830210016155095e-06,
|
291061 |
+
"loss": 0.3395,
|
291062 |
+
"step": 150670
|
291063 |
+
},
|
291064 |
+
{
|
291065 |
+
"epoch": 1214.78,
|
291066 |
+
"learning_rate": 7.5829402261712445e-06,
|
291067 |
+
"loss": 0.5267,
|
291068 |
+
"step": 150675
|
291069 |
+
},
|
291070 |
+
{
|
291071 |
+
"epoch": 1214.82,
|
291072 |
+
"learning_rate": 7.5828594507269795e-06,
|
291073 |
+
"loss": 1.1767,
|
291074 |
+
"step": 150680
|
291075 |
+
},
|
291076 |
+
{
|
291077 |
+
"epoch": 1214.86,
|
291078 |
+
"learning_rate": 7.5827786752827145e-06,
|
291079 |
+
"loss": 0.2499,
|
291080 |
+
"step": 150685
|
291081 |
+
},
|
291082 |
+
{
|
291083 |
+
"epoch": 1214.9,
|
291084 |
+
"learning_rate": 7.5826978998384495e-06,
|
291085 |
+
"loss": 0.3152,
|
291086 |
+
"step": 150690
|
291087 |
+
},
|
291088 |
+
{
|
291089 |
+
"epoch": 1214.94,
|
291090 |
+
"learning_rate": 7.5826171243941845e-06,
|
291091 |
+
"loss": 0.3273,
|
291092 |
+
"step": 150695
|
291093 |
+
},
|
291094 |
+
{
|
291095 |
+
"epoch": 1214.98,
|
291096 |
+
"learning_rate": 7.5825363489499195e-06,
|
291097 |
+
"loss": 0.5964,
|
291098 |
+
"step": 150700
|
291099 |
+
},
|
291100 |
+
{
|
291101 |
+
"epoch": 1215.0,
|
291102 |
+
"eval_loss": 0.3669835031032562,
|
291103 |
+
"eval_runtime": 42.421,
|
291104 |
+
"eval_samples_per_second": 19.707,
|
291105 |
+
"eval_steps_per_second": 0.636,
|
291106 |
+
"eval_wer": 0.16976108232584916,
|
291107 |
+
"step": 150702
|
291108 |
+
},
|
291109 |
+
{
|
291110 |
+
"epoch": 1205.02,
|
291111 |
+
"learning_rate": 7.5824555735056544e-06,
|
291112 |
+
"loss": 0.3077,
|
291113 |
+
"step": 150705
|
291114 |
+
},
|
291115 |
+
{
|
291116 |
+
"epoch": 1205.06,
|
291117 |
+
"learning_rate": 7.58237479806139e-06,
|
291118 |
+
"loss": 0.294,
|
291119 |
+
"step": 150710
|
291120 |
+
},
|
291121 |
+
{
|
291122 |
+
"epoch": 1205.1,
|
291123 |
+
"learning_rate": 7.582294022617125e-06,
|
291124 |
+
"loss": 0.3353,
|
291125 |
+
"step": 150715
|
291126 |
+
},
|
291127 |
+
{
|
291128 |
+
"epoch": 1205.14,
|
291129 |
+
"learning_rate": 7.58221324717286e-06,
|
291130 |
+
"loss": 0.3076,
|
291131 |
+
"step": 150720
|
291132 |
+
},
|
291133 |
+
{
|
291134 |
+
"epoch": 1205.18,
|
291135 |
+
"learning_rate": 7.582132471728595e-06,
|
291136 |
+
"loss": 0.6049,
|
291137 |
+
"step": 150725
|
291138 |
+
},
|
291139 |
+
{
|
291140 |
+
"epoch": 1205.22,
|
291141 |
+
"learning_rate": 7.58205169628433e-06,
|
291142 |
+
"loss": 0.8056,
|
291143 |
+
"step": 150730
|
291144 |
+
},
|
291145 |
+
{
|
291146 |
+
"epoch": 1205.26,
|
291147 |
+
"learning_rate": 7.581970920840065e-06,
|
291148 |
+
"loss": 0.2615,
|
291149 |
+
"step": 150735
|
291150 |
+
},
|
291151 |
+
{
|
291152 |
+
"epoch": 1205.3,
|
291153 |
+
"learning_rate": 7.5818901453958e-06,
|
291154 |
+
"loss": 0.2697,
|
291155 |
+
"step": 150740
|
291156 |
+
},
|
291157 |
+
{
|
291158 |
+
"epoch": 1205.34,
|
291159 |
+
"learning_rate": 7.581809369951535e-06,
|
291160 |
+
"loss": 0.3135,
|
291161 |
+
"step": 150745
|
291162 |
+
},
|
291163 |
+
{
|
291164 |
+
"epoch": 1205.38,
|
291165 |
+
"learning_rate": 7.58172859450727e-06,
|
291166 |
+
"loss": 0.7361,
|
291167 |
+
"step": 150750
|
291168 |
+
},
|
291169 |
+
{
|
291170 |
+
"epoch": 1205.42,
|
291171 |
+
"learning_rate": 7.581647819063005e-06,
|
291172 |
+
"loss": 0.778,
|
291173 |
+
"step": 150755
|
291174 |
+
},
|
291175 |
+
{
|
291176 |
+
"epoch": 1205.46,
|
291177 |
+
"learning_rate": 7.58156704361874e-06,
|
291178 |
+
"loss": 0.2723,
|
291179 |
+
"step": 150760
|
291180 |
+
},
|
291181 |
+
{
|
291182 |
+
"epoch": 1205.5,
|
291183 |
+
"learning_rate": 7.581486268174476e-06,
|
291184 |
+
"loss": 0.3067,
|
291185 |
+
"step": 150765
|
291186 |
+
},
|
291187 |
+
{
|
291188 |
+
"epoch": 1205.54,
|
291189 |
+
"learning_rate": 7.58140549273021e-06,
|
291190 |
+
"loss": 0.2966,
|
291191 |
+
"step": 150770
|
291192 |
+
},
|
291193 |
+
{
|
291194 |
+
"epoch": 1205.58,
|
291195 |
+
"learning_rate": 7.581324717285946e-06,
|
291196 |
+
"loss": 0.7155,
|
291197 |
+
"step": 150775
|
291198 |
+
},
|
291199 |
+
{
|
291200 |
+
"epoch": 1205.62,
|
291201 |
+
"learning_rate": 7.581243941841681e-06,
|
291202 |
+
"loss": 0.9073,
|
291203 |
+
"step": 150780
|
291204 |
+
},
|
291205 |
+
{
|
291206 |
+
"epoch": 1205.66,
|
291207 |
+
"learning_rate": 7.581163166397416e-06,
|
291208 |
+
"loss": 0.2664,
|
291209 |
+
"step": 150785
|
291210 |
+
},
|
291211 |
+
{
|
291212 |
+
"epoch": 1205.7,
|
291213 |
+
"learning_rate": 7.581082390953151e-06,
|
291214 |
+
"loss": 0.2695,
|
291215 |
+
"step": 150790
|
291216 |
+
},
|
291217 |
+
{
|
291218 |
+
"epoch": 1205.74,
|
291219 |
+
"learning_rate": 7.581001615508886e-06,
|
291220 |
+
"loss": 0.3299,
|
291221 |
+
"step": 150795
|
291222 |
+
},
|
291223 |
+
{
|
291224 |
+
"epoch": 1205.78,
|
291225 |
+
"learning_rate": 7.580920840064621e-06,
|
291226 |
+
"loss": 0.7078,
|
291227 |
+
"step": 150800
|
291228 |
+
},
|
291229 |
+
{
|
291230 |
+
"epoch": 1205.82,
|
291231 |
+
"learning_rate": 7.580840064620356e-06,
|
291232 |
+
"loss": 0.9233,
|
291233 |
+
"step": 150805
|
291234 |
+
},
|
291235 |
+
{
|
291236 |
+
"epoch": 1205.86,
|
291237 |
+
"learning_rate": 7.580759289176091e-06,
|
291238 |
+
"loss": 0.2724,
|
291239 |
+
"step": 150810
|
291240 |
+
},
|
291241 |
+
{
|
291242 |
+
"epoch": 1205.9,
|
291243 |
+
"learning_rate": 7.580678513731826e-06,
|
291244 |
+
"loss": 0.3251,
|
291245 |
+
"step": 150815
|
291246 |
+
},
|
291247 |
+
{
|
291248 |
+
"epoch": 1205.94,
|
291249 |
+
"learning_rate": 7.580597738287562e-06,
|
291250 |
+
"loss": 0.3044,
|
291251 |
+
"step": 150820
|
291252 |
+
},
|
291253 |
+
{
|
291254 |
+
"epoch": 1205.98,
|
291255 |
+
"learning_rate": 7.580516962843296e-06,
|
291256 |
+
"loss": 0.614,
|
291257 |
+
"step": 150825
|
291258 |
+
},
|
291259 |
+
{
|
291260 |
+
"epoch": 1206.0,
|
291261 |
+
"eval_loss": 0.3523024320602417,
|
291262 |
+
"eval_runtime": 40.2575,
|
291263 |
+
"eval_samples_per_second": 20.766,
|
291264 |
+
"eval_steps_per_second": 0.671,
|
291265 |
+
"eval_wer": 0.182829413960433,
|
291266 |
+
"step": 150827
|
291267 |
}
|
291268 |
],
|
291269 |
+
"max_steps": 625000,
|
291270 |
"num_train_epochs": 5000,
|
291271 |
+
"total_flos": 4.2444315493114164e+20,
|
291272 |
"trial_name": null,
|
291273 |
"trial_params": null
|
291274 |
}
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630244012.7530751/events.out.tfevents.1630244012.cc93b136ebf5.1086.179
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be26c9daead194841c91de81f386a51de6ca433a7e1908af359da8e170158788
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630244448.5792234/events.out.tfevents.1630244448.cc93b136ebf5.1086.181
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:52d266b5f028fc38c9d1a6a414c1c393506f6832953cb7488df1efa39b8bf893
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630244993.356596/events.out.tfevents.1630244993.cc93b136ebf5.1086.183
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db7a02b0a48cfebf6119ec88b66e2fb707cbbd8d04c1f4c7556763fea904dbd8
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630245427.5586958/events.out.tfevents.1630245427.cc93b136ebf5.1086.185
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c397d04b557890d4c7955c897cc3fe2286c2fbf5b49d17e2f205594fa1ca5726
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630245858.8737807/events.out.tfevents.1630245858.cc93b136ebf5.1086.187
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:816366c9a07099ac79c7220c098da88f70aabc2b8cfd69ef728187b81a69a4f1
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630244012.cc93b136ebf5.1086.178
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0aa3f0de0578fd1d28673ee4fdbb0ebfb66f41ea2387c6285d9df7e4d33ecad0
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630244448.cc93b136ebf5.1086.180
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e96fea5cd10e9f7da659de83310ebaee8c9de0e99e1f2c10547f556dcea96d12
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630244993.cc93b136ebf5.1086.182
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:afc99b9f992b08112289a38952d17cabf09acb3de0e8da1ccd48ac875887989b
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630245427.cc93b136ebf5.1086.184
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d7955bc7c27f04e99fcd1c98830cccfa12ac8c2b0edbaff0d6a30315c0b3e1d
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630245858.cc93b136ebf5.1086.186
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef4e6f78f20a1befa18e737a50ec4c4d90bf6a887d10552ee89424c85ea0395d
|
3 |
+
size 8622
|