"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630115798.4136324/events.out.tfevents.1630115798.86bb0ddabf9b.1042.41 +3 -0
- model-bin/finetune/base/log/1630116207.9281766/events.out.tfevents.1630116207.86bb0ddabf9b.1042.43 +3 -0
- model-bin/finetune/base/log/1630116621.3491957/events.out.tfevents.1630116621.86bb0ddabf9b.1042.45 +3 -0
- model-bin/finetune/base/log/1630117036.3824139/events.out.tfevents.1630117036.86bb0ddabf9b.1042.47 +3 -0
- model-bin/finetune/base/log/1630117449.777196/events.out.tfevents.1630117449.86bb0ddabf9b.1042.49 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630115798.86bb0ddabf9b.1042.40 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630116207.86bb0ddabf9b.1042.42 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630116621.86bb0ddabf9b.1042.44 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630117036.86bb0ddabf9b.1042.46 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630117449.86bb0ddabf9b.1042.48 +3 -0
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce19a2e5b74de571f58c07d682570d2210d3daa6eb857675fefc28ee8ffe8cfe
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe6f837af1e6098ca87218f9f6f489ef7326656822686625f997aa9e4da06051
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:310447b7017194990c17a691e647cf4393cfa4912178125648656448f89be377
|
3 |
+
size 14567
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1caec4dc20dfb31b5977c6031ec782c60b5f7b286f94be7ec1097ec7544efc9
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:747fd4894013b519f10367353486f379ccd36aaffe81e92b6be8bd41396f7093
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -247560,11 +247560,806 @@
|
|
247560 |
"eval_steps_per_second": 0.686,
|
247561 |
"eval_wer": 0.18822154026011984,
|
247562 |
"step": 116478
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
247563 |
}
|
247564 |
],
|
247565 |
"max_steps": 625000,
|
247566 |
"num_train_epochs": 5000,
|
247567 |
-
"total_flos": 3.
|
247568 |
"trial_name": null,
|
247569 |
"trial_params": null
|
247570 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 935.9960159362549,
|
5 |
+
"global_step": 117100,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
247560 |
"eval_steps_per_second": 0.686,
|
247561 |
"eval_wer": 0.18822154026011984,
|
247562 |
"step": 116478
|
247563 |
+
},
|
247564 |
+
{
|
247565 |
+
"epoch": 931.02,
|
247566 |
+
"learning_rate": 8.15011217948718e-06,
|
247567 |
+
"loss": 0.2933,
|
247568 |
+
"step": 116480
|
247569 |
+
},
|
247570 |
+
{
|
247571 |
+
"epoch": 931.06,
|
247572 |
+
"learning_rate": 8.150032051282052e-06,
|
247573 |
+
"loss": 0.292,
|
247574 |
+
"step": 116485
|
247575 |
+
},
|
247576 |
+
{
|
247577 |
+
"epoch": 931.1,
|
247578 |
+
"learning_rate": 8.149951923076923e-06,
|
247579 |
+
"loss": 0.2725,
|
247580 |
+
"step": 116490
|
247581 |
+
},
|
247582 |
+
{
|
247583 |
+
"epoch": 931.14,
|
247584 |
+
"learning_rate": 8.149871794871795e-06,
|
247585 |
+
"loss": 0.383,
|
247586 |
+
"step": 116495
|
247587 |
+
},
|
247588 |
+
{
|
247589 |
+
"epoch": 931.18,
|
247590 |
+
"learning_rate": 8.149791666666668e-06,
|
247591 |
+
"loss": 0.6222,
|
247592 |
+
"step": 116500
|
247593 |
+
},
|
247594 |
+
{
|
247595 |
+
"epoch": 931.22,
|
247596 |
+
"learning_rate": 8.14971153846154e-06,
|
247597 |
+
"loss": 0.9348,
|
247598 |
+
"step": 116505
|
247599 |
+
},
|
247600 |
+
{
|
247601 |
+
"epoch": 931.26,
|
247602 |
+
"learning_rate": 8.14963141025641e-06,
|
247603 |
+
"loss": 0.2812,
|
247604 |
+
"step": 116510
|
247605 |
+
},
|
247606 |
+
{
|
247607 |
+
"epoch": 931.3,
|
247608 |
+
"learning_rate": 8.149551282051282e-06,
|
247609 |
+
"loss": 0.2711,
|
247610 |
+
"step": 116515
|
247611 |
+
},
|
247612 |
+
{
|
247613 |
+
"epoch": 931.34,
|
247614 |
+
"learning_rate": 8.149471153846155e-06,
|
247615 |
+
"loss": 0.4673,
|
247616 |
+
"step": 116520
|
247617 |
+
},
|
247618 |
+
{
|
247619 |
+
"epoch": 931.38,
|
247620 |
+
"learning_rate": 8.149391025641026e-06,
|
247621 |
+
"loss": 0.7297,
|
247622 |
+
"step": 116525
|
247623 |
+
},
|
247624 |
+
{
|
247625 |
+
"epoch": 931.42,
|
247626 |
+
"learning_rate": 8.149310897435898e-06,
|
247627 |
+
"loss": 1.0439,
|
247628 |
+
"step": 116530
|
247629 |
+
},
|
247630 |
+
{
|
247631 |
+
"epoch": 931.46,
|
247632 |
+
"learning_rate": 8.14923076923077e-06,
|
247633 |
+
"loss": 0.3004,
|
247634 |
+
"step": 116535
|
247635 |
+
},
|
247636 |
+
{
|
247637 |
+
"epoch": 931.5,
|
247638 |
+
"learning_rate": 8.149150641025642e-06,
|
247639 |
+
"loss": 0.289,
|
247640 |
+
"step": 116540
|
247641 |
+
},
|
247642 |
+
{
|
247643 |
+
"epoch": 931.54,
|
247644 |
+
"learning_rate": 8.149070512820513e-06,
|
247645 |
+
"loss": 0.532,
|
247646 |
+
"step": 116545
|
247647 |
+
},
|
247648 |
+
{
|
247649 |
+
"epoch": 931.58,
|
247650 |
+
"learning_rate": 8.148990384615385e-06,
|
247651 |
+
"loss": 0.6062,
|
247652 |
+
"step": 116550
|
247653 |
+
},
|
247654 |
+
{
|
247655 |
+
"epoch": 931.62,
|
247656 |
+
"learning_rate": 8.148910256410258e-06,
|
247657 |
+
"loss": 1.037,
|
247658 |
+
"step": 116555
|
247659 |
+
},
|
247660 |
+
{
|
247661 |
+
"epoch": 931.66,
|
247662 |
+
"learning_rate": 8.14883012820513e-06,
|
247663 |
+
"loss": 0.2608,
|
247664 |
+
"step": 116560
|
247665 |
+
},
|
247666 |
+
{
|
247667 |
+
"epoch": 931.7,
|
247668 |
+
"learning_rate": 8.14875e-06,
|
247669 |
+
"loss": 0.2704,
|
247670 |
+
"step": 116565
|
247671 |
+
},
|
247672 |
+
{
|
247673 |
+
"epoch": 931.74,
|
247674 |
+
"learning_rate": 8.148669871794872e-06,
|
247675 |
+
"loss": 0.3608,
|
247676 |
+
"step": 116570
|
247677 |
+
},
|
247678 |
+
{
|
247679 |
+
"epoch": 931.78,
|
247680 |
+
"learning_rate": 8.148589743589745e-06,
|
247681 |
+
"loss": 0.615,
|
247682 |
+
"step": 116575
|
247683 |
+
},
|
247684 |
+
{
|
247685 |
+
"epoch": 931.82,
|
247686 |
+
"learning_rate": 8.148509615384616e-06,
|
247687 |
+
"loss": 0.9766,
|
247688 |
+
"step": 116580
|
247689 |
+
},
|
247690 |
+
{
|
247691 |
+
"epoch": 931.86,
|
247692 |
+
"learning_rate": 8.148429487179488e-06,
|
247693 |
+
"loss": 0.3059,
|
247694 |
+
"step": 116585
|
247695 |
+
},
|
247696 |
+
{
|
247697 |
+
"epoch": 931.9,
|
247698 |
+
"learning_rate": 8.14834935897436e-06,
|
247699 |
+
"loss": 0.2919,
|
247700 |
+
"step": 116590
|
247701 |
+
},
|
247702 |
+
{
|
247703 |
+
"epoch": 931.94,
|
247704 |
+
"learning_rate": 8.148269230769232e-06,
|
247705 |
+
"loss": 0.3559,
|
247706 |
+
"step": 116595
|
247707 |
+
},
|
247708 |
+
{
|
247709 |
+
"epoch": 931.98,
|
247710 |
+
"learning_rate": 8.148189102564103e-06,
|
247711 |
+
"loss": 0.7633,
|
247712 |
+
"step": 116600
|
247713 |
+
},
|
247714 |
+
{
|
247715 |
+
"epoch": 932.0,
|
247716 |
+
"eval_loss": 0.35194557905197144,
|
247717 |
+
"eval_runtime": 39.5509,
|
247718 |
+
"eval_samples_per_second": 21.314,
|
247719 |
+
"eval_steps_per_second": 0.683,
|
247720 |
+
"eval_wer": 0.18437974134816848,
|
247721 |
+
"step": 116603
|
247722 |
+
},
|
247723 |
+
{
|
247724 |
+
"epoch": 940.02,
|
247725 |
+
"learning_rate": 8.148108974358975e-06,
|
247726 |
+
"loss": 0.3737,
|
247727 |
+
"step": 116605
|
247728 |
+
},
|
247729 |
+
{
|
247730 |
+
"epoch": 940.06,
|
247731 |
+
"learning_rate": 8.148028846153848e-06,
|
247732 |
+
"loss": 0.2789,
|
247733 |
+
"step": 116610
|
247734 |
+
},
|
247735 |
+
{
|
247736 |
+
"epoch": 940.1,
|
247737 |
+
"learning_rate": 8.147948717948718e-06,
|
247738 |
+
"loss": 0.2851,
|
247739 |
+
"step": 116615
|
247740 |
+
},
|
247741 |
+
{
|
247742 |
+
"epoch": 940.14,
|
247743 |
+
"learning_rate": 8.14786858974359e-06,
|
247744 |
+
"loss": 0.3367,
|
247745 |
+
"step": 116620
|
247746 |
+
},
|
247747 |
+
{
|
247748 |
+
"epoch": 940.18,
|
247749 |
+
"learning_rate": 8.147788461538462e-06,
|
247750 |
+
"loss": 0.679,
|
247751 |
+
"step": 116625
|
247752 |
+
},
|
247753 |
+
{
|
247754 |
+
"epoch": 940.22,
|
247755 |
+
"learning_rate": 8.147708333333333e-06,
|
247756 |
+
"loss": 0.9981,
|
247757 |
+
"step": 116630
|
247758 |
+
},
|
247759 |
+
{
|
247760 |
+
"epoch": 940.26,
|
247761 |
+
"learning_rate": 8.147628205128206e-06,
|
247762 |
+
"loss": 0.3318,
|
247763 |
+
"step": 116635
|
247764 |
+
},
|
247765 |
+
{
|
247766 |
+
"epoch": 940.3,
|
247767 |
+
"learning_rate": 8.147548076923078e-06,
|
247768 |
+
"loss": 0.3239,
|
247769 |
+
"step": 116640
|
247770 |
+
},
|
247771 |
+
{
|
247772 |
+
"epoch": 940.34,
|
247773 |
+
"learning_rate": 8.147467948717949e-06,
|
247774 |
+
"loss": 0.3073,
|
247775 |
+
"step": 116645
|
247776 |
+
},
|
247777 |
+
{
|
247778 |
+
"epoch": 940.38,
|
247779 |
+
"learning_rate": 8.14738782051282e-06,
|
247780 |
+
"loss": 0.6717,
|
247781 |
+
"step": 116650
|
247782 |
+
},
|
247783 |
+
{
|
247784 |
+
"epoch": 940.42,
|
247785 |
+
"learning_rate": 8.147307692307693e-06,
|
247786 |
+
"loss": 1.1175,
|
247787 |
+
"step": 116655
|
247788 |
+
},
|
247789 |
+
{
|
247790 |
+
"epoch": 940.46,
|
247791 |
+
"learning_rate": 8.147227564102565e-06,
|
247792 |
+
"loss": 0.2822,
|
247793 |
+
"step": 116660
|
247794 |
+
},
|
247795 |
+
{
|
247796 |
+
"epoch": 940.5,
|
247797 |
+
"learning_rate": 8.147147435897436e-06,
|
247798 |
+
"loss": 0.2605,
|
247799 |
+
"step": 116665
|
247800 |
+
},
|
247801 |
+
{
|
247802 |
+
"epoch": 940.54,
|
247803 |
+
"learning_rate": 8.147067307692308e-06,
|
247804 |
+
"loss": 0.3236,
|
247805 |
+
"step": 116670
|
247806 |
+
},
|
247807 |
+
{
|
247808 |
+
"epoch": 940.58,
|
247809 |
+
"learning_rate": 8.14698717948718e-06,
|
247810 |
+
"loss": 0.6434,
|
247811 |
+
"step": 116675
|
247812 |
+
},
|
247813 |
+
{
|
247814 |
+
"epoch": 940.62,
|
247815 |
+
"learning_rate": 8.146907051282052e-06,
|
247816 |
+
"loss": 1.0242,
|
247817 |
+
"step": 116680
|
247818 |
+
},
|
247819 |
+
{
|
247820 |
+
"epoch": 940.66,
|
247821 |
+
"learning_rate": 8.146826923076923e-06,
|
247822 |
+
"loss": 0.3286,
|
247823 |
+
"step": 116685
|
247824 |
+
},
|
247825 |
+
{
|
247826 |
+
"epoch": 940.7,
|
247827 |
+
"learning_rate": 8.146746794871796e-06,
|
247828 |
+
"loss": 0.2694,
|
247829 |
+
"step": 116690
|
247830 |
+
},
|
247831 |
+
{
|
247832 |
+
"epoch": 940.74,
|
247833 |
+
"learning_rate": 8.146666666666668e-06,
|
247834 |
+
"loss": 0.2618,
|
247835 |
+
"step": 116695
|
247836 |
+
},
|
247837 |
+
{
|
247838 |
+
"epoch": 940.78,
|
247839 |
+
"learning_rate": 8.146586538461539e-06,
|
247840 |
+
"loss": 0.6336,
|
247841 |
+
"step": 116700
|
247842 |
+
},
|
247843 |
+
{
|
247844 |
+
"epoch": 940.82,
|
247845 |
+
"learning_rate": 8.14650641025641e-06,
|
247846 |
+
"loss": 0.9335,
|
247847 |
+
"step": 116705
|
247848 |
+
},
|
247849 |
+
{
|
247850 |
+
"epoch": 940.86,
|
247851 |
+
"learning_rate": 8.146426282051283e-06,
|
247852 |
+
"loss": 0.2617,
|
247853 |
+
"step": 116710
|
247854 |
+
},
|
247855 |
+
{
|
247856 |
+
"epoch": 940.9,
|
247857 |
+
"learning_rate": 8.146346153846155e-06,
|
247858 |
+
"loss": 0.3125,
|
247859 |
+
"step": 116715
|
247860 |
+
},
|
247861 |
+
{
|
247862 |
+
"epoch": 940.94,
|
247863 |
+
"learning_rate": 8.146266025641026e-06,
|
247864 |
+
"loss": 0.3524,
|
247865 |
+
"step": 116720
|
247866 |
+
},
|
247867 |
+
{
|
247868 |
+
"epoch": 940.98,
|
247869 |
+
"learning_rate": 8.146185897435898e-06,
|
247870 |
+
"loss": 0.6945,
|
247871 |
+
"step": 116725
|
247872 |
+
},
|
247873 |
+
{
|
247874 |
+
"epoch": 941.0,
|
247875 |
+
"eval_loss": 0.3652080297470093,
|
247876 |
+
"eval_runtime": 39.2072,
|
247877 |
+
"eval_samples_per_second": 21.501,
|
247878 |
+
"eval_steps_per_second": 0.689,
|
247879 |
+
"eval_wer": 0.18559156861329676,
|
247880 |
+
"step": 116727
|
247881 |
+
},
|
247882 |
+
{
|
247883 |
+
"epoch": 941.02,
|
247884 |
+
"learning_rate": 8.14610576923077e-06,
|
247885 |
+
"loss": 0.3172,
|
247886 |
+
"step": 116730
|
247887 |
+
},
|
247888 |
+
{
|
247889 |
+
"epoch": 941.06,
|
247890 |
+
"learning_rate": 8.146025641025642e-06,
|
247891 |
+
"loss": 0.2391,
|
247892 |
+
"step": 116735
|
247893 |
+
},
|
247894 |
+
{
|
247895 |
+
"epoch": 941.1,
|
247896 |
+
"learning_rate": 8.145945512820513e-06,
|
247897 |
+
"loss": 0.2877,
|
247898 |
+
"step": 116740
|
247899 |
+
},
|
247900 |
+
{
|
247901 |
+
"epoch": 941.14,
|
247902 |
+
"learning_rate": 8.145865384615386e-06,
|
247903 |
+
"loss": 0.3587,
|
247904 |
+
"step": 116745
|
247905 |
+
},
|
247906 |
+
{
|
247907 |
+
"epoch": 941.18,
|
247908 |
+
"learning_rate": 8.145785256410256e-06,
|
247909 |
+
"loss": 0.7811,
|
247910 |
+
"step": 116750
|
247911 |
+
},
|
247912 |
+
{
|
247913 |
+
"epoch": 941.22,
|
247914 |
+
"learning_rate": 8.145705128205129e-06,
|
247915 |
+
"loss": 0.7006,
|
247916 |
+
"step": 116755
|
247917 |
+
},
|
247918 |
+
{
|
247919 |
+
"epoch": 941.27,
|
247920 |
+
"learning_rate": 8.145625e-06,
|
247921 |
+
"loss": 0.327,
|
247922 |
+
"step": 116760
|
247923 |
+
},
|
247924 |
+
{
|
247925 |
+
"epoch": 941.31,
|
247926 |
+
"learning_rate": 8.145544871794872e-06,
|
247927 |
+
"loss": 0.3295,
|
247928 |
+
"step": 116765
|
247929 |
+
},
|
247930 |
+
{
|
247931 |
+
"epoch": 941.35,
|
247932 |
+
"learning_rate": 8.145464743589743e-06,
|
247933 |
+
"loss": 0.3501,
|
247934 |
+
"step": 116770
|
247935 |
+
},
|
247936 |
+
{
|
247937 |
+
"epoch": 941.39,
|
247938 |
+
"learning_rate": 8.145384615384616e-06,
|
247939 |
+
"loss": 0.9432,
|
247940 |
+
"step": 116775
|
247941 |
+
},
|
247942 |
+
{
|
247943 |
+
"epoch": 941.43,
|
247944 |
+
"learning_rate": 8.145304487179488e-06,
|
247945 |
+
"loss": 0.8745,
|
247946 |
+
"step": 116780
|
247947 |
+
},
|
247948 |
+
{
|
247949 |
+
"epoch": 941.47,
|
247950 |
+
"learning_rate": 8.145224358974359e-06,
|
247951 |
+
"loss": 0.2931,
|
247952 |
+
"step": 116785
|
247953 |
+
},
|
247954 |
+
{
|
247955 |
+
"epoch": 941.51,
|
247956 |
+
"learning_rate": 8.145144230769232e-06,
|
247957 |
+
"loss": 0.2615,
|
247958 |
+
"step": 116790
|
247959 |
+
},
|
247960 |
+
{
|
247961 |
+
"epoch": 941.55,
|
247962 |
+
"learning_rate": 8.145064102564103e-06,
|
247963 |
+
"loss": 0.3783,
|
247964 |
+
"step": 116795
|
247965 |
+
},
|
247966 |
+
{
|
247967 |
+
"epoch": 941.59,
|
247968 |
+
"learning_rate": 8.144983974358975e-06,
|
247969 |
+
"loss": 0.6995,
|
247970 |
+
"step": 116800
|
247971 |
+
},
|
247972 |
+
{
|
247973 |
+
"epoch": 941.63,
|
247974 |
+
"learning_rate": 8.144903846153846e-06,
|
247975 |
+
"loss": 0.8687,
|
247976 |
+
"step": 116805
|
247977 |
+
},
|
247978 |
+
{
|
247979 |
+
"epoch": 941.67,
|
247980 |
+
"learning_rate": 8.144823717948719e-06,
|
247981 |
+
"loss": 0.2578,
|
247982 |
+
"step": 116810
|
247983 |
+
},
|
247984 |
+
{
|
247985 |
+
"epoch": 941.71,
|
247986 |
+
"learning_rate": 8.14474358974359e-06,
|
247987 |
+
"loss": 0.2837,
|
247988 |
+
"step": 116815
|
247989 |
+
},
|
247990 |
+
{
|
247991 |
+
"epoch": 941.75,
|
247992 |
+
"learning_rate": 8.144663461538462e-06,
|
247993 |
+
"loss": 0.3907,
|
247994 |
+
"step": 116820
|
247995 |
+
},
|
247996 |
+
{
|
247997 |
+
"epoch": 941.79,
|
247998 |
+
"learning_rate": 8.144583333333333e-06,
|
247999 |
+
"loss": 0.6813,
|
248000 |
+
"step": 116825
|
248001 |
+
},
|
248002 |
+
{
|
248003 |
+
"epoch": 941.83,
|
248004 |
+
"learning_rate": 8.144503205128206e-06,
|
248005 |
+
"loss": 0.8454,
|
248006 |
+
"step": 116830
|
248007 |
+
},
|
248008 |
+
{
|
248009 |
+
"epoch": 941.87,
|
248010 |
+
"learning_rate": 8.144423076923078e-06,
|
248011 |
+
"loss": 0.2834,
|
248012 |
+
"step": 116835
|
248013 |
+
},
|
248014 |
+
{
|
248015 |
+
"epoch": 941.91,
|
248016 |
+
"learning_rate": 8.144342948717949e-06,
|
248017 |
+
"loss": 0.3404,
|
248018 |
+
"step": 116840
|
248019 |
+
},
|
248020 |
+
{
|
248021 |
+
"epoch": 941.95,
|
248022 |
+
"learning_rate": 8.144262820512822e-06,
|
248023 |
+
"loss": 0.3596,
|
248024 |
+
"step": 116845
|
248025 |
+
},
|
248026 |
+
{
|
248027 |
+
"epoch": 941.99,
|
248028 |
+
"learning_rate": 8.144182692307693e-06,
|
248029 |
+
"loss": 0.8048,
|
248030 |
+
"step": 116850
|
248031 |
+
},
|
248032 |
+
{
|
248033 |
+
"epoch": 942.0,
|
248034 |
+
"eval_loss": 0.37358829379081726,
|
248035 |
+
"eval_runtime": 38.8227,
|
248036 |
+
"eval_samples_per_second": 21.714,
|
248037 |
+
"eval_steps_per_second": 0.695,
|
248038 |
+
"eval_wer": 0.181710652768537,
|
248039 |
+
"step": 116851
|
248040 |
+
},
|
248041 |
+
{
|
248042 |
+
"epoch": 942.03,
|
248043 |
+
"learning_rate": 8.144102564102565e-06,
|
248044 |
+
"loss": 0.3532,
|
248045 |
+
"step": 116855
|
248046 |
+
},
|
248047 |
+
{
|
248048 |
+
"epoch": 942.07,
|
248049 |
+
"learning_rate": 8.144022435897436e-06,
|
248050 |
+
"loss": 0.2954,
|
248051 |
+
"step": 116860
|
248052 |
+
},
|
248053 |
+
{
|
248054 |
+
"epoch": 942.11,
|
248055 |
+
"learning_rate": 8.14394230769231e-06,
|
248056 |
+
"loss": 0.321,
|
248057 |
+
"step": 116865
|
248058 |
+
},
|
248059 |
+
{
|
248060 |
+
"epoch": 942.15,
|
248061 |
+
"learning_rate": 8.14386217948718e-06,
|
248062 |
+
"loss": 0.4422,
|
248063 |
+
"step": 116870
|
248064 |
+
},
|
248065 |
+
{
|
248066 |
+
"epoch": 942.19,
|
248067 |
+
"learning_rate": 8.143782051282052e-06,
|
248068 |
+
"loss": 0.8507,
|
248069 |
+
"step": 116875
|
248070 |
+
},
|
248071 |
+
{
|
248072 |
+
"epoch": 942.23,
|
248073 |
+
"learning_rate": 8.143701923076923e-06,
|
248074 |
+
"loss": 0.6194,
|
248075 |
+
"step": 116880
|
248076 |
+
},
|
248077 |
+
{
|
248078 |
+
"epoch": 942.27,
|
248079 |
+
"learning_rate": 8.143621794871796e-06,
|
248080 |
+
"loss": 0.3064,
|
248081 |
+
"step": 116885
|
248082 |
+
},
|
248083 |
+
{
|
248084 |
+
"epoch": 942.31,
|
248085 |
+
"learning_rate": 8.143541666666668e-06,
|
248086 |
+
"loss": 0.2845,
|
248087 |
+
"step": 116890
|
248088 |
+
},
|
248089 |
+
{
|
248090 |
+
"epoch": 942.35,
|
248091 |
+
"learning_rate": 8.143461538461539e-06,
|
248092 |
+
"loss": 0.3395,
|
248093 |
+
"step": 116895
|
248094 |
+
},
|
248095 |
+
{
|
248096 |
+
"epoch": 942.39,
|
248097 |
+
"learning_rate": 8.143381410256412e-06,
|
248098 |
+
"loss": 0.8709,
|
248099 |
+
"step": 116900
|
248100 |
+
},
|
248101 |
+
{
|
248102 |
+
"epoch": 942.43,
|
248103 |
+
"learning_rate": 8.143301282051282e-06,
|
248104 |
+
"loss": 0.5708,
|
248105 |
+
"step": 116905
|
248106 |
+
},
|
248107 |
+
{
|
248108 |
+
"epoch": 942.47,
|
248109 |
+
"learning_rate": 8.143221153846155e-06,
|
248110 |
+
"loss": 0.257,
|
248111 |
+
"step": 116910
|
248112 |
+
},
|
248113 |
+
{
|
248114 |
+
"epoch": 942.51,
|
248115 |
+
"learning_rate": 8.143141025641026e-06,
|
248116 |
+
"loss": 0.2881,
|
248117 |
+
"step": 116915
|
248118 |
+
},
|
248119 |
+
{
|
248120 |
+
"epoch": 942.55,
|
248121 |
+
"learning_rate": 8.143060897435897e-06,
|
248122 |
+
"loss": 0.3739,
|
248123 |
+
"step": 116920
|
248124 |
+
},
|
248125 |
+
{
|
248126 |
+
"epoch": 942.59,
|
248127 |
+
"learning_rate": 8.142980769230769e-06,
|
248128 |
+
"loss": 1.091,
|
248129 |
+
"step": 116925
|
248130 |
+
},
|
248131 |
+
{
|
248132 |
+
"epoch": 942.63,
|
248133 |
+
"learning_rate": 8.142900641025642e-06,
|
248134 |
+
"loss": 0.571,
|
248135 |
+
"step": 116930
|
248136 |
+
},
|
248137 |
+
{
|
248138 |
+
"epoch": 942.67,
|
248139 |
+
"learning_rate": 8.142820512820513e-06,
|
248140 |
+
"loss": 0.2523,
|
248141 |
+
"step": 116935
|
248142 |
+
},
|
248143 |
+
{
|
248144 |
+
"epoch": 942.71,
|
248145 |
+
"learning_rate": 8.142740384615385e-06,
|
248146 |
+
"loss": 0.3116,
|
248147 |
+
"step": 116940
|
248148 |
+
},
|
248149 |
+
{
|
248150 |
+
"epoch": 942.76,
|
248151 |
+
"learning_rate": 8.142660256410258e-06,
|
248152 |
+
"loss": 0.4294,
|
248153 |
+
"step": 116945
|
248154 |
+
},
|
248155 |
+
{
|
248156 |
+
"epoch": 942.8,
|
248157 |
+
"learning_rate": 8.142580128205129e-06,
|
248158 |
+
"loss": 0.8474,
|
248159 |
+
"step": 116950
|
248160 |
+
},
|
248161 |
+
{
|
248162 |
+
"epoch": 942.84,
|
248163 |
+
"learning_rate": 8.1425e-06,
|
248164 |
+
"loss": 0.6085,
|
248165 |
+
"step": 116955
|
248166 |
+
},
|
248167 |
+
{
|
248168 |
+
"epoch": 942.88,
|
248169 |
+
"learning_rate": 8.142419871794872e-06,
|
248170 |
+
"loss": 0.4619,
|
248171 |
+
"step": 116960
|
248172 |
+
},
|
248173 |
+
{
|
248174 |
+
"epoch": 942.92,
|
248175 |
+
"learning_rate": 8.142339743589745e-06,
|
248176 |
+
"loss": 0.4268,
|
248177 |
+
"step": 116965
|
248178 |
+
},
|
248179 |
+
{
|
248180 |
+
"epoch": 942.96,
|
248181 |
+
"learning_rate": 8.142259615384616e-06,
|
248182 |
+
"loss": 0.4165,
|
248183 |
+
"step": 116970
|
248184 |
+
},
|
248185 |
+
{
|
248186 |
+
"epoch": 943.0,
|
248187 |
+
"learning_rate": 8.142179487179488e-06,
|
248188 |
+
"loss": 1.148,
|
248189 |
+
"step": 116975
|
248190 |
+
},
|
248191 |
+
{
|
248192 |
+
"epoch": 943.0,
|
248193 |
+
"eval_loss": 0.3594907820224762,
|
248194 |
+
"eval_runtime": 39.3165,
|
248195 |
+
"eval_samples_per_second": 21.441,
|
248196 |
+
"eval_steps_per_second": 0.687,
|
248197 |
+
"eval_wer": 0.18600070348223707,
|
248198 |
+
"step": 116975
|
248199 |
+
},
|
248200 |
+
{
|
248201 |
+
"epoch": 935.04,
|
248202 |
+
"learning_rate": 8.142099358974359e-06,
|
248203 |
+
"loss": 0.323,
|
248204 |
+
"step": 116980
|
248205 |
+
},
|
248206 |
+
{
|
248207 |
+
"epoch": 935.08,
|
248208 |
+
"learning_rate": 8.142019230769232e-06,
|
248209 |
+
"loss": 0.284,
|
248210 |
+
"step": 116985
|
248211 |
+
},
|
248212 |
+
{
|
248213 |
+
"epoch": 935.12,
|
248214 |
+
"learning_rate": 8.141939102564103e-06,
|
248215 |
+
"loss": 0.2959,
|
248216 |
+
"step": 116990
|
248217 |
+
},
|
248218 |
+
{
|
248219 |
+
"epoch": 935.16,
|
248220 |
+
"learning_rate": 8.141858974358975e-06,
|
248221 |
+
"loss": 0.4164,
|
248222 |
+
"step": 116995
|
248223 |
+
},
|
248224 |
+
{
|
248225 |
+
"epoch": 935.2,
|
248226 |
+
"learning_rate": 8.141778846153848e-06,
|
248227 |
+
"loss": 1.1566,
|
248228 |
+
"step": 117000
|
248229 |
+
},
|
248230 |
+
{
|
248231 |
+
"epoch": 935.24,
|
248232 |
+
"learning_rate": 8.141698717948719e-06,
|
248233 |
+
"loss": 0.3514,
|
248234 |
+
"step": 117005
|
248235 |
+
},
|
248236 |
+
{
|
248237 |
+
"epoch": 935.28,
|
248238 |
+
"learning_rate": 8.14161858974359e-06,
|
248239 |
+
"loss": 0.2887,
|
248240 |
+
"step": 117010
|
248241 |
+
},
|
248242 |
+
{
|
248243 |
+
"epoch": 935.32,
|
248244 |
+
"learning_rate": 8.141538461538462e-06,
|
248245 |
+
"loss": 0.3248,
|
248246 |
+
"step": 117015
|
248247 |
+
},
|
248248 |
+
{
|
248249 |
+
"epoch": 935.36,
|
248250 |
+
"learning_rate": 8.141458333333335e-06,
|
248251 |
+
"loss": 0.5457,
|
248252 |
+
"step": 117020
|
248253 |
+
},
|
248254 |
+
{
|
248255 |
+
"epoch": 935.4,
|
248256 |
+
"learning_rate": 8.141378205128204e-06,
|
248257 |
+
"loss": 1.2681,
|
248258 |
+
"step": 117025
|
248259 |
+
},
|
248260 |
+
{
|
248261 |
+
"epoch": 935.44,
|
248262 |
+
"learning_rate": 8.141298076923078e-06,
|
248263 |
+
"loss": 0.2972,
|
248264 |
+
"step": 117030
|
248265 |
+
},
|
248266 |
+
{
|
248267 |
+
"epoch": 935.48,
|
248268 |
+
"learning_rate": 8.14121794871795e-06,
|
248269 |
+
"loss": 0.2916,
|
248270 |
+
"step": 117035
|
248271 |
+
},
|
248272 |
+
{
|
248273 |
+
"epoch": 935.52,
|
248274 |
+
"learning_rate": 8.14113782051282e-06,
|
248275 |
+
"loss": 0.3855,
|
248276 |
+
"step": 117040
|
248277 |
+
},
|
248278 |
+
{
|
248279 |
+
"epoch": 935.56,
|
248280 |
+
"learning_rate": 8.141057692307693e-06,
|
248281 |
+
"loss": 0.3866,
|
248282 |
+
"step": 117045
|
248283 |
+
},
|
248284 |
+
{
|
248285 |
+
"epoch": 935.6,
|
248286 |
+
"learning_rate": 8.140977564102565e-06,
|
248287 |
+
"loss": 1.0996,
|
248288 |
+
"step": 117050
|
248289 |
+
},
|
248290 |
+
{
|
248291 |
+
"epoch": 935.64,
|
248292 |
+
"learning_rate": 8.140897435897436e-06,
|
248293 |
+
"loss": 0.3514,
|
248294 |
+
"step": 117055
|
248295 |
+
},
|
248296 |
+
{
|
248297 |
+
"epoch": 935.68,
|
248298 |
+
"learning_rate": 8.140817307692307e-06,
|
248299 |
+
"loss": 0.2344,
|
248300 |
+
"step": 117060
|
248301 |
+
},
|
248302 |
+
{
|
248303 |
+
"epoch": 935.72,
|
248304 |
+
"learning_rate": 8.14073717948718e-06,
|
248305 |
+
"loss": 0.3085,
|
248306 |
+
"step": 117065
|
248307 |
+
},
|
248308 |
+
{
|
248309 |
+
"epoch": 935.76,
|
248310 |
+
"learning_rate": 8.140657051282052e-06,
|
248311 |
+
"loss": 0.413,
|
248312 |
+
"step": 117070
|
248313 |
+
},
|
248314 |
+
{
|
248315 |
+
"epoch": 935.8,
|
248316 |
+
"learning_rate": 8.140576923076923e-06,
|
248317 |
+
"loss": 1.332,
|
248318 |
+
"step": 117075
|
248319 |
+
},
|
248320 |
+
{
|
248321 |
+
"epoch": 935.84,
|
248322 |
+
"learning_rate": 8.140496794871795e-06,
|
248323 |
+
"loss": 0.301,
|
248324 |
+
"step": 117080
|
248325 |
+
},
|
248326 |
+
{
|
248327 |
+
"epoch": 935.88,
|
248328 |
+
"learning_rate": 8.140416666666668e-06,
|
248329 |
+
"loss": 0.2786,
|
248330 |
+
"step": 117085
|
248331 |
+
},
|
248332 |
+
{
|
248333 |
+
"epoch": 935.92,
|
248334 |
+
"learning_rate": 8.140336538461539e-06,
|
248335 |
+
"loss": 0.3301,
|
248336 |
+
"step": 117090
|
248337 |
+
},
|
248338 |
+
{
|
248339 |
+
"epoch": 935.96,
|
248340 |
+
"learning_rate": 8.14025641025641e-06,
|
248341 |
+
"loss": 0.557,
|
248342 |
+
"step": 117095
|
248343 |
+
},
|
248344 |
+
{
|
248345 |
+
"epoch": 936.0,
|
248346 |
+
"learning_rate": 8.140176282051283e-06,
|
248347 |
+
"loss": 1.4011,
|
248348 |
+
"step": 117100
|
248349 |
+
},
|
248350 |
+
{
|
248351 |
+
"epoch": 936.0,
|
248352 |
+
"eval_loss": 0.34906426072120667,
|
248353 |
+
"eval_runtime": 39.7868,
|
248354 |
+
"eval_samples_per_second": 21.188,
|
248355 |
+
"eval_steps_per_second": 0.679,
|
248356 |
+
"eval_wer": 0.1851399856424982,
|
248357 |
+
"step": 117100
|
248358 |
}
|
248359 |
],
|
248360 |
"max_steps": 625000,
|
248361 |
"num_train_epochs": 5000,
|
248362 |
+
"total_flos": 3.29550208819677e+20,
|
248363 |
"trial_name": null,
|
248364 |
"trial_params": null
|
248365 |
}
|
model-bin/finetune/base/{checkpoint-116478 β checkpoint-117100}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630115798.4136324/events.out.tfevents.1630115798.86bb0ddabf9b.1042.41
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1b074a755f2daa10761fb515413ac8941c898fbb3c07592e5fda37594331551
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630116207.9281766/events.out.tfevents.1630116207.86bb0ddabf9b.1042.43
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:68f3715bd0ab3d6f1ef9528a3b33c373e6085f271f727d16032a3c8f800f5c32
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630116621.3491957/events.out.tfevents.1630116621.86bb0ddabf9b.1042.45
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b61e75b6b5eedd92c5c3ec7d2c6390e778c14598efdc20e5064d0427e336885a
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630117036.3824139/events.out.tfevents.1630117036.86bb0ddabf9b.1042.47
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88ab210bdc5285d925d2e47a8308f6851a592c9f596915f05ffec0a75f9de165
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630117449.777196/events.out.tfevents.1630117449.86bb0ddabf9b.1042.49
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c8abf16a78934443be86f9165efeb04aef6c69ed81bc56bc9d7f82f4abe26ba9
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630115798.86bb0ddabf9b.1042.40
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11e43862968b4f8a128db20b0bdfb5c095f3a586a6dd95906ead603ea7a306a3
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630116207.86bb0ddabf9b.1042.42
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b0d26ec2fa4bd9832fa8d496a62c8dded8ab637811b0ff8e9731ec6b08accc1
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630116621.86bb0ddabf9b.1042.44
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe9d89f47418ab426a2df3d71b21fbae52017b55817c17fbc00d6fd3dfbf7344
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630117036.86bb0ddabf9b.1042.46
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ace9ba494ce37d956652650695ce1ffb4ec148d656c10db037ae03ec311edb80
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630117449.86bb0ddabf9b.1042.48
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:251d71865605240570d7ac2db4f9cd67a49e2ad8ffbe1d36f89a65c721298fec
|
3 |
+
size 8622
|