"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629872387.216304/events.out.tfevents.1629872387.7e498afd5545.905.63 +3 -0
- model-bin/finetune/base/log/1629873040.9577622/events.out.tfevents.1629873040.7e498afd5545.905.65 +3 -0
- model-bin/finetune/base/log/1629873815.6117036/events.out.tfevents.1629873815.7e498afd5545.905.67 +3 -0
- model-bin/finetune/base/log/1629874460.1476715/events.out.tfevents.1629874460.7e498afd5545.905.69 +3 -0
- model-bin/finetune/base/log/1629875104.9909084/events.out.tfevents.1629875104.7e498afd5545.905.71 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629872387.7e498afd5545.905.62 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629873040.7e498afd5545.905.64 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629873815.7e498afd5545.905.66 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629874460.7e498afd5545.905.68 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629875104.7e498afd5545.905.70 +3 -0
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfd36b6ee499d562847b385dc3ea784f1e658ef42e11f2762942b6000097d6a6
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:633059da288f26b3e326a5988ea125d0c4538291da52f14c3625a68fb22478fa
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e563717b5c623c6ea44320c2c602e0fe01a69436b866f3fa268004b71402a438
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb98e84d72fc9445ed9adb6eb66f1ee1e64e6fd672c8cf163f23ed10f15dd0e7
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d5509a63f3d04d5bbd7c28685f03d0c9d6c508593d58909a296e0f060d7e0530
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.18412114350410416,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -192615,11 +192615,806 @@
|
|
192615 |
"eval_steps_per_second": 0.654,
|
192616 |
"eval_wer": 0.19331724793448857,
|
192617 |
"step": 73294
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
192618 |
}
|
192619 |
],
|
192620 |
-
"max_steps":
|
192621 |
"num_train_epochs": 5000,
|
192622 |
-
"total_flos": 2.
|
192623 |
"trial_name": null,
|
192624 |
"trial_params": null
|
192625 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.18412114350410416,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
4 |
+
"epoch": 595.995983935743,
|
5 |
+
"global_step": 73916,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
192615 |
"eval_steps_per_second": 0.654,
|
192616 |
"eval_wer": 0.19331724793448857,
|
192617 |
"step": 73294
|
192618 |
+
},
|
192619 |
+
{
|
192620 |
+
"epoch": 591.01,
|
192621 |
+
"learning_rate": 8.841858974358975e-06,
|
192622 |
+
"loss": 0.3308,
|
192623 |
+
"step": 73295
|
192624 |
+
},
|
192625 |
+
{
|
192626 |
+
"epoch": 591.05,
|
192627 |
+
"learning_rate": 8.841778846153846e-06,
|
192628 |
+
"loss": 0.313,
|
192629 |
+
"step": 73300
|
192630 |
+
},
|
192631 |
+
{
|
192632 |
+
"epoch": 591.09,
|
192633 |
+
"learning_rate": 8.841698717948719e-06,
|
192634 |
+
"loss": 0.279,
|
192635 |
+
"step": 73305
|
192636 |
+
},
|
192637 |
+
{
|
192638 |
+
"epoch": 591.13,
|
192639 |
+
"learning_rate": 8.84161858974359e-06,
|
192640 |
+
"loss": 0.3329,
|
192641 |
+
"step": 73310
|
192642 |
+
},
|
192643 |
+
{
|
192644 |
+
"epoch": 591.17,
|
192645 |
+
"learning_rate": 8.841538461538462e-06,
|
192646 |
+
"loss": 0.6186,
|
192647 |
+
"step": 73315
|
192648 |
+
},
|
192649 |
+
{
|
192650 |
+
"epoch": 591.21,
|
192651 |
+
"learning_rate": 8.841458333333333e-06,
|
192652 |
+
"loss": 1.1625,
|
192653 |
+
"step": 73320
|
192654 |
+
},
|
192655 |
+
{
|
192656 |
+
"epoch": 591.25,
|
192657 |
+
"learning_rate": 8.841378205128206e-06,
|
192658 |
+
"loss": 0.5247,
|
192659 |
+
"step": 73325
|
192660 |
+
},
|
192661 |
+
{
|
192662 |
+
"epoch": 591.29,
|
192663 |
+
"learning_rate": 8.841298076923078e-06,
|
192664 |
+
"loss": 0.2802,
|
192665 |
+
"step": 73330
|
192666 |
+
},
|
192667 |
+
{
|
192668 |
+
"epoch": 591.33,
|
192669 |
+
"learning_rate": 8.841217948717949e-06,
|
192670 |
+
"loss": 0.3827,
|
192671 |
+
"step": 73335
|
192672 |
+
},
|
192673 |
+
{
|
192674 |
+
"epoch": 591.37,
|
192675 |
+
"learning_rate": 8.841137820512822e-06,
|
192676 |
+
"loss": 0.5569,
|
192677 |
+
"step": 73340
|
192678 |
+
},
|
192679 |
+
{
|
192680 |
+
"epoch": 591.41,
|
192681 |
+
"learning_rate": 8.841057692307693e-06,
|
192682 |
+
"loss": 1.2389,
|
192683 |
+
"step": 73345
|
192684 |
+
},
|
192685 |
+
{
|
192686 |
+
"epoch": 591.45,
|
192687 |
+
"learning_rate": 8.840977564102565e-06,
|
192688 |
+
"loss": 0.3074,
|
192689 |
+
"step": 73350
|
192690 |
+
},
|
192691 |
+
{
|
192692 |
+
"epoch": 591.49,
|
192693 |
+
"learning_rate": 8.840897435897436e-06,
|
192694 |
+
"loss": 0.2857,
|
192695 |
+
"step": 73355
|
192696 |
+
},
|
192697 |
+
{
|
192698 |
+
"epoch": 591.53,
|
192699 |
+
"learning_rate": 8.84081730769231e-06,
|
192700 |
+
"loss": 0.3403,
|
192701 |
+
"step": 73360
|
192702 |
+
},
|
192703 |
+
{
|
192704 |
+
"epoch": 591.57,
|
192705 |
+
"learning_rate": 8.84073717948718e-06,
|
192706 |
+
"loss": 0.5408,
|
192707 |
+
"step": 73365
|
192708 |
+
},
|
192709 |
+
{
|
192710 |
+
"epoch": 591.61,
|
192711 |
+
"learning_rate": 8.840657051282052e-06,
|
192712 |
+
"loss": 1.0292,
|
192713 |
+
"step": 73370
|
192714 |
+
},
|
192715 |
+
{
|
192716 |
+
"epoch": 591.65,
|
192717 |
+
"learning_rate": 8.840576923076923e-06,
|
192718 |
+
"loss": 0.3484,
|
192719 |
+
"step": 73375
|
192720 |
+
},
|
192721 |
+
{
|
192722 |
+
"epoch": 591.69,
|
192723 |
+
"learning_rate": 8.840496794871796e-06,
|
192724 |
+
"loss": 0.3111,
|
192725 |
+
"step": 73380
|
192726 |
+
},
|
192727 |
+
{
|
192728 |
+
"epoch": 591.73,
|
192729 |
+
"learning_rate": 8.840416666666668e-06,
|
192730 |
+
"loss": 0.331,
|
192731 |
+
"step": 73385
|
192732 |
+
},
|
192733 |
+
{
|
192734 |
+
"epoch": 591.77,
|
192735 |
+
"learning_rate": 8.840336538461539e-06,
|
192736 |
+
"loss": 0.6024,
|
192737 |
+
"step": 73390
|
192738 |
+
},
|
192739 |
+
{
|
192740 |
+
"epoch": 591.81,
|
192741 |
+
"learning_rate": 8.840256410256412e-06,
|
192742 |
+
"loss": 1.2207,
|
192743 |
+
"step": 73395
|
192744 |
+
},
|
192745 |
+
{
|
192746 |
+
"epoch": 591.85,
|
192747 |
+
"learning_rate": 8.840176282051283e-06,
|
192748 |
+
"loss": 0.3448,
|
192749 |
+
"step": 73400
|
192750 |
+
},
|
192751 |
+
{
|
192752 |
+
"epoch": 591.9,
|
192753 |
+
"learning_rate": 8.840096153846155e-06,
|
192754 |
+
"loss": 0.3189,
|
192755 |
+
"step": 73405
|
192756 |
+
},
|
192757 |
+
{
|
192758 |
+
"epoch": 591.94,
|
192759 |
+
"learning_rate": 8.840016025641026e-06,
|
192760 |
+
"loss": 0.3932,
|
192761 |
+
"step": 73410
|
192762 |
+
},
|
192763 |
+
{
|
192764 |
+
"epoch": 591.98,
|
192765 |
+
"learning_rate": 8.8399358974359e-06,
|
192766 |
+
"loss": 0.6275,
|
192767 |
+
"step": 73415
|
192768 |
+
},
|
192769 |
+
{
|
192770 |
+
"epoch": 592.0,
|
192771 |
+
"eval_loss": 0.3852952718734741,
|
192772 |
+
"eval_runtime": 40.5465,
|
192773 |
+
"eval_samples_per_second": 20.717,
|
192774 |
+
"eval_steps_per_second": 0.666,
|
192775 |
+
"eval_wer": 0.18575874333135745,
|
192776 |
+
"step": 73418
|
192777 |
+
},
|
192778 |
+
{
|
192779 |
+
"epoch": 587.02,
|
192780 |
+
"learning_rate": 8.839855769230769e-06,
|
192781 |
+
"loss": 0.4346,
|
192782 |
+
"step": 73420
|
192783 |
+
},
|
192784 |
+
{
|
192785 |
+
"epoch": 587.06,
|
192786 |
+
"learning_rate": 8.839775641025642e-06,
|
192787 |
+
"loss": 0.3137,
|
192788 |
+
"step": 73425
|
192789 |
+
},
|
192790 |
+
{
|
192791 |
+
"epoch": 587.1,
|
192792 |
+
"learning_rate": 8.839695512820513e-06,
|
192793 |
+
"loss": 0.3072,
|
192794 |
+
"step": 73430
|
192795 |
+
},
|
192796 |
+
{
|
192797 |
+
"epoch": 587.14,
|
192798 |
+
"learning_rate": 8.839615384615385e-06,
|
192799 |
+
"loss": 0.3478,
|
192800 |
+
"step": 73435
|
192801 |
+
},
|
192802 |
+
{
|
192803 |
+
"epoch": 587.18,
|
192804 |
+
"learning_rate": 8.839535256410258e-06,
|
192805 |
+
"loss": 0.5605,
|
192806 |
+
"step": 73440
|
192807 |
+
},
|
192808 |
+
{
|
192809 |
+
"epoch": 587.22,
|
192810 |
+
"learning_rate": 8.839455128205129e-06,
|
192811 |
+
"loss": 1.1978,
|
192812 |
+
"step": 73445
|
192813 |
+
},
|
192814 |
+
{
|
192815 |
+
"epoch": 587.26,
|
192816 |
+
"learning_rate": 8.839375e-06,
|
192817 |
+
"loss": 0.308,
|
192818 |
+
"step": 73450
|
192819 |
+
},
|
192820 |
+
{
|
192821 |
+
"epoch": 587.3,
|
192822 |
+
"learning_rate": 8.839294871794872e-06,
|
192823 |
+
"loss": 0.2844,
|
192824 |
+
"step": 73455
|
192825 |
+
},
|
192826 |
+
{
|
192827 |
+
"epoch": 587.34,
|
192828 |
+
"learning_rate": 8.839214743589745e-06,
|
192829 |
+
"loss": 0.3765,
|
192830 |
+
"step": 73460
|
192831 |
+
},
|
192832 |
+
{
|
192833 |
+
"epoch": 587.38,
|
192834 |
+
"learning_rate": 8.839134615384616e-06,
|
192835 |
+
"loss": 0.6947,
|
192836 |
+
"step": 73465
|
192837 |
+
},
|
192838 |
+
{
|
192839 |
+
"epoch": 587.42,
|
192840 |
+
"learning_rate": 8.839054487179488e-06,
|
192841 |
+
"loss": 0.9649,
|
192842 |
+
"step": 73470
|
192843 |
+
},
|
192844 |
+
{
|
192845 |
+
"epoch": 587.46,
|
192846 |
+
"learning_rate": 8.838974358974359e-06,
|
192847 |
+
"loss": 0.2923,
|
192848 |
+
"step": 73475
|
192849 |
+
},
|
192850 |
+
{
|
192851 |
+
"epoch": 587.5,
|
192852 |
+
"learning_rate": 8.838894230769232e-06,
|
192853 |
+
"loss": 0.3179,
|
192854 |
+
"step": 73480
|
192855 |
+
},
|
192856 |
+
{
|
192857 |
+
"epoch": 587.54,
|
192858 |
+
"learning_rate": 8.838814102564103e-06,
|
192859 |
+
"loss": 0.3462,
|
192860 |
+
"step": 73485
|
192861 |
+
},
|
192862 |
+
{
|
192863 |
+
"epoch": 587.58,
|
192864 |
+
"learning_rate": 8.838733974358975e-06,
|
192865 |
+
"loss": 0.7362,
|
192866 |
+
"step": 73490
|
192867 |
+
},
|
192868 |
+
{
|
192869 |
+
"epoch": 587.62,
|
192870 |
+
"learning_rate": 8.838653846153848e-06,
|
192871 |
+
"loss": 1.1776,
|
192872 |
+
"step": 73495
|
192873 |
+
},
|
192874 |
+
{
|
192875 |
+
"epoch": 587.66,
|
192876 |
+
"learning_rate": 8.838573717948719e-06,
|
192877 |
+
"loss": 0.3046,
|
192878 |
+
"step": 73500
|
192879 |
+
},
|
192880 |
+
{
|
192881 |
+
"epoch": 587.7,
|
192882 |
+
"learning_rate": 8.83849358974359e-06,
|
192883 |
+
"loss": 0.3213,
|
192884 |
+
"step": 73505
|
192885 |
+
},
|
192886 |
+
{
|
192887 |
+
"epoch": 587.74,
|
192888 |
+
"learning_rate": 8.838413461538462e-06,
|
192889 |
+
"loss": 0.3917,
|
192890 |
+
"step": 73510
|
192891 |
+
},
|
192892 |
+
{
|
192893 |
+
"epoch": 587.78,
|
192894 |
+
"learning_rate": 8.838333333333335e-06,
|
192895 |
+
"loss": 0.8034,
|
192896 |
+
"step": 73515
|
192897 |
+
},
|
192898 |
+
{
|
192899 |
+
"epoch": 587.82,
|
192900 |
+
"learning_rate": 8.838253205128206e-06,
|
192901 |
+
"loss": 0.8914,
|
192902 |
+
"step": 73520
|
192903 |
+
},
|
192904 |
+
{
|
192905 |
+
"epoch": 587.86,
|
192906 |
+
"learning_rate": 8.838173076923078e-06,
|
192907 |
+
"loss": 0.2794,
|
192908 |
+
"step": 73525
|
192909 |
+
},
|
192910 |
+
{
|
192911 |
+
"epoch": 587.9,
|
192912 |
+
"learning_rate": 8.838092948717949e-06,
|
192913 |
+
"loss": 0.5335,
|
192914 |
+
"step": 73530
|
192915 |
+
},
|
192916 |
+
{
|
192917 |
+
"epoch": 587.94,
|
192918 |
+
"learning_rate": 8.838012820512822e-06,
|
192919 |
+
"loss": 0.3818,
|
192920 |
+
"step": 73535
|
192921 |
+
},
|
192922 |
+
{
|
192923 |
+
"epoch": 587.98,
|
192924 |
+
"learning_rate": 8.837932692307693e-06,
|
192925 |
+
"loss": 0.6372,
|
192926 |
+
"step": 73540
|
192927 |
+
},
|
192928 |
+
{
|
192929 |
+
"epoch": 588.0,
|
192930 |
+
"eval_loss": 0.40962567925453186,
|
192931 |
+
"eval_runtime": 39.6457,
|
192932 |
+
"eval_samples_per_second": 21.188,
|
192933 |
+
"eval_steps_per_second": 0.681,
|
192934 |
+
"eval_wer": 0.1885777450257922,
|
192935 |
+
"step": 73543
|
192936 |
+
},
|
192937 |
+
{
|
192938 |
+
"epoch": 593.02,
|
192939 |
+
"learning_rate": 8.83786858974359e-06,
|
192940 |
+
"loss": 0.4499,
|
192941 |
+
"step": 73545
|
192942 |
+
},
|
192943 |
+
{
|
192944 |
+
"epoch": 593.06,
|
192945 |
+
"learning_rate": 8.837788461538462e-06,
|
192946 |
+
"loss": 0.3547,
|
192947 |
+
"step": 73550
|
192948 |
+
},
|
192949 |
+
{
|
192950 |
+
"epoch": 593.1,
|
192951 |
+
"learning_rate": 8.837708333333333e-06,
|
192952 |
+
"loss": 0.3296,
|
192953 |
+
"step": 73555
|
192954 |
+
},
|
192955 |
+
{
|
192956 |
+
"epoch": 593.14,
|
192957 |
+
"learning_rate": 8.837628205128207e-06,
|
192958 |
+
"loss": 0.3436,
|
192959 |
+
"step": 73560
|
192960 |
+
},
|
192961 |
+
{
|
192962 |
+
"epoch": 593.18,
|
192963 |
+
"learning_rate": 8.837548076923078e-06,
|
192964 |
+
"loss": 0.6371,
|
192965 |
+
"step": 73565
|
192966 |
+
},
|
192967 |
+
{
|
192968 |
+
"epoch": 593.22,
|
192969 |
+
"learning_rate": 8.83746794871795e-06,
|
192970 |
+
"loss": 1.066,
|
192971 |
+
"step": 73570
|
192972 |
+
},
|
192973 |
+
{
|
192974 |
+
"epoch": 593.26,
|
192975 |
+
"learning_rate": 8.837387820512822e-06,
|
192976 |
+
"loss": 0.3452,
|
192977 |
+
"step": 73575
|
192978 |
+
},
|
192979 |
+
{
|
192980 |
+
"epoch": 593.3,
|
192981 |
+
"learning_rate": 8.837307692307694e-06,
|
192982 |
+
"loss": 0.3008,
|
192983 |
+
"step": 73580
|
192984 |
+
},
|
192985 |
+
{
|
192986 |
+
"epoch": 593.34,
|
192987 |
+
"learning_rate": 8.837227564102565e-06,
|
192988 |
+
"loss": 0.3361,
|
192989 |
+
"step": 73585
|
192990 |
+
},
|
192991 |
+
{
|
192992 |
+
"epoch": 593.38,
|
192993 |
+
"learning_rate": 8.837147435897436e-06,
|
192994 |
+
"loss": 0.6165,
|
192995 |
+
"step": 73590
|
192996 |
+
},
|
192997 |
+
{
|
192998 |
+
"epoch": 593.42,
|
192999 |
+
"learning_rate": 8.83706730769231e-06,
|
193000 |
+
"loss": 0.9722,
|
193001 |
+
"step": 73595
|
193002 |
+
},
|
193003 |
+
{
|
193004 |
+
"epoch": 593.46,
|
193005 |
+
"learning_rate": 8.836987179487179e-06,
|
193006 |
+
"loss": 0.3145,
|
193007 |
+
"step": 73600
|
193008 |
+
},
|
193009 |
+
{
|
193010 |
+
"epoch": 593.5,
|
193011 |
+
"learning_rate": 8.836907051282052e-06,
|
193012 |
+
"loss": 0.2744,
|
193013 |
+
"step": 73605
|
193014 |
+
},
|
193015 |
+
{
|
193016 |
+
"epoch": 593.54,
|
193017 |
+
"learning_rate": 8.836826923076923e-06,
|
193018 |
+
"loss": 0.3529,
|
193019 |
+
"step": 73610
|
193020 |
+
},
|
193021 |
+
{
|
193022 |
+
"epoch": 593.58,
|
193023 |
+
"learning_rate": 8.836746794871795e-06,
|
193024 |
+
"loss": 0.588,
|
193025 |
+
"step": 73615
|
193026 |
+
},
|
193027 |
+
{
|
193028 |
+
"epoch": 593.62,
|
193029 |
+
"learning_rate": 8.836666666666668e-06,
|
193030 |
+
"loss": 1.2817,
|
193031 |
+
"step": 73620
|
193032 |
+
},
|
193033 |
+
{
|
193034 |
+
"epoch": 593.66,
|
193035 |
+
"learning_rate": 8.83658653846154e-06,
|
193036 |
+
"loss": 0.2927,
|
193037 |
+
"step": 73625
|
193038 |
+
},
|
193039 |
+
{
|
193040 |
+
"epoch": 593.7,
|
193041 |
+
"learning_rate": 8.83650641025641e-06,
|
193042 |
+
"loss": 0.3421,
|
193043 |
+
"step": 73630
|
193044 |
+
},
|
193045 |
+
{
|
193046 |
+
"epoch": 593.74,
|
193047 |
+
"learning_rate": 8.836426282051282e-06,
|
193048 |
+
"loss": 0.3321,
|
193049 |
+
"step": 73635
|
193050 |
+
},
|
193051 |
+
{
|
193052 |
+
"epoch": 593.78,
|
193053 |
+
"learning_rate": 8.836346153846155e-06,
|
193054 |
+
"loss": 0.7771,
|
193055 |
+
"step": 73640
|
193056 |
+
},
|
193057 |
+
{
|
193058 |
+
"epoch": 593.82,
|
193059 |
+
"learning_rate": 8.836282051282052e-06,
|
193060 |
+
"loss": Infinity,
|
193061 |
+
"step": 73645
|
193062 |
+
},
|
193063 |
+
{
|
193064 |
+
"epoch": 593.86,
|
193065 |
+
"learning_rate": 8.836201923076924e-06,
|
193066 |
+
"loss": NaN,
|
193067 |
+
"step": 73650
|
193068 |
+
},
|
193069 |
+
{
|
193070 |
+
"epoch": 593.9,
|
193071 |
+
"learning_rate": 8.836121794871795e-06,
|
193072 |
+
"loss": NaN,
|
193073 |
+
"step": 73655
|
193074 |
+
},
|
193075 |
+
{
|
193076 |
+
"epoch": 593.94,
|
193077 |
+
"learning_rate": 8.836041666666667e-06,
|
193078 |
+
"loss": NaN,
|
193079 |
+
"step": 73660
|
193080 |
+
},
|
193081 |
+
{
|
193082 |
+
"epoch": 593.98,
|
193083 |
+
"learning_rate": 8.83596153846154e-06,
|
193084 |
+
"loss": NaN,
|
193085 |
+
"step": 73665
|
193086 |
+
},
|
193087 |
+
{
|
193088 |
+
"epoch": 594.0,
|
193089 |
+
"eval_loss": 0.3711722195148468,
|
193090 |
+
"eval_runtime": 38.93,
|
193091 |
+
"eval_samples_per_second": 21.577,
|
193092 |
+
"eval_steps_per_second": 0.694,
|
193093 |
+
"eval_wer": 0.19046610169491526,
|
193094 |
+
"step": 73667
|
193095 |
+
},
|
193096 |
+
{
|
193097 |
+
"epoch": 589.02,
|
193098 |
+
"learning_rate": 8.835881410256411e-06,
|
193099 |
+
"loss": 0.4863,
|
193100 |
+
"step": 73670
|
193101 |
+
},
|
193102 |
+
{
|
193103 |
+
"epoch": 589.06,
|
193104 |
+
"learning_rate": 8.835801282051282e-06,
|
193105 |
+
"loss": 0.2866,
|
193106 |
+
"step": 73675
|
193107 |
+
},
|
193108 |
+
{
|
193109 |
+
"epoch": 589.1,
|
193110 |
+
"learning_rate": 8.835721153846154e-06,
|
193111 |
+
"loss": 0.3515,
|
193112 |
+
"step": 73680
|
193113 |
+
},
|
193114 |
+
{
|
193115 |
+
"epoch": 589.14,
|
193116 |
+
"learning_rate": 8.835641025641027e-06,
|
193117 |
+
"loss": 0.4241,
|
193118 |
+
"step": 73685
|
193119 |
+
},
|
193120 |
+
{
|
193121 |
+
"epoch": 589.18,
|
193122 |
+
"learning_rate": 8.835560897435898e-06,
|
193123 |
+
"loss": 0.7713,
|
193124 |
+
"step": 73690
|
193125 |
+
},
|
193126 |
+
{
|
193127 |
+
"epoch": 589.22,
|
193128 |
+
"learning_rate": 8.83548076923077e-06,
|
193129 |
+
"loss": 0.9977,
|
193130 |
+
"step": 73695
|
193131 |
+
},
|
193132 |
+
{
|
193133 |
+
"epoch": 589.26,
|
193134 |
+
"learning_rate": 8.835400641025642e-06,
|
193135 |
+
"loss": 0.3414,
|
193136 |
+
"step": 73700
|
193137 |
+
},
|
193138 |
+
{
|
193139 |
+
"epoch": 589.3,
|
193140 |
+
"learning_rate": 8.835320512820514e-06,
|
193141 |
+
"loss": 0.319,
|
193142 |
+
"step": 73705
|
193143 |
+
},
|
193144 |
+
{
|
193145 |
+
"epoch": 589.34,
|
193146 |
+
"learning_rate": 8.835240384615385e-06,
|
193147 |
+
"loss": 0.4791,
|
193148 |
+
"step": 73710
|
193149 |
+
},
|
193150 |
+
{
|
193151 |
+
"epoch": 589.38,
|
193152 |
+
"learning_rate": 8.835160256410257e-06,
|
193153 |
+
"loss": 0.8164,
|
193154 |
+
"step": 73715
|
193155 |
+
},
|
193156 |
+
{
|
193157 |
+
"epoch": 589.42,
|
193158 |
+
"learning_rate": 8.83508012820513e-06,
|
193159 |
+
"loss": 0.8898,
|
193160 |
+
"step": 73720
|
193161 |
+
},
|
193162 |
+
{
|
193163 |
+
"epoch": 589.46,
|
193164 |
+
"learning_rate": 8.835000000000001e-06,
|
193165 |
+
"loss": 0.3485,
|
193166 |
+
"step": 73725
|
193167 |
+
},
|
193168 |
+
{
|
193169 |
+
"epoch": 589.5,
|
193170 |
+
"learning_rate": 8.834919871794872e-06,
|
193171 |
+
"loss": 0.3385,
|
193172 |
+
"step": 73730
|
193173 |
+
},
|
193174 |
+
{
|
193175 |
+
"epoch": 589.54,
|
193176 |
+
"learning_rate": 8.834839743589745e-06,
|
193177 |
+
"loss": 0.4159,
|
193178 |
+
"step": 73735
|
193179 |
+
},
|
193180 |
+
{
|
193181 |
+
"epoch": 589.58,
|
193182 |
+
"learning_rate": 8.834759615384617e-06,
|
193183 |
+
"loss": 0.8097,
|
193184 |
+
"step": 73740
|
193185 |
+
},
|
193186 |
+
{
|
193187 |
+
"epoch": 589.62,
|
193188 |
+
"learning_rate": 8.834679487179488e-06,
|
193189 |
+
"loss": 0.9052,
|
193190 |
+
"step": 73745
|
193191 |
+
},
|
193192 |
+
{
|
193193 |
+
"epoch": 589.66,
|
193194 |
+
"learning_rate": 8.83459935897436e-06,
|
193195 |
+
"loss": 0.2916,
|
193196 |
+
"step": 73750
|
193197 |
+
},
|
193198 |
+
{
|
193199 |
+
"epoch": 589.7,
|
193200 |
+
"learning_rate": 8.834519230769233e-06,
|
193201 |
+
"loss": 0.3518,
|
193202 |
+
"step": 73755
|
193203 |
+
},
|
193204 |
+
{
|
193205 |
+
"epoch": 589.74,
|
193206 |
+
"learning_rate": 8.834439102564102e-06,
|
193207 |
+
"loss": 0.4072,
|
193208 |
+
"step": 73760
|
193209 |
+
},
|
193210 |
+
{
|
193211 |
+
"epoch": 589.78,
|
193212 |
+
"learning_rate": 8.834358974358975e-06,
|
193213 |
+
"loss": 0.7854,
|
193214 |
+
"step": 73765
|
193215 |
+
},
|
193216 |
+
{
|
193217 |
+
"epoch": 589.82,
|
193218 |
+
"learning_rate": 8.834278846153847e-06,
|
193219 |
+
"loss": 0.8298,
|
193220 |
+
"step": 73770
|
193221 |
+
},
|
193222 |
+
{
|
193223 |
+
"epoch": 589.86,
|
193224 |
+
"learning_rate": 8.834198717948718e-06,
|
193225 |
+
"loss": 0.3155,
|
193226 |
+
"step": 73775
|
193227 |
+
},
|
193228 |
+
{
|
193229 |
+
"epoch": 589.9,
|
193230 |
+
"learning_rate": 8.83411858974359e-06,
|
193231 |
+
"loss": 0.342,
|
193232 |
+
"step": 73780
|
193233 |
+
},
|
193234 |
+
{
|
193235 |
+
"epoch": 589.94,
|
193236 |
+
"learning_rate": 8.834038461538462e-06,
|
193237 |
+
"loss": 0.3902,
|
193238 |
+
"step": 73785
|
193239 |
+
},
|
193240 |
+
{
|
193241 |
+
"epoch": 589.98,
|
193242 |
+
"learning_rate": 8.833958333333334e-06,
|
193243 |
+
"loss": 0.7749,
|
193244 |
+
"step": 73790
|
193245 |
+
},
|
193246 |
+
{
|
193247 |
+
"epoch": 590.0,
|
193248 |
+
"eval_loss": 0.4138658046722412,
|
193249 |
+
"eval_runtime": 40.8862,
|
193250 |
+
"eval_samples_per_second": 20.545,
|
193251 |
+
"eval_steps_per_second": 0.66,
|
193252 |
+
"eval_wer": 0.19385140905209222,
|
193253 |
+
"step": 73792
|
193254 |
+
},
|
193255 |
+
{
|
193256 |
+
"epoch": 595.02,
|
193257 |
+
"learning_rate": 8.833878205128205e-06,
|
193258 |
+
"loss": 0.3728,
|
193259 |
+
"step": 73795
|
193260 |
+
},
|
193261 |
+
{
|
193262 |
+
"epoch": 595.06,
|
193263 |
+
"learning_rate": 8.833798076923078e-06,
|
193264 |
+
"loss": 0.2469,
|
193265 |
+
"step": 73800
|
193266 |
+
},
|
193267 |
+
{
|
193268 |
+
"epoch": 595.1,
|
193269 |
+
"learning_rate": 8.83371794871795e-06,
|
193270 |
+
"loss": 0.3617,
|
193271 |
+
"step": 73805
|
193272 |
+
},
|
193273 |
+
{
|
193274 |
+
"epoch": 595.14,
|
193275 |
+
"learning_rate": 8.83363782051282e-06,
|
193276 |
+
"loss": 0.5017,
|
193277 |
+
"step": 73810
|
193278 |
+
},
|
193279 |
+
{
|
193280 |
+
"epoch": 595.18,
|
193281 |
+
"learning_rate": 8.833557692307692e-06,
|
193282 |
+
"loss": 0.7071,
|
193283 |
+
"step": 73815
|
193284 |
+
},
|
193285 |
+
{
|
193286 |
+
"epoch": 595.22,
|
193287 |
+
"learning_rate": 8.833477564102565e-06,
|
193288 |
+
"loss": 0.8614,
|
193289 |
+
"step": 73820
|
193290 |
+
},
|
193291 |
+
{
|
193292 |
+
"epoch": 595.27,
|
193293 |
+
"learning_rate": 8.833397435897437e-06,
|
193294 |
+
"loss": 0.3067,
|
193295 |
+
"step": 73825
|
193296 |
+
},
|
193297 |
+
{
|
193298 |
+
"epoch": 595.31,
|
193299 |
+
"learning_rate": 8.833317307692308e-06,
|
193300 |
+
"loss": 0.3224,
|
193301 |
+
"step": 73830
|
193302 |
+
},
|
193303 |
+
{
|
193304 |
+
"epoch": 595.35,
|
193305 |
+
"learning_rate": 8.833237179487181e-06,
|
193306 |
+
"loss": 0.4127,
|
193307 |
+
"step": 73835
|
193308 |
+
},
|
193309 |
+
{
|
193310 |
+
"epoch": 595.39,
|
193311 |
+
"learning_rate": 8.833157051282052e-06,
|
193312 |
+
"loss": 0.8873,
|
193313 |
+
"step": 73840
|
193314 |
+
},
|
193315 |
+
{
|
193316 |
+
"epoch": 595.43,
|
193317 |
+
"learning_rate": 8.833076923076924e-06,
|
193318 |
+
"loss": 0.9318,
|
193319 |
+
"step": 73845
|
193320 |
+
},
|
193321 |
+
{
|
193322 |
+
"epoch": 595.47,
|
193323 |
+
"learning_rate": 8.832996794871795e-06,
|
193324 |
+
"loss": 0.3216,
|
193325 |
+
"step": 73850
|
193326 |
+
},
|
193327 |
+
{
|
193328 |
+
"epoch": 595.51,
|
193329 |
+
"learning_rate": 8.832916666666668e-06,
|
193330 |
+
"loss": 0.3496,
|
193331 |
+
"step": 73855
|
193332 |
+
},
|
193333 |
+
{
|
193334 |
+
"epoch": 595.55,
|
193335 |
+
"learning_rate": 8.83283653846154e-06,
|
193336 |
+
"loss": 0.3752,
|
193337 |
+
"step": 73860
|
193338 |
+
},
|
193339 |
+
{
|
193340 |
+
"epoch": 595.59,
|
193341 |
+
"learning_rate": 8.832756410256411e-06,
|
193342 |
+
"loss": 0.7592,
|
193343 |
+
"step": 73865
|
193344 |
+
},
|
193345 |
+
{
|
193346 |
+
"epoch": 595.63,
|
193347 |
+
"learning_rate": 8.832676282051282e-06,
|
193348 |
+
"loss": 0.927,
|
193349 |
+
"step": 73870
|
193350 |
+
},
|
193351 |
+
{
|
193352 |
+
"epoch": 595.67,
|
193353 |
+
"learning_rate": 8.832596153846155e-06,
|
193354 |
+
"loss": 0.3161,
|
193355 |
+
"step": 73875
|
193356 |
+
},
|
193357 |
+
{
|
193358 |
+
"epoch": 595.71,
|
193359 |
+
"learning_rate": 8.832516025641027e-06,
|
193360 |
+
"loss": 0.361,
|
193361 |
+
"step": 73880
|
193362 |
+
},
|
193363 |
+
{
|
193364 |
+
"epoch": 595.75,
|
193365 |
+
"learning_rate": 8.832435897435898e-06,
|
193366 |
+
"loss": 0.3999,
|
193367 |
+
"step": 73885
|
193368 |
+
},
|
193369 |
+
{
|
193370 |
+
"epoch": 595.79,
|
193371 |
+
"learning_rate": 8.832355769230771e-06,
|
193372 |
+
"loss": 0.7727,
|
193373 |
+
"step": 73890
|
193374 |
+
},
|
193375 |
+
{
|
193376 |
+
"epoch": 595.83,
|
193377 |
+
"learning_rate": 8.832275641025642e-06,
|
193378 |
+
"loss": 0.9198,
|
193379 |
+
"step": 73895
|
193380 |
+
},
|
193381 |
+
{
|
193382 |
+
"epoch": 595.87,
|
193383 |
+
"learning_rate": 8.832195512820514e-06,
|
193384 |
+
"loss": 0.4088,
|
193385 |
+
"step": 73900
|
193386 |
+
},
|
193387 |
+
{
|
193388 |
+
"epoch": 595.91,
|
193389 |
+
"learning_rate": 8.832115384615385e-06,
|
193390 |
+
"loss": 0.3442,
|
193391 |
+
"step": 73905
|
193392 |
+
},
|
193393 |
+
{
|
193394 |
+
"epoch": 595.95,
|
193395 |
+
"learning_rate": 8.832035256410258e-06,
|
193396 |
+
"loss": 0.3927,
|
193397 |
+
"step": 73910
|
193398 |
+
},
|
193399 |
+
{
|
193400 |
+
"epoch": 595.99,
|
193401 |
+
"learning_rate": 8.831955128205128e-06,
|
193402 |
+
"loss": 0.8166,
|
193403 |
+
"step": 73915
|
193404 |
+
},
|
193405 |
+
{
|
193406 |
+
"epoch": 596.0,
|
193407 |
+
"eval_loss": 0.4341099262237549,
|
193408 |
+
"eval_runtime": 39.1108,
|
193409 |
+
"eval_samples_per_second": 21.477,
|
193410 |
+
"eval_steps_per_second": 0.69,
|
193411 |
+
"eval_wer": 0.1899564928840056,
|
193412 |
+
"step": 73916
|
193413 |
}
|
193414 |
],
|
193415 |
+
"max_steps": 620000,
|
193416 |
"num_train_epochs": 5000,
|
193417 |
+
"total_flos": 2.08014258346129e+20,
|
193418 |
"trial_name": null,
|
193419 |
"trial_params": null
|
193420 |
}
|
model-bin/finetune/base/{checkpoint-73294 β checkpoint-73916}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1629872387.216304/events.out.tfevents.1629872387.7e498afd5545.905.63
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fab1213ee83b8ab20ed67b0a41b3a7119718916a3e5fc6cccf397fe0a158bb8a
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629873040.9577622/events.out.tfevents.1629873040.7e498afd5545.905.65
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfe40c82647fbf57e2fd91377769c78c9405cf05d96cd1b805e9a0f8ce9062bd
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629873815.6117036/events.out.tfevents.1629873815.7e498afd5545.905.67
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c5b7bdee9da9550fb64c64cdf7bdf3102bda3f4693fd0fe1c270d805a39e816f
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629874460.1476715/events.out.tfevents.1629874460.7e498afd5545.905.69
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7d36d0c7d2cfa270e17d1103291cafe23e90e17e5fca22ebae077f8d20a4d8f
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1629875104.9909084/events.out.tfevents.1629875104.7e498afd5545.905.71
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7483bed69bd1f30ee7f80da8fb30aa906d7d000597764347fbd2dae5f0d7529
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629872387.7e498afd5545.905.62
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc690dba20d9dc7f9438804ccd3bf313426dfd646917e757957a7e1c86f46562
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629873040.7e498afd5545.905.64
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49ba294d0a143840803fe76ce3b2051519f8a957a43ea2fe6fde9a03154bc358
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629873815.7e498afd5545.905.66
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5098ab750bc4e7c2bde49ac1cd790815cded43f504e1ddb933d7da48d49af9f3
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629874460.7e498afd5545.905.68
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2aee550599220ed9f542bac07d9127929e530ee706fa895d29696cd84afd3e4c
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629875104.7e498afd5545.905.70
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9cfb270bf0972cef4eface5ce1c36791ac9766aff700c59d6b0455da12c2b290
|
3 |
+
size 8622
|