"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630252977.4699829/events.out.tfevents.1630252977.cc93b136ebf5.1086.219 +3 -0
- model-bin/finetune/base/log/1630253413.1850004/events.out.tfevents.1630253413.cc93b136ebf5.1086.221 +3 -0
- model-bin/finetune/base/log/1630253854.71079/events.out.tfevents.1630253854.cc93b136ebf5.1086.223 +3 -0
- model-bin/finetune/base/log/1630254282.042753/events.out.tfevents.1630254282.cc93b136ebf5.1086.225 +3 -0
- model-bin/finetune/base/log/1630254714.536368/events.out.tfevents.1630254714.cc93b136ebf5.1086.227 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630252977.cc93b136ebf5.1086.218 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630253413.cc93b136ebf5.1086.220 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630253854.cc93b136ebf5.1086.222 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630254282.cc93b136ebf5.1086.224 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630254714.cc93b136ebf5.1086.226 +3 -0
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46ba5bc9ebc2d9cb5c039dedb6841c756b21ad0680230520f9190e7d9c8ee4d1
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:751bbf777839d159698b17abeaf9812e82ed6c85907fe60a3193f08e4f8a1146
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eff6a839f60f322cdb5909f5073b57b6d1ded874a541816583c3d41ecfaf7177
|
3 |
+
size 14567
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:054e7466e8875cd689e7563a044cae340ceb7830af56e188fce91c945fd5de6e
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11001b1c63c88e4fa645949c229ef32727a43e7b4b8e96937eec6a7dd2e7766e
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -293637,11 +293637,800 @@
|
|
293637 |
"eval_steps_per_second": 0.627,
|
293638 |
"eval_wer": 0.17048071387449626,
|
293639 |
"step": 152690
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
293640 |
}
|
293641 |
],
|
293642 |
-
"max_steps":
|
293643 |
"num_train_epochs": 5000,
|
293644 |
-
"total_flos": 4.
|
293645 |
"trial_name": null,
|
293646 |
"trial_params": null
|
293647 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1235.995983935743,
|
5 |
+
"global_step": 153312,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
293637 |
"eval_steps_per_second": 0.627,
|
293638 |
"eval_wer": 0.17048071387449626,
|
293639 |
"step": 152690
|
293640 |
+
},
|
293641 |
+
{
|
293642 |
+
"epoch": 1231.04,
|
293643 |
+
"learning_rate": 7.5503069466882074e-06,
|
293644 |
+
"loss": 0.2499,
|
293645 |
+
"step": 152695
|
293646 |
+
},
|
293647 |
+
{
|
293648 |
+
"epoch": 1231.08,
|
293649 |
+
"learning_rate": 7.550226171243943e-06,
|
293650 |
+
"loss": 0.2478,
|
293651 |
+
"step": 152700
|
293652 |
+
},
|
293653 |
+
{
|
293654 |
+
"epoch": 1231.12,
|
293655 |
+
"learning_rate": 7.550145395799677e-06,
|
293656 |
+
"loss": 0.3172,
|
293657 |
+
"step": 152705
|
293658 |
+
},
|
293659 |
+
{
|
293660 |
+
"epoch": 1231.16,
|
293661 |
+
"learning_rate": 7.550064620355413e-06,
|
293662 |
+
"loss": 0.4319,
|
293663 |
+
"step": 152710
|
293664 |
+
},
|
293665 |
+
{
|
293666 |
+
"epoch": 1231.2,
|
293667 |
+
"learning_rate": 7.549983844911147e-06,
|
293668 |
+
"loss": 1.2308,
|
293669 |
+
"step": 152715
|
293670 |
+
},
|
293671 |
+
{
|
293672 |
+
"epoch": 1231.24,
|
293673 |
+
"learning_rate": 7.549903069466883e-06,
|
293674 |
+
"loss": 0.4007,
|
293675 |
+
"step": 152720
|
293676 |
+
},
|
293677 |
+
{
|
293678 |
+
"epoch": 1231.28,
|
293679 |
+
"learning_rate": 7.549822294022617e-06,
|
293680 |
+
"loss": 0.2828,
|
293681 |
+
"step": 152725
|
293682 |
+
},
|
293683 |
+
{
|
293684 |
+
"epoch": 1231.32,
|
293685 |
+
"learning_rate": 7.549741518578353e-06,
|
293686 |
+
"loss": 0.2824,
|
293687 |
+
"step": 152730
|
293688 |
+
},
|
293689 |
+
{
|
293690 |
+
"epoch": 1231.36,
|
293691 |
+
"learning_rate": 7.549660743134087e-06,
|
293692 |
+
"loss": 0.4561,
|
293693 |
+
"step": 152735
|
293694 |
+
},
|
293695 |
+
{
|
293696 |
+
"epoch": 1231.4,
|
293697 |
+
"learning_rate": 7.549579967689823e-06,
|
293698 |
+
"loss": 1.1774,
|
293699 |
+
"step": 152740
|
293700 |
+
},
|
293701 |
+
{
|
293702 |
+
"epoch": 1231.44,
|
293703 |
+
"learning_rate": 7.549499192245558e-06,
|
293704 |
+
"loss": 0.3069,
|
293705 |
+
"step": 152745
|
293706 |
+
},
|
293707 |
+
{
|
293708 |
+
"epoch": 1231.48,
|
293709 |
+
"learning_rate": 7.549418416801293e-06,
|
293710 |
+
"loss": 0.2546,
|
293711 |
+
"step": 152750
|
293712 |
+
},
|
293713 |
+
{
|
293714 |
+
"epoch": 1231.52,
|
293715 |
+
"learning_rate": 7.549337641357028e-06,
|
293716 |
+
"loss": 0.3069,
|
293717 |
+
"step": 152755
|
293718 |
+
},
|
293719 |
+
{
|
293720 |
+
"epoch": 1231.56,
|
293721 |
+
"learning_rate": 7.549256865912763e-06,
|
293722 |
+
"loss": 0.5205,
|
293723 |
+
"step": 152760
|
293724 |
+
},
|
293725 |
+
{
|
293726 |
+
"epoch": 1231.6,
|
293727 |
+
"learning_rate": 7.549176090468498e-06,
|
293728 |
+
"loss": 1.0464,
|
293729 |
+
"step": 152765
|
293730 |
+
},
|
293731 |
+
{
|
293732 |
+
"epoch": 1231.64,
|
293733 |
+
"learning_rate": 7.549095315024233e-06,
|
293734 |
+
"loss": 0.3116,
|
293735 |
+
"step": 152770
|
293736 |
+
},
|
293737 |
+
{
|
293738 |
+
"epoch": 1231.68,
|
293739 |
+
"learning_rate": 7.549014539579969e-06,
|
293740 |
+
"loss": 0.3058,
|
293741 |
+
"step": 152775
|
293742 |
+
},
|
293743 |
+
{
|
293744 |
+
"epoch": 1231.72,
|
293745 |
+
"learning_rate": 7.548933764135703e-06,
|
293746 |
+
"loss": 0.2748,
|
293747 |
+
"step": 152780
|
293748 |
+
},
|
293749 |
+
{
|
293750 |
+
"epoch": 1231.76,
|
293751 |
+
"learning_rate": 7.548852988691439e-06,
|
293752 |
+
"loss": 0.4329,
|
293753 |
+
"step": 152785
|
293754 |
+
},
|
293755 |
+
{
|
293756 |
+
"epoch": 1231.8,
|
293757 |
+
"learning_rate": 7.548772213247173e-06,
|
293758 |
+
"loss": 1.1518,
|
293759 |
+
"step": 152790
|
293760 |
+
},
|
293761 |
+
{
|
293762 |
+
"epoch": 1231.84,
|
293763 |
+
"learning_rate": 7.548691437802909e-06,
|
293764 |
+
"loss": 0.2424,
|
293765 |
+
"step": 152795
|
293766 |
+
},
|
293767 |
+
{
|
293768 |
+
"epoch": 1231.88,
|
293769 |
+
"learning_rate": 7.548610662358643e-06,
|
293770 |
+
"loss": 0.2967,
|
293771 |
+
"step": 152800
|
293772 |
+
},
|
293773 |
+
{
|
293774 |
+
"epoch": 1231.92,
|
293775 |
+
"learning_rate": 7.548529886914379e-06,
|
293776 |
+
"loss": 0.3013,
|
293777 |
+
"step": 152805
|
293778 |
+
},
|
293779 |
+
{
|
293780 |
+
"epoch": 1231.96,
|
293781 |
+
"learning_rate": 7.548449111470114e-06,
|
293782 |
+
"loss": 0.427,
|
293783 |
+
"step": 152810
|
293784 |
+
},
|
293785 |
+
{
|
293786 |
+
"epoch": 1232.0,
|
293787 |
+
"eval_loss": 0.3473464250564575,
|
293788 |
+
"eval_runtime": 40.4236,
|
293789 |
+
"eval_samples_per_second": 20.681,
|
293790 |
+
"eval_steps_per_second": 0.668,
|
293791 |
+
"eval_wer": 0.1835759611795446,
|
293792 |
+
"step": 152814
|
293793 |
+
},
|
293794 |
+
{
|
293795 |
+
"epoch": 1222.01,
|
293796 |
+
"learning_rate": 7.548368336025849e-06,
|
293797 |
+
"loss": 0.4101,
|
293798 |
+
"step": 152815
|
293799 |
+
},
|
293800 |
+
{
|
293801 |
+
"epoch": 1222.05,
|
293802 |
+
"learning_rate": 7.548287560581584e-06,
|
293803 |
+
"loss": 0.2653,
|
293804 |
+
"step": 152820
|
293805 |
+
},
|
293806 |
+
{
|
293807 |
+
"epoch": 1222.09,
|
293808 |
+
"learning_rate": 7.548206785137319e-06,
|
293809 |
+
"loss": 0.2575,
|
293810 |
+
"step": 152825
|
293811 |
+
},
|
293812 |
+
{
|
293813 |
+
"epoch": 1222.13,
|
293814 |
+
"learning_rate": 7.548126009693054e-06,
|
293815 |
+
"loss": 0.2939,
|
293816 |
+
"step": 152830
|
293817 |
+
},
|
293818 |
+
{
|
293819 |
+
"epoch": 1222.17,
|
293820 |
+
"learning_rate": 7.548045234248789e-06,
|
293821 |
+
"loss": 0.4049,
|
293822 |
+
"step": 152835
|
293823 |
+
},
|
293824 |
+
{
|
293825 |
+
"epoch": 1222.21,
|
293826 |
+
"learning_rate": 7.547964458804525e-06,
|
293827 |
+
"loss": 1.1922,
|
293828 |
+
"step": 152840
|
293829 |
+
},
|
293830 |
+
{
|
293831 |
+
"epoch": 1222.25,
|
293832 |
+
"learning_rate": 7.547883683360259e-06,
|
293833 |
+
"loss": 0.2904,
|
293834 |
+
"step": 152845
|
293835 |
+
},
|
293836 |
+
{
|
293837 |
+
"epoch": 1222.29,
|
293838 |
+
"learning_rate": 7.547802907915995e-06,
|
293839 |
+
"loss": 0.4204,
|
293840 |
+
"step": 152850
|
293841 |
+
},
|
293842 |
+
{
|
293843 |
+
"epoch": 1222.33,
|
293844 |
+
"learning_rate": 7.547722132471729e-06,
|
293845 |
+
"loss": 0.4682,
|
293846 |
+
"step": 152855
|
293847 |
+
},
|
293848 |
+
{
|
293849 |
+
"epoch": 1222.37,
|
293850 |
+
"learning_rate": 7.5476413570274646e-06,
|
293851 |
+
"loss": 0.5323,
|
293852 |
+
"step": 152860
|
293853 |
+
},
|
293854 |
+
{
|
293855 |
+
"epoch": 1222.41,
|
293856 |
+
"learning_rate": 7.5475605815831996e-06,
|
293857 |
+
"loss": 1.0863,
|
293858 |
+
"step": 152865
|
293859 |
+
},
|
293860 |
+
{
|
293861 |
+
"epoch": 1222.45,
|
293862 |
+
"learning_rate": 7.5474798061389345e-06,
|
293863 |
+
"loss": 0.3252,
|
293864 |
+
"step": 152870
|
293865 |
+
},
|
293866 |
+
{
|
293867 |
+
"epoch": 1222.49,
|
293868 |
+
"learning_rate": 7.5473990306946695e-06,
|
293869 |
+
"loss": 0.2891,
|
293870 |
+
"step": 152875
|
293871 |
+
},
|
293872 |
+
{
|
293873 |
+
"epoch": 1222.53,
|
293874 |
+
"learning_rate": 7.5473182552504045e-06,
|
293875 |
+
"loss": 0.442,
|
293876 |
+
"step": 152880
|
293877 |
+
},
|
293878 |
+
{
|
293879 |
+
"epoch": 1222.57,
|
293880 |
+
"learning_rate": 7.5472374798061395e-06,
|
293881 |
+
"loss": 0.4414,
|
293882 |
+
"step": 152885
|
293883 |
+
},
|
293884 |
+
{
|
293885 |
+
"epoch": 1222.61,
|
293886 |
+
"learning_rate": 7.5471567043618745e-06,
|
293887 |
+
"loss": 1.0553,
|
293888 |
+
"step": 152890
|
293889 |
+
},
|
293890 |
+
{
|
293891 |
+
"epoch": 1222.65,
|
293892 |
+
"learning_rate": 7.5470759289176095e-06,
|
293893 |
+
"loss": 0.2663,
|
293894 |
+
"step": 152895
|
293895 |
+
},
|
293896 |
+
{
|
293897 |
+
"epoch": 1222.69,
|
293898 |
+
"learning_rate": 7.5469951534733445e-06,
|
293899 |
+
"loss": 0.3036,
|
293900 |
+
"step": 152900
|
293901 |
+
},
|
293902 |
+
{
|
293903 |
+
"epoch": 1222.73,
|
293904 |
+
"learning_rate": 7.54691437802908e-06,
|
293905 |
+
"loss": 0.2624,
|
293906 |
+
"step": 152905
|
293907 |
+
},
|
293908 |
+
{
|
293909 |
+
"epoch": 1222.77,
|
293910 |
+
"learning_rate": 7.5468336025848144e-06,
|
293911 |
+
"loss": 0.5299,
|
293912 |
+
"step": 152910
|
293913 |
+
},
|
293914 |
+
{
|
293915 |
+
"epoch": 1222.81,
|
293916 |
+
"learning_rate": 7.54675282714055e-06,
|
293917 |
+
"loss": 1.1371,
|
293918 |
+
"step": 152915
|
293919 |
+
},
|
293920 |
+
{
|
293921 |
+
"epoch": 1222.85,
|
293922 |
+
"learning_rate": 7.546672051696285e-06,
|
293923 |
+
"loss": 0.2932,
|
293924 |
+
"step": 152920
|
293925 |
+
},
|
293926 |
+
{
|
293927 |
+
"epoch": 1222.89,
|
293928 |
+
"learning_rate": 7.54659127625202e-06,
|
293929 |
+
"loss": 0.3077,
|
293930 |
+
"step": 152925
|
293931 |
+
},
|
293932 |
+
{
|
293933 |
+
"epoch": 1222.93,
|
293934 |
+
"learning_rate": 7.546510500807755e-06,
|
293935 |
+
"loss": 0.3004,
|
293936 |
+
"step": 152930
|
293937 |
+
},
|
293938 |
+
{
|
293939 |
+
"epoch": 1222.97,
|
293940 |
+
"learning_rate": 7.54642972536349e-06,
|
293941 |
+
"loss": 0.5459,
|
293942 |
+
"step": 152935
|
293943 |
+
},
|
293944 |
+
{
|
293945 |
+
"epoch": 1223.0,
|
293946 |
+
"eval_loss": 0.35849061608314514,
|
293947 |
+
"eval_runtime": 40.5447,
|
293948 |
+
"eval_samples_per_second": 20.644,
|
293949 |
+
"eval_steps_per_second": 0.666,
|
293950 |
+
"eval_wer": 0.1767148683722655,
|
293951 |
+
"step": 152939
|
293952 |
+
},
|
293953 |
+
{
|
293954 |
+
"epoch": 1233.01,
|
293955 |
+
"learning_rate": 7.546348949919225e-06,
|
293956 |
+
"loss": 0.4571,
|
293957 |
+
"step": 152940
|
293958 |
+
},
|
293959 |
+
{
|
293960 |
+
"epoch": 1233.05,
|
293961 |
+
"learning_rate": 7.54626817447496e-06,
|
293962 |
+
"loss": 0.3129,
|
293963 |
+
"step": 152945
|
293964 |
+
},
|
293965 |
+
{
|
293966 |
+
"epoch": 1233.09,
|
293967 |
+
"learning_rate": 7.546187399030695e-06,
|
293968 |
+
"loss": 0.2331,
|
293969 |
+
"step": 152950
|
293970 |
+
},
|
293971 |
+
{
|
293972 |
+
"epoch": 1233.13,
|
293973 |
+
"learning_rate": 7.54610662358643e-06,
|
293974 |
+
"loss": 0.2974,
|
293975 |
+
"step": 152955
|
293976 |
+
},
|
293977 |
+
{
|
293978 |
+
"epoch": 1233.17,
|
293979 |
+
"learning_rate": 7.546025848142165e-06,
|
293980 |
+
"loss": 0.4769,
|
293981 |
+
"step": 152960
|
293982 |
+
},
|
293983 |
+
{
|
293984 |
+
"epoch": 1233.21,
|
293985 |
+
"learning_rate": 7.5459450726979e-06,
|
293986 |
+
"loss": 0.9929,
|
293987 |
+
"step": 152965
|
293988 |
+
},
|
293989 |
+
{
|
293990 |
+
"epoch": 1233.25,
|
293991 |
+
"learning_rate": 7.545864297253635e-06,
|
293992 |
+
"loss": 0.3494,
|
293993 |
+
"step": 152970
|
293994 |
+
},
|
293995 |
+
{
|
293996 |
+
"epoch": 1233.29,
|
293997 |
+
"learning_rate": 7.54578352180937e-06,
|
293998 |
+
"loss": 0.2682,
|
293999 |
+
"step": 152975
|
294000 |
+
},
|
294001 |
+
{
|
294002 |
+
"epoch": 1233.33,
|
294003 |
+
"learning_rate": 7.545702746365106e-06,
|
294004 |
+
"loss": 0.2763,
|
294005 |
+
"step": 152980
|
294006 |
+
},
|
294007 |
+
{
|
294008 |
+
"epoch": 1233.37,
|
294009 |
+
"learning_rate": 7.545621970920841e-06,
|
294010 |
+
"loss": 0.5025,
|
294011 |
+
"step": 152985
|
294012 |
+
},
|
294013 |
+
{
|
294014 |
+
"epoch": 1233.41,
|
294015 |
+
"learning_rate": 7.545541195476576e-06,
|
294016 |
+
"loss": 1.0533,
|
294017 |
+
"step": 152990
|
294018 |
+
},
|
294019 |
+
{
|
294020 |
+
"epoch": 1233.45,
|
294021 |
+
"learning_rate": 7.545460420032311e-06,
|
294022 |
+
"loss": 0.3089,
|
294023 |
+
"step": 152995
|
294024 |
+
},
|
294025 |
+
{
|
294026 |
+
"epoch": 1233.49,
|
294027 |
+
"learning_rate": 7.545379644588046e-06,
|
294028 |
+
"loss": 0.271,
|
294029 |
+
"step": 153000
|
294030 |
+
},
|
294031 |
+
{
|
294032 |
+
"epoch": 1233.53,
|
294033 |
+
"learning_rate": 7.545298869143781e-06,
|
294034 |
+
"loss": 0.2785,
|
294035 |
+
"step": 153005
|
294036 |
+
},
|
294037 |
+
{
|
294038 |
+
"epoch": 1233.57,
|
294039 |
+
"learning_rate": 7.545218093699516e-06,
|
294040 |
+
"loss": 0.4914,
|
294041 |
+
"step": 153010
|
294042 |
+
},
|
294043 |
+
{
|
294044 |
+
"epoch": 1233.61,
|
294045 |
+
"learning_rate": 7.545137318255251e-06,
|
294046 |
+
"loss": 1.182,
|
294047 |
+
"step": 153015
|
294048 |
+
},
|
294049 |
+
{
|
294050 |
+
"epoch": 1233.65,
|
294051 |
+
"learning_rate": 7.545056542810986e-06,
|
294052 |
+
"loss": 0.2868,
|
294053 |
+
"step": 153020
|
294054 |
+
},
|
294055 |
+
{
|
294056 |
+
"epoch": 1233.69,
|
294057 |
+
"learning_rate": 7.544975767366721e-06,
|
294058 |
+
"loss": 0.2879,
|
294059 |
+
"step": 153025
|
294060 |
+
},
|
294061 |
+
{
|
294062 |
+
"epoch": 1233.73,
|
294063 |
+
"learning_rate": 7.544894991922456e-06,
|
294064 |
+
"loss": 0.2492,
|
294065 |
+
"step": 153030
|
294066 |
+
},
|
294067 |
+
{
|
294068 |
+
"epoch": 1233.77,
|
294069 |
+
"learning_rate": 7.544814216478191e-06,
|
294070 |
+
"loss": 0.4884,
|
294071 |
+
"step": 153035
|
294072 |
+
},
|
294073 |
+
{
|
294074 |
+
"epoch": 1233.81,
|
294075 |
+
"learning_rate": 7.544733441033927e-06,
|
294076 |
+
"loss": 1.0584,
|
294077 |
+
"step": 153040
|
294078 |
+
},
|
294079 |
+
{
|
294080 |
+
"epoch": 1233.85,
|
294081 |
+
"learning_rate": 7.544652665589662e-06,
|
294082 |
+
"loss": 0.2708,
|
294083 |
+
"step": 153045
|
294084 |
+
},
|
294085 |
+
{
|
294086 |
+
"epoch": 1233.89,
|
294087 |
+
"learning_rate": 7.544571890145397e-06,
|
294088 |
+
"loss": 0.2559,
|
294089 |
+
"step": 153050
|
294090 |
+
},
|
294091 |
+
{
|
294092 |
+
"epoch": 1233.93,
|
294093 |
+
"learning_rate": 7.544491114701132e-06,
|
294094 |
+
"loss": 0.3759,
|
294095 |
+
"step": 153055
|
294096 |
+
},
|
294097 |
+
{
|
294098 |
+
"epoch": 1233.97,
|
294099 |
+
"learning_rate": 7.544410339256867e-06,
|
294100 |
+
"loss": 0.6144,
|
294101 |
+
"step": 153060
|
294102 |
+
},
|
294103 |
+
{
|
294104 |
+
"epoch": 1234.0,
|
294105 |
+
"eval_loss": 0.3701328635215759,
|
294106 |
+
"eval_runtime": 40.1133,
|
294107 |
+
"eval_samples_per_second": 20.866,
|
294108 |
+
"eval_steps_per_second": 0.673,
|
294109 |
+
"eval_wer": 0.17889839411676423,
|
294110 |
+
"step": 153063
|
294111 |
+
},
|
294112 |
+
{
|
294113 |
+
"epoch": 1224.02,
|
294114 |
+
"learning_rate": 7.544329563812602e-06,
|
294115 |
+
"loss": 0.3558,
|
294116 |
+
"step": 153065
|
294117 |
+
},
|
294118 |
+
{
|
294119 |
+
"epoch": 1224.06,
|
294120 |
+
"learning_rate": 7.544248788368337e-06,
|
294121 |
+
"loss": 0.2894,
|
294122 |
+
"step": 153070
|
294123 |
+
},
|
294124 |
+
{
|
294125 |
+
"epoch": 1224.1,
|
294126 |
+
"learning_rate": 7.5441680129240716e-06,
|
294127 |
+
"loss": 0.3037,
|
294128 |
+
"step": 153075
|
294129 |
+
},
|
294130 |
+
{
|
294131 |
+
"epoch": 1224.14,
|
294132 |
+
"learning_rate": 7.5440872374798066e-06,
|
294133 |
+
"loss": 0.2791,
|
294134 |
+
"step": 153080
|
294135 |
+
},
|
294136 |
+
{
|
294137 |
+
"epoch": 1224.18,
|
294138 |
+
"learning_rate": 7.5440064620355415e-06,
|
294139 |
+
"loss": 0.5449,
|
294140 |
+
"step": 153085
|
294141 |
+
},
|
294142 |
+
{
|
294143 |
+
"epoch": 1224.22,
|
294144 |
+
"learning_rate": 7.5439256865912765e-06,
|
294145 |
+
"loss": 0.9683,
|
294146 |
+
"step": 153090
|
294147 |
+
},
|
294148 |
+
{
|
294149 |
+
"epoch": 1224.26,
|
294150 |
+
"learning_rate": 7.543844911147012e-06,
|
294151 |
+
"loss": 0.324,
|
294152 |
+
"step": 153095
|
294153 |
+
},
|
294154 |
+
{
|
294155 |
+
"epoch": 1224.3,
|
294156 |
+
"learning_rate": 7.5437641357027465e-06,
|
294157 |
+
"loss": 0.282,
|
294158 |
+
"step": 153100
|
294159 |
+
},
|
294160 |
+
{
|
294161 |
+
"epoch": 1224.34,
|
294162 |
+
"learning_rate": 7.543683360258482e-06,
|
294163 |
+
"loss": 0.3395,
|
294164 |
+
"step": 153105
|
294165 |
+
},
|
294166 |
+
{
|
294167 |
+
"epoch": 1224.38,
|
294168 |
+
"learning_rate": 7.543602584814217e-06,
|
294169 |
+
"loss": 0.7002,
|
294170 |
+
"step": 153110
|
294171 |
+
},
|
294172 |
+
{
|
294173 |
+
"epoch": 1224.42,
|
294174 |
+
"learning_rate": 7.543521809369952e-06,
|
294175 |
+
"loss": 0.9072,
|
294176 |
+
"step": 153115
|
294177 |
+
},
|
294178 |
+
{
|
294179 |
+
"epoch": 1224.46,
|
294180 |
+
"learning_rate": 7.543441033925687e-06,
|
294181 |
+
"loss": 0.3604,
|
294182 |
+
"step": 153120
|
294183 |
+
},
|
294184 |
+
{
|
294185 |
+
"epoch": 1224.5,
|
294186 |
+
"learning_rate": 7.543360258481422e-06,
|
294187 |
+
"loss": 0.2485,
|
294188 |
+
"step": 153125
|
294189 |
+
},
|
294190 |
+
{
|
294191 |
+
"epoch": 1224.54,
|
294192 |
+
"learning_rate": 7.543279483037157e-06,
|
294193 |
+
"loss": 0.2851,
|
294194 |
+
"step": 153130
|
294195 |
+
},
|
294196 |
+
{
|
294197 |
+
"epoch": 1224.58,
|
294198 |
+
"learning_rate": 7.543198707592892e-06,
|
294199 |
+
"loss": 0.5522,
|
294200 |
+
"step": 153135
|
294201 |
+
},
|
294202 |
+
{
|
294203 |
+
"epoch": 1224.62,
|
294204 |
+
"learning_rate": 7.543117932148627e-06,
|
294205 |
+
"loss": 0.8635,
|
294206 |
+
"step": 153140
|
294207 |
+
},
|
294208 |
+
{
|
294209 |
+
"epoch": 1224.66,
|
294210 |
+
"learning_rate": 7.543037156704362e-06,
|
294211 |
+
"loss": 0.2533,
|
294212 |
+
"step": 153145
|
294213 |
+
},
|
294214 |
+
{
|
294215 |
+
"epoch": 1224.7,
|
294216 |
+
"learning_rate": 7.542956381260097e-06,
|
294217 |
+
"loss": 0.2469,
|
294218 |
+
"step": 153150
|
294219 |
+
},
|
294220 |
+
{
|
294221 |
+
"epoch": 1224.74,
|
294222 |
+
"learning_rate": 7.542875605815832e-06,
|
294223 |
+
"loss": 0.3325,
|
294224 |
+
"step": 153155
|
294225 |
+
},
|
294226 |
+
{
|
294227 |
+
"epoch": 1224.78,
|
294228 |
+
"learning_rate": 7.542794830371568e-06,
|
294229 |
+
"loss": 0.6239,
|
294230 |
+
"step": 153160
|
294231 |
+
},
|
294232 |
+
{
|
294233 |
+
"epoch": 1224.82,
|
294234 |
+
"learning_rate": 7.542714054927302e-06,
|
294235 |
+
"loss": 0.9895,
|
294236 |
+
"step": 153165
|
294237 |
+
},
|
294238 |
+
{
|
294239 |
+
"epoch": 1224.86,
|
294240 |
+
"learning_rate": 7.542633279483038e-06,
|
294241 |
+
"loss": 0.2356,
|
294242 |
+
"step": 153170
|
294243 |
+
},
|
294244 |
+
{
|
294245 |
+
"epoch": 1224.9,
|
294246 |
+
"learning_rate": 7.542552504038773e-06,
|
294247 |
+
"loss": 0.3209,
|
294248 |
+
"step": 153175
|
294249 |
+
},
|
294250 |
+
{
|
294251 |
+
"epoch": 1224.94,
|
294252 |
+
"learning_rate": 7.542471728594508e-06,
|
294253 |
+
"loss": 0.293,
|
294254 |
+
"step": 153180
|
294255 |
+
},
|
294256 |
+
{
|
294257 |
+
"epoch": 1224.98,
|
294258 |
+
"learning_rate": 7.542390953150243e-06,
|
294259 |
+
"loss": 0.5575,
|
294260 |
+
"step": 153185
|
294261 |
+
},
|
294262 |
+
{
|
294263 |
+
"epoch": 1225.0,
|
294264 |
+
"eval_loss": 0.4036131501197815,
|
294265 |
+
"eval_runtime": 41.1248,
|
294266 |
+
"eval_samples_per_second": 20.328,
|
294267 |
+
"eval_steps_per_second": 0.657,
|
294268 |
+
"eval_wer": 0.18242702191162555,
|
294269 |
+
"step": 153188
|
294270 |
+
},
|
294271 |
+
{
|
294272 |
+
"epoch": 1235.02,
|
294273 |
+
"learning_rate": 7.542310177705978e-06,
|
294274 |
+
"loss": 0.3768,
|
294275 |
+
"step": 153190
|
294276 |
+
},
|
294277 |
+
{
|
294278 |
+
"epoch": 1235.06,
|
294279 |
+
"learning_rate": 7.542229402261713e-06,
|
294280 |
+
"loss": 0.3256,
|
294281 |
+
"step": 153195
|
294282 |
+
},
|
294283 |
+
{
|
294284 |
+
"epoch": 1235.1,
|
294285 |
+
"learning_rate": 7.542148626817448e-06,
|
294286 |
+
"loss": 0.2672,
|
294287 |
+
"step": 153200
|
294288 |
+
},
|
294289 |
+
{
|
294290 |
+
"epoch": 1235.14,
|
294291 |
+
"learning_rate": 7.542067851373183e-06,
|
294292 |
+
"loss": 0.4272,
|
294293 |
+
"step": 153205
|
294294 |
+
},
|
294295 |
+
{
|
294296 |
+
"epoch": 1235.18,
|
294297 |
+
"learning_rate": 7.541987075928918e-06,
|
294298 |
+
"loss": 0.6648,
|
294299 |
+
"step": 153210
|
294300 |
+
},
|
294301 |
+
{
|
294302 |
+
"epoch": 1235.22,
|
294303 |
+
"learning_rate": 7.541906300484654e-06,
|
294304 |
+
"loss": 1.0655,
|
294305 |
+
"step": 153215
|
294306 |
+
},
|
294307 |
+
{
|
294308 |
+
"epoch": 1235.26,
|
294309 |
+
"learning_rate": 7.541825525040388e-06,
|
294310 |
+
"loss": 0.277,
|
294311 |
+
"step": 153220
|
294312 |
+
},
|
294313 |
+
{
|
294314 |
+
"epoch": 1235.3,
|
294315 |
+
"learning_rate": 7.541744749596124e-06,
|
294316 |
+
"loss": 0.3115,
|
294317 |
+
"step": 153225
|
294318 |
+
},
|
294319 |
+
{
|
294320 |
+
"epoch": 1235.34,
|
294321 |
+
"learning_rate": 7.541663974151858e-06,
|
294322 |
+
"loss": 0.3355,
|
294323 |
+
"step": 153230
|
294324 |
+
},
|
294325 |
+
{
|
294326 |
+
"epoch": 1235.38,
|
294327 |
+
"learning_rate": 7.541583198707594e-06,
|
294328 |
+
"loss": 0.5795,
|
294329 |
+
"step": 153235
|
294330 |
+
},
|
294331 |
+
{
|
294332 |
+
"epoch": 1235.42,
|
294333 |
+
"learning_rate": 7.541502423263328e-06,
|
294334 |
+
"loss": 0.9555,
|
294335 |
+
"step": 153240
|
294336 |
+
},
|
294337 |
+
{
|
294338 |
+
"epoch": 1235.46,
|
294339 |
+
"learning_rate": 7.541421647819064e-06,
|
294340 |
+
"loss": 0.247,
|
294341 |
+
"step": 153245
|
294342 |
+
},
|
294343 |
+
{
|
294344 |
+
"epoch": 1235.5,
|
294345 |
+
"learning_rate": 7.541340872374799e-06,
|
294346 |
+
"loss": 0.3165,
|
294347 |
+
"step": 153250
|
294348 |
+
},
|
294349 |
+
{
|
294350 |
+
"epoch": 1235.54,
|
294351 |
+
"learning_rate": 7.541260096930534e-06,
|
294352 |
+
"loss": 0.3246,
|
294353 |
+
"step": 153255
|
294354 |
+
},
|
294355 |
+
{
|
294356 |
+
"epoch": 1235.58,
|
294357 |
+
"learning_rate": 7.541179321486269e-06,
|
294358 |
+
"loss": 0.6086,
|
294359 |
+
"step": 153260
|
294360 |
+
},
|
294361 |
+
{
|
294362 |
+
"epoch": 1235.62,
|
294363 |
+
"learning_rate": 7.541114701130857e-06,
|
294364 |
+
"loss": 1.0104,
|
294365 |
+
"step": 153265
|
294366 |
+
},
|
294367 |
+
{
|
294368 |
+
"epoch": 1235.66,
|
294369 |
+
"learning_rate": 7.541033925686592e-06,
|
294370 |
+
"loss": 0.3032,
|
294371 |
+
"step": 153270
|
294372 |
+
},
|
294373 |
+
{
|
294374 |
+
"epoch": 1235.7,
|
294375 |
+
"learning_rate": 7.540953150242328e-06,
|
294376 |
+
"loss": 0.2819,
|
294377 |
+
"step": 153275
|
294378 |
+
},
|
294379 |
+
{
|
294380 |
+
"epoch": 1235.74,
|
294381 |
+
"learning_rate": 7.540872374798062e-06,
|
294382 |
+
"loss": 0.3722,
|
294383 |
+
"step": 153280
|
294384 |
+
},
|
294385 |
+
{
|
294386 |
+
"epoch": 1235.78,
|
294387 |
+
"learning_rate": 7.540791599353798e-06,
|
294388 |
+
"loss": 0.5377,
|
294389 |
+
"step": 153285
|
294390 |
+
},
|
294391 |
+
{
|
294392 |
+
"epoch": 1235.82,
|
294393 |
+
"learning_rate": 7.540710823909532e-06,
|
294394 |
+
"loss": 0.9107,
|
294395 |
+
"step": 153290
|
294396 |
+
},
|
294397 |
+
{
|
294398 |
+
"epoch": 1235.86,
|
294399 |
+
"learning_rate": 7.540630048465268e-06,
|
294400 |
+
"loss": 0.2784,
|
294401 |
+
"step": 153295
|
294402 |
+
},
|
294403 |
+
{
|
294404 |
+
"epoch": 1235.9,
|
294405 |
+
"learning_rate": 7.540549273021002e-06,
|
294406 |
+
"loss": 0.4669,
|
294407 |
+
"step": 153300
|
294408 |
+
},
|
294409 |
+
{
|
294410 |
+
"epoch": 1235.94,
|
294411 |
+
"learning_rate": 7.5404684975767376e-06,
|
294412 |
+
"loss": 0.3216,
|
294413 |
+
"step": 153305
|
294414 |
+
},
|
294415 |
+
{
|
294416 |
+
"epoch": 1235.98,
|
294417 |
+
"learning_rate": 7.5403877221324725e-06,
|
294418 |
+
"loss": 0.7395,
|
294419 |
+
"step": 153310
|
294420 |
+
},
|
294421 |
+
{
|
294422 |
+
"epoch": 1236.0,
|
294423 |
+
"eval_loss": 0.4267891049385071,
|
294424 |
+
"eval_runtime": 41.6163,
|
294425 |
+
"eval_samples_per_second": 20.088,
|
294426 |
+
"eval_steps_per_second": 0.649,
|
294427 |
+
"eval_wer": 0.18537154524903376,
|
294428 |
+
"step": 153312
|
294429 |
}
|
294430 |
],
|
294431 |
+
"max_steps": 620000,
|
294432 |
"num_train_epochs": 5000,
|
294433 |
+
"total_flos": 4.3143346529907927e+20,
|
294434 |
"trial_name": null,
|
294435 |
"trial_params": null
|
294436 |
}
|
model-bin/finetune/base/{checkpoint-152690 β checkpoint-153312}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630252977.4699829/events.out.tfevents.1630252977.cc93b136ebf5.1086.219
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fce9bb6550dccfa92fd5070248c1eaa0adea6090228719306318bd35b393a7c1
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630253413.1850004/events.out.tfevents.1630253413.cc93b136ebf5.1086.221
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fcea54c33cc78347bd664ed7c545b006b39afae6d2f4c3938c518b8329b28413
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630253854.71079/events.out.tfevents.1630253854.cc93b136ebf5.1086.223
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb4e577d8a61a76383dcc972b4f6525bf89977d277aa96d3919390a96cf688c9
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630254282.042753/events.out.tfevents.1630254282.cc93b136ebf5.1086.225
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cccea25948a48b81c97d539880e4bff20b7201e890f3e60e28007074868d0389
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630254714.536368/events.out.tfevents.1630254714.cc93b136ebf5.1086.227
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb5faa028afecfaee81f4f30b722dc5ea2e181d02a29038dce872c8139f4a5a0
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630252977.cc93b136ebf5.1086.218
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e0dcc183616aa7c66d93997c2d0ca7ba1c5707044486378177096a43f9e5d806
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630253413.cc93b136ebf5.1086.220
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:54994f1e36b6ac1cd95dfcf6f8cdbd35f46d295cf24bfb8ca94ff3178845926a
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630253854.cc93b136ebf5.1086.222
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2cafc06fc47e8533d459af37e7fdcf4c5d6ecad960d99266a01edfe7a2035280
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630254282.cc93b136ebf5.1086.224
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eccb18ba1edf2eb52cf7e2b4d06a3239eb6c88d4d0cbf250abf6a9b148936020
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630254714.cc93b136ebf5.1086.226
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:774cf42f53da5f8c1a19ed1311290dc3788db045704b96a17f3f90c4964a3fad
|
3 |
+
size 8622
|