"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/trainer_state.json +164 -5
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/trainer_state.json +1902 -6
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630230484.4148061/events.out.tfevents.1630230484.cc93b136ebf5.1086.119 +3 -0
- model-bin/finetune/base/log/1630230917.0829825/events.out.tfevents.1630230917.cc93b136ebf5.1086.121 +3 -0
- model-bin/finetune/base/log/1630231362.9249244/events.out.tfevents.1630231362.cc93b136ebf5.1086.123 +3 -0
- model-bin/finetune/base/log/1630231799.7950127/events.out.tfevents.1630231799.cc93b136ebf5.1086.125 +3 -0
- model-bin/finetune/base/log/1630232234.2438745/events.out.tfevents.1630232234.cc93b136ebf5.1086.127 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630230484.cc93b136ebf5.1086.118 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630230917.cc93b136ebf5.1086.120 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630231362.cc93b136ebf5.1086.122 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630231799.cc93b136ebf5.1086.124 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630232234.cc93b136ebf5.1086.126 +3 -0
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a0ea9552f899d2d8f535fc434267063e77290df22b3ea0d489d5f3bcb93f46d0
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8dc25b2405ed70ef3243f1c3b26f7665ff79883dac75f72b56452985fad0a79b
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9340a78af637a8c0ce5c4b0fa3f2905a42f8cd1ccd9b627b31325f5eaab00977
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:998b6be1a485973ded56fe9124d5ea9cfa3bd59e3499608d8b1f8a68c1d7c92c
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:70ffda1a735fe44bfbcdb5fdd2fcf7dbc1ddead2a83c5e723492366f2b5337c5
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -285723,11 +285723,170 @@
|
|
285723 |
"eval_steps_per_second": 0.645,
|
285724 |
"eval_wer": 0.1758169934640523,
|
285725 |
"step": 146472
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
285726 |
}
|
285727 |
],
|
285728 |
"max_steps": 620000,
|
285729 |
"num_train_epochs": 5000,
|
285730 |
-
"total_flos": 4.
|
285731 |
"trial_name": null,
|
285732 |
"trial_params": null
|
285733 |
}
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.1668011442822563,
|
3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1181.995983935743,
|
5 |
+
"global_step": 146596,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
285723 |
"eval_steps_per_second": 0.645,
|
285724 |
"eval_wer": 0.1758169934640523,
|
285725 |
"step": 146472
|
285726 |
+
},
|
285727 |
+
{
|
285728 |
+
"epoch": 1181.02,
|
285729 |
+
"learning_rate": 7.65075928917609e-06,
|
285730 |
+
"loss": 0.306,
|
285731 |
+
"step": 146475
|
285732 |
+
},
|
285733 |
+
{
|
285734 |
+
"epoch": 1181.06,
|
285735 |
+
"learning_rate": 7.650678513731826e-06,
|
285736 |
+
"loss": 0.3149,
|
285737 |
+
"step": 146480
|
285738 |
+
},
|
285739 |
+
{
|
285740 |
+
"epoch": 1181.1,
|
285741 |
+
"learning_rate": 7.65059773828756e-06,
|
285742 |
+
"loss": 0.2999,
|
285743 |
+
"step": 146485
|
285744 |
+
},
|
285745 |
+
{
|
285746 |
+
"epoch": 1181.14,
|
285747 |
+
"learning_rate": 7.650516962843296e-06,
|
285748 |
+
"loss": 0.3347,
|
285749 |
+
"step": 146490
|
285750 |
+
},
|
285751 |
+
{
|
285752 |
+
"epoch": 1181.18,
|
285753 |
+
"learning_rate": 7.650436187399032e-06,
|
285754 |
+
"loss": 0.6613,
|
285755 |
+
"step": 146495
|
285756 |
+
},
|
285757 |
+
{
|
285758 |
+
"epoch": 1181.22,
|
285759 |
+
"learning_rate": 7.650355411954766e-06,
|
285760 |
+
"loss": 0.8353,
|
285761 |
+
"step": 146500
|
285762 |
+
},
|
285763 |
+
{
|
285764 |
+
"epoch": 1181.27,
|
285765 |
+
"learning_rate": 7.650274636510502e-06,
|
285766 |
+
"loss": 0.3035,
|
285767 |
+
"step": 146505
|
285768 |
+
},
|
285769 |
+
{
|
285770 |
+
"epoch": 1181.31,
|
285771 |
+
"learning_rate": 7.650193861066236e-06,
|
285772 |
+
"loss": 0.3257,
|
285773 |
+
"step": 146510
|
285774 |
+
},
|
285775 |
+
{
|
285776 |
+
"epoch": 1181.35,
|
285777 |
+
"learning_rate": 7.650113085621972e-06,
|
285778 |
+
"loss": 0.4288,
|
285779 |
+
"step": 146515
|
285780 |
+
},
|
285781 |
+
{
|
285782 |
+
"epoch": 1181.39,
|
285783 |
+
"learning_rate": 7.650032310177706e-06,
|
285784 |
+
"loss": 0.6413,
|
285785 |
+
"step": 146520
|
285786 |
+
},
|
285787 |
+
{
|
285788 |
+
"epoch": 1181.43,
|
285789 |
+
"learning_rate": 7.649951534733442e-06,
|
285790 |
+
"loss": 0.8146,
|
285791 |
+
"step": 146525
|
285792 |
+
},
|
285793 |
+
{
|
285794 |
+
"epoch": 1181.47,
|
285795 |
+
"learning_rate": 7.649870759289176e-06,
|
285796 |
+
"loss": 0.2767,
|
285797 |
+
"step": 146530
|
285798 |
+
},
|
285799 |
+
{
|
285800 |
+
"epoch": 1181.51,
|
285801 |
+
"learning_rate": 7.649789983844912e-06,
|
285802 |
+
"loss": 0.2579,
|
285803 |
+
"step": 146535
|
285804 |
+
},
|
285805 |
+
{
|
285806 |
+
"epoch": 1181.55,
|
285807 |
+
"learning_rate": 7.649709208400646e-06,
|
285808 |
+
"loss": 0.3754,
|
285809 |
+
"step": 146540
|
285810 |
+
},
|
285811 |
+
{
|
285812 |
+
"epoch": 1181.59,
|
285813 |
+
"learning_rate": 7.649628432956382e-06,
|
285814 |
+
"loss": 0.7602,
|
285815 |
+
"step": 146545
|
285816 |
+
},
|
285817 |
+
{
|
285818 |
+
"epoch": 1181.63,
|
285819 |
+
"learning_rate": 7.649547657512118e-06,
|
285820 |
+
"loss": 0.879,
|
285821 |
+
"step": 146550
|
285822 |
+
},
|
285823 |
+
{
|
285824 |
+
"epoch": 1181.67,
|
285825 |
+
"learning_rate": 7.649466882067852e-06,
|
285826 |
+
"loss": 0.2737,
|
285827 |
+
"step": 146555
|
285828 |
+
},
|
285829 |
+
{
|
285830 |
+
"epoch": 1181.71,
|
285831 |
+
"learning_rate": 7.649386106623588e-06,
|
285832 |
+
"loss": 0.2831,
|
285833 |
+
"step": 146560
|
285834 |
+
},
|
285835 |
+
{
|
285836 |
+
"epoch": 1181.75,
|
285837 |
+
"learning_rate": 7.649305331179322e-06,
|
285838 |
+
"loss": 0.3641,
|
285839 |
+
"step": 146565
|
285840 |
+
},
|
285841 |
+
{
|
285842 |
+
"epoch": 1181.79,
|
285843 |
+
"learning_rate": 7.649224555735058e-06,
|
285844 |
+
"loss": 0.7179,
|
285845 |
+
"step": 146570
|
285846 |
+
},
|
285847 |
+
{
|
285848 |
+
"epoch": 1181.83,
|
285849 |
+
"learning_rate": 7.649143780290792e-06,
|
285850 |
+
"loss": 0.8389,
|
285851 |
+
"step": 146575
|
285852 |
+
},
|
285853 |
+
{
|
285854 |
+
"epoch": 1181.87,
|
285855 |
+
"learning_rate": 7.649063004846527e-06,
|
285856 |
+
"loss": 0.2527,
|
285857 |
+
"step": 146580
|
285858 |
+
},
|
285859 |
+
{
|
285860 |
+
"epoch": 1181.91,
|
285861 |
+
"learning_rate": 7.648982229402262e-06,
|
285862 |
+
"loss": 0.2547,
|
285863 |
+
"step": 146585
|
285864 |
+
},
|
285865 |
+
{
|
285866 |
+
"epoch": 1181.95,
|
285867 |
+
"learning_rate": 7.648901453957997e-06,
|
285868 |
+
"loss": 0.3927,
|
285869 |
+
"step": 146590
|
285870 |
+
},
|
285871 |
+
{
|
285872 |
+
"epoch": 1181.99,
|
285873 |
+
"learning_rate": 7.648820678513732e-06,
|
285874 |
+
"loss": 0.8367,
|
285875 |
+
"step": 146595
|
285876 |
+
},
|
285877 |
+
{
|
285878 |
+
"epoch": 1182.0,
|
285879 |
+
"eval_loss": 0.34015288949012756,
|
285880 |
+
"eval_runtime": 41.2224,
|
285881 |
+
"eval_samples_per_second": 20.304,
|
285882 |
+
"eval_steps_per_second": 0.655,
|
285883 |
+
"eval_wer": 0.1668011442822563,
|
285884 |
+
"step": 146596
|
285885 |
}
|
285886 |
],
|
285887 |
"max_steps": 620000,
|
285888 |
"num_train_epochs": 5000,
|
285889 |
+
"total_flos": 4.125294303144466e+20,
|
285890 |
"trial_name": null,
|
285891 |
"trial_params": null
|
285892 |
}
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0bd46150aabe97ab3943950ceac270eaf7ba5a905ee8fc857f765d37689b06b8
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e642ff64d498eff54009c654ed787928c5cbb4d47b0eda3d26f1a951e1f4ec32
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6232982667952bbc19502a4f0ef987e311dcdf6d628a958e5239d2cb0215ece
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c47fa02eeeae7ce0eb2caa75c3dd36cc9207f4a30770a12e539672a58d4fadb9
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e7f4fe6d1acaa457416128487b7a022a2db612c7ffc4ebe77644069b686eb9a
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -284616,11 +284616,1907 @@
|
|
284616 |
"eval_steps_per_second": 0.687,
|
284617 |
"eval_wer": 0.16716790141568252,
|
284618 |
"step": 145602
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
284619 |
}
|
284620 |
],
|
284621 |
-
"max_steps":
|
284622 |
"num_train_epochs": 5000,
|
284623 |
-
"total_flos": 4.
|
284624 |
"trial_name": null,
|
284625 |
"trial_params": null
|
284626 |
}
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.1668011442822563,
|
3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1185.995983935743,
|
5 |
+
"global_step": 147094,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
284616 |
"eval_steps_per_second": 0.687,
|
284617 |
"eval_wer": 0.16716790141568252,
|
284618 |
"step": 145602
|
284619 |
+
},
|
284620 |
+
{
|
284621 |
+
"epoch": 1164.02,
|
284622 |
+
"learning_rate": 7.66481421647819e-06,
|
284623 |
+
"loss": 0.3273,
|
284624 |
+
"step": 145605
|
284625 |
+
},
|
284626 |
+
{
|
284627 |
+
"epoch": 1164.06,
|
284628 |
+
"learning_rate": 7.664733441033926e-06,
|
284629 |
+
"loss": 0.2867,
|
284630 |
+
"step": 145610
|
284631 |
+
},
|
284632 |
+
{
|
284633 |
+
"epoch": 1164.1,
|
284634 |
+
"learning_rate": 7.664652665589662e-06,
|
284635 |
+
"loss": 0.3066,
|
284636 |
+
"step": 145615
|
284637 |
+
},
|
284638 |
+
{
|
284639 |
+
"epoch": 1164.14,
|
284640 |
+
"learning_rate": 7.664571890145396e-06,
|
284641 |
+
"loss": 0.3405,
|
284642 |
+
"step": 145620
|
284643 |
+
},
|
284644 |
+
{
|
284645 |
+
"epoch": 1164.18,
|
284646 |
+
"learning_rate": 7.664491114701132e-06,
|
284647 |
+
"loss": 0.7475,
|
284648 |
+
"step": 145625
|
284649 |
+
},
|
284650 |
+
{
|
284651 |
+
"epoch": 1164.22,
|
284652 |
+
"learning_rate": 7.664410339256866e-06,
|
284653 |
+
"loss": 0.9464,
|
284654 |
+
"step": 145630
|
284655 |
+
},
|
284656 |
+
{
|
284657 |
+
"epoch": 1164.26,
|
284658 |
+
"learning_rate": 7.664329563812602e-06,
|
284659 |
+
"loss": 0.3219,
|
284660 |
+
"step": 145635
|
284661 |
+
},
|
284662 |
+
{
|
284663 |
+
"epoch": 1164.3,
|
284664 |
+
"learning_rate": 7.664248788368336e-06,
|
284665 |
+
"loss": 0.3095,
|
284666 |
+
"step": 145640
|
284667 |
+
},
|
284668 |
+
{
|
284669 |
+
"epoch": 1164.34,
|
284670 |
+
"learning_rate": 7.664168012924072e-06,
|
284671 |
+
"loss": 0.3189,
|
284672 |
+
"step": 145645
|
284673 |
+
},
|
284674 |
+
{
|
284675 |
+
"epoch": 1164.38,
|
284676 |
+
"learning_rate": 7.664087237479806e-06,
|
284677 |
+
"loss": 0.7207,
|
284678 |
+
"step": 145650
|
284679 |
+
},
|
284680 |
+
{
|
284681 |
+
"epoch": 1164.42,
|
284682 |
+
"learning_rate": 7.664006462035542e-06,
|
284683 |
+
"loss": 0.8142,
|
284684 |
+
"step": 145655
|
284685 |
+
},
|
284686 |
+
{
|
284687 |
+
"epoch": 1164.46,
|
284688 |
+
"learning_rate": 7.663925686591276e-06,
|
284689 |
+
"loss": 0.2481,
|
284690 |
+
"step": 145660
|
284691 |
+
},
|
284692 |
+
{
|
284693 |
+
"epoch": 1164.5,
|
284694 |
+
"learning_rate": 7.663844911147012e-06,
|
284695 |
+
"loss": 0.2636,
|
284696 |
+
"step": 145665
|
284697 |
+
},
|
284698 |
+
{
|
284699 |
+
"epoch": 1164.54,
|
284700 |
+
"learning_rate": 7.663764135702746e-06,
|
284701 |
+
"loss": 0.3821,
|
284702 |
+
"step": 145670
|
284703 |
+
},
|
284704 |
+
{
|
284705 |
+
"epoch": 1164.58,
|
284706 |
+
"learning_rate": 7.663683360258482e-06,
|
284707 |
+
"loss": 0.7099,
|
284708 |
+
"step": 145675
|
284709 |
+
},
|
284710 |
+
{
|
284711 |
+
"epoch": 1164.62,
|
284712 |
+
"learning_rate": 7.663602584814218e-06,
|
284713 |
+
"loss": 0.7693,
|
284714 |
+
"step": 145680
|
284715 |
+
},
|
284716 |
+
{
|
284717 |
+
"epoch": 1164.66,
|
284718 |
+
"learning_rate": 7.663521809369952e-06,
|
284719 |
+
"loss": 0.3184,
|
284720 |
+
"step": 145685
|
284721 |
+
},
|
284722 |
+
{
|
284723 |
+
"epoch": 1164.7,
|
284724 |
+
"learning_rate": 7.663441033925688e-06,
|
284725 |
+
"loss": 0.2638,
|
284726 |
+
"step": 145690
|
284727 |
+
},
|
284728 |
+
{
|
284729 |
+
"epoch": 1164.74,
|
284730 |
+
"learning_rate": 7.663360258481422e-06,
|
284731 |
+
"loss": 0.4138,
|
284732 |
+
"step": 145695
|
284733 |
+
},
|
284734 |
+
{
|
284735 |
+
"epoch": 1164.78,
|
284736 |
+
"learning_rate": 7.663279483037158e-06,
|
284737 |
+
"loss": 0.7006,
|
284738 |
+
"step": 145700
|
284739 |
+
},
|
284740 |
+
{
|
284741 |
+
"epoch": 1164.82,
|
284742 |
+
"learning_rate": 7.663198707592892e-06,
|
284743 |
+
"loss": 0.9005,
|
284744 |
+
"step": 145705
|
284745 |
+
},
|
284746 |
+
{
|
284747 |
+
"epoch": 1164.86,
|
284748 |
+
"learning_rate": 7.663117932148628e-06,
|
284749 |
+
"loss": 0.2489,
|
284750 |
+
"step": 145710
|
284751 |
+
},
|
284752 |
+
{
|
284753 |
+
"epoch": 1164.9,
|
284754 |
+
"learning_rate": 7.663037156704362e-06,
|
284755 |
+
"loss": 0.2708,
|
284756 |
+
"step": 145715
|
284757 |
+
},
|
284758 |
+
{
|
284759 |
+
"epoch": 1164.94,
|
284760 |
+
"learning_rate": 7.662956381260098e-06,
|
284761 |
+
"loss": 0.3742,
|
284762 |
+
"step": 145720
|
284763 |
+
},
|
284764 |
+
{
|
284765 |
+
"epoch": 1164.98,
|
284766 |
+
"learning_rate": 7.662875605815832e-06,
|
284767 |
+
"loss": 0.7858,
|
284768 |
+
"step": 145725
|
284769 |
+
},
|
284770 |
+
{
|
284771 |
+
"epoch": 1165.0,
|
284772 |
+
"eval_loss": 0.38368138670921326,
|
284773 |
+
"eval_runtime": 40.6993,
|
284774 |
+
"eval_samples_per_second": 20.565,
|
284775 |
+
"eval_steps_per_second": 0.663,
|
284776 |
+
"eval_wer": 0.1798013485101138,
|
284777 |
+
"step": 145727
|
284778 |
+
},
|
284779 |
+
{
|
284780 |
+
"epoch": 1165.02,
|
284781 |
+
"learning_rate": 7.662794830371568e-06,
|
284782 |
+
"loss": 0.4956,
|
284783 |
+
"step": 145730
|
284784 |
+
},
|
284785 |
+
{
|
284786 |
+
"epoch": 1165.06,
|
284787 |
+
"learning_rate": 7.662714054927302e-06,
|
284788 |
+
"loss": 0.2444,
|
284789 |
+
"step": 145735
|
284790 |
+
},
|
284791 |
+
{
|
284792 |
+
"epoch": 1165.1,
|
284793 |
+
"learning_rate": 7.662633279483038e-06,
|
284794 |
+
"loss": 0.2835,
|
284795 |
+
"step": 145740
|
284796 |
+
},
|
284797 |
+
{
|
284798 |
+
"epoch": 1165.14,
|
284799 |
+
"learning_rate": 7.662552504038772e-06,
|
284800 |
+
"loss": 0.4172,
|
284801 |
+
"step": 145745
|
284802 |
+
},
|
284803 |
+
{
|
284804 |
+
"epoch": 1165.18,
|
284805 |
+
"learning_rate": 7.662471728594508e-06,
|
284806 |
+
"loss": 0.656,
|
284807 |
+
"step": 145750
|
284808 |
+
},
|
284809 |
+
{
|
284810 |
+
"epoch": 1165.22,
|
284811 |
+
"learning_rate": 7.662390953150244e-06,
|
284812 |
+
"loss": 0.8101,
|
284813 |
+
"step": 145755
|
284814 |
+
},
|
284815 |
+
{
|
284816 |
+
"epoch": 1165.26,
|
284817 |
+
"learning_rate": 7.662310177705978e-06,
|
284818 |
+
"loss": 0.2991,
|
284819 |
+
"step": 145760
|
284820 |
+
},
|
284821 |
+
{
|
284822 |
+
"epoch": 1165.3,
|
284823 |
+
"learning_rate": 7.662229402261714e-06,
|
284824 |
+
"loss": 0.2851,
|
284825 |
+
"step": 145765
|
284826 |
+
},
|
284827 |
+
{
|
284828 |
+
"epoch": 1165.34,
|
284829 |
+
"learning_rate": 7.662148626817448e-06,
|
284830 |
+
"loss": 0.3243,
|
284831 |
+
"step": 145770
|
284832 |
+
},
|
284833 |
+
{
|
284834 |
+
"epoch": 1165.38,
|
284835 |
+
"learning_rate": 7.662067851373183e-06,
|
284836 |
+
"loss": 0.7754,
|
284837 |
+
"step": 145775
|
284838 |
+
},
|
284839 |
+
{
|
284840 |
+
"epoch": 1165.42,
|
284841 |
+
"learning_rate": 7.661987075928918e-06,
|
284842 |
+
"loss": 0.9141,
|
284843 |
+
"step": 145780
|
284844 |
+
},
|
284845 |
+
{
|
284846 |
+
"epoch": 1165.46,
|
284847 |
+
"learning_rate": 7.661906300484653e-06,
|
284848 |
+
"loss": 0.2886,
|
284849 |
+
"step": 145785
|
284850 |
+
},
|
284851 |
+
{
|
284852 |
+
"epoch": 1165.5,
|
284853 |
+
"learning_rate": 7.661825525040388e-06,
|
284854 |
+
"loss": 0.305,
|
284855 |
+
"step": 145790
|
284856 |
+
},
|
284857 |
+
{
|
284858 |
+
"epoch": 1165.54,
|
284859 |
+
"learning_rate": 7.661744749596123e-06,
|
284860 |
+
"loss": 0.5382,
|
284861 |
+
"step": 145795
|
284862 |
+
},
|
284863 |
+
{
|
284864 |
+
"epoch": 1165.58,
|
284865 |
+
"learning_rate": 7.661663974151858e-06,
|
284866 |
+
"loss": 0.7651,
|
284867 |
+
"step": 145800
|
284868 |
+
},
|
284869 |
+
{
|
284870 |
+
"epoch": 1165.62,
|
284871 |
+
"learning_rate": 7.661583198707593e-06,
|
284872 |
+
"loss": 0.8388,
|
284873 |
+
"step": 145805
|
284874 |
+
},
|
284875 |
+
{
|
284876 |
+
"epoch": 1165.66,
|
284877 |
+
"learning_rate": 7.661502423263328e-06,
|
284878 |
+
"loss": 0.2597,
|
284879 |
+
"step": 145810
|
284880 |
+
},
|
284881 |
+
{
|
284882 |
+
"epoch": 1165.7,
|
284883 |
+
"learning_rate": 7.661421647819063e-06,
|
284884 |
+
"loss": 0.3229,
|
284885 |
+
"step": 145815
|
284886 |
+
},
|
284887 |
+
{
|
284888 |
+
"epoch": 1165.74,
|
284889 |
+
"learning_rate": 7.6613408723748e-06,
|
284890 |
+
"loss": 0.376,
|
284891 |
+
"step": 145820
|
284892 |
+
},
|
284893 |
+
{
|
284894 |
+
"epoch": 1165.78,
|
284895 |
+
"learning_rate": 7.661260096930533e-06,
|
284896 |
+
"loss": 0.6114,
|
284897 |
+
"step": 145825
|
284898 |
+
},
|
284899 |
+
{
|
284900 |
+
"epoch": 1165.82,
|
284901 |
+
"learning_rate": 7.66117932148627e-06,
|
284902 |
+
"loss": 0.8276,
|
284903 |
+
"step": 145830
|
284904 |
+
},
|
284905 |
+
{
|
284906 |
+
"epoch": 1165.86,
|
284907 |
+
"learning_rate": 7.661098546042003e-06,
|
284908 |
+
"loss": 0.2959,
|
284909 |
+
"step": 145835
|
284910 |
+
},
|
284911 |
+
{
|
284912 |
+
"epoch": 1165.9,
|
284913 |
+
"learning_rate": 7.66101777059774e-06,
|
284914 |
+
"loss": 0.2697,
|
284915 |
+
"step": 145840
|
284916 |
+
},
|
284917 |
+
{
|
284918 |
+
"epoch": 1165.94,
|
284919 |
+
"learning_rate": 7.660936995153473e-06,
|
284920 |
+
"loss": 0.3293,
|
284921 |
+
"step": 145845
|
284922 |
+
},
|
284923 |
+
{
|
284924 |
+
"epoch": 1165.98,
|
284925 |
+
"learning_rate": 7.660856219709209e-06,
|
284926 |
+
"loss": 0.7433,
|
284927 |
+
"step": 145850
|
284928 |
+
},
|
284929 |
+
{
|
284930 |
+
"epoch": 1166.0,
|
284931 |
+
"eval_loss": 0.33688074350357056,
|
284932 |
+
"eval_runtime": 41.2402,
|
284933 |
+
"eval_samples_per_second": 20.296,
|
284934 |
+
"eval_steps_per_second": 0.655,
|
284935 |
+
"eval_wer": 0.17012599228024178,
|
284936 |
+
"step": 145852
|
284937 |
+
},
|
284938 |
+
{
|
284939 |
+
"epoch": 1176.02,
|
284940 |
+
"learning_rate": 7.660775444264943e-06,
|
284941 |
+
"loss": 0.4127,
|
284942 |
+
"step": 145855
|
284943 |
+
},
|
284944 |
+
{
|
284945 |
+
"epoch": 1176.06,
|
284946 |
+
"learning_rate": 7.660694668820679e-06,
|
284947 |
+
"loss": 0.2818,
|
284948 |
+
"step": 145860
|
284949 |
+
},
|
284950 |
+
{
|
284951 |
+
"epoch": 1176.1,
|
284952 |
+
"learning_rate": 7.660613893376413e-06,
|
284953 |
+
"loss": 0.2966,
|
284954 |
+
"step": 145865
|
284955 |
+
},
|
284956 |
+
{
|
284957 |
+
"epoch": 1176.15,
|
284958 |
+
"learning_rate": 7.660533117932149e-06,
|
284959 |
+
"loss": 0.3262,
|
284960 |
+
"step": 145870
|
284961 |
+
},
|
284962 |
+
{
|
284963 |
+
"epoch": 1176.19,
|
284964 |
+
"learning_rate": 7.660452342487883e-06,
|
284965 |
+
"loss": 0.744,
|
284966 |
+
"step": 145875
|
284967 |
+
},
|
284968 |
+
{
|
284969 |
+
"epoch": 1176.23,
|
284970 |
+
"learning_rate": 7.660371567043619e-06,
|
284971 |
+
"loss": 0.8761,
|
284972 |
+
"step": 145880
|
284973 |
+
},
|
284974 |
+
{
|
284975 |
+
"epoch": 1176.27,
|
284976 |
+
"learning_rate": 7.660290791599355e-06,
|
284977 |
+
"loss": 0.2513,
|
284978 |
+
"step": 145885
|
284979 |
+
},
|
284980 |
+
{
|
284981 |
+
"epoch": 1176.31,
|
284982 |
+
"learning_rate": 7.660210016155089e-06,
|
284983 |
+
"loss": 0.2938,
|
284984 |
+
"step": 145890
|
284985 |
+
},
|
284986 |
+
{
|
284987 |
+
"epoch": 1176.35,
|
284988 |
+
"learning_rate": 7.660129240710825e-06,
|
284989 |
+
"loss": 0.3237,
|
284990 |
+
"step": 145895
|
284991 |
+
},
|
284992 |
+
{
|
284993 |
+
"epoch": 1176.39,
|
284994 |
+
"learning_rate": 7.660048465266559e-06,
|
284995 |
+
"loss": 0.7375,
|
284996 |
+
"step": 145900
|
284997 |
+
},
|
284998 |
+
{
|
284999 |
+
"epoch": 1176.43,
|
285000 |
+
"learning_rate": 7.659967689822295e-06,
|
285001 |
+
"loss": 0.8132,
|
285002 |
+
"step": 145905
|
285003 |
+
},
|
285004 |
+
{
|
285005 |
+
"epoch": 1176.47,
|
285006 |
+
"learning_rate": 7.659886914378029e-06,
|
285007 |
+
"loss": 0.3004,
|
285008 |
+
"step": 145910
|
285009 |
+
},
|
285010 |
+
{
|
285011 |
+
"epoch": 1176.51,
|
285012 |
+
"learning_rate": 7.659806138933765e-06,
|
285013 |
+
"loss": 0.2726,
|
285014 |
+
"step": 145915
|
285015 |
+
},
|
285016 |
+
{
|
285017 |
+
"epoch": 1176.55,
|
285018 |
+
"learning_rate": 7.659725363489499e-06,
|
285019 |
+
"loss": 0.3235,
|
285020 |
+
"step": 145920
|
285021 |
+
},
|
285022 |
+
{
|
285023 |
+
"epoch": 1176.59,
|
285024 |
+
"learning_rate": 7.659644588045235e-06,
|
285025 |
+
"loss": 0.7467,
|
285026 |
+
"step": 145925
|
285027 |
+
},
|
285028 |
+
{
|
285029 |
+
"epoch": 1176.63,
|
285030 |
+
"learning_rate": 7.659563812600969e-06,
|
285031 |
+
"loss": 0.7942,
|
285032 |
+
"step": 145930
|
285033 |
+
},
|
285034 |
+
{
|
285035 |
+
"epoch": 1176.67,
|
285036 |
+
"learning_rate": 7.659483037156705e-06,
|
285037 |
+
"loss": 0.2774,
|
285038 |
+
"step": 145935
|
285039 |
+
},
|
285040 |
+
{
|
285041 |
+
"epoch": 1176.71,
|
285042 |
+
"learning_rate": 7.659402261712439e-06,
|
285043 |
+
"loss": 0.2752,
|
285044 |
+
"step": 145940
|
285045 |
+
},
|
285046 |
+
{
|
285047 |
+
"epoch": 1176.75,
|
285048 |
+
"learning_rate": 7.659321486268175e-06,
|
285049 |
+
"loss": 0.3578,
|
285050 |
+
"step": 145945
|
285051 |
+
},
|
285052 |
+
{
|
285053 |
+
"epoch": 1176.79,
|
285054 |
+
"learning_rate": 7.659240710823909e-06,
|
285055 |
+
"loss": 0.6986,
|
285056 |
+
"step": 145950
|
285057 |
+
},
|
285058 |
+
{
|
285059 |
+
"epoch": 1176.83,
|
285060 |
+
"learning_rate": 7.659159935379645e-06,
|
285061 |
+
"loss": 0.7739,
|
285062 |
+
"step": 145955
|
285063 |
+
},
|
285064 |
+
{
|
285065 |
+
"epoch": 1176.87,
|
285066 |
+
"learning_rate": 7.65907915993538e-06,
|
285067 |
+
"loss": 0.2683,
|
285068 |
+
"step": 145960
|
285069 |
+
},
|
285070 |
+
{
|
285071 |
+
"epoch": 1176.91,
|
285072 |
+
"learning_rate": 7.658998384491115e-06,
|
285073 |
+
"loss": 0.3132,
|
285074 |
+
"step": 145965
|
285075 |
+
},
|
285076 |
+
{
|
285077 |
+
"epoch": 1176.95,
|
285078 |
+
"learning_rate": 7.65891760904685e-06,
|
285079 |
+
"loss": 0.373,
|
285080 |
+
"step": 145970
|
285081 |
+
},
|
285082 |
+
{
|
285083 |
+
"epoch": 1176.99,
|
285084 |
+
"learning_rate": 7.658836833602585e-06,
|
285085 |
+
"loss": 1.049,
|
285086 |
+
"step": 145975
|
285087 |
+
},
|
285088 |
+
{
|
285089 |
+
"epoch": 1177.0,
|
285090 |
+
"eval_loss": 0.3793235719203949,
|
285091 |
+
"eval_runtime": 40.3575,
|
285092 |
+
"eval_samples_per_second": 20.74,
|
285093 |
+
"eval_steps_per_second": 0.669,
|
285094 |
+
"eval_wer": 0.17174677608440797,
|
285095 |
+
"step": 145976
|
285096 |
+
},
|
285097 |
+
{
|
285098 |
+
"epoch": 1177.03,
|
285099 |
+
"learning_rate": 7.65875605815832e-06,
|
285100 |
+
"loss": 0.3211,
|
285101 |
+
"step": 145980
|
285102 |
+
},
|
285103 |
+
{
|
285104 |
+
"epoch": 1177.07,
|
285105 |
+
"learning_rate": 7.658675282714055e-06,
|
285106 |
+
"loss": 0.2587,
|
285107 |
+
"step": 145985
|
285108 |
+
},
|
285109 |
+
{
|
285110 |
+
"epoch": 1177.11,
|
285111 |
+
"learning_rate": 7.65859450726979e-06,
|
285112 |
+
"loss": 0.3043,
|
285113 |
+
"step": 145990
|
285114 |
+
},
|
285115 |
+
{
|
285116 |
+
"epoch": 1177.15,
|
285117 |
+
"learning_rate": 7.658513731825525e-06,
|
285118 |
+
"loss": 0.374,
|
285119 |
+
"step": 145995
|
285120 |
+
},
|
285121 |
+
{
|
285122 |
+
"epoch": 1177.19,
|
285123 |
+
"learning_rate": 7.65843295638126e-06,
|
285124 |
+
"loss": 0.9153,
|
285125 |
+
"step": 146000
|
285126 |
+
},
|
285127 |
+
{
|
285128 |
+
"epoch": 1177.23,
|
285129 |
+
"learning_rate": 7.658352180936995e-06,
|
285130 |
+
"loss": 0.7194,
|
285131 |
+
"step": 146005
|
285132 |
+
},
|
285133 |
+
{
|
285134 |
+
"epoch": 1177.27,
|
285135 |
+
"learning_rate": 7.65827140549273e-06,
|
285136 |
+
"loss": 0.2684,
|
285137 |
+
"step": 146010
|
285138 |
+
},
|
285139 |
+
{
|
285140 |
+
"epoch": 1177.31,
|
285141 |
+
"learning_rate": 7.658190630048465e-06,
|
285142 |
+
"loss": 0.2866,
|
285143 |
+
"step": 146015
|
285144 |
+
},
|
285145 |
+
{
|
285146 |
+
"epoch": 1177.35,
|
285147 |
+
"learning_rate": 7.6581098546042e-06,
|
285148 |
+
"loss": 0.3972,
|
285149 |
+
"step": 146020
|
285150 |
+
},
|
285151 |
+
{
|
285152 |
+
"epoch": 1177.39,
|
285153 |
+
"learning_rate": 7.658029079159936e-06,
|
285154 |
+
"loss": 1.0221,
|
285155 |
+
"step": 146025
|
285156 |
+
},
|
285157 |
+
{
|
285158 |
+
"epoch": 1177.43,
|
285159 |
+
"learning_rate": 7.65794830371567e-06,
|
285160 |
+
"loss": 0.5779,
|
285161 |
+
"step": 146030
|
285162 |
+
},
|
285163 |
+
{
|
285164 |
+
"epoch": 1177.47,
|
285165 |
+
"learning_rate": 7.657867528271406e-06,
|
285166 |
+
"loss": 0.3161,
|
285167 |
+
"step": 146035
|
285168 |
+
},
|
285169 |
+
{
|
285170 |
+
"epoch": 1177.51,
|
285171 |
+
"learning_rate": 7.65778675282714e-06,
|
285172 |
+
"loss": 0.2923,
|
285173 |
+
"step": 146040
|
285174 |
+
},
|
285175 |
+
{
|
285176 |
+
"epoch": 1177.55,
|
285177 |
+
"learning_rate": 7.657705977382876e-06,
|
285178 |
+
"loss": 0.4056,
|
285179 |
+
"step": 146045
|
285180 |
+
},
|
285181 |
+
{
|
285182 |
+
"epoch": 1177.59,
|
285183 |
+
"learning_rate": 7.65762520193861e-06,
|
285184 |
+
"loss": 0.8064,
|
285185 |
+
"step": 146050
|
285186 |
+
},
|
285187 |
+
{
|
285188 |
+
"epoch": 1177.63,
|
285189 |
+
"learning_rate": 7.657544426494346e-06,
|
285190 |
+
"loss": 0.6026,
|
285191 |
+
"step": 146055
|
285192 |
+
},
|
285193 |
+
{
|
285194 |
+
"epoch": 1177.67,
|
285195 |
+
"learning_rate": 7.65746365105008e-06,
|
285196 |
+
"loss": 0.2731,
|
285197 |
+
"step": 146060
|
285198 |
+
},
|
285199 |
+
{
|
285200 |
+
"epoch": 1177.71,
|
285201 |
+
"learning_rate": 7.657382875605816e-06,
|
285202 |
+
"loss": 0.2761,
|
285203 |
+
"step": 146065
|
285204 |
+
},
|
285205 |
+
{
|
285206 |
+
"epoch": 1177.76,
|
285207 |
+
"learning_rate": 7.65730210016155e-06,
|
285208 |
+
"loss": 0.3563,
|
285209 |
+
"step": 146070
|
285210 |
+
},
|
285211 |
+
{
|
285212 |
+
"epoch": 1177.8,
|
285213 |
+
"learning_rate": 7.657221324717286e-06,
|
285214 |
+
"loss": 0.8102,
|
285215 |
+
"step": 146075
|
285216 |
+
},
|
285217 |
+
{
|
285218 |
+
"epoch": 1177.84,
|
285219 |
+
"learning_rate": 7.65714054927302e-06,
|
285220 |
+
"loss": 0.5813,
|
285221 |
+
"step": 146080
|
285222 |
+
},
|
285223 |
+
{
|
285224 |
+
"epoch": 1177.88,
|
285225 |
+
"learning_rate": 7.657059773828756e-06,
|
285226 |
+
"loss": 0.3342,
|
285227 |
+
"step": 146085
|
285228 |
+
},
|
285229 |
+
{
|
285230 |
+
"epoch": 1177.92,
|
285231 |
+
"learning_rate": 7.656978998384492e-06,
|
285232 |
+
"loss": 0.3355,
|
285233 |
+
"step": 146090
|
285234 |
+
},
|
285235 |
+
{
|
285236 |
+
"epoch": 1177.96,
|
285237 |
+
"learning_rate": 7.656898222940226e-06,
|
285238 |
+
"loss": 0.4923,
|
285239 |
+
"step": 146095
|
285240 |
+
},
|
285241 |
+
{
|
285242 |
+
"epoch": 1178.0,
|
285243 |
+
"learning_rate": 7.656817447495962e-06,
|
285244 |
+
"loss": 1.2601,
|
285245 |
+
"step": 146100
|
285246 |
+
},
|
285247 |
+
{
|
285248 |
+
"epoch": 1178.0,
|
285249 |
+
"eval_loss": 0.4012053608894348,
|
285250 |
+
"eval_runtime": 40.577,
|
285251 |
+
"eval_samples_per_second": 20.627,
|
285252 |
+
"eval_steps_per_second": 0.665,
|
285253 |
+
"eval_wer": 0.1786023678891135,
|
285254 |
+
"step": 146100
|
285255 |
+
},
|
285256 |
+
{
|
285257 |
+
"epoch": 1178.04,
|
285258 |
+
"learning_rate": 7.656736672051696e-06,
|
285259 |
+
"loss": 0.2932,
|
285260 |
+
"step": 146105
|
285261 |
+
},
|
285262 |
+
{
|
285263 |
+
"epoch": 1178.08,
|
285264 |
+
"learning_rate": 7.656655896607432e-06,
|
285265 |
+
"loss": 0.2603,
|
285266 |
+
"step": 146110
|
285267 |
+
},
|
285268 |
+
{
|
285269 |
+
"epoch": 1178.12,
|
285270 |
+
"learning_rate": 7.656575121163166e-06,
|
285271 |
+
"loss": 0.2993,
|
285272 |
+
"step": 146115
|
285273 |
+
},
|
285274 |
+
{
|
285275 |
+
"epoch": 1178.16,
|
285276 |
+
"learning_rate": 7.656494345718902e-06,
|
285277 |
+
"loss": 0.4776,
|
285278 |
+
"step": 146120
|
285279 |
+
},
|
285280 |
+
{
|
285281 |
+
"epoch": 1178.2,
|
285282 |
+
"learning_rate": 7.656413570274636e-06,
|
285283 |
+
"loss": 1.1354,
|
285284 |
+
"step": 146125
|
285285 |
+
},
|
285286 |
+
{
|
285287 |
+
"epoch": 1178.24,
|
285288 |
+
"learning_rate": 7.656332794830372e-06,
|
285289 |
+
"loss": 0.2897,
|
285290 |
+
"step": 146130
|
285291 |
+
},
|
285292 |
+
{
|
285293 |
+
"epoch": 1178.28,
|
285294 |
+
"learning_rate": 7.656252019386106e-06,
|
285295 |
+
"loss": 0.2256,
|
285296 |
+
"step": 146135
|
285297 |
+
},
|
285298 |
+
{
|
285299 |
+
"epoch": 1178.32,
|
285300 |
+
"learning_rate": 7.656171243941842e-06,
|
285301 |
+
"loss": 0.2845,
|
285302 |
+
"step": 146140
|
285303 |
+
},
|
285304 |
+
{
|
285305 |
+
"epoch": 1178.36,
|
285306 |
+
"learning_rate": 7.656090468497578e-06,
|
285307 |
+
"loss": 0.4063,
|
285308 |
+
"step": 146145
|
285309 |
+
},
|
285310 |
+
{
|
285311 |
+
"epoch": 1178.4,
|
285312 |
+
"learning_rate": 7.656009693053312e-06,
|
285313 |
+
"loss": 1.1799,
|
285314 |
+
"step": 146150
|
285315 |
+
},
|
285316 |
+
{
|
285317 |
+
"epoch": 1178.44,
|
285318 |
+
"learning_rate": 7.655928917609048e-06,
|
285319 |
+
"loss": 0.3141,
|
285320 |
+
"step": 146155
|
285321 |
+
},
|
285322 |
+
{
|
285323 |
+
"epoch": 1178.48,
|
285324 |
+
"learning_rate": 7.655848142164782e-06,
|
285325 |
+
"loss": 0.2586,
|
285326 |
+
"step": 146160
|
285327 |
+
},
|
285328 |
+
{
|
285329 |
+
"epoch": 1178.52,
|
285330 |
+
"learning_rate": 7.655767366720518e-06,
|
285331 |
+
"loss": 0.3502,
|
285332 |
+
"step": 146165
|
285333 |
+
},
|
285334 |
+
{
|
285335 |
+
"epoch": 1178.56,
|
285336 |
+
"learning_rate": 7.655686591276252e-06,
|
285337 |
+
"loss": 0.4228,
|
285338 |
+
"step": 146170
|
285339 |
+
},
|
285340 |
+
{
|
285341 |
+
"epoch": 1178.6,
|
285342 |
+
"learning_rate": 7.655605815831988e-06,
|
285343 |
+
"loss": 1.2601,
|
285344 |
+
"step": 146175
|
285345 |
+
},
|
285346 |
+
{
|
285347 |
+
"epoch": 1178.64,
|
285348 |
+
"learning_rate": 7.655525040387722e-06,
|
285349 |
+
"loss": 0.2679,
|
285350 |
+
"step": 146180
|
285351 |
+
},
|
285352 |
+
{
|
285353 |
+
"epoch": 1178.68,
|
285354 |
+
"learning_rate": 7.655444264943458e-06,
|
285355 |
+
"loss": 0.2357,
|
285356 |
+
"step": 146185
|
285357 |
+
},
|
285358 |
+
{
|
285359 |
+
"epoch": 1178.72,
|
285360 |
+
"learning_rate": 7.655363489499192e-06,
|
285361 |
+
"loss": 0.3045,
|
285362 |
+
"step": 146190
|
285363 |
+
},
|
285364 |
+
{
|
285365 |
+
"epoch": 1178.76,
|
285366 |
+
"learning_rate": 7.655282714054928e-06,
|
285367 |
+
"loss": 0.5178,
|
285368 |
+
"step": 146195
|
285369 |
+
},
|
285370 |
+
{
|
285371 |
+
"epoch": 1178.8,
|
285372 |
+
"learning_rate": 7.655201938610663e-06,
|
285373 |
+
"loss": 1.2614,
|
285374 |
+
"step": 146200
|
285375 |
+
},
|
285376 |
+
{
|
285377 |
+
"epoch": 1178.84,
|
285378 |
+
"learning_rate": 7.655121163166397e-06,
|
285379 |
+
"loss": 0.3759,
|
285380 |
+
"step": 146205
|
285381 |
+
},
|
285382 |
+
{
|
285383 |
+
"epoch": 1178.88,
|
285384 |
+
"learning_rate": 7.655040387722133e-06,
|
285385 |
+
"loss": 0.2394,
|
285386 |
+
"step": 146210
|
285387 |
+
},
|
285388 |
+
{
|
285389 |
+
"epoch": 1178.92,
|
285390 |
+
"learning_rate": 7.654959612277867e-06,
|
285391 |
+
"loss": 0.275,
|
285392 |
+
"step": 146215
|
285393 |
+
},
|
285394 |
+
{
|
285395 |
+
"epoch": 1178.96,
|
285396 |
+
"learning_rate": 7.654878836833603e-06,
|
285397 |
+
"loss": 0.5791,
|
285398 |
+
"step": 146220
|
285399 |
+
},
|
285400 |
+
{
|
285401 |
+
"epoch": 1179.0,
|
285402 |
+
"eval_loss": 0.358010470867157,
|
285403 |
+
"eval_runtime": 40.5201,
|
285404 |
+
"eval_samples_per_second": 20.656,
|
285405 |
+
"eval_steps_per_second": 0.666,
|
285406 |
+
"eval_wer": 0.176530911221799,
|
285407 |
+
"step": 146224
|
285408 |
+
},
|
285409 |
+
{
|
285410 |
+
"epoch": 1179.01,
|
285411 |
+
"learning_rate": 7.654798061389337e-06,
|
285412 |
+
"loss": 0.5013,
|
285413 |
+
"step": 146225
|
285414 |
+
},
|
285415 |
+
{
|
285416 |
+
"epoch": 1179.05,
|
285417 |
+
"learning_rate": 7.654717285945073e-06,
|
285418 |
+
"loss": 0.2654,
|
285419 |
+
"step": 146230
|
285420 |
+
},
|
285421 |
+
{
|
285422 |
+
"epoch": 1179.09,
|
285423 |
+
"learning_rate": 7.654636510500807e-06,
|
285424 |
+
"loss": 0.2647,
|
285425 |
+
"step": 146235
|
285426 |
+
},
|
285427 |
+
{
|
285428 |
+
"epoch": 1179.13,
|
285429 |
+
"learning_rate": 7.654555735056543e-06,
|
285430 |
+
"loss": 0.3248,
|
285431 |
+
"step": 146240
|
285432 |
+
},
|
285433 |
+
{
|
285434 |
+
"epoch": 1179.17,
|
285435 |
+
"learning_rate": 7.654474959612277e-06,
|
285436 |
+
"loss": 0.4694,
|
285437 |
+
"step": 146245
|
285438 |
+
},
|
285439 |
+
{
|
285440 |
+
"epoch": 1179.21,
|
285441 |
+
"learning_rate": 7.654394184168013e-06,
|
285442 |
+
"loss": 1.0148,
|
285443 |
+
"step": 146250
|
285444 |
+
},
|
285445 |
+
{
|
285446 |
+
"epoch": 1179.25,
|
285447 |
+
"learning_rate": 7.654313408723747e-06,
|
285448 |
+
"loss": 0.2833,
|
285449 |
+
"step": 146255
|
285450 |
+
},
|
285451 |
+
{
|
285452 |
+
"epoch": 1179.29,
|
285453 |
+
"learning_rate": 7.654232633279483e-06,
|
285454 |
+
"loss": 0.2983,
|
285455 |
+
"step": 146260
|
285456 |
+
},
|
285457 |
+
{
|
285458 |
+
"epoch": 1179.33,
|
285459 |
+
"learning_rate": 7.654151857835219e-06,
|
285460 |
+
"loss": 0.3522,
|
285461 |
+
"step": 146265
|
285462 |
+
},
|
285463 |
+
{
|
285464 |
+
"epoch": 1179.37,
|
285465 |
+
"learning_rate": 7.654071082390953e-06,
|
285466 |
+
"loss": 0.5271,
|
285467 |
+
"step": 146270
|
285468 |
+
},
|
285469 |
+
{
|
285470 |
+
"epoch": 1179.41,
|
285471 |
+
"learning_rate": 7.653990306946689e-06,
|
285472 |
+
"loss": 1.1747,
|
285473 |
+
"step": 146275
|
285474 |
+
},
|
285475 |
+
{
|
285476 |
+
"epoch": 1179.45,
|
285477 |
+
"learning_rate": 7.653909531502423e-06,
|
285478 |
+
"loss": 0.2796,
|
285479 |
+
"step": 146280
|
285480 |
+
},
|
285481 |
+
{
|
285482 |
+
"epoch": 1179.49,
|
285483 |
+
"learning_rate": 7.653828756058159e-06,
|
285484 |
+
"loss": 0.2518,
|
285485 |
+
"step": 146285
|
285486 |
+
},
|
285487 |
+
{
|
285488 |
+
"epoch": 1179.53,
|
285489 |
+
"learning_rate": 7.653747980613893e-06,
|
285490 |
+
"loss": 0.307,
|
285491 |
+
"step": 146290
|
285492 |
+
},
|
285493 |
+
{
|
285494 |
+
"epoch": 1179.57,
|
285495 |
+
"learning_rate": 7.653667205169629e-06,
|
285496 |
+
"loss": 0.4902,
|
285497 |
+
"step": 146295
|
285498 |
+
},
|
285499 |
+
{
|
285500 |
+
"epoch": 1179.61,
|
285501 |
+
"learning_rate": 7.653586429725363e-06,
|
285502 |
+
"loss": 1.0019,
|
285503 |
+
"step": 146300
|
285504 |
+
},
|
285505 |
+
{
|
285506 |
+
"epoch": 1179.65,
|
285507 |
+
"learning_rate": 7.653505654281099e-06,
|
285508 |
+
"loss": 0.278,
|
285509 |
+
"step": 146305
|
285510 |
+
},
|
285511 |
+
{
|
285512 |
+
"epoch": 1179.69,
|
285513 |
+
"learning_rate": 7.653424878836833e-06,
|
285514 |
+
"loss": 0.3384,
|
285515 |
+
"step": 146310
|
285516 |
+
},
|
285517 |
+
{
|
285518 |
+
"epoch": 1179.73,
|
285519 |
+
"learning_rate": 7.653344103392569e-06,
|
285520 |
+
"loss": 0.3353,
|
285521 |
+
"step": 146315
|
285522 |
+
},
|
285523 |
+
{
|
285524 |
+
"epoch": 1179.77,
|
285525 |
+
"learning_rate": 7.653263327948305e-06,
|
285526 |
+
"loss": 0.5559,
|
285527 |
+
"step": 146320
|
285528 |
+
},
|
285529 |
+
{
|
285530 |
+
"epoch": 1179.81,
|
285531 |
+
"learning_rate": 7.653182552504039e-06,
|
285532 |
+
"loss": 1.4385,
|
285533 |
+
"step": 146325
|
285534 |
+
},
|
285535 |
+
{
|
285536 |
+
"epoch": 1179.85,
|
285537 |
+
"learning_rate": 7.653101777059775e-06,
|
285538 |
+
"loss": 0.2435,
|
285539 |
+
"step": 146330
|
285540 |
+
},
|
285541 |
+
{
|
285542 |
+
"epoch": 1179.9,
|
285543 |
+
"learning_rate": 7.653021001615509e-06,
|
285544 |
+
"loss": 0.2579,
|
285545 |
+
"step": 146335
|
285546 |
+
},
|
285547 |
+
{
|
285548 |
+
"epoch": 1179.94,
|
285549 |
+
"learning_rate": 7.652940226171245e-06,
|
285550 |
+
"loss": 0.345,
|
285551 |
+
"step": 146340
|
285552 |
+
},
|
285553 |
+
{
|
285554 |
+
"epoch": 1179.98,
|
285555 |
+
"learning_rate": 7.652859450726979e-06,
|
285556 |
+
"loss": 0.5621,
|
285557 |
+
"step": 146345
|
285558 |
+
},
|
285559 |
+
{
|
285560 |
+
"epoch": 1180.0,
|
285561 |
+
"eval_loss": 0.449453204870224,
|
285562 |
+
"eval_runtime": 40.0278,
|
285563 |
+
"eval_samples_per_second": 20.91,
|
285564 |
+
"eval_steps_per_second": 0.675,
|
285565 |
+
"eval_wer": 0.18728573557907288,
|
285566 |
+
"step": 146348
|
285567 |
+
},
|
285568 |
+
{
|
285569 |
+
"epoch": 1180.02,
|
285570 |
+
"learning_rate": 7.652778675282715e-06,
|
285571 |
+
"loss": 0.2882,
|
285572 |
+
"step": 146350
|
285573 |
+
},
|
285574 |
+
{
|
285575 |
+
"epoch": 1180.06,
|
285576 |
+
"learning_rate": 7.652697899838449e-06,
|
285577 |
+
"loss": 0.2908,
|
285578 |
+
"step": 146355
|
285579 |
+
},
|
285580 |
+
{
|
285581 |
+
"epoch": 1180.1,
|
285582 |
+
"learning_rate": 7.652617124394185e-06,
|
285583 |
+
"loss": 0.2317,
|
285584 |
+
"step": 146360
|
285585 |
+
},
|
285586 |
+
{
|
285587 |
+
"epoch": 1180.14,
|
285588 |
+
"learning_rate": 7.652536348949919e-06,
|
285589 |
+
"loss": 0.3416,
|
285590 |
+
"step": 146365
|
285591 |
+
},
|
285592 |
+
{
|
285593 |
+
"epoch": 1180.18,
|
285594 |
+
"learning_rate": 7.652455573505655e-06,
|
285595 |
+
"loss": 0.6812,
|
285596 |
+
"step": 146370
|
285597 |
+
},
|
285598 |
+
{
|
285599 |
+
"epoch": 1180.22,
|
285600 |
+
"learning_rate": 7.65237479806139e-06,
|
285601 |
+
"loss": 0.8339,
|
285602 |
+
"step": 146375
|
285603 |
+
},
|
285604 |
+
{
|
285605 |
+
"epoch": 1180.26,
|
285606 |
+
"learning_rate": 7.652294022617125e-06,
|
285607 |
+
"loss": 0.431,
|
285608 |
+
"step": 146380
|
285609 |
+
},
|
285610 |
+
{
|
285611 |
+
"epoch": 1180.3,
|
285612 |
+
"learning_rate": 7.65221324717286e-06,
|
285613 |
+
"loss": 0.3809,
|
285614 |
+
"step": 146385
|
285615 |
+
},
|
285616 |
+
{
|
285617 |
+
"epoch": 1180.34,
|
285618 |
+
"learning_rate": 7.652132471728595e-06,
|
285619 |
+
"loss": 0.3004,
|
285620 |
+
"step": 146390
|
285621 |
+
},
|
285622 |
+
{
|
285623 |
+
"epoch": 1180.38,
|
285624 |
+
"learning_rate": 7.65205169628433e-06,
|
285625 |
+
"loss": 0.578,
|
285626 |
+
"step": 146395
|
285627 |
+
},
|
285628 |
+
{
|
285629 |
+
"epoch": 1180.42,
|
285630 |
+
"learning_rate": 7.651970920840065e-06,
|
285631 |
+
"loss": 0.9797,
|
285632 |
+
"step": 146400
|
285633 |
+
},
|
285634 |
+
{
|
285635 |
+
"epoch": 1180.46,
|
285636 |
+
"learning_rate": 7.6518901453958e-06,
|
285637 |
+
"loss": 0.2545,
|
285638 |
+
"step": 146405
|
285639 |
+
},
|
285640 |
+
{
|
285641 |
+
"epoch": 1180.5,
|
285642 |
+
"learning_rate": 7.651809369951535e-06,
|
285643 |
+
"loss": 0.2621,
|
285644 |
+
"step": 146410
|
285645 |
+
},
|
285646 |
+
{
|
285647 |
+
"epoch": 1180.54,
|
285648 |
+
"learning_rate": 7.65172859450727e-06,
|
285649 |
+
"loss": 0.3193,
|
285650 |
+
"step": 146415
|
285651 |
+
},
|
285652 |
+
{
|
285653 |
+
"epoch": 1180.58,
|
285654 |
+
"learning_rate": 7.651647819063005e-06,
|
285655 |
+
"loss": 0.6539,
|
285656 |
+
"step": 146420
|
285657 |
+
},
|
285658 |
+
{
|
285659 |
+
"epoch": 1180.62,
|
285660 |
+
"learning_rate": 7.65156704361874e-06,
|
285661 |
+
"loss": 1.0436,
|
285662 |
+
"step": 146425
|
285663 |
+
},
|
285664 |
+
{
|
285665 |
+
"epoch": 1180.66,
|
285666 |
+
"learning_rate": 7.651486268174474e-06,
|
285667 |
+
"loss": 0.2563,
|
285668 |
+
"step": 146430
|
285669 |
+
},
|
285670 |
+
{
|
285671 |
+
"epoch": 1180.7,
|
285672 |
+
"learning_rate": 7.65140549273021e-06,
|
285673 |
+
"loss": 0.3142,
|
285674 |
+
"step": 146435
|
285675 |
+
},
|
285676 |
+
{
|
285677 |
+
"epoch": 1180.74,
|
285678 |
+
"learning_rate": 7.651324717285946e-06,
|
285679 |
+
"loss": 0.3735,
|
285680 |
+
"step": 146440
|
285681 |
+
},
|
285682 |
+
{
|
285683 |
+
"epoch": 1180.78,
|
285684 |
+
"learning_rate": 7.65124394184168e-06,
|
285685 |
+
"loss": 0.6211,
|
285686 |
+
"step": 146445
|
285687 |
+
},
|
285688 |
+
{
|
285689 |
+
"epoch": 1180.82,
|
285690 |
+
"learning_rate": 7.651163166397416e-06,
|
285691 |
+
"loss": 1.0659,
|
285692 |
+
"step": 146450
|
285693 |
+
},
|
285694 |
+
{
|
285695 |
+
"epoch": 1180.86,
|
285696 |
+
"learning_rate": 7.65108239095315e-06,
|
285697 |
+
"loss": 0.3007,
|
285698 |
+
"step": 146455
|
285699 |
+
},
|
285700 |
+
{
|
285701 |
+
"epoch": 1180.9,
|
285702 |
+
"learning_rate": 7.651001615508886e-06,
|
285703 |
+
"loss": 0.2761,
|
285704 |
+
"step": 146460
|
285705 |
+
},
|
285706 |
+
{
|
285707 |
+
"epoch": 1180.94,
|
285708 |
+
"learning_rate": 7.65092084006462e-06,
|
285709 |
+
"loss": 0.3487,
|
285710 |
+
"step": 146465
|
285711 |
+
},
|
285712 |
+
{
|
285713 |
+
"epoch": 1180.98,
|
285714 |
+
"learning_rate": 7.650840064620356e-06,
|
285715 |
+
"loss": 0.7013,
|
285716 |
+
"step": 146470
|
285717 |
+
},
|
285718 |
+
{
|
285719 |
+
"epoch": 1181.0,
|
285720 |
+
"eval_loss": 0.34985774755477905,
|
285721 |
+
"eval_runtime": 41.8541,
|
285722 |
+
"eval_samples_per_second": 19.998,
|
285723 |
+
"eval_steps_per_second": 0.645,
|
285724 |
+
"eval_wer": 0.1758169934640523,
|
285725 |
+
"step": 146472
|
285726 |
+
},
|
285727 |
+
{
|
285728 |
+
"epoch": 1181.02,
|
285729 |
+
"learning_rate": 7.65075928917609e-06,
|
285730 |
+
"loss": 0.306,
|
285731 |
+
"step": 146475
|
285732 |
+
},
|
285733 |
+
{
|
285734 |
+
"epoch": 1181.06,
|
285735 |
+
"learning_rate": 7.650678513731826e-06,
|
285736 |
+
"loss": 0.3149,
|
285737 |
+
"step": 146480
|
285738 |
+
},
|
285739 |
+
{
|
285740 |
+
"epoch": 1181.1,
|
285741 |
+
"learning_rate": 7.65059773828756e-06,
|
285742 |
+
"loss": 0.2999,
|
285743 |
+
"step": 146485
|
285744 |
+
},
|
285745 |
+
{
|
285746 |
+
"epoch": 1181.14,
|
285747 |
+
"learning_rate": 7.650516962843296e-06,
|
285748 |
+
"loss": 0.3347,
|
285749 |
+
"step": 146490
|
285750 |
+
},
|
285751 |
+
{
|
285752 |
+
"epoch": 1181.18,
|
285753 |
+
"learning_rate": 7.650436187399032e-06,
|
285754 |
+
"loss": 0.6613,
|
285755 |
+
"step": 146495
|
285756 |
+
},
|
285757 |
+
{
|
285758 |
+
"epoch": 1181.22,
|
285759 |
+
"learning_rate": 7.650355411954766e-06,
|
285760 |
+
"loss": 0.8353,
|
285761 |
+
"step": 146500
|
285762 |
+
},
|
285763 |
+
{
|
285764 |
+
"epoch": 1181.27,
|
285765 |
+
"learning_rate": 7.650274636510502e-06,
|
285766 |
+
"loss": 0.3035,
|
285767 |
+
"step": 146505
|
285768 |
+
},
|
285769 |
+
{
|
285770 |
+
"epoch": 1181.31,
|
285771 |
+
"learning_rate": 7.650193861066236e-06,
|
285772 |
+
"loss": 0.3257,
|
285773 |
+
"step": 146510
|
285774 |
+
},
|
285775 |
+
{
|
285776 |
+
"epoch": 1181.35,
|
285777 |
+
"learning_rate": 7.650113085621972e-06,
|
285778 |
+
"loss": 0.4288,
|
285779 |
+
"step": 146515
|
285780 |
+
},
|
285781 |
+
{
|
285782 |
+
"epoch": 1181.39,
|
285783 |
+
"learning_rate": 7.650032310177706e-06,
|
285784 |
+
"loss": 0.6413,
|
285785 |
+
"step": 146520
|
285786 |
+
},
|
285787 |
+
{
|
285788 |
+
"epoch": 1181.43,
|
285789 |
+
"learning_rate": 7.649951534733442e-06,
|
285790 |
+
"loss": 0.8146,
|
285791 |
+
"step": 146525
|
285792 |
+
},
|
285793 |
+
{
|
285794 |
+
"epoch": 1181.47,
|
285795 |
+
"learning_rate": 7.649870759289176e-06,
|
285796 |
+
"loss": 0.2767,
|
285797 |
+
"step": 146530
|
285798 |
+
},
|
285799 |
+
{
|
285800 |
+
"epoch": 1181.51,
|
285801 |
+
"learning_rate": 7.649789983844912e-06,
|
285802 |
+
"loss": 0.2579,
|
285803 |
+
"step": 146535
|
285804 |
+
},
|
285805 |
+
{
|
285806 |
+
"epoch": 1181.55,
|
285807 |
+
"learning_rate": 7.649709208400646e-06,
|
285808 |
+
"loss": 0.3754,
|
285809 |
+
"step": 146540
|
285810 |
+
},
|
285811 |
+
{
|
285812 |
+
"epoch": 1181.59,
|
285813 |
+
"learning_rate": 7.649628432956382e-06,
|
285814 |
+
"loss": 0.7602,
|
285815 |
+
"step": 146545
|
285816 |
+
},
|
285817 |
+
{
|
285818 |
+
"epoch": 1181.63,
|
285819 |
+
"learning_rate": 7.649547657512118e-06,
|
285820 |
+
"loss": 0.879,
|
285821 |
+
"step": 146550
|
285822 |
+
},
|
285823 |
+
{
|
285824 |
+
"epoch": 1181.67,
|
285825 |
+
"learning_rate": 7.649466882067852e-06,
|
285826 |
+
"loss": 0.2737,
|
285827 |
+
"step": 146555
|
285828 |
+
},
|
285829 |
+
{
|
285830 |
+
"epoch": 1181.71,
|
285831 |
+
"learning_rate": 7.649386106623588e-06,
|
285832 |
+
"loss": 0.2831,
|
285833 |
+
"step": 146560
|
285834 |
+
},
|
285835 |
+
{
|
285836 |
+
"epoch": 1181.75,
|
285837 |
+
"learning_rate": 7.649305331179322e-06,
|
285838 |
+
"loss": 0.3641,
|
285839 |
+
"step": 146565
|
285840 |
+
},
|
285841 |
+
{
|
285842 |
+
"epoch": 1181.79,
|
285843 |
+
"learning_rate": 7.649224555735058e-06,
|
285844 |
+
"loss": 0.7179,
|
285845 |
+
"step": 146570
|
285846 |
+
},
|
285847 |
+
{
|
285848 |
+
"epoch": 1181.83,
|
285849 |
+
"learning_rate": 7.649143780290792e-06,
|
285850 |
+
"loss": 0.8389,
|
285851 |
+
"step": 146575
|
285852 |
+
},
|
285853 |
+
{
|
285854 |
+
"epoch": 1181.87,
|
285855 |
+
"learning_rate": 7.649063004846527e-06,
|
285856 |
+
"loss": 0.2527,
|
285857 |
+
"step": 146580
|
285858 |
+
},
|
285859 |
+
{
|
285860 |
+
"epoch": 1181.91,
|
285861 |
+
"learning_rate": 7.648982229402262e-06,
|
285862 |
+
"loss": 0.2547,
|
285863 |
+
"step": 146585
|
285864 |
+
},
|
285865 |
+
{
|
285866 |
+
"epoch": 1181.95,
|
285867 |
+
"learning_rate": 7.648901453957997e-06,
|
285868 |
+
"loss": 0.3927,
|
285869 |
+
"step": 146590
|
285870 |
+
},
|
285871 |
+
{
|
285872 |
+
"epoch": 1181.99,
|
285873 |
+
"learning_rate": 7.648820678513732e-06,
|
285874 |
+
"loss": 0.8367,
|
285875 |
+
"step": 146595
|
285876 |
+
},
|
285877 |
+
{
|
285878 |
+
"epoch": 1182.0,
|
285879 |
+
"eval_loss": 0.34015288949012756,
|
285880 |
+
"eval_runtime": 41.2224,
|
285881 |
+
"eval_samples_per_second": 20.304,
|
285882 |
+
"eval_steps_per_second": 0.655,
|
285883 |
+
"eval_wer": 0.1668011442822563,
|
285884 |
+
"step": 146596
|
285885 |
+
},
|
285886 |
+
{
|
285887 |
+
"epoch": 1172.03,
|
285888 |
+
"learning_rate": 7.648739903069467e-06,
|
285889 |
+
"loss": 0.3097,
|
285890 |
+
"step": 146600
|
285891 |
+
},
|
285892 |
+
{
|
285893 |
+
"epoch": 1172.07,
|
285894 |
+
"learning_rate": 7.648659127625202e-06,
|
285895 |
+
"loss": 0.2985,
|
285896 |
+
"step": 146605
|
285897 |
+
},
|
285898 |
+
{
|
285899 |
+
"epoch": 1172.11,
|
285900 |
+
"learning_rate": 7.648578352180937e-06,
|
285901 |
+
"loss": 0.2698,
|
285902 |
+
"step": 146610
|
285903 |
+
},
|
285904 |
+
{
|
285905 |
+
"epoch": 1172.15,
|
285906 |
+
"learning_rate": 7.648497576736673e-06,
|
285907 |
+
"loss": 0.3778,
|
285908 |
+
"step": 146615
|
285909 |
+
},
|
285910 |
+
{
|
285911 |
+
"epoch": 1172.19,
|
285912 |
+
"learning_rate": 7.648416801292407e-06,
|
285913 |
+
"loss": 1.0694,
|
285914 |
+
"step": 146620
|
285915 |
+
},
|
285916 |
+
{
|
285917 |
+
"epoch": 1172.23,
|
285918 |
+
"learning_rate": 7.648336025848143e-06,
|
285919 |
+
"loss": 0.5919,
|
285920 |
+
"step": 146625
|
285921 |
+
},
|
285922 |
+
{
|
285923 |
+
"epoch": 1172.27,
|
285924 |
+
"learning_rate": 7.648255250403877e-06,
|
285925 |
+
"loss": 0.2701,
|
285926 |
+
"step": 146630
|
285927 |
+
},
|
285928 |
+
{
|
285929 |
+
"epoch": 1172.31,
|
285930 |
+
"learning_rate": 7.648174474959613e-06,
|
285931 |
+
"loss": 0.2998,
|
285932 |
+
"step": 146635
|
285933 |
+
},
|
285934 |
+
{
|
285935 |
+
"epoch": 1172.35,
|
285936 |
+
"learning_rate": 7.648093699515347e-06,
|
285937 |
+
"loss": 0.3737,
|
285938 |
+
"step": 146640
|
285939 |
+
},
|
285940 |
+
{
|
285941 |
+
"epoch": 1172.39,
|
285942 |
+
"learning_rate": 7.648012924071083e-06,
|
285943 |
+
"loss": 1.0077,
|
285944 |
+
"step": 146645
|
285945 |
+
},
|
285946 |
+
{
|
285947 |
+
"epoch": 1172.43,
|
285948 |
+
"learning_rate": 7.647932148626817e-06,
|
285949 |
+
"loss": 0.6127,
|
285950 |
+
"step": 146650
|
285951 |
+
},
|
285952 |
+
{
|
285953 |
+
"epoch": 1172.47,
|
285954 |
+
"learning_rate": 7.647851373182553e-06,
|
285955 |
+
"loss": 0.3308,
|
285956 |
+
"step": 146655
|
285957 |
+
},
|
285958 |
+
{
|
285959 |
+
"epoch": 1172.51,
|
285960 |
+
"learning_rate": 7.647770597738287e-06,
|
285961 |
+
"loss": 0.312,
|
285962 |
+
"step": 146660
|
285963 |
+
},
|
285964 |
+
{
|
285965 |
+
"epoch": 1172.55,
|
285966 |
+
"learning_rate": 7.647689822294023e-06,
|
285967 |
+
"loss": 0.4248,
|
285968 |
+
"step": 146665
|
285969 |
+
},
|
285970 |
+
{
|
285971 |
+
"epoch": 1172.59,
|
285972 |
+
"learning_rate": 7.647609046849759e-06,
|
285973 |
+
"loss": 0.8503,
|
285974 |
+
"step": 146670
|
285975 |
+
},
|
285976 |
+
{
|
285977 |
+
"epoch": 1172.63,
|
285978 |
+
"learning_rate": 7.647528271405493e-06,
|
285979 |
+
"loss": 0.7404,
|
285980 |
+
"step": 146675
|
285981 |
+
},
|
285982 |
+
{
|
285983 |
+
"epoch": 1172.67,
|
285984 |
+
"learning_rate": 7.647447495961229e-06,
|
285985 |
+
"loss": 0.3356,
|
285986 |
+
"step": 146680
|
285987 |
+
},
|
285988 |
+
{
|
285989 |
+
"epoch": 1172.71,
|
285990 |
+
"learning_rate": 7.647366720516963e-06,
|
285991 |
+
"loss": 0.2776,
|
285992 |
+
"step": 146685
|
285993 |
+
},
|
285994 |
+
{
|
285995 |
+
"epoch": 1172.75,
|
285996 |
+
"learning_rate": 7.647285945072699e-06,
|
285997 |
+
"loss": 0.4439,
|
285998 |
+
"step": 146690
|
285999 |
+
},
|
286000 |
+
{
|
286001 |
+
"epoch": 1172.79,
|
286002 |
+
"learning_rate": 7.647205169628433e-06,
|
286003 |
+
"loss": 0.957,
|
286004 |
+
"step": 146695
|
286005 |
+
},
|
286006 |
+
{
|
286007 |
+
"epoch": 1172.83,
|
286008 |
+
"learning_rate": 7.647124394184169e-06,
|
286009 |
+
"loss": 0.5475,
|
286010 |
+
"step": 146700
|
286011 |
+
},
|
286012 |
+
{
|
286013 |
+
"epoch": 1172.87,
|
286014 |
+
"learning_rate": 7.647043618739903e-06,
|
286015 |
+
"loss": 0.2893,
|
286016 |
+
"step": 146705
|
286017 |
+
},
|
286018 |
+
{
|
286019 |
+
"epoch": 1172.91,
|
286020 |
+
"learning_rate": 7.646962843295639e-06,
|
286021 |
+
"loss": 0.2689,
|
286022 |
+
"step": 146710
|
286023 |
+
},
|
286024 |
+
{
|
286025 |
+
"epoch": 1172.95,
|
286026 |
+
"learning_rate": 7.646882067851373e-06,
|
286027 |
+
"loss": 0.3002,
|
286028 |
+
"step": 146715
|
286029 |
+
},
|
286030 |
+
{
|
286031 |
+
"epoch": 1172.99,
|
286032 |
+
"learning_rate": 7.646801292407109e-06,
|
286033 |
+
"loss": 0.9223,
|
286034 |
+
"step": 146720
|
286035 |
+
},
|
286036 |
+
{
|
286037 |
+
"epoch": 1173.0,
|
286038 |
+
"eval_loss": 0.3900996744632721,
|
286039 |
+
"eval_runtime": 41.2507,
|
286040 |
+
"eval_samples_per_second": 20.291,
|
286041 |
+
"eval_steps_per_second": 0.655,
|
286042 |
+
"eval_wer": 0.1805263539476546,
|
286043 |
+
"step": 146721
|
286044 |
+
},
|
286045 |
+
{
|
286046 |
+
"epoch": 1183.03,
|
286047 |
+
"learning_rate": 7.646720516962845e-06,
|
286048 |
+
"loss": 0.3527,
|
286049 |
+
"step": 146725
|
286050 |
+
},
|
286051 |
+
{
|
286052 |
+
"epoch": 1183.07,
|
286053 |
+
"learning_rate": 7.646639741518579e-06,
|
286054 |
+
"loss": 0.2931,
|
286055 |
+
"step": 146730
|
286056 |
+
},
|
286057 |
+
{
|
286058 |
+
"epoch": 1183.11,
|
286059 |
+
"learning_rate": 7.646558966074315e-06,
|
286060 |
+
"loss": 0.2955,
|
286061 |
+
"step": 146735
|
286062 |
+
},
|
286063 |
+
{
|
286064 |
+
"epoch": 1183.15,
|
286065 |
+
"learning_rate": 7.646478190630049e-06,
|
286066 |
+
"loss": 0.3639,
|
286067 |
+
"step": 146740
|
286068 |
+
},
|
286069 |
+
{
|
286070 |
+
"epoch": 1183.19,
|
286071 |
+
"learning_rate": 7.646397415185785e-06,
|
286072 |
+
"loss": 0.8325,
|
286073 |
+
"step": 146745
|
286074 |
+
},
|
286075 |
+
{
|
286076 |
+
"epoch": 1183.23,
|
286077 |
+
"learning_rate": 7.646316639741519e-06,
|
286078 |
+
"loss": 0.5682,
|
286079 |
+
"step": 146750
|
286080 |
+
},
|
286081 |
+
{
|
286082 |
+
"epoch": 1183.27,
|
286083 |
+
"learning_rate": 7.646235864297255e-06,
|
286084 |
+
"loss": 0.3174,
|
286085 |
+
"step": 146755
|
286086 |
+
},
|
286087 |
+
{
|
286088 |
+
"epoch": 1183.31,
|
286089 |
+
"learning_rate": 7.646155088852989e-06,
|
286090 |
+
"loss": 0.3424,
|
286091 |
+
"step": 146760
|
286092 |
+
},
|
286093 |
+
{
|
286094 |
+
"epoch": 1183.35,
|
286095 |
+
"learning_rate": 7.646074313408725e-06,
|
286096 |
+
"loss": 0.3159,
|
286097 |
+
"step": 146765
|
286098 |
+
},
|
286099 |
+
{
|
286100 |
+
"epoch": 1183.39,
|
286101 |
+
"learning_rate": 7.645993537964459e-06,
|
286102 |
+
"loss": 0.7583,
|
286103 |
+
"step": 146770
|
286104 |
+
},
|
286105 |
+
{
|
286106 |
+
"epoch": 1183.43,
|
286107 |
+
"learning_rate": 7.645912762520195e-06,
|
286108 |
+
"loss": 0.5913,
|
286109 |
+
"step": 146775
|
286110 |
+
},
|
286111 |
+
{
|
286112 |
+
"epoch": 1183.47,
|
286113 |
+
"learning_rate": 7.645831987075929e-06,
|
286114 |
+
"loss": 0.2356,
|
286115 |
+
"step": 146780
|
286116 |
+
},
|
286117 |
+
{
|
286118 |
+
"epoch": 1183.51,
|
286119 |
+
"learning_rate": 7.645751211631665e-06,
|
286120 |
+
"loss": 0.3036,
|
286121 |
+
"step": 146785
|
286122 |
+
},
|
286123 |
+
{
|
286124 |
+
"epoch": 1183.55,
|
286125 |
+
"learning_rate": 7.6456704361874e-06,
|
286126 |
+
"loss": 0.4368,
|
286127 |
+
"step": 146790
|
286128 |
+
},
|
286129 |
+
{
|
286130 |
+
"epoch": 1183.59,
|
286131 |
+
"learning_rate": 7.645589660743134e-06,
|
286132 |
+
"loss": 0.8289,
|
286133 |
+
"step": 146795
|
286134 |
+
},
|
286135 |
+
{
|
286136 |
+
"epoch": 1183.63,
|
286137 |
+
"learning_rate": 7.64550888529887e-06,
|
286138 |
+
"loss": 0.6256,
|
286139 |
+
"step": 146800
|
286140 |
+
},
|
286141 |
+
{
|
286142 |
+
"epoch": 1183.67,
|
286143 |
+
"learning_rate": 7.645428109854604e-06,
|
286144 |
+
"loss": 0.2853,
|
286145 |
+
"step": 146805
|
286146 |
+
},
|
286147 |
+
{
|
286148 |
+
"epoch": 1183.71,
|
286149 |
+
"learning_rate": 7.64534733441034e-06,
|
286150 |
+
"loss": 0.3083,
|
286151 |
+
"step": 146810
|
286152 |
+
},
|
286153 |
+
{
|
286154 |
+
"epoch": 1183.76,
|
286155 |
+
"learning_rate": 7.645266558966074e-06,
|
286156 |
+
"loss": 0.4051,
|
286157 |
+
"step": 146815
|
286158 |
+
},
|
286159 |
+
{
|
286160 |
+
"epoch": 1183.8,
|
286161 |
+
"learning_rate": 7.64518578352181e-06,
|
286162 |
+
"loss": 1.006,
|
286163 |
+
"step": 146820
|
286164 |
+
},
|
286165 |
+
{
|
286166 |
+
"epoch": 1183.84,
|
286167 |
+
"learning_rate": 7.645105008077544e-06,
|
286168 |
+
"loss": 0.7374,
|
286169 |
+
"step": 146825
|
286170 |
+
},
|
286171 |
+
{
|
286172 |
+
"epoch": 1183.88,
|
286173 |
+
"learning_rate": 7.64502423263328e-06,
|
286174 |
+
"loss": 0.3358,
|
286175 |
+
"step": 146830
|
286176 |
+
},
|
286177 |
+
{
|
286178 |
+
"epoch": 1183.92,
|
286179 |
+
"learning_rate": 7.644943457189014e-06,
|
286180 |
+
"loss": 0.2254,
|
286181 |
+
"step": 146835
|
286182 |
+
},
|
286183 |
+
{
|
286184 |
+
"epoch": 1183.96,
|
286185 |
+
"learning_rate": 7.64486268174475e-06,
|
286186 |
+
"loss": 0.4519,
|
286187 |
+
"step": 146840
|
286188 |
+
},
|
286189 |
+
{
|
286190 |
+
"epoch": 1184.0,
|
286191 |
+
"learning_rate": 7.644781906300486e-06,
|
286192 |
+
"loss": 1.0456,
|
286193 |
+
"step": 146845
|
286194 |
+
},
|
286195 |
+
{
|
286196 |
+
"epoch": 1184.0,
|
286197 |
+
"eval_loss": 0.3610660433769226,
|
286198 |
+
"eval_runtime": 41.7394,
|
286199 |
+
"eval_samples_per_second": 20.053,
|
286200 |
+
"eval_steps_per_second": 0.647,
|
286201 |
+
"eval_wer": 0.16990750855727915,
|
286202 |
+
"step": 146845
|
286203 |
+
},
|
286204 |
+
{
|
286205 |
+
"epoch": 1174.04,
|
286206 |
+
"learning_rate": 7.64470113085622e-06,
|
286207 |
+
"loss": 0.2998,
|
286208 |
+
"step": 146850
|
286209 |
+
},
|
286210 |
+
{
|
286211 |
+
"epoch": 1174.08,
|
286212 |
+
"learning_rate": 7.644620355411956e-06,
|
286213 |
+
"loss": 0.369,
|
286214 |
+
"step": 146855
|
286215 |
+
},
|
286216 |
+
{
|
286217 |
+
"epoch": 1174.12,
|
286218 |
+
"learning_rate": 7.64453957996769e-06,
|
286219 |
+
"loss": 0.3144,
|
286220 |
+
"step": 146860
|
286221 |
+
},
|
286222 |
+
{
|
286223 |
+
"epoch": 1174.16,
|
286224 |
+
"learning_rate": 7.644458804523426e-06,
|
286225 |
+
"loss": 0.4513,
|
286226 |
+
"step": 146865
|
286227 |
+
},
|
286228 |
+
{
|
286229 |
+
"epoch": 1174.2,
|
286230 |
+
"learning_rate": 7.64437802907916e-06,
|
286231 |
+
"loss": 1.1218,
|
286232 |
+
"step": 146870
|
286233 |
+
},
|
286234 |
+
{
|
286235 |
+
"epoch": 1174.24,
|
286236 |
+
"learning_rate": 7.644297253634896e-06,
|
286237 |
+
"loss": 0.2873,
|
286238 |
+
"step": 146875
|
286239 |
+
},
|
286240 |
+
{
|
286241 |
+
"epoch": 1174.28,
|
286242 |
+
"learning_rate": 7.64421647819063e-06,
|
286243 |
+
"loss": 0.2713,
|
286244 |
+
"step": 146880
|
286245 |
+
},
|
286246 |
+
{
|
286247 |
+
"epoch": 1174.32,
|
286248 |
+
"learning_rate": 7.644135702746366e-06,
|
286249 |
+
"loss": 0.2787,
|
286250 |
+
"step": 146885
|
286251 |
+
},
|
286252 |
+
{
|
286253 |
+
"epoch": 1174.36,
|
286254 |
+
"learning_rate": 7.6440549273021e-06,
|
286255 |
+
"loss": 0.4724,
|
286256 |
+
"step": 146890
|
286257 |
+
},
|
286258 |
+
{
|
286259 |
+
"epoch": 1174.4,
|
286260 |
+
"learning_rate": 7.643974151857836e-06,
|
286261 |
+
"loss": 1.149,
|
286262 |
+
"step": 146895
|
286263 |
+
},
|
286264 |
+
{
|
286265 |
+
"epoch": 1174.44,
|
286266 |
+
"learning_rate": 7.643893376413572e-06,
|
286267 |
+
"loss": 0.3304,
|
286268 |
+
"step": 146900
|
286269 |
+
},
|
286270 |
+
{
|
286271 |
+
"epoch": 1174.48,
|
286272 |
+
"learning_rate": 7.643812600969306e-06,
|
286273 |
+
"loss": 0.2587,
|
286274 |
+
"step": 146905
|
286275 |
+
},
|
286276 |
+
{
|
286277 |
+
"epoch": 1174.52,
|
286278 |
+
"learning_rate": 7.643731825525042e-06,
|
286279 |
+
"loss": 0.2845,
|
286280 |
+
"step": 146910
|
286281 |
+
},
|
286282 |
+
{
|
286283 |
+
"epoch": 1174.56,
|
286284 |
+
"learning_rate": 7.643651050080776e-06,
|
286285 |
+
"loss": 0.4071,
|
286286 |
+
"step": 146915
|
286287 |
+
},
|
286288 |
+
{
|
286289 |
+
"epoch": 1174.6,
|
286290 |
+
"learning_rate": 7.643570274636512e-06,
|
286291 |
+
"loss": 1.1262,
|
286292 |
+
"step": 146920
|
286293 |
+
},
|
286294 |
+
{
|
286295 |
+
"epoch": 1174.64,
|
286296 |
+
"learning_rate": 7.643489499192246e-06,
|
286297 |
+
"loss": 0.3429,
|
286298 |
+
"step": 146925
|
286299 |
+
},
|
286300 |
+
{
|
286301 |
+
"epoch": 1174.68,
|
286302 |
+
"learning_rate": 7.643408723747982e-06,
|
286303 |
+
"loss": 0.3127,
|
286304 |
+
"step": 146930
|
286305 |
+
},
|
286306 |
+
{
|
286307 |
+
"epoch": 1174.72,
|
286308 |
+
"learning_rate": 7.643327948303716e-06,
|
286309 |
+
"loss": 0.2756,
|
286310 |
+
"step": 146935
|
286311 |
+
},
|
286312 |
+
{
|
286313 |
+
"epoch": 1174.76,
|
286314 |
+
"learning_rate": 7.643247172859452e-06,
|
286315 |
+
"loss": 0.5054,
|
286316 |
+
"step": 146940
|
286317 |
+
},
|
286318 |
+
{
|
286319 |
+
"epoch": 1174.8,
|
286320 |
+
"learning_rate": 7.643166397415186e-06,
|
286321 |
+
"loss": 1.0996,
|
286322 |
+
"step": 146945
|
286323 |
+
},
|
286324 |
+
{
|
286325 |
+
"epoch": 1174.84,
|
286326 |
+
"learning_rate": 7.643085621970922e-06,
|
286327 |
+
"loss": 0.3266,
|
286328 |
+
"step": 146950
|
286329 |
+
},
|
286330 |
+
{
|
286331 |
+
"epoch": 1174.88,
|
286332 |
+
"learning_rate": 7.643004846526656e-06,
|
286333 |
+
"loss": 0.2945,
|
286334 |
+
"step": 146955
|
286335 |
+
},
|
286336 |
+
{
|
286337 |
+
"epoch": 1174.92,
|
286338 |
+
"learning_rate": 7.642924071082392e-06,
|
286339 |
+
"loss": 0.3572,
|
286340 |
+
"step": 146960
|
286341 |
+
},
|
286342 |
+
{
|
286343 |
+
"epoch": 1174.96,
|
286344 |
+
"learning_rate": 7.642843295638127e-06,
|
286345 |
+
"loss": 0.4849,
|
286346 |
+
"step": 146965
|
286347 |
+
},
|
286348 |
+
{
|
286349 |
+
"epoch": 1175.0,
|
286350 |
+
"learning_rate": 7.642762520193862e-06,
|
286351 |
+
"loss": 1.3137,
|
286352 |
+
"step": 146970
|
286353 |
+
},
|
286354 |
+
{
|
286355 |
+
"epoch": 1175.0,
|
286356 |
+
"eval_loss": 0.3988247811794281,
|
286357 |
+
"eval_runtime": 41.5097,
|
286358 |
+
"eval_samples_per_second": 20.164,
|
286359 |
+
"eval_steps_per_second": 0.65,
|
286360 |
+
"eval_wer": 0.1729923798358734,
|
286361 |
+
"step": 146970
|
286362 |
+
},
|
286363 |
+
{
|
286364 |
+
"epoch": 1185.04,
|
286365 |
+
"learning_rate": 7.642681744749597e-06,
|
286366 |
+
"loss": 0.3428,
|
286367 |
+
"step": 146975
|
286368 |
+
},
|
286369 |
+
{
|
286370 |
+
"epoch": 1185.08,
|
286371 |
+
"learning_rate": 7.642600969305332e-06,
|
286372 |
+
"loss": 0.2385,
|
286373 |
+
"step": 146980
|
286374 |
+
},
|
286375 |
+
{
|
286376 |
+
"epoch": 1185.12,
|
286377 |
+
"learning_rate": 7.642520193861067e-06,
|
286378 |
+
"loss": 0.3427,
|
286379 |
+
"step": 146985
|
286380 |
+
},
|
286381 |
+
{
|
286382 |
+
"epoch": 1185.16,
|
286383 |
+
"learning_rate": 7.642439418416802e-06,
|
286384 |
+
"loss": 0.3936,
|
286385 |
+
"step": 146990
|
286386 |
+
},
|
286387 |
+
{
|
286388 |
+
"epoch": 1185.2,
|
286389 |
+
"learning_rate": 7.642358642972537e-06,
|
286390 |
+
"loss": 1.0638,
|
286391 |
+
"step": 146995
|
286392 |
+
},
|
286393 |
+
{
|
286394 |
+
"epoch": 1185.24,
|
286395 |
+
"learning_rate": 7.642277867528272e-06,
|
286396 |
+
"loss": 0.304,
|
286397 |
+
"step": 147000
|
286398 |
+
},
|
286399 |
+
{
|
286400 |
+
"epoch": 1185.28,
|
286401 |
+
"learning_rate": 7.642197092084007e-06,
|
286402 |
+
"loss": 0.2536,
|
286403 |
+
"step": 147005
|
286404 |
+
},
|
286405 |
+
{
|
286406 |
+
"epoch": 1185.32,
|
286407 |
+
"learning_rate": 7.642116316639741e-06,
|
286408 |
+
"loss": 0.3188,
|
286409 |
+
"step": 147010
|
286410 |
+
},
|
286411 |
+
{
|
286412 |
+
"epoch": 1185.36,
|
286413 |
+
"learning_rate": 7.642035541195477e-06,
|
286414 |
+
"loss": 0.5025,
|
286415 |
+
"step": 147015
|
286416 |
+
},
|
286417 |
+
{
|
286418 |
+
"epoch": 1185.4,
|
286419 |
+
"learning_rate": 7.641954765751213e-06,
|
286420 |
+
"loss": 1.2492,
|
286421 |
+
"step": 147020
|
286422 |
+
},
|
286423 |
+
{
|
286424 |
+
"epoch": 1185.44,
|
286425 |
+
"learning_rate": 7.641873990306947e-06,
|
286426 |
+
"loss": 0.2905,
|
286427 |
+
"step": 147025
|
286428 |
+
},
|
286429 |
+
{
|
286430 |
+
"epoch": 1185.48,
|
286431 |
+
"learning_rate": 7.641793214862683e-06,
|
286432 |
+
"loss": 0.3111,
|
286433 |
+
"step": 147030
|
286434 |
+
},
|
286435 |
+
{
|
286436 |
+
"epoch": 1185.52,
|
286437 |
+
"learning_rate": 7.641712439418417e-06,
|
286438 |
+
"loss": 0.3018,
|
286439 |
+
"step": 147035
|
286440 |
+
},
|
286441 |
+
{
|
286442 |
+
"epoch": 1185.56,
|
286443 |
+
"learning_rate": 7.641631663974153e-06,
|
286444 |
+
"loss": 0.4451,
|
286445 |
+
"step": 147040
|
286446 |
+
},
|
286447 |
+
{
|
286448 |
+
"epoch": 1185.6,
|
286449 |
+
"learning_rate": 7.641550888529887e-06,
|
286450 |
+
"loss": 1.0976,
|
286451 |
+
"step": 147045
|
286452 |
+
},
|
286453 |
+
{
|
286454 |
+
"epoch": 1185.64,
|
286455 |
+
"learning_rate": 7.641470113085623e-06,
|
286456 |
+
"loss": 0.3226,
|
286457 |
+
"step": 147050
|
286458 |
+
},
|
286459 |
+
{
|
286460 |
+
"epoch": 1185.68,
|
286461 |
+
"learning_rate": 7.641389337641357e-06,
|
286462 |
+
"loss": 0.2418,
|
286463 |
+
"step": 147055
|
286464 |
+
},
|
286465 |
+
{
|
286466 |
+
"epoch": 1185.72,
|
286467 |
+
"learning_rate": 7.641308562197093e-06,
|
286468 |
+
"loss": 0.3772,
|
286469 |
+
"step": 147060
|
286470 |
+
},
|
286471 |
+
{
|
286472 |
+
"epoch": 1185.76,
|
286473 |
+
"learning_rate": 7.641227786752827e-06,
|
286474 |
+
"loss": 0.4226,
|
286475 |
+
"step": 147065
|
286476 |
+
},
|
286477 |
+
{
|
286478 |
+
"epoch": 1185.8,
|
286479 |
+
"learning_rate": 7.641147011308563e-06,
|
286480 |
+
"loss": 1.1949,
|
286481 |
+
"step": 147070
|
286482 |
+
},
|
286483 |
+
{
|
286484 |
+
"epoch": 1185.84,
|
286485 |
+
"learning_rate": 7.641066235864299e-06,
|
286486 |
+
"loss": 0.3455,
|
286487 |
+
"step": 147075
|
286488 |
+
},
|
286489 |
+
{
|
286490 |
+
"epoch": 1185.88,
|
286491 |
+
"learning_rate": 7.640985460420033e-06,
|
286492 |
+
"loss": 0.2664,
|
286493 |
+
"step": 147080
|
286494 |
+
},
|
286495 |
+
{
|
286496 |
+
"epoch": 1185.92,
|
286497 |
+
"learning_rate": 7.640904684975769e-06,
|
286498 |
+
"loss": 0.3259,
|
286499 |
+
"step": 147085
|
286500 |
+
},
|
286501 |
+
{
|
286502 |
+
"epoch": 1185.96,
|
286503 |
+
"learning_rate": 7.640823909531503e-06,
|
286504 |
+
"loss": 0.4639,
|
286505 |
+
"step": 147090
|
286506 |
+
},
|
286507 |
+
{
|
286508 |
+
"epoch": 1186.0,
|
286509 |
+
"eval_loss": 0.41105982661247253,
|
286510 |
+
"eval_runtime": 41.11,
|
286511 |
+
"eval_samples_per_second": 20.36,
|
286512 |
+
"eval_steps_per_second": 0.657,
|
286513 |
+
"eval_wer": 0.1767975743574935,
|
286514 |
+
"step": 147094
|
286515 |
}
|
286516 |
],
|
286517 |
+
"max_steps": 620000,
|
286518 |
"num_train_epochs": 5000,
|
286519 |
+
"total_flos": 4.1393121623086616e+20,
|
286520 |
"trial_name": null,
|
286521 |
"trial_params": null
|
286522 |
}
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630230484.4148061/events.out.tfevents.1630230484.cc93b136ebf5.1086.119
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:373e03a0065b2e921eefd867e202f84f548054ff66160ac42f034369030f3220
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630230917.0829825/events.out.tfevents.1630230917.cc93b136ebf5.1086.121
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4941891e313e1a07773f5f32179f173fb50469156bae41861b1c5fdb30df5d6
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630231362.9249244/events.out.tfevents.1630231362.cc93b136ebf5.1086.123
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a30f575e5d92c7236fc30cd24fb7cea6c129bed326a9687dcf47af1671d3698
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630231799.7950127/events.out.tfevents.1630231799.cc93b136ebf5.1086.125
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d581e2ac1cb2ca93023856f14c8f46437e5fbfd5ed33ec467f2dd73f4f8a4907
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630232234.2438745/events.out.tfevents.1630232234.cc93b136ebf5.1086.127
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b656bb8581cc3b0e8e81c57ed8bf041161d884c948fa9c079028a2323c3785ce
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630230484.cc93b136ebf5.1086.118
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21341d712d1a1d11a64108a5815e19580ec3c3b91f3adb0fcfa9f363754af853
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630230917.cc93b136ebf5.1086.120
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9207bff65ec5f514cf53a6f839848b0d88cffaba3a0673819f5b98e62a4475cd
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630231362.cc93b136ebf5.1086.122
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e143394f3a4ff3d7ad1b06893d174d817d4c9b3c2fd9b14acd48ed85157171bd
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630231799.cc93b136ebf5.1086.124
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13d1c476968ed905c4cb56413c983b39e1d176290a9c27dfd7a277e45fbe8511
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630232234.cc93b136ebf5.1086.126
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17df0783581cbdfcc60e0903bf27fdb281f33af1809e9e2e65cabef2ab06909c
|
3 |
+
size 8462
|