"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630082884.969348/events.out.tfevents.1630082884.dea8124df033.1391.7 +3 -0
- model-bin/finetune/base/log/1630083307.1487231/events.out.tfevents.1630083307.dea8124df033.1391.9 +3 -0
- model-bin/finetune/base/log/1630083735.6256015/events.out.tfevents.1630083735.dea8124df033.1391.11 +3 -0
- model-bin/finetune/base/log/1630084174.1170754/events.out.tfevents.1630084174.dea8124df033.1391.13 +3 -0
- model-bin/finetune/base/log/1630084598.2761755/events.out.tfevents.1630084598.dea8124df033.1391.15 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630082884.dea8124df033.1391.6 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630083307.dea8124df033.1391.8 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630083735.dea8124df033.1391.10 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630084174.dea8124df033.1391.12 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630084598.dea8124df033.1391.14 +3 -0
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1d348482d17ba9e738c11c659795d83ff2bcf3cfd11baf8797570f6c9b26411a
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc2b6659e86d951e27ce212a257a0b916a94a565fd3d84e430bfde3718539cb4
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df743e69dcda2e4cf4a9f3f8ef82556dde39a400b2c9ae8b01c8bdc7894d03fb
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7fe07346d923a1da088bea107f5ab09f94c498e2865fabad8db6fc04e90f2d85
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1905b108e5460ced8e245231c20dc93cf00ce38aab3458596eda343809ebdd48
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -242820,11 +242820,800 @@
|
|
242820 |
"eval_steps_per_second": 0.674,
|
242821 |
"eval_wer": 0.19159501927067893,
|
242822 |
"step": 112750
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
242823 |
}
|
242824 |
],
|
242825 |
-
"max_steps":
|
242826 |
"num_train_epochs": 5000,
|
242827 |
-
"total_flos": 3.
|
242828 |
"trial_name": null,
|
242829 |
"trial_params": null
|
242830 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 913.995983935743,
|
5 |
+
"global_step": 113370,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
242820 |
"eval_steps_per_second": 0.674,
|
242821 |
"eval_wer": 0.19159501927067893,
|
242822 |
"step": 112750
|
242823 |
+
},
|
242824 |
+
{
|
242825 |
+
"epoch": 909.04,
|
242826 |
+
"learning_rate": 8.19529886914378e-06,
|
242827 |
+
"loss": 0.3194,
|
242828 |
+
"step": 112755
|
242829 |
+
},
|
242830 |
+
{
|
242831 |
+
"epoch": 909.08,
|
242832 |
+
"learning_rate": 8.195218093699516e-06,
|
242833 |
+
"loss": 0.2594,
|
242834 |
+
"step": 112760
|
242835 |
+
},
|
242836 |
+
{
|
242837 |
+
"epoch": 909.12,
|
242838 |
+
"learning_rate": 8.19513731825525e-06,
|
242839 |
+
"loss": 0.323,
|
242840 |
+
"step": 112765
|
242841 |
+
},
|
242842 |
+
{
|
242843 |
+
"epoch": 909.16,
|
242844 |
+
"learning_rate": 8.195056542810986e-06,
|
242845 |
+
"loss": 0.3869,
|
242846 |
+
"step": 112770
|
242847 |
+
},
|
242848 |
+
{
|
242849 |
+
"epoch": 909.2,
|
242850 |
+
"learning_rate": 8.19497576736672e-06,
|
242851 |
+
"loss": 1.2147,
|
242852 |
+
"step": 112775
|
242853 |
+
},
|
242854 |
+
{
|
242855 |
+
"epoch": 909.24,
|
242856 |
+
"learning_rate": 8.194894991922456e-06,
|
242857 |
+
"loss": 0.2554,
|
242858 |
+
"step": 112780
|
242859 |
+
},
|
242860 |
+
{
|
242861 |
+
"epoch": 909.28,
|
242862 |
+
"learning_rate": 8.19481421647819e-06,
|
242863 |
+
"loss": 0.2766,
|
242864 |
+
"step": 112785
|
242865 |
+
},
|
242866 |
+
{
|
242867 |
+
"epoch": 909.32,
|
242868 |
+
"learning_rate": 8.194733441033926e-06,
|
242869 |
+
"loss": 0.3144,
|
242870 |
+
"step": 112790
|
242871 |
+
},
|
242872 |
+
{
|
242873 |
+
"epoch": 909.36,
|
242874 |
+
"learning_rate": 8.194652665589662e-06,
|
242875 |
+
"loss": 0.4586,
|
242876 |
+
"step": 112795
|
242877 |
+
},
|
242878 |
+
{
|
242879 |
+
"epoch": 909.4,
|
242880 |
+
"learning_rate": 8.194571890145396e-06,
|
242881 |
+
"loss": 1.1297,
|
242882 |
+
"step": 112800
|
242883 |
+
},
|
242884 |
+
{
|
242885 |
+
"epoch": 909.44,
|
242886 |
+
"learning_rate": 8.194491114701132e-06,
|
242887 |
+
"loss": 0.3203,
|
242888 |
+
"step": 112805
|
242889 |
+
},
|
242890 |
+
{
|
242891 |
+
"epoch": 909.48,
|
242892 |
+
"learning_rate": 8.194410339256866e-06,
|
242893 |
+
"loss": 0.2947,
|
242894 |
+
"step": 112810
|
242895 |
+
},
|
242896 |
+
{
|
242897 |
+
"epoch": 909.52,
|
242898 |
+
"learning_rate": 8.194329563812602e-06,
|
242899 |
+
"loss": 0.3166,
|
242900 |
+
"step": 112815
|
242901 |
+
},
|
242902 |
+
{
|
242903 |
+
"epoch": 909.56,
|
242904 |
+
"learning_rate": 8.194248788368336e-06,
|
242905 |
+
"loss": 0.4555,
|
242906 |
+
"step": 112820
|
242907 |
+
},
|
242908 |
+
{
|
242909 |
+
"epoch": 909.6,
|
242910 |
+
"learning_rate": 8.194168012924072e-06,
|
242911 |
+
"loss": 1.2846,
|
242912 |
+
"step": 112825
|
242913 |
+
},
|
242914 |
+
{
|
242915 |
+
"epoch": 909.64,
|
242916 |
+
"learning_rate": 8.194087237479806e-06,
|
242917 |
+
"loss": 0.3921,
|
242918 |
+
"step": 112830
|
242919 |
+
},
|
242920 |
+
{
|
242921 |
+
"epoch": 909.68,
|
242922 |
+
"learning_rate": 8.194006462035542e-06,
|
242923 |
+
"loss": 0.3258,
|
242924 |
+
"step": 112835
|
242925 |
+
},
|
242926 |
+
{
|
242927 |
+
"epoch": 909.72,
|
242928 |
+
"learning_rate": 8.193925686591276e-06,
|
242929 |
+
"loss": 0.3415,
|
242930 |
+
"step": 112840
|
242931 |
+
},
|
242932 |
+
{
|
242933 |
+
"epoch": 909.76,
|
242934 |
+
"learning_rate": 8.193844911147012e-06,
|
242935 |
+
"loss": 0.4425,
|
242936 |
+
"step": 112845
|
242937 |
+
},
|
242938 |
+
{
|
242939 |
+
"epoch": 909.8,
|
242940 |
+
"learning_rate": 8.193764135702748e-06,
|
242941 |
+
"loss": 1.1291,
|
242942 |
+
"step": 112850
|
242943 |
+
},
|
242944 |
+
{
|
242945 |
+
"epoch": 909.84,
|
242946 |
+
"learning_rate": 8.193683360258482e-06,
|
242947 |
+
"loss": 0.3087,
|
242948 |
+
"step": 112855
|
242949 |
+
},
|
242950 |
+
{
|
242951 |
+
"epoch": 909.88,
|
242952 |
+
"learning_rate": 8.193602584814218e-06,
|
242953 |
+
"loss": 0.3314,
|
242954 |
+
"step": 112860
|
242955 |
+
},
|
242956 |
+
{
|
242957 |
+
"epoch": 909.92,
|
242958 |
+
"learning_rate": 8.193521809369952e-06,
|
242959 |
+
"loss": 0.329,
|
242960 |
+
"step": 112865
|
242961 |
+
},
|
242962 |
+
{
|
242963 |
+
"epoch": 909.96,
|
242964 |
+
"learning_rate": 8.193441033925688e-06,
|
242965 |
+
"loss": 0.4559,
|
242966 |
+
"step": 112870
|
242967 |
+
},
|
242968 |
+
{
|
242969 |
+
"epoch": 910.0,
|
242970 |
+
"eval_loss": 0.3853907585144043,
|
242971 |
+
"eval_runtime": 39.1333,
|
242972 |
+
"eval_samples_per_second": 21.363,
|
242973 |
+
"eval_steps_per_second": 0.69,
|
242974 |
+
"eval_wer": 0.18328861061419202,
|
242975 |
+
"step": 112874
|
242976 |
+
},
|
242977 |
+
{
|
242978 |
+
"epoch": 910.01,
|
242979 |
+
"learning_rate": 8.193360258481422e-06,
|
242980 |
+
"loss": 0.3418,
|
242981 |
+
"step": 112875
|
242982 |
+
},
|
242983 |
+
{
|
242984 |
+
"epoch": 910.05,
|
242985 |
+
"learning_rate": 8.193279483037158e-06,
|
242986 |
+
"loss": 0.3218,
|
242987 |
+
"step": 112880
|
242988 |
+
},
|
242989 |
+
{
|
242990 |
+
"epoch": 910.09,
|
242991 |
+
"learning_rate": 8.193198707592892e-06,
|
242992 |
+
"loss": 0.2755,
|
242993 |
+
"step": 112885
|
242994 |
+
},
|
242995 |
+
{
|
242996 |
+
"epoch": 910.13,
|
242997 |
+
"learning_rate": 8.193117932148628e-06,
|
242998 |
+
"loss": 0.3143,
|
242999 |
+
"step": 112890
|
243000 |
+
},
|
243001 |
+
{
|
243002 |
+
"epoch": 910.17,
|
243003 |
+
"learning_rate": 8.193037156704362e-06,
|
243004 |
+
"loss": 0.4879,
|
243005 |
+
"step": 112895
|
243006 |
+
},
|
243007 |
+
{
|
243008 |
+
"epoch": 910.21,
|
243009 |
+
"learning_rate": 8.192956381260098e-06,
|
243010 |
+
"loss": 1.1204,
|
243011 |
+
"step": 112900
|
243012 |
+
},
|
243013 |
+
{
|
243014 |
+
"epoch": 910.25,
|
243015 |
+
"learning_rate": 8.192875605815834e-06,
|
243016 |
+
"loss": 0.3345,
|
243017 |
+
"step": 112905
|
243018 |
+
},
|
243019 |
+
{
|
243020 |
+
"epoch": 910.29,
|
243021 |
+
"learning_rate": 8.192794830371568e-06,
|
243022 |
+
"loss": 0.2979,
|
243023 |
+
"step": 112910
|
243024 |
+
},
|
243025 |
+
{
|
243026 |
+
"epoch": 910.33,
|
243027 |
+
"learning_rate": 8.192714054927304e-06,
|
243028 |
+
"loss": 0.3332,
|
243029 |
+
"step": 112915
|
243030 |
+
},
|
243031 |
+
{
|
243032 |
+
"epoch": 910.37,
|
243033 |
+
"learning_rate": 8.192633279483038e-06,
|
243034 |
+
"loss": 0.5393,
|
243035 |
+
"step": 112920
|
243036 |
+
},
|
243037 |
+
{
|
243038 |
+
"epoch": 910.41,
|
243039 |
+
"learning_rate": 8.192552504038774e-06,
|
243040 |
+
"loss": 1.2886,
|
243041 |
+
"step": 112925
|
243042 |
+
},
|
243043 |
+
{
|
243044 |
+
"epoch": 910.45,
|
243045 |
+
"learning_rate": 8.192471728594508e-06,
|
243046 |
+
"loss": 0.3061,
|
243047 |
+
"step": 112930
|
243048 |
+
},
|
243049 |
+
{
|
243050 |
+
"epoch": 910.49,
|
243051 |
+
"learning_rate": 8.192390953150244e-06,
|
243052 |
+
"loss": 0.2628,
|
243053 |
+
"step": 112935
|
243054 |
+
},
|
243055 |
+
{
|
243056 |
+
"epoch": 910.53,
|
243057 |
+
"learning_rate": 8.192310177705978e-06,
|
243058 |
+
"loss": 0.2704,
|
243059 |
+
"step": 112940
|
243060 |
+
},
|
243061 |
+
{
|
243062 |
+
"epoch": 910.57,
|
243063 |
+
"learning_rate": 8.192229402261714e-06,
|
243064 |
+
"loss": 0.5995,
|
243065 |
+
"step": 112945
|
243066 |
+
},
|
243067 |
+
{
|
243068 |
+
"epoch": 910.61,
|
243069 |
+
"learning_rate": 8.192148626817448e-06,
|
243070 |
+
"loss": 1.1335,
|
243071 |
+
"step": 112950
|
243072 |
+
},
|
243073 |
+
{
|
243074 |
+
"epoch": 910.65,
|
243075 |
+
"learning_rate": 8.192067851373183e-06,
|
243076 |
+
"loss": 0.3048,
|
243077 |
+
"step": 112955
|
243078 |
+
},
|
243079 |
+
{
|
243080 |
+
"epoch": 910.69,
|
243081 |
+
"learning_rate": 8.191987075928918e-06,
|
243082 |
+
"loss": 0.3209,
|
243083 |
+
"step": 112960
|
243084 |
+
},
|
243085 |
+
{
|
243086 |
+
"epoch": 910.73,
|
243087 |
+
"learning_rate": 8.191906300484653e-06,
|
243088 |
+
"loss": 0.3437,
|
243089 |
+
"step": 112965
|
243090 |
+
},
|
243091 |
+
{
|
243092 |
+
"epoch": 910.77,
|
243093 |
+
"learning_rate": 8.19182552504039e-06,
|
243094 |
+
"loss": 0.6269,
|
243095 |
+
"step": 112970
|
243096 |
+
},
|
243097 |
+
{
|
243098 |
+
"epoch": 910.81,
|
243099 |
+
"learning_rate": 8.191744749596123e-06,
|
243100 |
+
"loss": 1.0891,
|
243101 |
+
"step": 112975
|
243102 |
+
},
|
243103 |
+
{
|
243104 |
+
"epoch": 910.85,
|
243105 |
+
"learning_rate": 8.19166397415186e-06,
|
243106 |
+
"loss": 0.3427,
|
243107 |
+
"step": 112980
|
243108 |
+
},
|
243109 |
+
{
|
243110 |
+
"epoch": 910.89,
|
243111 |
+
"learning_rate": 8.191583198707593e-06,
|
243112 |
+
"loss": 0.2962,
|
243113 |
+
"step": 112985
|
243114 |
+
},
|
243115 |
+
{
|
243116 |
+
"epoch": 910.93,
|
243117 |
+
"learning_rate": 8.19150242326333e-06,
|
243118 |
+
"loss": 0.3852,
|
243119 |
+
"step": 112990
|
243120 |
+
},
|
243121 |
+
{
|
243122 |
+
"epoch": 910.97,
|
243123 |
+
"learning_rate": 8.191421647819063e-06,
|
243124 |
+
"loss": 0.663,
|
243125 |
+
"step": 112995
|
243126 |
+
},
|
243127 |
+
{
|
243128 |
+
"epoch": 911.0,
|
243129 |
+
"eval_loss": 0.3766191005706787,
|
243130 |
+
"eval_runtime": 41.0289,
|
243131 |
+
"eval_samples_per_second": 20.376,
|
243132 |
+
"eval_steps_per_second": 0.658,
|
243133 |
+
"eval_wer": 0.18356084809758932,
|
243134 |
+
"step": 112998
|
243135 |
+
},
|
243136 |
+
{
|
243137 |
+
"epoch": 911.02,
|
243138 |
+
"learning_rate": 8.1913408723748e-06,
|
243139 |
+
"loss": 0.3573,
|
243140 |
+
"step": 113000
|
243141 |
+
},
|
243142 |
+
{
|
243143 |
+
"epoch": 911.06,
|
243144 |
+
"learning_rate": 8.191260096930533e-06,
|
243145 |
+
"loss": 0.2916,
|
243146 |
+
"step": 113005
|
243147 |
+
},
|
243148 |
+
{
|
243149 |
+
"epoch": 911.1,
|
243150 |
+
"learning_rate": 8.19117932148627e-06,
|
243151 |
+
"loss": 0.2908,
|
243152 |
+
"step": 113010
|
243153 |
+
},
|
243154 |
+
{
|
243155 |
+
"epoch": 911.14,
|
243156 |
+
"learning_rate": 8.191098546042003e-06,
|
243157 |
+
"loss": 0.322,
|
243158 |
+
"step": 113015
|
243159 |
+
},
|
243160 |
+
{
|
243161 |
+
"epoch": 911.18,
|
243162 |
+
"learning_rate": 8.19101777059774e-06,
|
243163 |
+
"loss": 0.5831,
|
243164 |
+
"step": 113020
|
243165 |
+
},
|
243166 |
+
{
|
243167 |
+
"epoch": 911.22,
|
243168 |
+
"learning_rate": 8.190936995153475e-06,
|
243169 |
+
"loss": 0.9917,
|
243170 |
+
"step": 113025
|
243171 |
+
},
|
243172 |
+
{
|
243173 |
+
"epoch": 911.26,
|
243174 |
+
"learning_rate": 8.19085621970921e-06,
|
243175 |
+
"loss": 0.2945,
|
243176 |
+
"step": 113030
|
243177 |
+
},
|
243178 |
+
{
|
243179 |
+
"epoch": 911.3,
|
243180 |
+
"learning_rate": 8.190775444264945e-06,
|
243181 |
+
"loss": 0.2512,
|
243182 |
+
"step": 113035
|
243183 |
+
},
|
243184 |
+
{
|
243185 |
+
"epoch": 911.34,
|
243186 |
+
"learning_rate": 8.190694668820679e-06,
|
243187 |
+
"loss": 0.3935,
|
243188 |
+
"step": 113040
|
243189 |
+
},
|
243190 |
+
{
|
243191 |
+
"epoch": 911.38,
|
243192 |
+
"learning_rate": 8.190613893376415e-06,
|
243193 |
+
"loss": 0.6559,
|
243194 |
+
"step": 113045
|
243195 |
+
},
|
243196 |
+
{
|
243197 |
+
"epoch": 911.42,
|
243198 |
+
"learning_rate": 8.190533117932149e-06,
|
243199 |
+
"loss": 1.014,
|
243200 |
+
"step": 113050
|
243201 |
+
},
|
243202 |
+
{
|
243203 |
+
"epoch": 911.46,
|
243204 |
+
"learning_rate": 8.190452342487885e-06,
|
243205 |
+
"loss": 0.3467,
|
243206 |
+
"step": 113055
|
243207 |
+
},
|
243208 |
+
{
|
243209 |
+
"epoch": 911.5,
|
243210 |
+
"learning_rate": 8.190371567043619e-06,
|
243211 |
+
"loss": 0.3123,
|
243212 |
+
"step": 113060
|
243213 |
+
},
|
243214 |
+
{
|
243215 |
+
"epoch": 911.54,
|
243216 |
+
"learning_rate": 8.190290791599355e-06,
|
243217 |
+
"loss": 0.4067,
|
243218 |
+
"step": 113065
|
243219 |
+
},
|
243220 |
+
{
|
243221 |
+
"epoch": 911.58,
|
243222 |
+
"learning_rate": 8.190210016155089e-06,
|
243223 |
+
"loss": 0.6399,
|
243224 |
+
"step": 113070
|
243225 |
+
},
|
243226 |
+
{
|
243227 |
+
"epoch": 911.62,
|
243228 |
+
"learning_rate": 8.190129240710825e-06,
|
243229 |
+
"loss": 0.9487,
|
243230 |
+
"step": 113075
|
243231 |
+
},
|
243232 |
+
{
|
243233 |
+
"epoch": 911.66,
|
243234 |
+
"learning_rate": 8.19004846526656e-06,
|
243235 |
+
"loss": 0.2608,
|
243236 |
+
"step": 113080
|
243237 |
+
},
|
243238 |
+
{
|
243239 |
+
"epoch": 911.7,
|
243240 |
+
"learning_rate": 8.189967689822295e-06,
|
243241 |
+
"loss": 0.229,
|
243242 |
+
"step": 113085
|
243243 |
+
},
|
243244 |
+
{
|
243245 |
+
"epoch": 911.74,
|
243246 |
+
"learning_rate": 8.18988691437803e-06,
|
243247 |
+
"loss": 0.3772,
|
243248 |
+
"step": 113090
|
243249 |
+
},
|
243250 |
+
{
|
243251 |
+
"epoch": 911.78,
|
243252 |
+
"learning_rate": 8.189806138933765e-06,
|
243253 |
+
"loss": 0.6131,
|
243254 |
+
"step": 113095
|
243255 |
+
},
|
243256 |
+
{
|
243257 |
+
"epoch": 911.82,
|
243258 |
+
"learning_rate": 8.189741518578352e-06,
|
243259 |
+
"loss": 1.0935,
|
243260 |
+
"step": 113100
|
243261 |
+
},
|
243262 |
+
{
|
243263 |
+
"epoch": 911.86,
|
243264 |
+
"learning_rate": 8.189660743134088e-06,
|
243265 |
+
"loss": 0.3042,
|
243266 |
+
"step": 113105
|
243267 |
+
},
|
243268 |
+
{
|
243269 |
+
"epoch": 911.9,
|
243270 |
+
"learning_rate": 8.189579967689822e-06,
|
243271 |
+
"loss": 0.2531,
|
243272 |
+
"step": 113110
|
243273 |
+
},
|
243274 |
+
{
|
243275 |
+
"epoch": 911.94,
|
243276 |
+
"learning_rate": 8.189499192245558e-06,
|
243277 |
+
"loss": 0.4252,
|
243278 |
+
"step": 113115
|
243279 |
+
},
|
243280 |
+
{
|
243281 |
+
"epoch": 911.98,
|
243282 |
+
"learning_rate": 8.189418416801294e-06,
|
243283 |
+
"loss": 0.6596,
|
243284 |
+
"step": 113120
|
243285 |
+
},
|
243286 |
+
{
|
243287 |
+
"epoch": 912.0,
|
243288 |
+
"eval_loss": 0.444323867559433,
|
243289 |
+
"eval_runtime": 40.7673,
|
243290 |
+
"eval_samples_per_second": 20.507,
|
243291 |
+
"eval_steps_per_second": 0.662,
|
243292 |
+
"eval_wer": 0.18970503181029497,
|
243293 |
+
"step": 113122
|
243294 |
+
},
|
243295 |
+
{
|
243296 |
+
"epoch": 912.02,
|
243297 |
+
"learning_rate": 8.189337641357028e-06,
|
243298 |
+
"loss": 0.3245,
|
243299 |
+
"step": 113125
|
243300 |
+
},
|
243301 |
+
{
|
243302 |
+
"epoch": 912.06,
|
243303 |
+
"learning_rate": 8.189256865912764e-06,
|
243304 |
+
"loss": 0.2808,
|
243305 |
+
"step": 113130
|
243306 |
+
},
|
243307 |
+
{
|
243308 |
+
"epoch": 912.1,
|
243309 |
+
"learning_rate": 8.189176090468498e-06,
|
243310 |
+
"loss": 0.3132,
|
243311 |
+
"step": 113135
|
243312 |
+
},
|
243313 |
+
{
|
243314 |
+
"epoch": 912.14,
|
243315 |
+
"learning_rate": 8.189095315024234e-06,
|
243316 |
+
"loss": 0.3814,
|
243317 |
+
"step": 113140
|
243318 |
+
},
|
243319 |
+
{
|
243320 |
+
"epoch": 912.18,
|
243321 |
+
"learning_rate": 8.189014539579968e-06,
|
243322 |
+
"loss": 0.7656,
|
243323 |
+
"step": 113145
|
243324 |
+
},
|
243325 |
+
{
|
243326 |
+
"epoch": 912.22,
|
243327 |
+
"learning_rate": 8.188933764135704e-06,
|
243328 |
+
"loss": 0.9321,
|
243329 |
+
"step": 113150
|
243330 |
+
},
|
243331 |
+
{
|
243332 |
+
"epoch": 912.27,
|
243333 |
+
"learning_rate": 8.188852988691438e-06,
|
243334 |
+
"loss": 0.2926,
|
243335 |
+
"step": 113155
|
243336 |
+
},
|
243337 |
+
{
|
243338 |
+
"epoch": 912.31,
|
243339 |
+
"learning_rate": 8.188772213247174e-06,
|
243340 |
+
"loss": 0.3008,
|
243341 |
+
"step": 113160
|
243342 |
+
},
|
243343 |
+
{
|
243344 |
+
"epoch": 912.35,
|
243345 |
+
"learning_rate": 8.188691437802908e-06,
|
243346 |
+
"loss": 0.3474,
|
243347 |
+
"step": 113165
|
243348 |
+
},
|
243349 |
+
{
|
243350 |
+
"epoch": 912.39,
|
243351 |
+
"learning_rate": 8.188610662358644e-06,
|
243352 |
+
"loss": 0.8603,
|
243353 |
+
"step": 113170
|
243354 |
+
},
|
243355 |
+
{
|
243356 |
+
"epoch": 912.43,
|
243357 |
+
"learning_rate": 8.18852988691438e-06,
|
243358 |
+
"loss": 0.9023,
|
243359 |
+
"step": 113175
|
243360 |
+
},
|
243361 |
+
{
|
243362 |
+
"epoch": 912.47,
|
243363 |
+
"learning_rate": 8.188449111470114e-06,
|
243364 |
+
"loss": 0.2997,
|
243365 |
+
"step": 113180
|
243366 |
+
},
|
243367 |
+
{
|
243368 |
+
"epoch": 912.51,
|
243369 |
+
"learning_rate": 8.18836833602585e-06,
|
243370 |
+
"loss": 0.3067,
|
243371 |
+
"step": 113185
|
243372 |
+
},
|
243373 |
+
{
|
243374 |
+
"epoch": 912.55,
|
243375 |
+
"learning_rate": 8.188287560581584e-06,
|
243376 |
+
"loss": 0.36,
|
243377 |
+
"step": 113190
|
243378 |
+
},
|
243379 |
+
{
|
243380 |
+
"epoch": 912.59,
|
243381 |
+
"learning_rate": 8.18820678513732e-06,
|
243382 |
+
"loss": 0.7879,
|
243383 |
+
"step": 113195
|
243384 |
+
},
|
243385 |
+
{
|
243386 |
+
"epoch": 912.63,
|
243387 |
+
"learning_rate": 8.188126009693054e-06,
|
243388 |
+
"loss": 0.9157,
|
243389 |
+
"step": 113200
|
243390 |
+
},
|
243391 |
+
{
|
243392 |
+
"epoch": 912.67,
|
243393 |
+
"learning_rate": 8.18804523424879e-06,
|
243394 |
+
"loss": 0.3224,
|
243395 |
+
"step": 113205
|
243396 |
+
},
|
243397 |
+
{
|
243398 |
+
"epoch": 912.71,
|
243399 |
+
"learning_rate": 8.187964458804524e-06,
|
243400 |
+
"loss": 0.2828,
|
243401 |
+
"step": 113210
|
243402 |
+
},
|
243403 |
+
{
|
243404 |
+
"epoch": 912.75,
|
243405 |
+
"learning_rate": 8.18788368336026e-06,
|
243406 |
+
"loss": 0.3505,
|
243407 |
+
"step": 113215
|
243408 |
+
},
|
243409 |
+
{
|
243410 |
+
"epoch": 912.79,
|
243411 |
+
"learning_rate": 8.187802907915994e-06,
|
243412 |
+
"loss": 0.7458,
|
243413 |
+
"step": 113220
|
243414 |
+
},
|
243415 |
+
{
|
243416 |
+
"epoch": 912.83,
|
243417 |
+
"learning_rate": 8.18772213247173e-06,
|
243418 |
+
"loss": 0.793,
|
243419 |
+
"step": 113225
|
243420 |
+
},
|
243421 |
+
{
|
243422 |
+
"epoch": 912.87,
|
243423 |
+
"learning_rate": 8.187641357027464e-06,
|
243424 |
+
"loss": 0.3386,
|
243425 |
+
"step": 113230
|
243426 |
+
},
|
243427 |
+
{
|
243428 |
+
"epoch": 912.91,
|
243429 |
+
"learning_rate": 8.1875605815832e-06,
|
243430 |
+
"loss": 0.3076,
|
243431 |
+
"step": 113235
|
243432 |
+
},
|
243433 |
+
{
|
243434 |
+
"epoch": 912.95,
|
243435 |
+
"learning_rate": 8.187479806138935e-06,
|
243436 |
+
"loss": 0.3701,
|
243437 |
+
"step": 113240
|
243438 |
+
},
|
243439 |
+
{
|
243440 |
+
"epoch": 912.99,
|
243441 |
+
"learning_rate": 8.18739903069467e-06,
|
243442 |
+
"loss": 0.807,
|
243443 |
+
"step": 113245
|
243444 |
+
},
|
243445 |
+
{
|
243446 |
+
"epoch": 913.0,
|
243447 |
+
"eval_loss": 0.4701347053050995,
|
243448 |
+
"eval_runtime": 41.2757,
|
243449 |
+
"eval_samples_per_second": 20.278,
|
243450 |
+
"eval_steps_per_second": 0.654,
|
243451 |
+
"eval_wer": 0.17971326164874551,
|
243452 |
+
"step": 113246
|
243453 |
+
},
|
243454 |
+
{
|
243455 |
+
"epoch": 913.03,
|
243456 |
+
"learning_rate": 8.187318255250405e-06,
|
243457 |
+
"loss": 0.3049,
|
243458 |
+
"step": 113250
|
243459 |
+
},
|
243460 |
+
{
|
243461 |
+
"epoch": 913.07,
|
243462 |
+
"learning_rate": 8.18723747980614e-06,
|
243463 |
+
"loss": 0.3326,
|
243464 |
+
"step": 113255
|
243465 |
+
},
|
243466 |
+
{
|
243467 |
+
"epoch": 913.11,
|
243468 |
+
"learning_rate": 8.187156704361875e-06,
|
243469 |
+
"loss": 0.2772,
|
243470 |
+
"step": 113260
|
243471 |
+
},
|
243472 |
+
{
|
243473 |
+
"epoch": 913.15,
|
243474 |
+
"learning_rate": 8.18707592891761e-06,
|
243475 |
+
"loss": 0.3754,
|
243476 |
+
"step": 113265
|
243477 |
+
},
|
243478 |
+
{
|
243479 |
+
"epoch": 913.19,
|
243480 |
+
"learning_rate": 8.186995153473345e-06,
|
243481 |
+
"loss": 0.8271,
|
243482 |
+
"step": 113270
|
243483 |
+
},
|
243484 |
+
{
|
243485 |
+
"epoch": 913.23,
|
243486 |
+
"learning_rate": 8.18691437802908e-06,
|
243487 |
+
"loss": 0.7523,
|
243488 |
+
"step": 113275
|
243489 |
+
},
|
243490 |
+
{
|
243491 |
+
"epoch": 913.27,
|
243492 |
+
"learning_rate": 8.186833602584815e-06,
|
243493 |
+
"loss": 0.2855,
|
243494 |
+
"step": 113280
|
243495 |
+
},
|
243496 |
+
{
|
243497 |
+
"epoch": 913.31,
|
243498 |
+
"learning_rate": 8.18675282714055e-06,
|
243499 |
+
"loss": 0.3099,
|
243500 |
+
"step": 113285
|
243501 |
+
},
|
243502 |
+
{
|
243503 |
+
"epoch": 913.35,
|
243504 |
+
"learning_rate": 8.186672051696285e-06,
|
243505 |
+
"loss": 0.3889,
|
243506 |
+
"step": 113290
|
243507 |
+
},
|
243508 |
+
{
|
243509 |
+
"epoch": 913.39,
|
243510 |
+
"learning_rate": 8.186591276252021e-06,
|
243511 |
+
"loss": 0.8476,
|
243512 |
+
"step": 113295
|
243513 |
+
},
|
243514 |
+
{
|
243515 |
+
"epoch": 913.43,
|
243516 |
+
"learning_rate": 8.186510500807755e-06,
|
243517 |
+
"loss": 0.6524,
|
243518 |
+
"step": 113300
|
243519 |
+
},
|
243520 |
+
{
|
243521 |
+
"epoch": 913.47,
|
243522 |
+
"learning_rate": 8.186429725363491e-06,
|
243523 |
+
"loss": 0.2498,
|
243524 |
+
"step": 113305
|
243525 |
+
},
|
243526 |
+
{
|
243527 |
+
"epoch": 913.51,
|
243528 |
+
"learning_rate": 8.186348949919225e-06,
|
243529 |
+
"loss": 0.3264,
|
243530 |
+
"step": 113310
|
243531 |
+
},
|
243532 |
+
{
|
243533 |
+
"epoch": 913.55,
|
243534 |
+
"learning_rate": 8.186268174474961e-06,
|
243535 |
+
"loss": 0.3934,
|
243536 |
+
"step": 113315
|
243537 |
+
},
|
243538 |
+
{
|
243539 |
+
"epoch": 913.59,
|
243540 |
+
"learning_rate": 8.186187399030695e-06,
|
243541 |
+
"loss": 0.9911,
|
243542 |
+
"step": 113320
|
243543 |
+
},
|
243544 |
+
{
|
243545 |
+
"epoch": 913.63,
|
243546 |
+
"learning_rate": 8.18610662358643e-06,
|
243547 |
+
"loss": 0.6645,
|
243548 |
+
"step": 113325
|
243549 |
+
},
|
243550 |
+
{
|
243551 |
+
"epoch": 913.67,
|
243552 |
+
"learning_rate": 8.186025848142165e-06,
|
243553 |
+
"loss": 0.2985,
|
243554 |
+
"step": 113330
|
243555 |
+
},
|
243556 |
+
{
|
243557 |
+
"epoch": 913.71,
|
243558 |
+
"learning_rate": 8.1859450726979e-06,
|
243559 |
+
"loss": 0.3302,
|
243560 |
+
"step": 113335
|
243561 |
+
},
|
243562 |
+
{
|
243563 |
+
"epoch": 913.76,
|
243564 |
+
"learning_rate": 8.185864297253635e-06,
|
243565 |
+
"loss": 0.3933,
|
243566 |
+
"step": 113340
|
243567 |
+
},
|
243568 |
+
{
|
243569 |
+
"epoch": 913.8,
|
243570 |
+
"learning_rate": 8.18578352180937e-06,
|
243571 |
+
"loss": 0.9581,
|
243572 |
+
"step": 113345
|
243573 |
+
},
|
243574 |
+
{
|
243575 |
+
"epoch": 913.84,
|
243576 |
+
"learning_rate": 8.185702746365107e-06,
|
243577 |
+
"loss": 0.7659,
|
243578 |
+
"step": 113350
|
243579 |
+
},
|
243580 |
+
{
|
243581 |
+
"epoch": 913.88,
|
243582 |
+
"learning_rate": 8.18562197092084e-06,
|
243583 |
+
"loss": 0.2852,
|
243584 |
+
"step": 113355
|
243585 |
+
},
|
243586 |
+
{
|
243587 |
+
"epoch": 913.92,
|
243588 |
+
"learning_rate": 8.185541195476577e-06,
|
243589 |
+
"loss": 0.2891,
|
243590 |
+
"step": 113360
|
243591 |
+
},
|
243592 |
+
{
|
243593 |
+
"epoch": 913.96,
|
243594 |
+
"learning_rate": 8.18546042003231e-06,
|
243595 |
+
"loss": 0.4714,
|
243596 |
+
"step": 113365
|
243597 |
+
},
|
243598 |
+
{
|
243599 |
+
"epoch": 914.0,
|
243600 |
+
"learning_rate": 8.185379644588047e-06,
|
243601 |
+
"loss": 1.1086,
|
243602 |
+
"step": 113370
|
243603 |
+
},
|
243604 |
+
{
|
243605 |
+
"epoch": 914.0,
|
243606 |
+
"eval_loss": 0.39893263578414917,
|
243607 |
+
"eval_runtime": 39.8088,
|
243608 |
+
"eval_samples_per_second": 21.025,
|
243609 |
+
"eval_steps_per_second": 0.678,
|
243610 |
+
"eval_wer": 0.18508635386553998,
|
243611 |
+
"step": 113370
|
243612 |
}
|
243613 |
],
|
243614 |
+
"max_steps": 620000,
|
243615 |
"num_train_epochs": 5000,
|
243616 |
+
"total_flos": 3.190555567482095e+20,
|
243617 |
"trial_name": null,
|
243618 |
"trial_params": null
|
243619 |
}
|
model-bin/finetune/base/{checkpoint-112750 β checkpoint-113370}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630082884.969348/events.out.tfevents.1630082884.dea8124df033.1391.7
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41fd86e6e7c1a5c77593c5bbda325bd5c975686ea34b46f2f3a6e958dbca7861
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630083307.1487231/events.out.tfevents.1630083307.dea8124df033.1391.9
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11391acdb58354ca2cac2c60fd34dff94d1f18b0bcbdfb0572682728e5a86432
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630083735.6256015/events.out.tfevents.1630083735.dea8124df033.1391.11
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e25a50c37d80004082e7bc62597d5bf40a647914e074c4a409928874ead615a
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630084174.1170754/events.out.tfevents.1630084174.dea8124df033.1391.13
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6063e6cb2ea6a249fc018a71d2bc14207d6991ca21c0e26094c96758b8d8827a
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630084598.2761755/events.out.tfevents.1630084598.dea8124df033.1391.15
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f620b324894ddc530420ee660ac32ef67a80a90e8eaf65637db8ee2548cf5cd
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630082884.dea8124df033.1391.6
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f9dc3f5f57f07b97c5848164fa8a44cf6407370c74df678c5318e16f5e9bd6e
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630083307.dea8124df033.1391.8
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81129d21b671588c0030205e0ce3ee50b437dcd581b30365dc937c7d279d6f2b
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630083735.dea8124df033.1391.10
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9203203993faf14718631ba75e93c0624f6aa871e966f57dfaa38274522d3aea
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630084174.dea8124df033.1391.12
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e1059774e54b65b2bea6167c90283c2debf4d7672d04144c264476154ca1eedc
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630084598.dea8124df033.1391.14
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c1ce76c497b5a7a4d19715c80fc0b5e0948b520e68b45153bb07b93b8e9a7f4
|
3 |
+
size 8622
|