"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630228288.1449773/events.out.tfevents.1630228288.cc93b136ebf5.1086.109 +3 -0
- model-bin/finetune/base/log/1630228721.9317095/events.out.tfevents.1630228721.cc93b136ebf5.1086.111 +3 -0
- model-bin/finetune/base/log/1630229152.9104557/events.out.tfevents.1630229152.cc93b136ebf5.1086.113 +3 -0
- model-bin/finetune/base/log/1630229586.5053954/events.out.tfevents.1630229586.cc93b136ebf5.1086.115 +3 -0
- model-bin/finetune/base/log/1630230014.08416/events.out.tfevents.1630230014.cc93b136ebf5.1086.117 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630228288.cc93b136ebf5.1086.108 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630228721.cc93b136ebf5.1086.110 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630229152.cc93b136ebf5.1086.112 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630229586.cc93b136ebf5.1086.114 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630230014.cc93b136ebf5.1086.116 +3 -0
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f7568a2befca944be90026427116813896320af24320ff0d6d6cce3345d3e75
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2a22ed2c49e0003da5e3bff5a28ee7e00afcf8ecb8920666d8355ae042273f4b
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49db60bb7be014c03e20f099f8b5b41640dd5b10966a8d382a375eaacf0dba1f
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e32e4b8d64538adb2e3a82fff0984d5e19137de747f31f2ccfe398a635069ee
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:75a35ac11ff6fdc1ff80807c5620a77c06912969c11e90901d0e650569700174
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.16716790141568252,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -284934,11 +284934,800 @@
|
|
284934 |
"eval_steps_per_second": 0.655,
|
284935 |
"eval_wer": 0.17012599228024178,
|
284936 |
"step": 145852
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
284937 |
}
|
284938 |
],
|
284939 |
-
"max_steps":
|
284940 |
"num_train_epochs": 5000,
|
284941 |
-
"total_flos": 4.
|
284942 |
"trial_name": null,
|
284943 |
"trial_params": null
|
284944 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.16716790141568252,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
4 |
+
"epoch": 1180.995983935743,
|
5 |
+
"global_step": 146472,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
284934 |
"eval_steps_per_second": 0.655,
|
284935 |
"eval_wer": 0.17012599228024178,
|
284936 |
"step": 145852
|
284937 |
+
},
|
284938 |
+
{
|
284939 |
+
"epoch": 1176.02,
|
284940 |
+
"learning_rate": 7.660775444264943e-06,
|
284941 |
+
"loss": 0.4127,
|
284942 |
+
"step": 145855
|
284943 |
+
},
|
284944 |
+
{
|
284945 |
+
"epoch": 1176.06,
|
284946 |
+
"learning_rate": 7.660694668820679e-06,
|
284947 |
+
"loss": 0.2818,
|
284948 |
+
"step": 145860
|
284949 |
+
},
|
284950 |
+
{
|
284951 |
+
"epoch": 1176.1,
|
284952 |
+
"learning_rate": 7.660613893376413e-06,
|
284953 |
+
"loss": 0.2966,
|
284954 |
+
"step": 145865
|
284955 |
+
},
|
284956 |
+
{
|
284957 |
+
"epoch": 1176.15,
|
284958 |
+
"learning_rate": 7.660533117932149e-06,
|
284959 |
+
"loss": 0.3262,
|
284960 |
+
"step": 145870
|
284961 |
+
},
|
284962 |
+
{
|
284963 |
+
"epoch": 1176.19,
|
284964 |
+
"learning_rate": 7.660452342487883e-06,
|
284965 |
+
"loss": 0.744,
|
284966 |
+
"step": 145875
|
284967 |
+
},
|
284968 |
+
{
|
284969 |
+
"epoch": 1176.23,
|
284970 |
+
"learning_rate": 7.660371567043619e-06,
|
284971 |
+
"loss": 0.8761,
|
284972 |
+
"step": 145880
|
284973 |
+
},
|
284974 |
+
{
|
284975 |
+
"epoch": 1176.27,
|
284976 |
+
"learning_rate": 7.660290791599355e-06,
|
284977 |
+
"loss": 0.2513,
|
284978 |
+
"step": 145885
|
284979 |
+
},
|
284980 |
+
{
|
284981 |
+
"epoch": 1176.31,
|
284982 |
+
"learning_rate": 7.660210016155089e-06,
|
284983 |
+
"loss": 0.2938,
|
284984 |
+
"step": 145890
|
284985 |
+
},
|
284986 |
+
{
|
284987 |
+
"epoch": 1176.35,
|
284988 |
+
"learning_rate": 7.660129240710825e-06,
|
284989 |
+
"loss": 0.3237,
|
284990 |
+
"step": 145895
|
284991 |
+
},
|
284992 |
+
{
|
284993 |
+
"epoch": 1176.39,
|
284994 |
+
"learning_rate": 7.660048465266559e-06,
|
284995 |
+
"loss": 0.7375,
|
284996 |
+
"step": 145900
|
284997 |
+
},
|
284998 |
+
{
|
284999 |
+
"epoch": 1176.43,
|
285000 |
+
"learning_rate": 7.659967689822295e-06,
|
285001 |
+
"loss": 0.8132,
|
285002 |
+
"step": 145905
|
285003 |
+
},
|
285004 |
+
{
|
285005 |
+
"epoch": 1176.47,
|
285006 |
+
"learning_rate": 7.659886914378029e-06,
|
285007 |
+
"loss": 0.3004,
|
285008 |
+
"step": 145910
|
285009 |
+
},
|
285010 |
+
{
|
285011 |
+
"epoch": 1176.51,
|
285012 |
+
"learning_rate": 7.659806138933765e-06,
|
285013 |
+
"loss": 0.2726,
|
285014 |
+
"step": 145915
|
285015 |
+
},
|
285016 |
+
{
|
285017 |
+
"epoch": 1176.55,
|
285018 |
+
"learning_rate": 7.659725363489499e-06,
|
285019 |
+
"loss": 0.3235,
|
285020 |
+
"step": 145920
|
285021 |
+
},
|
285022 |
+
{
|
285023 |
+
"epoch": 1176.59,
|
285024 |
+
"learning_rate": 7.659644588045235e-06,
|
285025 |
+
"loss": 0.7467,
|
285026 |
+
"step": 145925
|
285027 |
+
},
|
285028 |
+
{
|
285029 |
+
"epoch": 1176.63,
|
285030 |
+
"learning_rate": 7.659563812600969e-06,
|
285031 |
+
"loss": 0.7942,
|
285032 |
+
"step": 145930
|
285033 |
+
},
|
285034 |
+
{
|
285035 |
+
"epoch": 1176.67,
|
285036 |
+
"learning_rate": 7.659483037156705e-06,
|
285037 |
+
"loss": 0.2774,
|
285038 |
+
"step": 145935
|
285039 |
+
},
|
285040 |
+
{
|
285041 |
+
"epoch": 1176.71,
|
285042 |
+
"learning_rate": 7.659402261712439e-06,
|
285043 |
+
"loss": 0.2752,
|
285044 |
+
"step": 145940
|
285045 |
+
},
|
285046 |
+
{
|
285047 |
+
"epoch": 1176.75,
|
285048 |
+
"learning_rate": 7.659321486268175e-06,
|
285049 |
+
"loss": 0.3578,
|
285050 |
+
"step": 145945
|
285051 |
+
},
|
285052 |
+
{
|
285053 |
+
"epoch": 1176.79,
|
285054 |
+
"learning_rate": 7.659240710823909e-06,
|
285055 |
+
"loss": 0.6986,
|
285056 |
+
"step": 145950
|
285057 |
+
},
|
285058 |
+
{
|
285059 |
+
"epoch": 1176.83,
|
285060 |
+
"learning_rate": 7.659159935379645e-06,
|
285061 |
+
"loss": 0.7739,
|
285062 |
+
"step": 145955
|
285063 |
+
},
|
285064 |
+
{
|
285065 |
+
"epoch": 1176.87,
|
285066 |
+
"learning_rate": 7.65907915993538e-06,
|
285067 |
+
"loss": 0.2683,
|
285068 |
+
"step": 145960
|
285069 |
+
},
|
285070 |
+
{
|
285071 |
+
"epoch": 1176.91,
|
285072 |
+
"learning_rate": 7.658998384491115e-06,
|
285073 |
+
"loss": 0.3132,
|
285074 |
+
"step": 145965
|
285075 |
+
},
|
285076 |
+
{
|
285077 |
+
"epoch": 1176.95,
|
285078 |
+
"learning_rate": 7.65891760904685e-06,
|
285079 |
+
"loss": 0.373,
|
285080 |
+
"step": 145970
|
285081 |
+
},
|
285082 |
+
{
|
285083 |
+
"epoch": 1176.99,
|
285084 |
+
"learning_rate": 7.658836833602585e-06,
|
285085 |
+
"loss": 1.049,
|
285086 |
+
"step": 145975
|
285087 |
+
},
|
285088 |
+
{
|
285089 |
+
"epoch": 1177.0,
|
285090 |
+
"eval_loss": 0.3793235719203949,
|
285091 |
+
"eval_runtime": 40.3575,
|
285092 |
+
"eval_samples_per_second": 20.74,
|
285093 |
+
"eval_steps_per_second": 0.669,
|
285094 |
+
"eval_wer": 0.17174677608440797,
|
285095 |
+
"step": 145976
|
285096 |
+
},
|
285097 |
+
{
|
285098 |
+
"epoch": 1177.03,
|
285099 |
+
"learning_rate": 7.65875605815832e-06,
|
285100 |
+
"loss": 0.3211,
|
285101 |
+
"step": 145980
|
285102 |
+
},
|
285103 |
+
{
|
285104 |
+
"epoch": 1177.07,
|
285105 |
+
"learning_rate": 7.658675282714055e-06,
|
285106 |
+
"loss": 0.2587,
|
285107 |
+
"step": 145985
|
285108 |
+
},
|
285109 |
+
{
|
285110 |
+
"epoch": 1177.11,
|
285111 |
+
"learning_rate": 7.65859450726979e-06,
|
285112 |
+
"loss": 0.3043,
|
285113 |
+
"step": 145990
|
285114 |
+
},
|
285115 |
+
{
|
285116 |
+
"epoch": 1177.15,
|
285117 |
+
"learning_rate": 7.658513731825525e-06,
|
285118 |
+
"loss": 0.374,
|
285119 |
+
"step": 145995
|
285120 |
+
},
|
285121 |
+
{
|
285122 |
+
"epoch": 1177.19,
|
285123 |
+
"learning_rate": 7.65843295638126e-06,
|
285124 |
+
"loss": 0.9153,
|
285125 |
+
"step": 146000
|
285126 |
+
},
|
285127 |
+
{
|
285128 |
+
"epoch": 1177.23,
|
285129 |
+
"learning_rate": 7.658352180936995e-06,
|
285130 |
+
"loss": 0.7194,
|
285131 |
+
"step": 146005
|
285132 |
+
},
|
285133 |
+
{
|
285134 |
+
"epoch": 1177.27,
|
285135 |
+
"learning_rate": 7.65827140549273e-06,
|
285136 |
+
"loss": 0.2684,
|
285137 |
+
"step": 146010
|
285138 |
+
},
|
285139 |
+
{
|
285140 |
+
"epoch": 1177.31,
|
285141 |
+
"learning_rate": 7.658190630048465e-06,
|
285142 |
+
"loss": 0.2866,
|
285143 |
+
"step": 146015
|
285144 |
+
},
|
285145 |
+
{
|
285146 |
+
"epoch": 1177.35,
|
285147 |
+
"learning_rate": 7.6581098546042e-06,
|
285148 |
+
"loss": 0.3972,
|
285149 |
+
"step": 146020
|
285150 |
+
},
|
285151 |
+
{
|
285152 |
+
"epoch": 1177.39,
|
285153 |
+
"learning_rate": 7.658029079159936e-06,
|
285154 |
+
"loss": 1.0221,
|
285155 |
+
"step": 146025
|
285156 |
+
},
|
285157 |
+
{
|
285158 |
+
"epoch": 1177.43,
|
285159 |
+
"learning_rate": 7.65794830371567e-06,
|
285160 |
+
"loss": 0.5779,
|
285161 |
+
"step": 146030
|
285162 |
+
},
|
285163 |
+
{
|
285164 |
+
"epoch": 1177.47,
|
285165 |
+
"learning_rate": 7.657867528271406e-06,
|
285166 |
+
"loss": 0.3161,
|
285167 |
+
"step": 146035
|
285168 |
+
},
|
285169 |
+
{
|
285170 |
+
"epoch": 1177.51,
|
285171 |
+
"learning_rate": 7.65778675282714e-06,
|
285172 |
+
"loss": 0.2923,
|
285173 |
+
"step": 146040
|
285174 |
+
},
|
285175 |
+
{
|
285176 |
+
"epoch": 1177.55,
|
285177 |
+
"learning_rate": 7.657705977382876e-06,
|
285178 |
+
"loss": 0.4056,
|
285179 |
+
"step": 146045
|
285180 |
+
},
|
285181 |
+
{
|
285182 |
+
"epoch": 1177.59,
|
285183 |
+
"learning_rate": 7.65762520193861e-06,
|
285184 |
+
"loss": 0.8064,
|
285185 |
+
"step": 146050
|
285186 |
+
},
|
285187 |
+
{
|
285188 |
+
"epoch": 1177.63,
|
285189 |
+
"learning_rate": 7.657544426494346e-06,
|
285190 |
+
"loss": 0.6026,
|
285191 |
+
"step": 146055
|
285192 |
+
},
|
285193 |
+
{
|
285194 |
+
"epoch": 1177.67,
|
285195 |
+
"learning_rate": 7.65746365105008e-06,
|
285196 |
+
"loss": 0.2731,
|
285197 |
+
"step": 146060
|
285198 |
+
},
|
285199 |
+
{
|
285200 |
+
"epoch": 1177.71,
|
285201 |
+
"learning_rate": 7.657382875605816e-06,
|
285202 |
+
"loss": 0.2761,
|
285203 |
+
"step": 146065
|
285204 |
+
},
|
285205 |
+
{
|
285206 |
+
"epoch": 1177.76,
|
285207 |
+
"learning_rate": 7.65730210016155e-06,
|
285208 |
+
"loss": 0.3563,
|
285209 |
+
"step": 146070
|
285210 |
+
},
|
285211 |
+
{
|
285212 |
+
"epoch": 1177.8,
|
285213 |
+
"learning_rate": 7.657221324717286e-06,
|
285214 |
+
"loss": 0.8102,
|
285215 |
+
"step": 146075
|
285216 |
+
},
|
285217 |
+
{
|
285218 |
+
"epoch": 1177.84,
|
285219 |
+
"learning_rate": 7.65714054927302e-06,
|
285220 |
+
"loss": 0.5813,
|
285221 |
+
"step": 146080
|
285222 |
+
},
|
285223 |
+
{
|
285224 |
+
"epoch": 1177.88,
|
285225 |
+
"learning_rate": 7.657059773828756e-06,
|
285226 |
+
"loss": 0.3342,
|
285227 |
+
"step": 146085
|
285228 |
+
},
|
285229 |
+
{
|
285230 |
+
"epoch": 1177.92,
|
285231 |
+
"learning_rate": 7.656978998384492e-06,
|
285232 |
+
"loss": 0.3355,
|
285233 |
+
"step": 146090
|
285234 |
+
},
|
285235 |
+
{
|
285236 |
+
"epoch": 1177.96,
|
285237 |
+
"learning_rate": 7.656898222940226e-06,
|
285238 |
+
"loss": 0.4923,
|
285239 |
+
"step": 146095
|
285240 |
+
},
|
285241 |
+
{
|
285242 |
+
"epoch": 1178.0,
|
285243 |
+
"learning_rate": 7.656817447495962e-06,
|
285244 |
+
"loss": 1.2601,
|
285245 |
+
"step": 146100
|
285246 |
+
},
|
285247 |
+
{
|
285248 |
+
"epoch": 1178.0,
|
285249 |
+
"eval_loss": 0.4012053608894348,
|
285250 |
+
"eval_runtime": 40.577,
|
285251 |
+
"eval_samples_per_second": 20.627,
|
285252 |
+
"eval_steps_per_second": 0.665,
|
285253 |
+
"eval_wer": 0.1786023678891135,
|
285254 |
+
"step": 146100
|
285255 |
+
},
|
285256 |
+
{
|
285257 |
+
"epoch": 1178.04,
|
285258 |
+
"learning_rate": 7.656736672051696e-06,
|
285259 |
+
"loss": 0.2932,
|
285260 |
+
"step": 146105
|
285261 |
+
},
|
285262 |
+
{
|
285263 |
+
"epoch": 1178.08,
|
285264 |
+
"learning_rate": 7.656655896607432e-06,
|
285265 |
+
"loss": 0.2603,
|
285266 |
+
"step": 146110
|
285267 |
+
},
|
285268 |
+
{
|
285269 |
+
"epoch": 1178.12,
|
285270 |
+
"learning_rate": 7.656575121163166e-06,
|
285271 |
+
"loss": 0.2993,
|
285272 |
+
"step": 146115
|
285273 |
+
},
|
285274 |
+
{
|
285275 |
+
"epoch": 1178.16,
|
285276 |
+
"learning_rate": 7.656494345718902e-06,
|
285277 |
+
"loss": 0.4776,
|
285278 |
+
"step": 146120
|
285279 |
+
},
|
285280 |
+
{
|
285281 |
+
"epoch": 1178.2,
|
285282 |
+
"learning_rate": 7.656413570274636e-06,
|
285283 |
+
"loss": 1.1354,
|
285284 |
+
"step": 146125
|
285285 |
+
},
|
285286 |
+
{
|
285287 |
+
"epoch": 1178.24,
|
285288 |
+
"learning_rate": 7.656332794830372e-06,
|
285289 |
+
"loss": 0.2897,
|
285290 |
+
"step": 146130
|
285291 |
+
},
|
285292 |
+
{
|
285293 |
+
"epoch": 1178.28,
|
285294 |
+
"learning_rate": 7.656252019386106e-06,
|
285295 |
+
"loss": 0.2256,
|
285296 |
+
"step": 146135
|
285297 |
+
},
|
285298 |
+
{
|
285299 |
+
"epoch": 1178.32,
|
285300 |
+
"learning_rate": 7.656171243941842e-06,
|
285301 |
+
"loss": 0.2845,
|
285302 |
+
"step": 146140
|
285303 |
+
},
|
285304 |
+
{
|
285305 |
+
"epoch": 1178.36,
|
285306 |
+
"learning_rate": 7.656090468497578e-06,
|
285307 |
+
"loss": 0.4063,
|
285308 |
+
"step": 146145
|
285309 |
+
},
|
285310 |
+
{
|
285311 |
+
"epoch": 1178.4,
|
285312 |
+
"learning_rate": 7.656009693053312e-06,
|
285313 |
+
"loss": 1.1799,
|
285314 |
+
"step": 146150
|
285315 |
+
},
|
285316 |
+
{
|
285317 |
+
"epoch": 1178.44,
|
285318 |
+
"learning_rate": 7.655928917609048e-06,
|
285319 |
+
"loss": 0.3141,
|
285320 |
+
"step": 146155
|
285321 |
+
},
|
285322 |
+
{
|
285323 |
+
"epoch": 1178.48,
|
285324 |
+
"learning_rate": 7.655848142164782e-06,
|
285325 |
+
"loss": 0.2586,
|
285326 |
+
"step": 146160
|
285327 |
+
},
|
285328 |
+
{
|
285329 |
+
"epoch": 1178.52,
|
285330 |
+
"learning_rate": 7.655767366720518e-06,
|
285331 |
+
"loss": 0.3502,
|
285332 |
+
"step": 146165
|
285333 |
+
},
|
285334 |
+
{
|
285335 |
+
"epoch": 1178.56,
|
285336 |
+
"learning_rate": 7.655686591276252e-06,
|
285337 |
+
"loss": 0.4228,
|
285338 |
+
"step": 146170
|
285339 |
+
},
|
285340 |
+
{
|
285341 |
+
"epoch": 1178.6,
|
285342 |
+
"learning_rate": 7.655605815831988e-06,
|
285343 |
+
"loss": 1.2601,
|
285344 |
+
"step": 146175
|
285345 |
+
},
|
285346 |
+
{
|
285347 |
+
"epoch": 1178.64,
|
285348 |
+
"learning_rate": 7.655525040387722e-06,
|
285349 |
+
"loss": 0.2679,
|
285350 |
+
"step": 146180
|
285351 |
+
},
|
285352 |
+
{
|
285353 |
+
"epoch": 1178.68,
|
285354 |
+
"learning_rate": 7.655444264943458e-06,
|
285355 |
+
"loss": 0.2357,
|
285356 |
+
"step": 146185
|
285357 |
+
},
|
285358 |
+
{
|
285359 |
+
"epoch": 1178.72,
|
285360 |
+
"learning_rate": 7.655363489499192e-06,
|
285361 |
+
"loss": 0.3045,
|
285362 |
+
"step": 146190
|
285363 |
+
},
|
285364 |
+
{
|
285365 |
+
"epoch": 1178.76,
|
285366 |
+
"learning_rate": 7.655282714054928e-06,
|
285367 |
+
"loss": 0.5178,
|
285368 |
+
"step": 146195
|
285369 |
+
},
|
285370 |
+
{
|
285371 |
+
"epoch": 1178.8,
|
285372 |
+
"learning_rate": 7.655201938610663e-06,
|
285373 |
+
"loss": 1.2614,
|
285374 |
+
"step": 146200
|
285375 |
+
},
|
285376 |
+
{
|
285377 |
+
"epoch": 1178.84,
|
285378 |
+
"learning_rate": 7.655121163166397e-06,
|
285379 |
+
"loss": 0.3759,
|
285380 |
+
"step": 146205
|
285381 |
+
},
|
285382 |
+
{
|
285383 |
+
"epoch": 1178.88,
|
285384 |
+
"learning_rate": 7.655040387722133e-06,
|
285385 |
+
"loss": 0.2394,
|
285386 |
+
"step": 146210
|
285387 |
+
},
|
285388 |
+
{
|
285389 |
+
"epoch": 1178.92,
|
285390 |
+
"learning_rate": 7.654959612277867e-06,
|
285391 |
+
"loss": 0.275,
|
285392 |
+
"step": 146215
|
285393 |
+
},
|
285394 |
+
{
|
285395 |
+
"epoch": 1178.96,
|
285396 |
+
"learning_rate": 7.654878836833603e-06,
|
285397 |
+
"loss": 0.5791,
|
285398 |
+
"step": 146220
|
285399 |
+
},
|
285400 |
+
{
|
285401 |
+
"epoch": 1179.0,
|
285402 |
+
"eval_loss": 0.358010470867157,
|
285403 |
+
"eval_runtime": 40.5201,
|
285404 |
+
"eval_samples_per_second": 20.656,
|
285405 |
+
"eval_steps_per_second": 0.666,
|
285406 |
+
"eval_wer": 0.176530911221799,
|
285407 |
+
"step": 146224
|
285408 |
+
},
|
285409 |
+
{
|
285410 |
+
"epoch": 1179.01,
|
285411 |
+
"learning_rate": 7.654798061389337e-06,
|
285412 |
+
"loss": 0.5013,
|
285413 |
+
"step": 146225
|
285414 |
+
},
|
285415 |
+
{
|
285416 |
+
"epoch": 1179.05,
|
285417 |
+
"learning_rate": 7.654717285945073e-06,
|
285418 |
+
"loss": 0.2654,
|
285419 |
+
"step": 146230
|
285420 |
+
},
|
285421 |
+
{
|
285422 |
+
"epoch": 1179.09,
|
285423 |
+
"learning_rate": 7.654636510500807e-06,
|
285424 |
+
"loss": 0.2647,
|
285425 |
+
"step": 146235
|
285426 |
+
},
|
285427 |
+
{
|
285428 |
+
"epoch": 1179.13,
|
285429 |
+
"learning_rate": 7.654555735056543e-06,
|
285430 |
+
"loss": 0.3248,
|
285431 |
+
"step": 146240
|
285432 |
+
},
|
285433 |
+
{
|
285434 |
+
"epoch": 1179.17,
|
285435 |
+
"learning_rate": 7.654474959612277e-06,
|
285436 |
+
"loss": 0.4694,
|
285437 |
+
"step": 146245
|
285438 |
+
},
|
285439 |
+
{
|
285440 |
+
"epoch": 1179.21,
|
285441 |
+
"learning_rate": 7.654394184168013e-06,
|
285442 |
+
"loss": 1.0148,
|
285443 |
+
"step": 146250
|
285444 |
+
},
|
285445 |
+
{
|
285446 |
+
"epoch": 1179.25,
|
285447 |
+
"learning_rate": 7.654313408723747e-06,
|
285448 |
+
"loss": 0.2833,
|
285449 |
+
"step": 146255
|
285450 |
+
},
|
285451 |
+
{
|
285452 |
+
"epoch": 1179.29,
|
285453 |
+
"learning_rate": 7.654232633279483e-06,
|
285454 |
+
"loss": 0.2983,
|
285455 |
+
"step": 146260
|
285456 |
+
},
|
285457 |
+
{
|
285458 |
+
"epoch": 1179.33,
|
285459 |
+
"learning_rate": 7.654151857835219e-06,
|
285460 |
+
"loss": 0.3522,
|
285461 |
+
"step": 146265
|
285462 |
+
},
|
285463 |
+
{
|
285464 |
+
"epoch": 1179.37,
|
285465 |
+
"learning_rate": 7.654071082390953e-06,
|
285466 |
+
"loss": 0.5271,
|
285467 |
+
"step": 146270
|
285468 |
+
},
|
285469 |
+
{
|
285470 |
+
"epoch": 1179.41,
|
285471 |
+
"learning_rate": 7.653990306946689e-06,
|
285472 |
+
"loss": 1.1747,
|
285473 |
+
"step": 146275
|
285474 |
+
},
|
285475 |
+
{
|
285476 |
+
"epoch": 1179.45,
|
285477 |
+
"learning_rate": 7.653909531502423e-06,
|
285478 |
+
"loss": 0.2796,
|
285479 |
+
"step": 146280
|
285480 |
+
},
|
285481 |
+
{
|
285482 |
+
"epoch": 1179.49,
|
285483 |
+
"learning_rate": 7.653828756058159e-06,
|
285484 |
+
"loss": 0.2518,
|
285485 |
+
"step": 146285
|
285486 |
+
},
|
285487 |
+
{
|
285488 |
+
"epoch": 1179.53,
|
285489 |
+
"learning_rate": 7.653747980613893e-06,
|
285490 |
+
"loss": 0.307,
|
285491 |
+
"step": 146290
|
285492 |
+
},
|
285493 |
+
{
|
285494 |
+
"epoch": 1179.57,
|
285495 |
+
"learning_rate": 7.653667205169629e-06,
|
285496 |
+
"loss": 0.4902,
|
285497 |
+
"step": 146295
|
285498 |
+
},
|
285499 |
+
{
|
285500 |
+
"epoch": 1179.61,
|
285501 |
+
"learning_rate": 7.653586429725363e-06,
|
285502 |
+
"loss": 1.0019,
|
285503 |
+
"step": 146300
|
285504 |
+
},
|
285505 |
+
{
|
285506 |
+
"epoch": 1179.65,
|
285507 |
+
"learning_rate": 7.653505654281099e-06,
|
285508 |
+
"loss": 0.278,
|
285509 |
+
"step": 146305
|
285510 |
+
},
|
285511 |
+
{
|
285512 |
+
"epoch": 1179.69,
|
285513 |
+
"learning_rate": 7.653424878836833e-06,
|
285514 |
+
"loss": 0.3384,
|
285515 |
+
"step": 146310
|
285516 |
+
},
|
285517 |
+
{
|
285518 |
+
"epoch": 1179.73,
|
285519 |
+
"learning_rate": 7.653344103392569e-06,
|
285520 |
+
"loss": 0.3353,
|
285521 |
+
"step": 146315
|
285522 |
+
},
|
285523 |
+
{
|
285524 |
+
"epoch": 1179.77,
|
285525 |
+
"learning_rate": 7.653263327948305e-06,
|
285526 |
+
"loss": 0.5559,
|
285527 |
+
"step": 146320
|
285528 |
+
},
|
285529 |
+
{
|
285530 |
+
"epoch": 1179.81,
|
285531 |
+
"learning_rate": 7.653182552504039e-06,
|
285532 |
+
"loss": 1.4385,
|
285533 |
+
"step": 146325
|
285534 |
+
},
|
285535 |
+
{
|
285536 |
+
"epoch": 1179.85,
|
285537 |
+
"learning_rate": 7.653101777059775e-06,
|
285538 |
+
"loss": 0.2435,
|
285539 |
+
"step": 146330
|
285540 |
+
},
|
285541 |
+
{
|
285542 |
+
"epoch": 1179.9,
|
285543 |
+
"learning_rate": 7.653021001615509e-06,
|
285544 |
+
"loss": 0.2579,
|
285545 |
+
"step": 146335
|
285546 |
+
},
|
285547 |
+
{
|
285548 |
+
"epoch": 1179.94,
|
285549 |
+
"learning_rate": 7.652940226171245e-06,
|
285550 |
+
"loss": 0.345,
|
285551 |
+
"step": 146340
|
285552 |
+
},
|
285553 |
+
{
|
285554 |
+
"epoch": 1179.98,
|
285555 |
+
"learning_rate": 7.652859450726979e-06,
|
285556 |
+
"loss": 0.5621,
|
285557 |
+
"step": 146345
|
285558 |
+
},
|
285559 |
+
{
|
285560 |
+
"epoch": 1180.0,
|
285561 |
+
"eval_loss": 0.449453204870224,
|
285562 |
+
"eval_runtime": 40.0278,
|
285563 |
+
"eval_samples_per_second": 20.91,
|
285564 |
+
"eval_steps_per_second": 0.675,
|
285565 |
+
"eval_wer": 0.18728573557907288,
|
285566 |
+
"step": 146348
|
285567 |
+
},
|
285568 |
+
{
|
285569 |
+
"epoch": 1180.02,
|
285570 |
+
"learning_rate": 7.652778675282715e-06,
|
285571 |
+
"loss": 0.2882,
|
285572 |
+
"step": 146350
|
285573 |
+
},
|
285574 |
+
{
|
285575 |
+
"epoch": 1180.06,
|
285576 |
+
"learning_rate": 7.652697899838449e-06,
|
285577 |
+
"loss": 0.2908,
|
285578 |
+
"step": 146355
|
285579 |
+
},
|
285580 |
+
{
|
285581 |
+
"epoch": 1180.1,
|
285582 |
+
"learning_rate": 7.652617124394185e-06,
|
285583 |
+
"loss": 0.2317,
|
285584 |
+
"step": 146360
|
285585 |
+
},
|
285586 |
+
{
|
285587 |
+
"epoch": 1180.14,
|
285588 |
+
"learning_rate": 7.652536348949919e-06,
|
285589 |
+
"loss": 0.3416,
|
285590 |
+
"step": 146365
|
285591 |
+
},
|
285592 |
+
{
|
285593 |
+
"epoch": 1180.18,
|
285594 |
+
"learning_rate": 7.652455573505655e-06,
|
285595 |
+
"loss": 0.6812,
|
285596 |
+
"step": 146370
|
285597 |
+
},
|
285598 |
+
{
|
285599 |
+
"epoch": 1180.22,
|
285600 |
+
"learning_rate": 7.65237479806139e-06,
|
285601 |
+
"loss": 0.8339,
|
285602 |
+
"step": 146375
|
285603 |
+
},
|
285604 |
+
{
|
285605 |
+
"epoch": 1180.26,
|
285606 |
+
"learning_rate": 7.652294022617125e-06,
|
285607 |
+
"loss": 0.431,
|
285608 |
+
"step": 146380
|
285609 |
+
},
|
285610 |
+
{
|
285611 |
+
"epoch": 1180.3,
|
285612 |
+
"learning_rate": 7.65221324717286e-06,
|
285613 |
+
"loss": 0.3809,
|
285614 |
+
"step": 146385
|
285615 |
+
},
|
285616 |
+
{
|
285617 |
+
"epoch": 1180.34,
|
285618 |
+
"learning_rate": 7.652132471728595e-06,
|
285619 |
+
"loss": 0.3004,
|
285620 |
+
"step": 146390
|
285621 |
+
},
|
285622 |
+
{
|
285623 |
+
"epoch": 1180.38,
|
285624 |
+
"learning_rate": 7.65205169628433e-06,
|
285625 |
+
"loss": 0.578,
|
285626 |
+
"step": 146395
|
285627 |
+
},
|
285628 |
+
{
|
285629 |
+
"epoch": 1180.42,
|
285630 |
+
"learning_rate": 7.651970920840065e-06,
|
285631 |
+
"loss": 0.9797,
|
285632 |
+
"step": 146400
|
285633 |
+
},
|
285634 |
+
{
|
285635 |
+
"epoch": 1180.46,
|
285636 |
+
"learning_rate": 7.6518901453958e-06,
|
285637 |
+
"loss": 0.2545,
|
285638 |
+
"step": 146405
|
285639 |
+
},
|
285640 |
+
{
|
285641 |
+
"epoch": 1180.5,
|
285642 |
+
"learning_rate": 7.651809369951535e-06,
|
285643 |
+
"loss": 0.2621,
|
285644 |
+
"step": 146410
|
285645 |
+
},
|
285646 |
+
{
|
285647 |
+
"epoch": 1180.54,
|
285648 |
+
"learning_rate": 7.65172859450727e-06,
|
285649 |
+
"loss": 0.3193,
|
285650 |
+
"step": 146415
|
285651 |
+
},
|
285652 |
+
{
|
285653 |
+
"epoch": 1180.58,
|
285654 |
+
"learning_rate": 7.651647819063005e-06,
|
285655 |
+
"loss": 0.6539,
|
285656 |
+
"step": 146420
|
285657 |
+
},
|
285658 |
+
{
|
285659 |
+
"epoch": 1180.62,
|
285660 |
+
"learning_rate": 7.65156704361874e-06,
|
285661 |
+
"loss": 1.0436,
|
285662 |
+
"step": 146425
|
285663 |
+
},
|
285664 |
+
{
|
285665 |
+
"epoch": 1180.66,
|
285666 |
+
"learning_rate": 7.651486268174474e-06,
|
285667 |
+
"loss": 0.2563,
|
285668 |
+
"step": 146430
|
285669 |
+
},
|
285670 |
+
{
|
285671 |
+
"epoch": 1180.7,
|
285672 |
+
"learning_rate": 7.65140549273021e-06,
|
285673 |
+
"loss": 0.3142,
|
285674 |
+
"step": 146435
|
285675 |
+
},
|
285676 |
+
{
|
285677 |
+
"epoch": 1180.74,
|
285678 |
+
"learning_rate": 7.651324717285946e-06,
|
285679 |
+
"loss": 0.3735,
|
285680 |
+
"step": 146440
|
285681 |
+
},
|
285682 |
+
{
|
285683 |
+
"epoch": 1180.78,
|
285684 |
+
"learning_rate": 7.65124394184168e-06,
|
285685 |
+
"loss": 0.6211,
|
285686 |
+
"step": 146445
|
285687 |
+
},
|
285688 |
+
{
|
285689 |
+
"epoch": 1180.82,
|
285690 |
+
"learning_rate": 7.651163166397416e-06,
|
285691 |
+
"loss": 1.0659,
|
285692 |
+
"step": 146450
|
285693 |
+
},
|
285694 |
+
{
|
285695 |
+
"epoch": 1180.86,
|
285696 |
+
"learning_rate": 7.65108239095315e-06,
|
285697 |
+
"loss": 0.3007,
|
285698 |
+
"step": 146455
|
285699 |
+
},
|
285700 |
+
{
|
285701 |
+
"epoch": 1180.9,
|
285702 |
+
"learning_rate": 7.651001615508886e-06,
|
285703 |
+
"loss": 0.2761,
|
285704 |
+
"step": 146460
|
285705 |
+
},
|
285706 |
+
{
|
285707 |
+
"epoch": 1180.94,
|
285708 |
+
"learning_rate": 7.65092084006462e-06,
|
285709 |
+
"loss": 0.3487,
|
285710 |
+
"step": 146465
|
285711 |
+
},
|
285712 |
+
{
|
285713 |
+
"epoch": 1180.98,
|
285714 |
+
"learning_rate": 7.650840064620356e-06,
|
285715 |
+
"loss": 0.7013,
|
285716 |
+
"step": 146470
|
285717 |
+
},
|
285718 |
+
{
|
285719 |
+
"epoch": 1181.0,
|
285720 |
+
"eval_loss": 0.34985774755477905,
|
285721 |
+
"eval_runtime": 41.8541,
|
285722 |
+
"eval_samples_per_second": 19.998,
|
285723 |
+
"eval_steps_per_second": 0.645,
|
285724 |
+
"eval_wer": 0.1758169934640523,
|
285725 |
+
"step": 146472
|
285726 |
}
|
285727 |
],
|
285728 |
+
"max_steps": 620000,
|
285729 |
"num_train_epochs": 5000,
|
285730 |
+
"total_flos": 4.1218134700506854e+20,
|
285731 |
"trial_name": null,
|
285732 |
"trial_params": null
|
285733 |
}
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630228288.1449773/events.out.tfevents.1630228288.cc93b136ebf5.1086.109
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfe38cec38c3d2adbc09d27a46a1e57a1f70fa126623f4f52627be155109468e
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630228721.9317095/events.out.tfevents.1630228721.cc93b136ebf5.1086.111
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28112e380a1179c13caf1bfbf7221311ccc001134b60d565f8f56c89d707889c
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630229152.9104557/events.out.tfevents.1630229152.cc93b136ebf5.1086.113
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6748d316e70868842a9f08d249dc07f834ef93882afb2c7c7fd28f72375b2422
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630229586.5053954/events.out.tfevents.1630229586.cc93b136ebf5.1086.115
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:feb8a302b12fa256e6d48cbf4c25ff505338ca01049c4f58a77b6d4f4b35098b
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630230014.08416/events.out.tfevents.1630230014.cc93b136ebf5.1086.117
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e6b0961bc142db049474b69ea745d4c3a673f78a77e8b18e020138d7fee8b38
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630228288.cc93b136ebf5.1086.108
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:630bfa7f41e9ac1296d8c100a479675d8878d9a89021acbe1d1e3fd54480eaec
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630228721.cc93b136ebf5.1086.110
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc375cfeb2958a56e334f99323d9a111298cd5b84f7f763f7c9f2e5c1b031302
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630229152.cc93b136ebf5.1086.112
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:716525fcb37cbd537930bceaeec45c754a919f6732b733a07571f7ff0a0ea616
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630229586.cc93b136ebf5.1086.114
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:933a727cbd0e343f0d73aec1a0317b5741e8e77cd32fdda735bd47490f98c084
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630230014.cc93b136ebf5.1086.116
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:931819ecb40a82e206200184d2b3d92970a599c9a49de08dcace11c58b3ff38f
|
3 |
+
size 8622
|