"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630048138.6581674/events.out.tfevents.1630048138.52f5c7e305a3.886.81 +3 -0
- model-bin/finetune/base/log/1630048619.9533648/events.out.tfevents.1630048619.52f5c7e305a3.886.83 +3 -0
- model-bin/finetune/base/log/1630049102.1531558/events.out.tfevents.1630049102.52f5c7e305a3.886.85 +3 -0
- model-bin/finetune/base/log/1630049572.6758537/events.out.tfevents.1630049572.52f5c7e305a3.886.87 +3 -0
- model-bin/finetune/base/log/1630050046.2035077/events.out.tfevents.1630050046.52f5c7e305a3.886.89 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630048138.52f5c7e305a3.886.80 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630048619.52f5c7e305a3.886.82 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630049102.52f5c7e305a3.886.84 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630049572.52f5c7e305a3.886.86 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630050046.52f5c7e305a3.886.88 +3 -0
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1573691ea7d0fc61236cc2759103b23e73e155f43b48959ab586bdb034de93db
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0ad107c141d3f7809507fa2921dcccdf0788b1defb6ed5e37e0ce2eaa7d50eec
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0940cfeead080e2f3ccd94ee15dab324a13e5461273c03d52367d56baa29b599
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd0c1baf2f5398bf54715f4be73eadcf661402972348e07ee31ecfc7ef64bd75
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcf1b5511c9d2f60343a826d533e62e2cd2de37fa4f207684f8625a938a917e1
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -241071,11 +241071,806 @@
|
|
241071 |
"eval_steps_per_second": 0.642,
|
241072 |
"eval_wer": 0.18447376051029285,
|
241073 |
"step": 111379
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
241074 |
}
|
241075 |
],
|
241076 |
"max_steps": 620000,
|
241077 |
"num_train_epochs": 5000,
|
241078 |
-
"total_flos": 3.
|
241079 |
"trial_name": null,
|
241080 |
"trial_params": null
|
241081 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 902.995983935743,
|
5 |
+
"global_step": 112002,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
241071 |
"eval_steps_per_second": 0.642,
|
241072 |
"eval_wer": 0.18447376051029285,
|
241073 |
"step": 111379
|
241074 |
+
},
|
241075 |
+
{
|
241076 |
+
"epoch": 898.01,
|
241077 |
+
"learning_rate": 8.217512116316641e-06,
|
241078 |
+
"loss": 0.3168,
|
241079 |
+
"step": 111380
|
241080 |
+
},
|
241081 |
+
{
|
241082 |
+
"epoch": 898.05,
|
241083 |
+
"learning_rate": 8.217431340872375e-06,
|
241084 |
+
"loss": 0.2956,
|
241085 |
+
"step": 111385
|
241086 |
+
},
|
241087 |
+
{
|
241088 |
+
"epoch": 898.09,
|
241089 |
+
"learning_rate": 8.217350565428111e-06,
|
241090 |
+
"loss": 0.2882,
|
241091 |
+
"step": 111390
|
241092 |
+
},
|
241093 |
+
{
|
241094 |
+
"epoch": 898.13,
|
241095 |
+
"learning_rate": 8.217269789983845e-06,
|
241096 |
+
"loss": 0.2941,
|
241097 |
+
"step": 111395
|
241098 |
+
},
|
241099 |
+
{
|
241100 |
+
"epoch": 898.17,
|
241101 |
+
"learning_rate": 8.217189014539581e-06,
|
241102 |
+
"loss": 0.5359,
|
241103 |
+
"step": 111400
|
241104 |
+
},
|
241105 |
+
{
|
241106 |
+
"epoch": 898.21,
|
241107 |
+
"learning_rate": 8.217108239095315e-06,
|
241108 |
+
"loss": 1.293,
|
241109 |
+
"step": 111405
|
241110 |
+
},
|
241111 |
+
{
|
241112 |
+
"epoch": 898.25,
|
241113 |
+
"learning_rate": 8.217027463651051e-06,
|
241114 |
+
"loss": 0.2586,
|
241115 |
+
"step": 111410
|
241116 |
+
},
|
241117 |
+
{
|
241118 |
+
"epoch": 898.29,
|
241119 |
+
"learning_rate": 8.216946688206785e-06,
|
241120 |
+
"loss": 0.2983,
|
241121 |
+
"step": 111415
|
241122 |
+
},
|
241123 |
+
{
|
241124 |
+
"epoch": 898.33,
|
241125 |
+
"learning_rate": 8.216865912762521e-06,
|
241126 |
+
"loss": 0.3252,
|
241127 |
+
"step": 111420
|
241128 |
+
},
|
241129 |
+
{
|
241130 |
+
"epoch": 898.37,
|
241131 |
+
"learning_rate": 8.216785137318255e-06,
|
241132 |
+
"loss": 0.4947,
|
241133 |
+
"step": 111425
|
241134 |
+
},
|
241135 |
+
{
|
241136 |
+
"epoch": 898.41,
|
241137 |
+
"learning_rate": 8.216704361873991e-06,
|
241138 |
+
"loss": 1.025,
|
241139 |
+
"step": 111430
|
241140 |
+
},
|
241141 |
+
{
|
241142 |
+
"epoch": 898.45,
|
241143 |
+
"learning_rate": 8.216623586429727e-06,
|
241144 |
+
"loss": 0.3136,
|
241145 |
+
"step": 111435
|
241146 |
+
},
|
241147 |
+
{
|
241148 |
+
"epoch": 898.49,
|
241149 |
+
"learning_rate": 8.216542810985461e-06,
|
241150 |
+
"loss": 0.2458,
|
241151 |
+
"step": 111440
|
241152 |
+
},
|
241153 |
+
{
|
241154 |
+
"epoch": 898.53,
|
241155 |
+
"learning_rate": 8.216462035541197e-06,
|
241156 |
+
"loss": 0.3212,
|
241157 |
+
"step": 111445
|
241158 |
+
},
|
241159 |
+
{
|
241160 |
+
"epoch": 898.57,
|
241161 |
+
"learning_rate": 8.216381260096931e-06,
|
241162 |
+
"loss": 0.5968,
|
241163 |
+
"step": 111450
|
241164 |
+
},
|
241165 |
+
{
|
241166 |
+
"epoch": 898.61,
|
241167 |
+
"learning_rate": 8.216300484652667e-06,
|
241168 |
+
"loss": 1.074,
|
241169 |
+
"step": 111455
|
241170 |
+
},
|
241171 |
+
{
|
241172 |
+
"epoch": 898.65,
|
241173 |
+
"learning_rate": 8.216219709208401e-06,
|
241174 |
+
"loss": 0.2987,
|
241175 |
+
"step": 111460
|
241176 |
+
},
|
241177 |
+
{
|
241178 |
+
"epoch": 898.69,
|
241179 |
+
"learning_rate": 8.216138933764137e-06,
|
241180 |
+
"loss": 0.3227,
|
241181 |
+
"step": 111465
|
241182 |
+
},
|
241183 |
+
{
|
241184 |
+
"epoch": 898.73,
|
241185 |
+
"learning_rate": 8.216058158319871e-06,
|
241186 |
+
"loss": 0.5044,
|
241187 |
+
"step": 111470
|
241188 |
+
},
|
241189 |
+
{
|
241190 |
+
"epoch": 898.77,
|
241191 |
+
"learning_rate": 8.215977382875607e-06,
|
241192 |
+
"loss": 0.5049,
|
241193 |
+
"step": 111475
|
241194 |
+
},
|
241195 |
+
{
|
241196 |
+
"epoch": 898.81,
|
241197 |
+
"learning_rate": 8.215896607431341e-06,
|
241198 |
+
"loss": 1.1103,
|
241199 |
+
"step": 111480
|
241200 |
+
},
|
241201 |
+
{
|
241202 |
+
"epoch": 898.85,
|
241203 |
+
"learning_rate": 8.215815831987077e-06,
|
241204 |
+
"loss": 0.292,
|
241205 |
+
"step": 111485
|
241206 |
+
},
|
241207 |
+
{
|
241208 |
+
"epoch": 898.89,
|
241209 |
+
"learning_rate": 8.215735056542811e-06,
|
241210 |
+
"loss": 0.2973,
|
241211 |
+
"step": 111490
|
241212 |
+
},
|
241213 |
+
{
|
241214 |
+
"epoch": 898.93,
|
241215 |
+
"learning_rate": 8.215654281098547e-06,
|
241216 |
+
"loss": 0.3051,
|
241217 |
+
"step": 111495
|
241218 |
+
},
|
241219 |
+
{
|
241220 |
+
"epoch": 898.97,
|
241221 |
+
"learning_rate": 8.215573505654283e-06,
|
241222 |
+
"loss": 0.5463,
|
241223 |
+
"step": 111500
|
241224 |
+
},
|
241225 |
+
{
|
241226 |
+
"epoch": 899.0,
|
241227 |
+
"eval_loss": 0.45212462544441223,
|
241228 |
+
"eval_runtime": 41.8965,
|
241229 |
+
"eval_samples_per_second": 19.954,
|
241230 |
+
"eval_steps_per_second": 0.644,
|
241231 |
+
"eval_wer": 0.18223921193854298,
|
241232 |
+
"step": 111503
|
241233 |
+
},
|
241234 |
+
{
|
241235 |
+
"epoch": 892.02,
|
241236 |
+
"learning_rate": 8.215492730210017e-06,
|
241237 |
+
"loss": 0.5205,
|
241238 |
+
"step": 111505
|
241239 |
+
},
|
241240 |
+
{
|
241241 |
+
"epoch": 892.06,
|
241242 |
+
"learning_rate": 8.215411954765753e-06,
|
241243 |
+
"loss": 0.2519,
|
241244 |
+
"step": 111510
|
241245 |
+
},
|
241246 |
+
{
|
241247 |
+
"epoch": 892.1,
|
241248 |
+
"learning_rate": 8.215331179321487e-06,
|
241249 |
+
"loss": 0.2812,
|
241250 |
+
"step": 111515
|
241251 |
+
},
|
241252 |
+
{
|
241253 |
+
"epoch": 892.14,
|
241254 |
+
"learning_rate": 8.215250403877223e-06,
|
241255 |
+
"loss": 0.3625,
|
241256 |
+
"step": 111520
|
241257 |
+
},
|
241258 |
+
{
|
241259 |
+
"epoch": 892.18,
|
241260 |
+
"learning_rate": 8.215169628432957e-06,
|
241261 |
+
"loss": 0.6445,
|
241262 |
+
"step": 111525
|
241263 |
+
},
|
241264 |
+
{
|
241265 |
+
"epoch": 892.22,
|
241266 |
+
"learning_rate": 8.215088852988693e-06,
|
241267 |
+
"loss": 0.9554,
|
241268 |
+
"step": 111530
|
241269 |
+
},
|
241270 |
+
{
|
241271 |
+
"epoch": 892.25,
|
241272 |
+
"learning_rate": 8.215008077544427e-06,
|
241273 |
+
"loss": 0.2955,
|
241274 |
+
"step": 111535
|
241275 |
+
},
|
241276 |
+
{
|
241277 |
+
"epoch": 892.29,
|
241278 |
+
"learning_rate": 8.214927302100163e-06,
|
241279 |
+
"loss": 0.2583,
|
241280 |
+
"step": 111540
|
241281 |
+
},
|
241282 |
+
{
|
241283 |
+
"epoch": 892.33,
|
241284 |
+
"learning_rate": 8.214846526655897e-06,
|
241285 |
+
"loss": 0.3004,
|
241286 |
+
"step": 111545
|
241287 |
+
},
|
241288 |
+
{
|
241289 |
+
"epoch": 892.37,
|
241290 |
+
"learning_rate": 8.214765751211633e-06,
|
241291 |
+
"loss": 0.6362,
|
241292 |
+
"step": 111550
|
241293 |
+
},
|
241294 |
+
{
|
241295 |
+
"epoch": 892.41,
|
241296 |
+
"learning_rate": 8.214684975767367e-06,
|
241297 |
+
"loss": 1.0824,
|
241298 |
+
"step": 111555
|
241299 |
+
},
|
241300 |
+
{
|
241301 |
+
"epoch": 892.45,
|
241302 |
+
"learning_rate": 8.214604200323102e-06,
|
241303 |
+
"loss": 0.2687,
|
241304 |
+
"step": 111560
|
241305 |
+
},
|
241306 |
+
{
|
241307 |
+
"epoch": 892.49,
|
241308 |
+
"learning_rate": 8.214523424878837e-06,
|
241309 |
+
"loss": 0.3912,
|
241310 |
+
"step": 111565
|
241311 |
+
},
|
241312 |
+
{
|
241313 |
+
"epoch": 892.53,
|
241314 |
+
"learning_rate": 8.214442649434572e-06,
|
241315 |
+
"loss": 0.4064,
|
241316 |
+
"step": 111570
|
241317 |
+
},
|
241318 |
+
{
|
241319 |
+
"epoch": 892.57,
|
241320 |
+
"learning_rate": 8.214361873990308e-06,
|
241321 |
+
"loss": 0.6121,
|
241322 |
+
"step": 111575
|
241323 |
+
},
|
241324 |
+
{
|
241325 |
+
"epoch": 892.61,
|
241326 |
+
"learning_rate": 8.214281098546042e-06,
|
241327 |
+
"loss": 1.0386,
|
241328 |
+
"step": 111580
|
241329 |
+
},
|
241330 |
+
{
|
241331 |
+
"epoch": 892.65,
|
241332 |
+
"learning_rate": 8.214200323101778e-06,
|
241333 |
+
"loss": 0.2815,
|
241334 |
+
"step": 111585
|
241335 |
+
},
|
241336 |
+
{
|
241337 |
+
"epoch": 892.69,
|
241338 |
+
"learning_rate": 8.214119547657512e-06,
|
241339 |
+
"loss": 0.3376,
|
241340 |
+
"step": 111590
|
241341 |
+
},
|
241342 |
+
{
|
241343 |
+
"epoch": 892.73,
|
241344 |
+
"learning_rate": 8.214038772213248e-06,
|
241345 |
+
"loss": 0.3432,
|
241346 |
+
"step": 111595
|
241347 |
+
},
|
241348 |
+
{
|
241349 |
+
"epoch": 892.77,
|
241350 |
+
"learning_rate": 8.213957996768982e-06,
|
241351 |
+
"loss": 0.6253,
|
241352 |
+
"step": 111600
|
241353 |
+
},
|
241354 |
+
{
|
241355 |
+
"epoch": 892.81,
|
241356 |
+
"learning_rate": 8.213877221324718e-06,
|
241357 |
+
"loss": 0.8157,
|
241358 |
+
"step": 111605
|
241359 |
+
},
|
241360 |
+
{
|
241361 |
+
"epoch": 892.85,
|
241362 |
+
"learning_rate": 8.213796445880452e-06,
|
241363 |
+
"loss": 0.282,
|
241364 |
+
"step": 111610
|
241365 |
+
},
|
241366 |
+
{
|
241367 |
+
"epoch": 892.89,
|
241368 |
+
"learning_rate": 8.213715670436188e-06,
|
241369 |
+
"loss": 0.2759,
|
241370 |
+
"step": 111615
|
241371 |
+
},
|
241372 |
+
{
|
241373 |
+
"epoch": 892.93,
|
241374 |
+
"learning_rate": 8.213634894991922e-06,
|
241375 |
+
"loss": 0.3045,
|
241376 |
+
"step": 111620
|
241377 |
+
},
|
241378 |
+
{
|
241379 |
+
"epoch": 892.97,
|
241380 |
+
"learning_rate": 8.213554119547658e-06,
|
241381 |
+
"loss": 0.5921,
|
241382 |
+
"step": 111625
|
241383 |
+
},
|
241384 |
+
{
|
241385 |
+
"epoch": 893.0,
|
241386 |
+
"eval_loss": 0.4683041274547577,
|
241387 |
+
"eval_runtime": 42.5438,
|
241388 |
+
"eval_samples_per_second": 19.65,
|
241389 |
+
"eval_steps_per_second": 0.635,
|
241390 |
+
"eval_wer": 0.18959383338274532,
|
241391 |
+
"step": 111628
|
241392 |
+
},
|
241393 |
+
{
|
241394 |
+
"epoch": 893.02,
|
241395 |
+
"learning_rate": 8.213473344103392e-06,
|
241396 |
+
"loss": 0.3051,
|
241397 |
+
"step": 111630
|
241398 |
+
},
|
241399 |
+
{
|
241400 |
+
"epoch": 893.06,
|
241401 |
+
"learning_rate": 8.213392568659128e-06,
|
241402 |
+
"loss": 0.2988,
|
241403 |
+
"step": 111635
|
241404 |
+
},
|
241405 |
+
{
|
241406 |
+
"epoch": 893.1,
|
241407 |
+
"learning_rate": 8.213311793214864e-06,
|
241408 |
+
"loss": 0.3022,
|
241409 |
+
"step": 111640
|
241410 |
+
},
|
241411 |
+
{
|
241412 |
+
"epoch": 893.14,
|
241413 |
+
"learning_rate": 8.213231017770598e-06,
|
241414 |
+
"loss": 0.3361,
|
241415 |
+
"step": 111645
|
241416 |
+
},
|
241417 |
+
{
|
241418 |
+
"epoch": 893.18,
|
241419 |
+
"learning_rate": 8.213150242326334e-06,
|
241420 |
+
"loss": 0.6711,
|
241421 |
+
"step": 111650
|
241422 |
+
},
|
241423 |
+
{
|
241424 |
+
"epoch": 893.22,
|
241425 |
+
"learning_rate": 8.213069466882068e-06,
|
241426 |
+
"loss": 1.1263,
|
241427 |
+
"step": 111655
|
241428 |
+
},
|
241429 |
+
{
|
241430 |
+
"epoch": 893.26,
|
241431 |
+
"learning_rate": 8.212988691437804e-06,
|
241432 |
+
"loss": 0.3122,
|
241433 |
+
"step": 111660
|
241434 |
+
},
|
241435 |
+
{
|
241436 |
+
"epoch": 893.3,
|
241437 |
+
"learning_rate": 8.212907915993538e-06,
|
241438 |
+
"loss": 0.2739,
|
241439 |
+
"step": 111665
|
241440 |
+
},
|
241441 |
+
{
|
241442 |
+
"epoch": 893.34,
|
241443 |
+
"learning_rate": 8.212827140549274e-06,
|
241444 |
+
"loss": 0.3829,
|
241445 |
+
"step": 111670
|
241446 |
+
},
|
241447 |
+
{
|
241448 |
+
"epoch": 893.38,
|
241449 |
+
"learning_rate": 8.212746365105008e-06,
|
241450 |
+
"loss": 0.6704,
|
241451 |
+
"step": 111675
|
241452 |
+
},
|
241453 |
+
{
|
241454 |
+
"epoch": 893.42,
|
241455 |
+
"learning_rate": 8.212665589660744e-06,
|
241456 |
+
"loss": 1.0724,
|
241457 |
+
"step": 111680
|
241458 |
+
},
|
241459 |
+
{
|
241460 |
+
"epoch": 893.46,
|
241461 |
+
"learning_rate": 8.212584814216478e-06,
|
241462 |
+
"loss": 0.2658,
|
241463 |
+
"step": 111685
|
241464 |
+
},
|
241465 |
+
{
|
241466 |
+
"epoch": 893.5,
|
241467 |
+
"learning_rate": 8.212504038772214e-06,
|
241468 |
+
"loss": 0.3109,
|
241469 |
+
"step": 111690
|
241470 |
+
},
|
241471 |
+
{
|
241472 |
+
"epoch": 893.54,
|
241473 |
+
"learning_rate": 8.212423263327948e-06,
|
241474 |
+
"loss": 0.3202,
|
241475 |
+
"step": 111695
|
241476 |
+
},
|
241477 |
+
{
|
241478 |
+
"epoch": 893.58,
|
241479 |
+
"learning_rate": 8.212342487883684e-06,
|
241480 |
+
"loss": 0.6632,
|
241481 |
+
"step": 111700
|
241482 |
+
},
|
241483 |
+
{
|
241484 |
+
"epoch": 893.62,
|
241485 |
+
"learning_rate": 8.21226171243942e-06,
|
241486 |
+
"loss": 1.0617,
|
241487 |
+
"step": 111705
|
241488 |
+
},
|
241489 |
+
{
|
241490 |
+
"epoch": 893.66,
|
241491 |
+
"learning_rate": 8.212180936995154e-06,
|
241492 |
+
"loss": 0.2693,
|
241493 |
+
"step": 111710
|
241494 |
+
},
|
241495 |
+
{
|
241496 |
+
"epoch": 893.7,
|
241497 |
+
"learning_rate": 8.21210016155089e-06,
|
241498 |
+
"loss": 0.32,
|
241499 |
+
"step": 111715
|
241500 |
+
},
|
241501 |
+
{
|
241502 |
+
"epoch": 893.74,
|
241503 |
+
"learning_rate": 8.212019386106624e-06,
|
241504 |
+
"loss": 0.4235,
|
241505 |
+
"step": 111720
|
241506 |
+
},
|
241507 |
+
{
|
241508 |
+
"epoch": 893.78,
|
241509 |
+
"learning_rate": 8.21193861066236e-06,
|
241510 |
+
"loss": 0.6357,
|
241511 |
+
"step": 111725
|
241512 |
+
},
|
241513 |
+
{
|
241514 |
+
"epoch": 893.82,
|
241515 |
+
"learning_rate": 8.211857835218094e-06,
|
241516 |
+
"loss": 0.9449,
|
241517 |
+
"step": 111730
|
241518 |
+
},
|
241519 |
+
{
|
241520 |
+
"epoch": 893.86,
|
241521 |
+
"learning_rate": 8.21177705977383e-06,
|
241522 |
+
"loss": 0.2906,
|
241523 |
+
"step": 111735
|
241524 |
+
},
|
241525 |
+
{
|
241526 |
+
"epoch": 893.9,
|
241527 |
+
"learning_rate": 8.211696284329564e-06,
|
241528 |
+
"loss": 0.298,
|
241529 |
+
"step": 111740
|
241530 |
+
},
|
241531 |
+
{
|
241532 |
+
"epoch": 893.94,
|
241533 |
+
"learning_rate": 8.2116155088853e-06,
|
241534 |
+
"loss": 0.3205,
|
241535 |
+
"step": 111745
|
241536 |
+
},
|
241537 |
+
{
|
241538 |
+
"epoch": 893.98,
|
241539 |
+
"learning_rate": 8.211534733441034e-06,
|
241540 |
+
"loss": 0.602,
|
241541 |
+
"step": 111750
|
241542 |
+
},
|
241543 |
+
{
|
241544 |
+
"epoch": 894.0,
|
241545 |
+
"eval_loss": 0.3670633137226105,
|
241546 |
+
"eval_runtime": 42.7136,
|
241547 |
+
"eval_samples_per_second": 19.572,
|
241548 |
+
"eval_steps_per_second": 0.632,
|
241549 |
+
"eval_wer": 0.18261776982707215,
|
241550 |
+
"step": 111753
|
241551 |
+
},
|
241552 |
+
{
|
241553 |
+
"epoch": 894.02,
|
241554 |
+
"learning_rate": 8.21145395799677e-06,
|
241555 |
+
"loss": 0.3963,
|
241556 |
+
"step": 111755
|
241557 |
+
},
|
241558 |
+
{
|
241559 |
+
"epoch": 894.06,
|
241560 |
+
"learning_rate": 8.211373182552504e-06,
|
241561 |
+
"loss": 0.3368,
|
241562 |
+
"step": 111760
|
241563 |
+
},
|
241564 |
+
{
|
241565 |
+
"epoch": 894.1,
|
241566 |
+
"learning_rate": 8.21129240710824e-06,
|
241567 |
+
"loss": 0.2995,
|
241568 |
+
"step": 111765
|
241569 |
+
},
|
241570 |
+
{
|
241571 |
+
"epoch": 894.14,
|
241572 |
+
"learning_rate": 8.211211631663974e-06,
|
241573 |
+
"loss": 0.3304,
|
241574 |
+
"step": 111770
|
241575 |
+
},
|
241576 |
+
{
|
241577 |
+
"epoch": 894.18,
|
241578 |
+
"learning_rate": 8.21113085621971e-06,
|
241579 |
+
"loss": 0.5743,
|
241580 |
+
"step": 111775
|
241581 |
+
},
|
241582 |
+
{
|
241583 |
+
"epoch": 894.22,
|
241584 |
+
"learning_rate": 8.211050080775445e-06,
|
241585 |
+
"loss": 0.9688,
|
241586 |
+
"step": 111780
|
241587 |
+
},
|
241588 |
+
{
|
241589 |
+
"epoch": 894.26,
|
241590 |
+
"learning_rate": 8.21096930533118e-06,
|
241591 |
+
"loss": 0.2755,
|
241592 |
+
"step": 111785
|
241593 |
+
},
|
241594 |
+
{
|
241595 |
+
"epoch": 894.3,
|
241596 |
+
"learning_rate": 8.210888529886915e-06,
|
241597 |
+
"loss": 0.3825,
|
241598 |
+
"step": 111790
|
241599 |
+
},
|
241600 |
+
{
|
241601 |
+
"epoch": 894.34,
|
241602 |
+
"learning_rate": 8.21080775444265e-06,
|
241603 |
+
"loss": 0.3727,
|
241604 |
+
"step": 111795
|
241605 |
+
},
|
241606 |
+
{
|
241607 |
+
"epoch": 894.38,
|
241608 |
+
"learning_rate": 8.210726978998385e-06,
|
241609 |
+
"loss": 0.6055,
|
241610 |
+
"step": 111800
|
241611 |
+
},
|
241612 |
+
{
|
241613 |
+
"epoch": 894.42,
|
241614 |
+
"learning_rate": 8.21064620355412e-06,
|
241615 |
+
"loss": 0.949,
|
241616 |
+
"step": 111805
|
241617 |
+
},
|
241618 |
+
{
|
241619 |
+
"epoch": 894.46,
|
241620 |
+
"learning_rate": 8.210565428109855e-06,
|
241621 |
+
"loss": 0.3053,
|
241622 |
+
"step": 111810
|
241623 |
+
},
|
241624 |
+
{
|
241625 |
+
"epoch": 894.5,
|
241626 |
+
"learning_rate": 8.21048465266559e-06,
|
241627 |
+
"loss": 0.4324,
|
241628 |
+
"step": 111815
|
241629 |
+
},
|
241630 |
+
{
|
241631 |
+
"epoch": 894.54,
|
241632 |
+
"learning_rate": 8.210403877221325e-06,
|
241633 |
+
"loss": 0.4127,
|
241634 |
+
"step": 111820
|
241635 |
+
},
|
241636 |
+
{
|
241637 |
+
"epoch": 894.58,
|
241638 |
+
"learning_rate": 8.21032310177706e-06,
|
241639 |
+
"loss": 0.5997,
|
241640 |
+
"step": 111825
|
241641 |
+
},
|
241642 |
+
{
|
241643 |
+
"epoch": 894.62,
|
241644 |
+
"learning_rate": 8.210242326332795e-06,
|
241645 |
+
"loss": 1.0255,
|
241646 |
+
"step": 111830
|
241647 |
+
},
|
241648 |
+
{
|
241649 |
+
"epoch": 894.66,
|
241650 |
+
"learning_rate": 8.21016155088853e-06,
|
241651 |
+
"loss": 0.2515,
|
241652 |
+
"step": 111835
|
241653 |
+
},
|
241654 |
+
{
|
241655 |
+
"epoch": 894.7,
|
241656 |
+
"learning_rate": 8.210080775444265e-06,
|
241657 |
+
"loss": 0.3091,
|
241658 |
+
"step": 111840
|
241659 |
+
},
|
241660 |
+
{
|
241661 |
+
"epoch": 894.74,
|
241662 |
+
"learning_rate": 8.210000000000001e-06,
|
241663 |
+
"loss": 0.3108,
|
241664 |
+
"step": 111845
|
241665 |
+
},
|
241666 |
+
{
|
241667 |
+
"epoch": 894.78,
|
241668 |
+
"learning_rate": 8.209919224555735e-06,
|
241669 |
+
"loss": 0.6145,
|
241670 |
+
"step": 111850
|
241671 |
+
},
|
241672 |
+
{
|
241673 |
+
"epoch": 894.82,
|
241674 |
+
"learning_rate": 8.209838449111471e-06,
|
241675 |
+
"loss": 1.1353,
|
241676 |
+
"step": 111855
|
241677 |
+
},
|
241678 |
+
{
|
241679 |
+
"epoch": 894.86,
|
241680 |
+
"learning_rate": 8.209757673667205e-06,
|
241681 |
+
"loss": 0.3686,
|
241682 |
+
"step": 111860
|
241683 |
+
},
|
241684 |
+
{
|
241685 |
+
"epoch": 894.9,
|
241686 |
+
"learning_rate": 8.209676898222941e-06,
|
241687 |
+
"loss": 0.281,
|
241688 |
+
"step": 111865
|
241689 |
+
},
|
241690 |
+
{
|
241691 |
+
"epoch": 894.94,
|
241692 |
+
"learning_rate": 8.209596122778675e-06,
|
241693 |
+
"loss": 0.4096,
|
241694 |
+
"step": 111870
|
241695 |
+
},
|
241696 |
+
{
|
241697 |
+
"epoch": 894.98,
|
241698 |
+
"learning_rate": 8.209515347334411e-06,
|
241699 |
+
"loss": 0.6653,
|
241700 |
+
"step": 111875
|
241701 |
+
},
|
241702 |
+
{
|
241703 |
+
"epoch": 895.0,
|
241704 |
+
"eval_loss": 0.3755977153778076,
|
241705 |
+
"eval_runtime": 42.4892,
|
241706 |
+
"eval_samples_per_second": 19.676,
|
241707 |
+
"eval_steps_per_second": 0.635,
|
241708 |
+
"eval_wer": 0.18443218123729305,
|
241709 |
+
"step": 111878
|
241710 |
+
},
|
241711 |
+
{
|
241712 |
+
"epoch": 902.02,
|
241713 |
+
"learning_rate": 8.209434571890145e-06,
|
241714 |
+
"loss": 0.3006,
|
241715 |
+
"step": 111880
|
241716 |
+
},
|
241717 |
+
{
|
241718 |
+
"epoch": 902.06,
|
241719 |
+
"learning_rate": 8.209353796445881e-06,
|
241720 |
+
"loss": 0.2412,
|
241721 |
+
"step": 111885
|
241722 |
+
},
|
241723 |
+
{
|
241724 |
+
"epoch": 902.1,
|
241725 |
+
"learning_rate": 8.209273021001615e-06,
|
241726 |
+
"loss": 0.3399,
|
241727 |
+
"step": 111890
|
241728 |
+
},
|
241729 |
+
{
|
241730 |
+
"epoch": 902.14,
|
241731 |
+
"learning_rate": 8.209192245557351e-06,
|
241732 |
+
"loss": 0.3268,
|
241733 |
+
"step": 111895
|
241734 |
+
},
|
241735 |
+
{
|
241736 |
+
"epoch": 902.18,
|
241737 |
+
"learning_rate": 8.209111470113085e-06,
|
241738 |
+
"loss": 0.6156,
|
241739 |
+
"step": 111900
|
241740 |
+
},
|
241741 |
+
{
|
241742 |
+
"epoch": 902.22,
|
241743 |
+
"learning_rate": 8.209030694668821e-06,
|
241744 |
+
"loss": 1.1313,
|
241745 |
+
"step": 111905
|
241746 |
+
},
|
241747 |
+
{
|
241748 |
+
"epoch": 902.26,
|
241749 |
+
"learning_rate": 8.208949919224557e-06,
|
241750 |
+
"loss": 0.3566,
|
241751 |
+
"step": 111910
|
241752 |
+
},
|
241753 |
+
{
|
241754 |
+
"epoch": 902.3,
|
241755 |
+
"learning_rate": 8.20886914378029e-06,
|
241756 |
+
"loss": 0.3318,
|
241757 |
+
"step": 111915
|
241758 |
+
},
|
241759 |
+
{
|
241760 |
+
"epoch": 902.34,
|
241761 |
+
"learning_rate": 8.208788368336027e-06,
|
241762 |
+
"loss": 0.3042,
|
241763 |
+
"step": 111920
|
241764 |
+
},
|
241765 |
+
{
|
241766 |
+
"epoch": 902.38,
|
241767 |
+
"learning_rate": 8.20870759289176e-06,
|
241768 |
+
"loss": 0.5518,
|
241769 |
+
"step": 111925
|
241770 |
+
},
|
241771 |
+
{
|
241772 |
+
"epoch": 902.42,
|
241773 |
+
"learning_rate": 8.208626817447497e-06,
|
241774 |
+
"loss": 0.9917,
|
241775 |
+
"step": 111930
|
241776 |
+
},
|
241777 |
+
{
|
241778 |
+
"epoch": 902.46,
|
241779 |
+
"learning_rate": 8.20854604200323e-06,
|
241780 |
+
"loss": 0.3356,
|
241781 |
+
"step": 111935
|
241782 |
+
},
|
241783 |
+
{
|
241784 |
+
"epoch": 902.5,
|
241785 |
+
"learning_rate": 8.208465266558967e-06,
|
241786 |
+
"loss": 0.286,
|
241787 |
+
"step": 111940
|
241788 |
+
},
|
241789 |
+
{
|
241790 |
+
"epoch": 902.54,
|
241791 |
+
"learning_rate": 8.2083844911147e-06,
|
241792 |
+
"loss": 0.3935,
|
241793 |
+
"step": 111945
|
241794 |
+
},
|
241795 |
+
{
|
241796 |
+
"epoch": 902.58,
|
241797 |
+
"learning_rate": 8.208303715670437e-06,
|
241798 |
+
"loss": 0.6371,
|
241799 |
+
"step": 111950
|
241800 |
+
},
|
241801 |
+
{
|
241802 |
+
"epoch": 902.62,
|
241803 |
+
"learning_rate": 8.20822294022617e-06,
|
241804 |
+
"loss": 0.9316,
|
241805 |
+
"step": 111955
|
241806 |
+
},
|
241807 |
+
{
|
241808 |
+
"epoch": 902.66,
|
241809 |
+
"learning_rate": 8.208142164781907e-06,
|
241810 |
+
"loss": 0.3458,
|
241811 |
+
"step": 111960
|
241812 |
+
},
|
241813 |
+
{
|
241814 |
+
"epoch": 902.7,
|
241815 |
+
"learning_rate": 8.20806138933764e-06,
|
241816 |
+
"loss": 0.3009,
|
241817 |
+
"step": 111965
|
241818 |
+
},
|
241819 |
+
{
|
241820 |
+
"epoch": 902.74,
|
241821 |
+
"learning_rate": 8.207980613893377e-06,
|
241822 |
+
"loss": 0.3419,
|
241823 |
+
"step": 111970
|
241824 |
+
},
|
241825 |
+
{
|
241826 |
+
"epoch": 902.78,
|
241827 |
+
"learning_rate": 8.207899838449112e-06,
|
241828 |
+
"loss": 0.5863,
|
241829 |
+
"step": 111975
|
241830 |
+
},
|
241831 |
+
{
|
241832 |
+
"epoch": 902.82,
|
241833 |
+
"learning_rate": 8.207819063004847e-06,
|
241834 |
+
"loss": 1.0017,
|
241835 |
+
"step": 111980
|
241836 |
+
},
|
241837 |
+
{
|
241838 |
+
"epoch": 902.86,
|
241839 |
+
"learning_rate": 8.207738287560582e-06,
|
241840 |
+
"loss": 0.2961,
|
241841 |
+
"step": 111985
|
241842 |
+
},
|
241843 |
+
{
|
241844 |
+
"epoch": 902.9,
|
241845 |
+
"learning_rate": 8.207657512116317e-06,
|
241846 |
+
"loss": 0.2509,
|
241847 |
+
"step": 111990
|
241848 |
+
},
|
241849 |
+
{
|
241850 |
+
"epoch": 902.94,
|
241851 |
+
"learning_rate": 8.207576736672052e-06,
|
241852 |
+
"loss": 0.33,
|
241853 |
+
"step": 111995
|
241854 |
+
},
|
241855 |
+
{
|
241856 |
+
"epoch": 902.98,
|
241857 |
+
"learning_rate": 8.207495961227786e-06,
|
241858 |
+
"loss": 0.6344,
|
241859 |
+
"step": 112000
|
241860 |
+
},
|
241861 |
+
{
|
241862 |
+
"epoch": 903.0,
|
241863 |
+
"eval_loss": 0.4465380311012268,
|
241864 |
+
"eval_runtime": 43.2217,
|
241865 |
+
"eval_samples_per_second": 19.342,
|
241866 |
+
"eval_steps_per_second": 0.625,
|
241867 |
+
"eval_wer": 0.1886928860613071,
|
241868 |
+
"step": 112002
|
241869 |
}
|
241870 |
],
|
241871 |
"max_steps": 620000,
|
241872 |
"num_train_epochs": 5000,
|
241873 |
+
"total_flos": 3.1520264569981986e+20,
|
241874 |
"trial_name": null,
|
241875 |
"trial_params": null
|
241876 |
}
|
model-bin/finetune/base/{checkpoint-111379 β checkpoint-112002}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630048138.6581674/events.out.tfevents.1630048138.52f5c7e305a3.886.81
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6fd565bfc4df649dc0f4e5aadcf1eead11a12fb058bf7d9f5390ea0410bd88c0
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630048619.9533648/events.out.tfevents.1630048619.52f5c7e305a3.886.83
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd30dccc3acb9b02a74544425cb00b04f97859f9d1abf3d1eb655dec856c326c
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630049102.1531558/events.out.tfevents.1630049102.52f5c7e305a3.886.85
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab07217fdffdd042e7a95682d29c5c31f2248936ef812bffa701cd9f704a18f3
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630049572.6758537/events.out.tfevents.1630049572.52f5c7e305a3.886.87
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d3d43cf4345ea0f605d5516e05ca68007bcf57f75ba4819cc09f69f91395822
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630050046.2035077/events.out.tfevents.1630050046.52f5c7e305a3.886.89
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:709b11cfce27e86178a99bf7524ea96c43320f71e12f0b0b627dfbda8169b128
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630048138.52f5c7e305a3.886.80
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c9452f94cea99f7d845a7fce5fd6275f9cb7a4fc00b5141501e09163cefe720
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630048619.52f5c7e305a3.886.82
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc002d98a39a9740f6a94d265791543e732e073653e36f3504b9b7813d90f821
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630049102.52f5c7e305a3.886.84
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7bac776b58ce1ed85f4af1b4fa33016f75c4b5c31fe37a4fa301132e607e7928
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630049572.52f5c7e305a3.886.86
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:298c644a488ca1e5101e65db1c62c62a2a587a66960116c40947d0643f9cc208
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630050046.52f5c7e305a3.886.88
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:afc1ed3a9e82f43336874596bd326ec16c14ee316e9cf41c1c14df3165618864
|
3 |
+
size 8622
|