"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630109431.8143556/events.out.tfevents.1630109431.86bb0ddabf9b.1042.11 +3 -0
- model-bin/finetune/base/log/1630109864.1808882/events.out.tfevents.1630109864.86bb0ddabf9b.1042.13 +3 -0
- model-bin/finetune/base/log/1630110291.1942098/events.out.tfevents.1630110291.86bb0ddabf9b.1042.15 +3 -0
- model-bin/finetune/base/log/1630110708.1539922/events.out.tfevents.1630110708.86bb0ddabf9b.1042.17 +3 -0
- model-bin/finetune/base/log/1630111124.0089736/events.out.tfevents.1630111124.86bb0ddabf9b.1042.19 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630109431.86bb0ddabf9b.1042.10 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630109864.86bb0ddabf9b.1042.12 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630110291.86bb0ddabf9b.1042.14 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630110708.86bb0ddabf9b.1042.16 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630111124.86bb0ddabf9b.1042.18 +3 -0
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3bc213f8b1a88e0b2512f257ed2494755fd2795f55af3712d785a4772256d44
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:845b93ea6994b83196fa59050206c61d3f4fc9ef2aa5ba6e72e95c173579ac5a
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14567
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3c7fcf8524438e496a57b73e453b24b2519c6f58941a6fb8da174143dac50f8d
|
3 |
size 14567
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e28e0530f2551ac6de66de5fb95c14533593794a980009f90dec1f828fe4986
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6eac4ec54a649aef54843448fdae717a79bc1578521fca8e21301c3cddb0022
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -245187,11 +245187,800 @@
|
|
245187 |
"eval_steps_per_second": 0.677,
|
245188 |
"eval_wer": 0.18582472364713531,
|
245189 |
"step": 114613
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
245190 |
}
|
245191 |
],
|
245192 |
"max_steps": 620000,
|
245193 |
"num_train_epochs": 5000,
|
245194 |
-
"total_flos": 3.
|
245195 |
"trial_name": null,
|
245196 |
"trial_params": null
|
245197 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 928.995983935743,
|
5 |
+
"global_step": 115234,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
245187 |
"eval_steps_per_second": 0.677,
|
245188 |
"eval_wer": 0.18582472364713531,
|
245189 |
"step": 114613
|
245190 |
+
},
|
245191 |
+
{
|
245192 |
+
"epoch": 924.02,
|
245193 |
+
"learning_rate": 8.179983974358975e-06,
|
245194 |
+
"loss": 0.3841,
|
245195 |
+
"step": 114615
|
245196 |
+
},
|
245197 |
+
{
|
245198 |
+
"epoch": 924.06,
|
245199 |
+
"learning_rate": 8.179903846153847e-06,
|
245200 |
+
"loss": 0.3366,
|
245201 |
+
"step": 114620
|
245202 |
+
},
|
245203 |
+
{
|
245204 |
+
"epoch": 924.1,
|
245205 |
+
"learning_rate": 8.179823717948718e-06,
|
245206 |
+
"loss": 0.2559,
|
245207 |
+
"step": 114625
|
245208 |
+
},
|
245209 |
+
{
|
245210 |
+
"epoch": 924.14,
|
245211 |
+
"learning_rate": 8.179743589743591e-06,
|
245212 |
+
"loss": 0.3411,
|
245213 |
+
"step": 114630
|
245214 |
+
},
|
245215 |
+
{
|
245216 |
+
"epoch": 924.18,
|
245217 |
+
"learning_rate": 8.179663461538462e-06,
|
245218 |
+
"loss": 0.6367,
|
245219 |
+
"step": 114635
|
245220 |
+
},
|
245221 |
+
{
|
245222 |
+
"epoch": 924.22,
|
245223 |
+
"learning_rate": 8.179583333333334e-06,
|
245224 |
+
"loss": 1.0442,
|
245225 |
+
"step": 114640
|
245226 |
+
},
|
245227 |
+
{
|
245228 |
+
"epoch": 924.26,
|
245229 |
+
"learning_rate": 8.179503205128205e-06,
|
245230 |
+
"loss": 0.3043,
|
245231 |
+
"step": 114645
|
245232 |
+
},
|
245233 |
+
{
|
245234 |
+
"epoch": 924.3,
|
245235 |
+
"learning_rate": 8.179423076923078e-06,
|
245236 |
+
"loss": 0.2539,
|
245237 |
+
"step": 114650
|
245238 |
+
},
|
245239 |
+
{
|
245240 |
+
"epoch": 924.34,
|
245241 |
+
"learning_rate": 8.17934294871795e-06,
|
245242 |
+
"loss": 0.3802,
|
245243 |
+
"step": 114655
|
245244 |
+
},
|
245245 |
+
{
|
245246 |
+
"epoch": 924.38,
|
245247 |
+
"learning_rate": 8.179262820512821e-06,
|
245248 |
+
"loss": 0.6318,
|
245249 |
+
"step": 114660
|
245250 |
+
},
|
245251 |
+
{
|
245252 |
+
"epoch": 924.42,
|
245253 |
+
"learning_rate": 8.179182692307694e-06,
|
245254 |
+
"loss": 0.9742,
|
245255 |
+
"step": 114665
|
245256 |
+
},
|
245257 |
+
{
|
245258 |
+
"epoch": 924.46,
|
245259 |
+
"learning_rate": 8.179102564102565e-06,
|
245260 |
+
"loss": 0.2686,
|
245261 |
+
"step": 114670
|
245262 |
+
},
|
245263 |
+
{
|
245264 |
+
"epoch": 924.5,
|
245265 |
+
"learning_rate": 8.179022435897437e-06,
|
245266 |
+
"loss": 0.3525,
|
245267 |
+
"step": 114675
|
245268 |
+
},
|
245269 |
+
{
|
245270 |
+
"epoch": 924.54,
|
245271 |
+
"learning_rate": 8.178942307692308e-06,
|
245272 |
+
"loss": 0.3878,
|
245273 |
+
"step": 114680
|
245274 |
+
},
|
245275 |
+
{
|
245276 |
+
"epoch": 924.58,
|
245277 |
+
"learning_rate": 8.178862179487181e-06,
|
245278 |
+
"loss": 0.6801,
|
245279 |
+
"step": 114685
|
245280 |
+
},
|
245281 |
+
{
|
245282 |
+
"epoch": 924.62,
|
245283 |
+
"learning_rate": 8.178798076923078e-06,
|
245284 |
+
"loss": 0.976,
|
245285 |
+
"step": 114690
|
245286 |
+
},
|
245287 |
+
{
|
245288 |
+
"epoch": 924.66,
|
245289 |
+
"learning_rate": 8.17871794871795e-06,
|
245290 |
+
"loss": 0.2602,
|
245291 |
+
"step": 114695
|
245292 |
+
},
|
245293 |
+
{
|
245294 |
+
"epoch": 924.7,
|
245295 |
+
"learning_rate": 8.178637820512821e-06,
|
245296 |
+
"loss": 0.2754,
|
245297 |
+
"step": 114700
|
245298 |
+
},
|
245299 |
+
{
|
245300 |
+
"epoch": 924.74,
|
245301 |
+
"learning_rate": 8.178557692307693e-06,
|
245302 |
+
"loss": 0.3819,
|
245303 |
+
"step": 114705
|
245304 |
+
},
|
245305 |
+
{
|
245306 |
+
"epoch": 924.78,
|
245307 |
+
"learning_rate": 8.178477564102566e-06,
|
245308 |
+
"loss": 0.5863,
|
245309 |
+
"step": 114710
|
245310 |
+
},
|
245311 |
+
{
|
245312 |
+
"epoch": 924.82,
|
245313 |
+
"learning_rate": 8.178397435897435e-06,
|
245314 |
+
"loss": 1.1332,
|
245315 |
+
"step": 114715
|
245316 |
+
},
|
245317 |
+
{
|
245318 |
+
"epoch": 924.86,
|
245319 |
+
"learning_rate": 8.178317307692308e-06,
|
245320 |
+
"loss": 0.2884,
|
245321 |
+
"step": 114720
|
245322 |
+
},
|
245323 |
+
{
|
245324 |
+
"epoch": 924.9,
|
245325 |
+
"learning_rate": 8.178237179487181e-06,
|
245326 |
+
"loss": 0.2688,
|
245327 |
+
"step": 114725
|
245328 |
+
},
|
245329 |
+
{
|
245330 |
+
"epoch": 924.94,
|
245331 |
+
"learning_rate": 8.178157051282051e-06,
|
245332 |
+
"loss": 0.4118,
|
245333 |
+
"step": 114730
|
245334 |
+
},
|
245335 |
+
{
|
245336 |
+
"epoch": 924.98,
|
245337 |
+
"learning_rate": 8.178076923076924e-06,
|
245338 |
+
"loss": 0.608,
|
245339 |
+
"step": 114735
|
245340 |
+
},
|
245341 |
+
{
|
245342 |
+
"epoch": 925.0,
|
245343 |
+
"eval_loss": 0.3686845004558563,
|
245344 |
+
"eval_runtime": 40.4476,
|
245345 |
+
"eval_samples_per_second": 20.842,
|
245346 |
+
"eval_steps_per_second": 0.668,
|
245347 |
+
"eval_wer": 0.18588031473327077,
|
245348 |
+
"step": 114737
|
245349 |
+
},
|
245350 |
+
{
|
245351 |
+
"epoch": 925.02,
|
245352 |
+
"learning_rate": 8.177996794871795e-06,
|
245353 |
+
"loss": 0.4001,
|
245354 |
+
"step": 114740
|
245355 |
+
},
|
245356 |
+
{
|
245357 |
+
"epoch": 925.06,
|
245358 |
+
"learning_rate": 8.177916666666667e-06,
|
245359 |
+
"loss": 0.2602,
|
245360 |
+
"step": 114745
|
245361 |
+
},
|
245362 |
+
{
|
245363 |
+
"epoch": 925.1,
|
245364 |
+
"learning_rate": 8.177836538461538e-06,
|
245365 |
+
"loss": 0.3255,
|
245366 |
+
"step": 114750
|
245367 |
+
},
|
245368 |
+
{
|
245369 |
+
"epoch": 925.14,
|
245370 |
+
"learning_rate": 8.177756410256411e-06,
|
245371 |
+
"loss": 0.3547,
|
245372 |
+
"step": 114755
|
245373 |
+
},
|
245374 |
+
{
|
245375 |
+
"epoch": 925.18,
|
245376 |
+
"learning_rate": 8.177676282051283e-06,
|
245377 |
+
"loss": 0.7224,
|
245378 |
+
"step": 114760
|
245379 |
+
},
|
245380 |
+
{
|
245381 |
+
"epoch": 925.22,
|
245382 |
+
"learning_rate": 8.177596153846154e-06,
|
245383 |
+
"loss": 0.8328,
|
245384 |
+
"step": 114765
|
245385 |
+
},
|
245386 |
+
{
|
245387 |
+
"epoch": 925.27,
|
245388 |
+
"learning_rate": 8.177516025641025e-06,
|
245389 |
+
"loss": 0.2706,
|
245390 |
+
"step": 114770
|
245391 |
+
},
|
245392 |
+
{
|
245393 |
+
"epoch": 925.31,
|
245394 |
+
"learning_rate": 8.177435897435898e-06,
|
245395 |
+
"loss": 0.2727,
|
245396 |
+
"step": 114775
|
245397 |
+
},
|
245398 |
+
{
|
245399 |
+
"epoch": 925.35,
|
245400 |
+
"learning_rate": 8.17735576923077e-06,
|
245401 |
+
"loss": 0.3431,
|
245402 |
+
"step": 114780
|
245403 |
+
},
|
245404 |
+
{
|
245405 |
+
"epoch": 925.39,
|
245406 |
+
"learning_rate": 8.177275641025641e-06,
|
245407 |
+
"loss": 0.7878,
|
245408 |
+
"step": 114785
|
245409 |
+
},
|
245410 |
+
{
|
245411 |
+
"epoch": 925.43,
|
245412 |
+
"learning_rate": 8.177195512820514e-06,
|
245413 |
+
"loss": 1.0417,
|
245414 |
+
"step": 114790
|
245415 |
+
},
|
245416 |
+
{
|
245417 |
+
"epoch": 925.47,
|
245418 |
+
"learning_rate": 8.177115384615385e-06,
|
245419 |
+
"loss": 0.3444,
|
245420 |
+
"step": 114795
|
245421 |
+
},
|
245422 |
+
{
|
245423 |
+
"epoch": 925.51,
|
245424 |
+
"learning_rate": 8.177035256410257e-06,
|
245425 |
+
"loss": 0.2683,
|
245426 |
+
"step": 114800
|
245427 |
+
},
|
245428 |
+
{
|
245429 |
+
"epoch": 925.55,
|
245430 |
+
"learning_rate": 8.176955128205128e-06,
|
245431 |
+
"loss": 0.3586,
|
245432 |
+
"step": 114805
|
245433 |
+
},
|
245434 |
+
{
|
245435 |
+
"epoch": 925.59,
|
245436 |
+
"learning_rate": 8.176875000000001e-06,
|
245437 |
+
"loss": 0.7673,
|
245438 |
+
"step": 114810
|
245439 |
+
},
|
245440 |
+
{
|
245441 |
+
"epoch": 925.63,
|
245442 |
+
"learning_rate": 8.176794871794873e-06,
|
245443 |
+
"loss": 0.8954,
|
245444 |
+
"step": 114815
|
245445 |
+
},
|
245446 |
+
{
|
245447 |
+
"epoch": 925.67,
|
245448 |
+
"learning_rate": 8.176714743589744e-06,
|
245449 |
+
"loss": 0.3164,
|
245450 |
+
"step": 114820
|
245451 |
+
},
|
245452 |
+
{
|
245453 |
+
"epoch": 925.71,
|
245454 |
+
"learning_rate": 8.176634615384617e-06,
|
245455 |
+
"loss": 0.3331,
|
245456 |
+
"step": 114825
|
245457 |
+
},
|
245458 |
+
{
|
245459 |
+
"epoch": 925.75,
|
245460 |
+
"learning_rate": 8.176554487179488e-06,
|
245461 |
+
"loss": 0.3298,
|
245462 |
+
"step": 114830
|
245463 |
+
},
|
245464 |
+
{
|
245465 |
+
"epoch": 925.79,
|
245466 |
+
"learning_rate": 8.17647435897436e-06,
|
245467 |
+
"loss": 0.6546,
|
245468 |
+
"step": 114835
|
245469 |
+
},
|
245470 |
+
{
|
245471 |
+
"epoch": 925.83,
|
245472 |
+
"learning_rate": 8.176394230769231e-06,
|
245473 |
+
"loss": 0.8006,
|
245474 |
+
"step": 114840
|
245475 |
+
},
|
245476 |
+
{
|
245477 |
+
"epoch": 925.87,
|
245478 |
+
"learning_rate": 8.176314102564104e-06,
|
245479 |
+
"loss": 0.3202,
|
245480 |
+
"step": 114845
|
245481 |
+
},
|
245482 |
+
{
|
245483 |
+
"epoch": 925.91,
|
245484 |
+
"learning_rate": 8.176233974358975e-06,
|
245485 |
+
"loss": 0.3295,
|
245486 |
+
"step": 114850
|
245487 |
+
},
|
245488 |
+
{
|
245489 |
+
"epoch": 925.95,
|
245490 |
+
"learning_rate": 8.176153846153847e-06,
|
245491 |
+
"loss": 0.3493,
|
245492 |
+
"step": 114855
|
245493 |
+
},
|
245494 |
+
{
|
245495 |
+
"epoch": 925.99,
|
245496 |
+
"learning_rate": 8.176073717948718e-06,
|
245497 |
+
"loss": 0.7909,
|
245498 |
+
"step": 114860
|
245499 |
+
},
|
245500 |
+
{
|
245501 |
+
"epoch": 926.0,
|
245502 |
+
"eval_loss": 0.38164782524108887,
|
245503 |
+
"eval_runtime": 38.7323,
|
245504 |
+
"eval_samples_per_second": 21.765,
|
245505 |
+
"eval_steps_per_second": 0.697,
|
245506 |
+
"eval_wer": 0.180527833222444,
|
245507 |
+
"step": 114861
|
245508 |
+
},
|
245509 |
+
{
|
245510 |
+
"epoch": 926.03,
|
245511 |
+
"learning_rate": 8.175993589743591e-06,
|
245512 |
+
"loss": 0.3478,
|
245513 |
+
"step": 114865
|
245514 |
+
},
|
245515 |
+
{
|
245516 |
+
"epoch": 926.07,
|
245517 |
+
"learning_rate": 8.175913461538461e-06,
|
245518 |
+
"loss": 0.3354,
|
245519 |
+
"step": 114870
|
245520 |
+
},
|
245521 |
+
{
|
245522 |
+
"epoch": 926.11,
|
245523 |
+
"learning_rate": 8.175833333333334e-06,
|
245524 |
+
"loss": 0.2706,
|
245525 |
+
"step": 114875
|
245526 |
+
},
|
245527 |
+
{
|
245528 |
+
"epoch": 926.15,
|
245529 |
+
"learning_rate": 8.175753205128207e-06,
|
245530 |
+
"loss": 0.3976,
|
245531 |
+
"step": 114880
|
245532 |
+
},
|
245533 |
+
{
|
245534 |
+
"epoch": 926.19,
|
245535 |
+
"learning_rate": 8.175673076923077e-06,
|
245536 |
+
"loss": 0.9176,
|
245537 |
+
"step": 114885
|
245538 |
+
},
|
245539 |
+
{
|
245540 |
+
"epoch": 926.23,
|
245541 |
+
"learning_rate": 8.17559294871795e-06,
|
245542 |
+
"loss": 0.6236,
|
245543 |
+
"step": 114890
|
245544 |
+
},
|
245545 |
+
{
|
245546 |
+
"epoch": 926.27,
|
245547 |
+
"learning_rate": 8.175512820512821e-06,
|
245548 |
+
"loss": 0.2932,
|
245549 |
+
"step": 114895
|
245550 |
+
},
|
245551 |
+
{
|
245552 |
+
"epoch": 926.31,
|
245553 |
+
"learning_rate": 8.175432692307692e-06,
|
245554 |
+
"loss": 0.3139,
|
245555 |
+
"step": 114900
|
245556 |
+
},
|
245557 |
+
{
|
245558 |
+
"epoch": 926.35,
|
245559 |
+
"learning_rate": 8.175352564102564e-06,
|
245560 |
+
"loss": 0.4103,
|
245561 |
+
"step": 114905
|
245562 |
+
},
|
245563 |
+
{
|
245564 |
+
"epoch": 926.39,
|
245565 |
+
"learning_rate": 8.175272435897437e-06,
|
245566 |
+
"loss": 0.8761,
|
245567 |
+
"step": 114910
|
245568 |
+
},
|
245569 |
+
{
|
245570 |
+
"epoch": 926.43,
|
245571 |
+
"learning_rate": 8.175192307692308e-06,
|
245572 |
+
"loss": 0.6826,
|
245573 |
+
"step": 114915
|
245574 |
+
},
|
245575 |
+
{
|
245576 |
+
"epoch": 926.47,
|
245577 |
+
"learning_rate": 8.17511217948718e-06,
|
245578 |
+
"loss": 0.268,
|
245579 |
+
"step": 114920
|
245580 |
+
},
|
245581 |
+
{
|
245582 |
+
"epoch": 926.51,
|
245583 |
+
"learning_rate": 8.175032051282053e-06,
|
245584 |
+
"loss": 0.2962,
|
245585 |
+
"step": 114925
|
245586 |
+
},
|
245587 |
+
{
|
245588 |
+
"epoch": 926.55,
|
245589 |
+
"learning_rate": 8.174951923076924e-06,
|
245590 |
+
"loss": 0.4022,
|
245591 |
+
"step": 114930
|
245592 |
+
},
|
245593 |
+
{
|
245594 |
+
"epoch": 926.59,
|
245595 |
+
"learning_rate": 8.174871794871795e-06,
|
245596 |
+
"loss": 0.8668,
|
245597 |
+
"step": 114935
|
245598 |
+
},
|
245599 |
+
{
|
245600 |
+
"epoch": 926.63,
|
245601 |
+
"learning_rate": 8.174791666666667e-06,
|
245602 |
+
"loss": 0.6838,
|
245603 |
+
"step": 114940
|
245604 |
+
},
|
245605 |
+
{
|
245606 |
+
"epoch": 926.67,
|
245607 |
+
"learning_rate": 8.17471153846154e-06,
|
245608 |
+
"loss": 0.2747,
|
245609 |
+
"step": 114945
|
245610 |
+
},
|
245611 |
+
{
|
245612 |
+
"epoch": 926.71,
|
245613 |
+
"learning_rate": 8.174631410256411e-06,
|
245614 |
+
"loss": 0.364,
|
245615 |
+
"step": 114950
|
245616 |
+
},
|
245617 |
+
{
|
245618 |
+
"epoch": 926.76,
|
245619 |
+
"learning_rate": 8.174551282051282e-06,
|
245620 |
+
"loss": 0.3743,
|
245621 |
+
"step": 114955
|
245622 |
+
},
|
245623 |
+
{
|
245624 |
+
"epoch": 926.8,
|
245625 |
+
"learning_rate": 8.174471153846154e-06,
|
245626 |
+
"loss": 0.9189,
|
245627 |
+
"step": 114960
|
245628 |
+
},
|
245629 |
+
{
|
245630 |
+
"epoch": 926.84,
|
245631 |
+
"learning_rate": 8.174391025641027e-06,
|
245632 |
+
"loss": 0.6867,
|
245633 |
+
"step": 114965
|
245634 |
+
},
|
245635 |
+
{
|
245636 |
+
"epoch": 926.88,
|
245637 |
+
"learning_rate": 8.174310897435898e-06,
|
245638 |
+
"loss": 0.3038,
|
245639 |
+
"step": 114970
|
245640 |
+
},
|
245641 |
+
{
|
245642 |
+
"epoch": 926.92,
|
245643 |
+
"learning_rate": 8.17423076923077e-06,
|
245644 |
+
"loss": 0.2775,
|
245645 |
+
"step": 114975
|
245646 |
+
},
|
245647 |
+
{
|
245648 |
+
"epoch": 926.96,
|
245649 |
+
"learning_rate": 8.174150641025643e-06,
|
245650 |
+
"loss": 0.4685,
|
245651 |
+
"step": 114980
|
245652 |
+
},
|
245653 |
+
{
|
245654 |
+
"epoch": 927.0,
|
245655 |
+
"learning_rate": 8.174070512820514e-06,
|
245656 |
+
"loss": 1.0971,
|
245657 |
+
"step": 114985
|
245658 |
+
},
|
245659 |
+
{
|
245660 |
+
"epoch": 927.0,
|
245661 |
+
"eval_loss": 0.3635014295578003,
|
245662 |
+
"eval_runtime": 39.7417,
|
245663 |
+
"eval_samples_per_second": 21.212,
|
245664 |
+
"eval_steps_per_second": 0.679,
|
245665 |
+
"eval_wer": 0.18649314104818854,
|
245666 |
+
"step": 114985
|
245667 |
+
},
|
245668 |
+
{
|
245669 |
+
"epoch": 919.04,
|
245670 |
+
"learning_rate": 8.173990384615385e-06,
|
245671 |
+
"loss": 0.302,
|
245672 |
+
"step": 114990
|
245673 |
+
},
|
245674 |
+
{
|
245675 |
+
"epoch": 919.08,
|
245676 |
+
"learning_rate": 8.173910256410257e-06,
|
245677 |
+
"loss": 0.3467,
|
245678 |
+
"step": 114995
|
245679 |
+
},
|
245680 |
+
{
|
245681 |
+
"epoch": 919.12,
|
245682 |
+
"learning_rate": 8.17383012820513e-06,
|
245683 |
+
"loss": 0.339,
|
245684 |
+
"step": 115000
|
245685 |
+
},
|
245686 |
+
{
|
245687 |
+
"epoch": 919.16,
|
245688 |
+
"learning_rate": 8.17375e-06,
|
245689 |
+
"loss": 0.4937,
|
245690 |
+
"step": 115005
|
245691 |
+
},
|
245692 |
+
{
|
245693 |
+
"epoch": 919.2,
|
245694 |
+
"learning_rate": 8.173669871794873e-06,
|
245695 |
+
"loss": 1.1532,
|
245696 |
+
"step": 115010
|
245697 |
+
},
|
245698 |
+
{
|
245699 |
+
"epoch": 919.24,
|
245700 |
+
"learning_rate": 8.173589743589744e-06,
|
245701 |
+
"loss": 0.3483,
|
245702 |
+
"step": 115015
|
245703 |
+
},
|
245704 |
+
{
|
245705 |
+
"epoch": 919.28,
|
245706 |
+
"learning_rate": 8.173509615384615e-06,
|
245707 |
+
"loss": 0.3049,
|
245708 |
+
"step": 115020
|
245709 |
+
},
|
245710 |
+
{
|
245711 |
+
"epoch": 919.32,
|
245712 |
+
"learning_rate": 8.173429487179488e-06,
|
245713 |
+
"loss": 0.3562,
|
245714 |
+
"step": 115025
|
245715 |
+
},
|
245716 |
+
{
|
245717 |
+
"epoch": 919.36,
|
245718 |
+
"learning_rate": 8.17334935897436e-06,
|
245719 |
+
"loss": 0.4262,
|
245720 |
+
"step": 115030
|
245721 |
+
},
|
245722 |
+
{
|
245723 |
+
"epoch": 919.4,
|
245724 |
+
"learning_rate": 8.173269230769233e-06,
|
245725 |
+
"loss": 1.2422,
|
245726 |
+
"step": 115035
|
245727 |
+
},
|
245728 |
+
{
|
245729 |
+
"epoch": 919.44,
|
245730 |
+
"learning_rate": 8.173189102564102e-06,
|
245731 |
+
"loss": 0.341,
|
245732 |
+
"step": 115040
|
245733 |
+
},
|
245734 |
+
{
|
245735 |
+
"epoch": 919.48,
|
245736 |
+
"learning_rate": 8.173108974358975e-06,
|
245737 |
+
"loss": 0.2614,
|
245738 |
+
"step": 115045
|
245739 |
+
},
|
245740 |
+
{
|
245741 |
+
"epoch": 919.52,
|
245742 |
+
"learning_rate": 8.173028846153847e-06,
|
245743 |
+
"loss": 0.2816,
|
245744 |
+
"step": 115050
|
245745 |
+
},
|
245746 |
+
{
|
245747 |
+
"epoch": 919.56,
|
245748 |
+
"learning_rate": 8.172948717948718e-06,
|
245749 |
+
"loss": 0.3867,
|
245750 |
+
"step": 115055
|
245751 |
+
},
|
245752 |
+
{
|
245753 |
+
"epoch": 919.6,
|
245754 |
+
"learning_rate": 8.17286858974359e-06,
|
245755 |
+
"loss": 1.1399,
|
245756 |
+
"step": 115060
|
245757 |
+
},
|
245758 |
+
{
|
245759 |
+
"epoch": 919.64,
|
245760 |
+
"learning_rate": 8.172788461538463e-06,
|
245761 |
+
"loss": 0.278,
|
245762 |
+
"step": 115065
|
245763 |
+
},
|
245764 |
+
{
|
245765 |
+
"epoch": 919.68,
|
245766 |
+
"learning_rate": 8.172708333333334e-06,
|
245767 |
+
"loss": 0.2704,
|
245768 |
+
"step": 115070
|
245769 |
+
},
|
245770 |
+
{
|
245771 |
+
"epoch": 919.72,
|
245772 |
+
"learning_rate": 8.172628205128205e-06,
|
245773 |
+
"loss": 0.2866,
|
245774 |
+
"step": 115075
|
245775 |
+
},
|
245776 |
+
{
|
245777 |
+
"epoch": 919.76,
|
245778 |
+
"learning_rate": 8.172548076923078e-06,
|
245779 |
+
"loss": 0.4154,
|
245780 |
+
"step": 115080
|
245781 |
+
},
|
245782 |
+
{
|
245783 |
+
"epoch": 919.8,
|
245784 |
+
"learning_rate": 8.17246794871795e-06,
|
245785 |
+
"loss": 1.0377,
|
245786 |
+
"step": 115085
|
245787 |
+
},
|
245788 |
+
{
|
245789 |
+
"epoch": 919.84,
|
245790 |
+
"learning_rate": 8.172387820512821e-06,
|
245791 |
+
"loss": 0.3569,
|
245792 |
+
"step": 115090
|
245793 |
+
},
|
245794 |
+
{
|
245795 |
+
"epoch": 919.88,
|
245796 |
+
"learning_rate": 8.172307692307692e-06,
|
245797 |
+
"loss": 0.2438,
|
245798 |
+
"step": 115095
|
245799 |
+
},
|
245800 |
+
{
|
245801 |
+
"epoch": 919.92,
|
245802 |
+
"learning_rate": 8.172227564102565e-06,
|
245803 |
+
"loss": 0.3913,
|
245804 |
+
"step": 115100
|
245805 |
+
},
|
245806 |
+
{
|
245807 |
+
"epoch": 919.96,
|
245808 |
+
"learning_rate": 8.172147435897437e-06,
|
245809 |
+
"loss": 0.4283,
|
245810 |
+
"step": 115105
|
245811 |
+
},
|
245812 |
+
{
|
245813 |
+
"epoch": 920.0,
|
245814 |
+
"learning_rate": 8.172067307692308e-06,
|
245815 |
+
"loss": 1.2519,
|
245816 |
+
"step": 115110
|
245817 |
+
},
|
245818 |
+
{
|
245819 |
+
"epoch": 920.0,
|
245820 |
+
"eval_loss": 0.36768418550491333,
|
245821 |
+
"eval_runtime": 39.8638,
|
245822 |
+
"eval_samples_per_second": 21.147,
|
245823 |
+
"eval_steps_per_second": 0.677,
|
245824 |
+
"eval_wer": 0.18435032304379037,
|
245825 |
+
"step": 115110
|
245826 |
+
},
|
245827 |
+
{
|
245828 |
+
"epoch": 928.04,
|
245829 |
+
"learning_rate": 8.17198717948718e-06,
|
245830 |
+
"loss": 0.3685,
|
245831 |
+
"step": 115115
|
245832 |
+
},
|
245833 |
+
{
|
245834 |
+
"epoch": 928.08,
|
245835 |
+
"learning_rate": 8.171907051282053e-06,
|
245836 |
+
"loss": 0.3035,
|
245837 |
+
"step": 115120
|
245838 |
+
},
|
245839 |
+
{
|
245840 |
+
"epoch": 928.12,
|
245841 |
+
"learning_rate": 8.171826923076924e-06,
|
245842 |
+
"loss": 0.3222,
|
245843 |
+
"step": 115125
|
245844 |
+
},
|
245845 |
+
{
|
245846 |
+
"epoch": 928.16,
|
245847 |
+
"learning_rate": 8.171746794871795e-06,
|
245848 |
+
"loss": 0.4358,
|
245849 |
+
"step": 115130
|
245850 |
+
},
|
245851 |
+
{
|
245852 |
+
"epoch": 928.2,
|
245853 |
+
"learning_rate": 8.171666666666668e-06,
|
245854 |
+
"loss": 1.1252,
|
245855 |
+
"step": 115135
|
245856 |
+
},
|
245857 |
+
{
|
245858 |
+
"epoch": 928.24,
|
245859 |
+
"learning_rate": 8.17158653846154e-06,
|
245860 |
+
"loss": 0.3178,
|
245861 |
+
"step": 115140
|
245862 |
+
},
|
245863 |
+
{
|
245864 |
+
"epoch": 928.28,
|
245865 |
+
"learning_rate": 8.171506410256411e-06,
|
245866 |
+
"loss": 0.3091,
|
245867 |
+
"step": 115145
|
245868 |
+
},
|
245869 |
+
{
|
245870 |
+
"epoch": 928.32,
|
245871 |
+
"learning_rate": 8.171426282051282e-06,
|
245872 |
+
"loss": 0.3167,
|
245873 |
+
"step": 115150
|
245874 |
+
},
|
245875 |
+
{
|
245876 |
+
"epoch": 928.36,
|
245877 |
+
"learning_rate": 8.171346153846155e-06,
|
245878 |
+
"loss": 0.4127,
|
245879 |
+
"step": 115155
|
245880 |
+
},
|
245881 |
+
{
|
245882 |
+
"epoch": 928.4,
|
245883 |
+
"learning_rate": 8.171266025641025e-06,
|
245884 |
+
"loss": 1.0843,
|
245885 |
+
"step": 115160
|
245886 |
+
},
|
245887 |
+
{
|
245888 |
+
"epoch": 928.44,
|
245889 |
+
"learning_rate": 8.171185897435898e-06,
|
245890 |
+
"loss": 0.3319,
|
245891 |
+
"step": 115165
|
245892 |
+
},
|
245893 |
+
{
|
245894 |
+
"epoch": 928.48,
|
245895 |
+
"learning_rate": 8.17110576923077e-06,
|
245896 |
+
"loss": 0.2568,
|
245897 |
+
"step": 115170
|
245898 |
+
},
|
245899 |
+
{
|
245900 |
+
"epoch": 928.52,
|
245901 |
+
"learning_rate": 8.171025641025641e-06,
|
245902 |
+
"loss": 0.3007,
|
245903 |
+
"step": 115175
|
245904 |
+
},
|
245905 |
+
{
|
245906 |
+
"epoch": 928.56,
|
245907 |
+
"learning_rate": 8.170945512820514e-06,
|
245908 |
+
"loss": 0.4638,
|
245909 |
+
"step": 115180
|
245910 |
+
},
|
245911 |
+
{
|
245912 |
+
"epoch": 928.6,
|
245913 |
+
"learning_rate": 8.170865384615385e-06,
|
245914 |
+
"loss": 1.3013,
|
245915 |
+
"step": 115185
|
245916 |
+
},
|
245917 |
+
{
|
245918 |
+
"epoch": 928.64,
|
245919 |
+
"learning_rate": 8.170785256410257e-06,
|
245920 |
+
"loss": 0.3193,
|
245921 |
+
"step": 115190
|
245922 |
+
},
|
245923 |
+
{
|
245924 |
+
"epoch": 928.68,
|
245925 |
+
"learning_rate": 8.170705128205128e-06,
|
245926 |
+
"loss": 0.273,
|
245927 |
+
"step": 115195
|
245928 |
+
},
|
245929 |
+
{
|
245930 |
+
"epoch": 928.72,
|
245931 |
+
"learning_rate": 8.170625000000001e-06,
|
245932 |
+
"loss": 0.3355,
|
245933 |
+
"step": 115200
|
245934 |
+
},
|
245935 |
+
{
|
245936 |
+
"epoch": 928.76,
|
245937 |
+
"learning_rate": 8.170544871794872e-06,
|
245938 |
+
"loss": 0.4521,
|
245939 |
+
"step": 115205
|
245940 |
+
},
|
245941 |
+
{
|
245942 |
+
"epoch": 928.8,
|
245943 |
+
"learning_rate": 8.170464743589744e-06,
|
245944 |
+
"loss": 1.1896,
|
245945 |
+
"step": 115210
|
245946 |
+
},
|
245947 |
+
{
|
245948 |
+
"epoch": 928.84,
|
245949 |
+
"learning_rate": 8.170384615384615e-06,
|
245950 |
+
"loss": 0.2887,
|
245951 |
+
"step": 115215
|
245952 |
+
},
|
245953 |
+
{
|
245954 |
+
"epoch": 928.88,
|
245955 |
+
"learning_rate": 8.170304487179488e-06,
|
245956 |
+
"loss": 0.27,
|
245957 |
+
"step": 115220
|
245958 |
+
},
|
245959 |
+
{
|
245960 |
+
"epoch": 928.92,
|
245961 |
+
"learning_rate": 8.17022435897436e-06,
|
245962 |
+
"loss": 0.3528,
|
245963 |
+
"step": 115225
|
245964 |
+
},
|
245965 |
+
{
|
245966 |
+
"epoch": 928.96,
|
245967 |
+
"learning_rate": 8.170144230769231e-06,
|
245968 |
+
"loss": 0.4421,
|
245969 |
+
"step": 115230
|
245970 |
+
},
|
245971 |
+
{
|
245972 |
+
"epoch": 929.0,
|
245973 |
+
"eval_loss": 0.38033702969551086,
|
245974 |
+
"eval_runtime": 39.6678,
|
245975 |
+
"eval_samples_per_second": 21.277,
|
245976 |
+
"eval_steps_per_second": 0.681,
|
245977 |
+
"eval_wer": 0.18934867596075827,
|
245978 |
+
"step": 115234
|
245979 |
}
|
245980 |
],
|
245981 |
"max_steps": 620000,
|
245982 |
"num_train_epochs": 5000,
|
245983 |
+
"total_flos": 3.243039409882843e+20,
|
245984 |
"trial_name": null,
|
245985 |
"trial_params": null
|
245986 |
}
|
model-bin/finetune/base/{checkpoint-114613 β checkpoint-115234}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630109431.8143556/events.out.tfevents.1630109431.86bb0ddabf9b.1042.11
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1752755bd47a1298986b325a70db8b95c0108718e4f7caea9973b49ff20223e
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630109864.1808882/events.out.tfevents.1630109864.86bb0ddabf9b.1042.13
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:492fcb0f04e16405c50032f9b9f601fbf6ccb916bc8c6f9c82b6cbf6072c2f51
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630110291.1942098/events.out.tfevents.1630110291.86bb0ddabf9b.1042.15
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9678335599331e6c8a5e95ac8ae114b537ca30f5aa61f90121c6ad0909316186
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630110708.1539922/events.out.tfevents.1630110708.86bb0ddabf9b.1042.17
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8789ebd4a0d049bcc6e7a6cfb131f3de072740ef62692b320e8d3b2457ea1f9
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630111124.0089736/events.out.tfevents.1630111124.86bb0ddabf9b.1042.19
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50220645914c1340a01a0938428ab7312dd5ae9ee446dcb2cf3f29d82df09f88
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630109431.86bb0ddabf9b.1042.10
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2f69acf780d6b84f0085f54961d3041b1ec46c0d7af1f152bfabb4b960ce1e9
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630109864.86bb0ddabf9b.1042.12
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dbf005686af30527fa466f8f955ccb86be67ba4c19c6e2a7a603c53d9ffe89c7
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630110291.86bb0ddabf9b.1042.14
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b0f307121ff5f987c2125b3494b029ead7320a9107cbbe1f662acbfb27ad23fa
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630110708.86bb0ddabf9b.1042.16
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b8c9cc1c211c48db928e1b7c27c884da18dd9b5870bc8b09f9fbb05f683bdcf5
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630111124.86bb0ddabf9b.1042.18
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b51ed23240af71ded7356cae20da843c27a40e6f060bd5a294754f7398910155
|
3 |
+
size 8462
|