"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630045620.5655968/events.out.tfevents.1630045620.52f5c7e305a3.886.71 +3 -0
- model-bin/finetune/base/log/1630046118.6803894/events.out.tfevents.1630046118.52f5c7e305a3.886.73 +3 -0
- model-bin/finetune/base/log/1630046681.8092928/events.out.tfevents.1630046681.52f5c7e305a3.886.75 +3 -0
- model-bin/finetune/base/log/1630047172.2078328/events.out.tfevents.1630047172.52f5c7e305a3.886.77 +3 -0
- model-bin/finetune/base/log/1630047639.1210005/events.out.tfevents.1630047639.52f5c7e305a3.886.79 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630045620.52f5c7e305a3.886.70 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630046118.52f5c7e305a3.886.72 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630046681.52f5c7e305a3.886.74 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630047172.52f5c7e305a3.886.76 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630047639.52f5c7e305a3.886.78 +3 -0
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1b8197993eb37d9ee36acbb12a1f9605ec9f468a36950f0b19cfe128de2b5f0
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8232bc1150a76590269afa478f75b4236d9220bb172b83e409418781d8efeebd
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a14f61f7dddaa53dbcb08f44b28b663a43e59c31dc4d069de7af3154f39d5215
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f60fd12782841de6f909cf814170025073e913f168ae702ff0cdd93272700668
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e9a559e52c3b4b08e9232cdf54f10d4fa2d1e564b84998252aef70d42eaf820
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -240282,11 +240282,800 @@
|
|
240282 |
"eval_steps_per_second": 0.631,
|
240283 |
"eval_wer": 0.19120699881376038,
|
240284 |
"step": 110758
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
240285 |
}
|
240286 |
],
|
240287 |
"max_steps": 620000,
|
240288 |
"num_train_epochs": 5000,
|
240289 |
-
"total_flos": 3.
|
240290 |
"trial_name": null,
|
240291 |
"trial_params": null
|
240292 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1743826049391605,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
4 |
+
"epoch": 898.0,
|
5 |
+
"global_step": 111379,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
240282 |
"eval_steps_per_second": 0.631,
|
240283 |
"eval_wer": 0.19120699881376038,
|
240284 |
"step": 110758
|
240285 |
+
},
|
240286 |
+
{
|
240287 |
+
"epoch": 900.02,
|
240288 |
+
"learning_rate": 8.227528271405494e-06,
|
240289 |
+
"loss": 0.3431,
|
240290 |
+
"step": 110760
|
240291 |
+
},
|
240292 |
+
{
|
240293 |
+
"epoch": 900.06,
|
240294 |
+
"learning_rate": 8.227447495961228e-06,
|
240295 |
+
"loss": 0.2264,
|
240296 |
+
"step": 110765
|
240297 |
+
},
|
240298 |
+
{
|
240299 |
+
"epoch": 900.1,
|
240300 |
+
"learning_rate": 8.227366720516964e-06,
|
240301 |
+
"loss": 0.3079,
|
240302 |
+
"step": 110770
|
240303 |
+
},
|
240304 |
+
{
|
240305 |
+
"epoch": 900.14,
|
240306 |
+
"learning_rate": 8.227285945072698e-06,
|
240307 |
+
"loss": 0.4818,
|
240308 |
+
"step": 110775
|
240309 |
+
},
|
240310 |
+
{
|
240311 |
+
"epoch": 900.18,
|
240312 |
+
"learning_rate": 8.227205169628434e-06,
|
240313 |
+
"loss": 0.7211,
|
240314 |
+
"step": 110780
|
240315 |
+
},
|
240316 |
+
{
|
240317 |
+
"epoch": 900.22,
|
240318 |
+
"learning_rate": 8.227124394184168e-06,
|
240319 |
+
"loss": 1.0533,
|
240320 |
+
"step": 110785
|
240321 |
+
},
|
240322 |
+
{
|
240323 |
+
"epoch": 900.26,
|
240324 |
+
"learning_rate": 8.227043618739904e-06,
|
240325 |
+
"loss": 0.3337,
|
240326 |
+
"step": 110790
|
240327 |
+
},
|
240328 |
+
{
|
240329 |
+
"epoch": 900.3,
|
240330 |
+
"learning_rate": 8.226962843295638e-06,
|
240331 |
+
"loss": 0.3322,
|
240332 |
+
"step": 110795
|
240333 |
+
},
|
240334 |
+
{
|
240335 |
+
"epoch": 900.34,
|
240336 |
+
"learning_rate": 8.226882067851374e-06,
|
240337 |
+
"loss": 0.3738,
|
240338 |
+
"step": 110800
|
240339 |
+
},
|
240340 |
+
{
|
240341 |
+
"epoch": 900.38,
|
240342 |
+
"learning_rate": 8.22680129240711e-06,
|
240343 |
+
"loss": 0.6992,
|
240344 |
+
"step": 110805
|
240345 |
+
},
|
240346 |
+
{
|
240347 |
+
"epoch": 900.42,
|
240348 |
+
"learning_rate": 8.226720516962844e-06,
|
240349 |
+
"loss": 0.9494,
|
240350 |
+
"step": 110810
|
240351 |
+
},
|
240352 |
+
{
|
240353 |
+
"epoch": 900.46,
|
240354 |
+
"learning_rate": 8.22663974151858e-06,
|
240355 |
+
"loss": 0.2978,
|
240356 |
+
"step": 110815
|
240357 |
+
},
|
240358 |
+
{
|
240359 |
+
"epoch": 900.5,
|
240360 |
+
"learning_rate": 8.226558966074314e-06,
|
240361 |
+
"loss": 0.3104,
|
240362 |
+
"step": 110820
|
240363 |
+
},
|
240364 |
+
{
|
240365 |
+
"epoch": 900.54,
|
240366 |
+
"learning_rate": 8.22647819063005e-06,
|
240367 |
+
"loss": 0.3612,
|
240368 |
+
"step": 110825
|
240369 |
+
},
|
240370 |
+
{
|
240371 |
+
"epoch": 900.58,
|
240372 |
+
"learning_rate": 8.226397415185784e-06,
|
240373 |
+
"loss": 0.6377,
|
240374 |
+
"step": 110830
|
240375 |
+
},
|
240376 |
+
{
|
240377 |
+
"epoch": 900.62,
|
240378 |
+
"learning_rate": 8.22631663974152e-06,
|
240379 |
+
"loss": 0.8597,
|
240380 |
+
"step": 110835
|
240381 |
+
},
|
240382 |
+
{
|
240383 |
+
"epoch": 900.66,
|
240384 |
+
"learning_rate": 8.226235864297254e-06,
|
240385 |
+
"loss": 0.3028,
|
240386 |
+
"step": 110840
|
240387 |
+
},
|
240388 |
+
{
|
240389 |
+
"epoch": 900.7,
|
240390 |
+
"learning_rate": 8.22615508885299e-06,
|
240391 |
+
"loss": 0.3288,
|
240392 |
+
"step": 110845
|
240393 |
+
},
|
240394 |
+
{
|
240395 |
+
"epoch": 900.74,
|
240396 |
+
"learning_rate": 8.226074313408724e-06,
|
240397 |
+
"loss": 0.3319,
|
240398 |
+
"step": 110850
|
240399 |
+
},
|
240400 |
+
{
|
240401 |
+
"epoch": 900.79,
|
240402 |
+
"learning_rate": 8.22599353796446e-06,
|
240403 |
+
"loss": 0.5342,
|
240404 |
+
"step": 110855
|
240405 |
+
},
|
240406 |
+
{
|
240407 |
+
"epoch": 900.83,
|
240408 |
+
"learning_rate": 8.225912762520194e-06,
|
240409 |
+
"loss": 0.994,
|
240410 |
+
"step": 110860
|
240411 |
+
},
|
240412 |
+
{
|
240413 |
+
"epoch": 900.87,
|
240414 |
+
"learning_rate": 8.22583198707593e-06,
|
240415 |
+
"loss": 0.3007,
|
240416 |
+
"step": 110865
|
240417 |
+
},
|
240418 |
+
{
|
240419 |
+
"epoch": 900.91,
|
240420 |
+
"learning_rate": 8.225751211631666e-06,
|
240421 |
+
"loss": 0.3091,
|
240422 |
+
"step": 110870
|
240423 |
+
},
|
240424 |
+
{
|
240425 |
+
"epoch": 900.95,
|
240426 |
+
"learning_rate": 8.2256704361874e-06,
|
240427 |
+
"loss": 0.3254,
|
240428 |
+
"step": 110875
|
240429 |
+
},
|
240430 |
+
{
|
240431 |
+
"epoch": 900.99,
|
240432 |
+
"learning_rate": 8.225589660743136e-06,
|
240433 |
+
"loss": 0.7219,
|
240434 |
+
"step": 110880
|
240435 |
+
},
|
240436 |
+
{
|
240437 |
+
"epoch": 901.0,
|
240438 |
+
"eval_loss": 0.36234134435653687,
|
240439 |
+
"eval_runtime": 43.3119,
|
240440 |
+
"eval_samples_per_second": 19.325,
|
240441 |
+
"eval_steps_per_second": 0.623,
|
240442 |
+
"eval_wer": 0.18013960819635216,
|
240443 |
+
"step": 110881
|
240444 |
+
},
|
240445 |
+
{
|
240446 |
+
"epoch": 894.03,
|
240447 |
+
"learning_rate": 8.22550888529887e-06,
|
240448 |
+
"loss": 0.4134,
|
240449 |
+
"step": 110885
|
240450 |
+
},
|
240451 |
+
{
|
240452 |
+
"epoch": 894.07,
|
240453 |
+
"learning_rate": 8.225428109854606e-06,
|
240454 |
+
"loss": 0.2736,
|
240455 |
+
"step": 110890
|
240456 |
+
},
|
240457 |
+
{
|
240458 |
+
"epoch": 894.11,
|
240459 |
+
"learning_rate": 8.22534733441034e-06,
|
240460 |
+
"loss": 0.3003,
|
240461 |
+
"step": 110895
|
240462 |
+
},
|
240463 |
+
{
|
240464 |
+
"epoch": 894.15,
|
240465 |
+
"learning_rate": 8.225266558966076e-06,
|
240466 |
+
"loss": 0.457,
|
240467 |
+
"step": 110900
|
240468 |
+
},
|
240469 |
+
{
|
240470 |
+
"epoch": 894.19,
|
240471 |
+
"learning_rate": 8.22518578352181e-06,
|
240472 |
+
"loss": 0.8826,
|
240473 |
+
"step": 110905
|
240474 |
+
},
|
240475 |
+
{
|
240476 |
+
"epoch": 894.23,
|
240477 |
+
"learning_rate": 8.225105008077546e-06,
|
240478 |
+
"loss": 0.6164,
|
240479 |
+
"step": 110910
|
240480 |
+
},
|
240481 |
+
{
|
240482 |
+
"epoch": 894.27,
|
240483 |
+
"learning_rate": 8.22502423263328e-06,
|
240484 |
+
"loss": 0.2989,
|
240485 |
+
"step": 110915
|
240486 |
+
},
|
240487 |
+
{
|
240488 |
+
"epoch": 894.31,
|
240489 |
+
"learning_rate": 8.224943457189016e-06,
|
240490 |
+
"loss": 0.3474,
|
240491 |
+
"step": 110920
|
240492 |
+
},
|
240493 |
+
{
|
240494 |
+
"epoch": 894.35,
|
240495 |
+
"learning_rate": 8.224862681744751e-06,
|
240496 |
+
"loss": 0.3896,
|
240497 |
+
"step": 110925
|
240498 |
+
},
|
240499 |
+
{
|
240500 |
+
"epoch": 894.4,
|
240501 |
+
"learning_rate": 8.224781906300486e-06,
|
240502 |
+
"loss": 0.8943,
|
240503 |
+
"step": 110930
|
240504 |
+
},
|
240505 |
+
{
|
240506 |
+
"epoch": 894.44,
|
240507 |
+
"learning_rate": 8.224701130856221e-06,
|
240508 |
+
"loss": 0.6775,
|
240509 |
+
"step": 110935
|
240510 |
+
},
|
240511 |
+
{
|
240512 |
+
"epoch": 894.48,
|
240513 |
+
"learning_rate": 8.224620355411956e-06,
|
240514 |
+
"loss": 0.3509,
|
240515 |
+
"step": 110940
|
240516 |
+
},
|
240517 |
+
{
|
240518 |
+
"epoch": 894.52,
|
240519 |
+
"learning_rate": 8.224539579967691e-06,
|
240520 |
+
"loss": 0.2667,
|
240521 |
+
"step": 110945
|
240522 |
+
},
|
240523 |
+
{
|
240524 |
+
"epoch": 894.56,
|
240525 |
+
"learning_rate": 8.224458804523426e-06,
|
240526 |
+
"loss": 0.3828,
|
240527 |
+
"step": 110950
|
240528 |
+
},
|
240529 |
+
{
|
240530 |
+
"epoch": 894.6,
|
240531 |
+
"learning_rate": 8.224378029079161e-06,
|
240532 |
+
"loss": 0.8308,
|
240533 |
+
"step": 110955
|
240534 |
+
},
|
240535 |
+
{
|
240536 |
+
"epoch": 894.64,
|
240537 |
+
"learning_rate": 8.224297253634896e-06,
|
240538 |
+
"loss": 0.6964,
|
240539 |
+
"step": 110960
|
240540 |
+
},
|
240541 |
+
{
|
240542 |
+
"epoch": 894.68,
|
240543 |
+
"learning_rate": 8.224216478190631e-06,
|
240544 |
+
"loss": 0.3014,
|
240545 |
+
"step": 110965
|
240546 |
+
},
|
240547 |
+
{
|
240548 |
+
"epoch": 894.72,
|
240549 |
+
"learning_rate": 8.224135702746365e-06,
|
240550 |
+
"loss": 0.3124,
|
240551 |
+
"step": 110970
|
240552 |
+
},
|
240553 |
+
{
|
240554 |
+
"epoch": 894.76,
|
240555 |
+
"learning_rate": 8.224054927302101e-06,
|
240556 |
+
"loss": 0.3854,
|
240557 |
+
"step": 110975
|
240558 |
+
},
|
240559 |
+
{
|
240560 |
+
"epoch": 894.8,
|
240561 |
+
"learning_rate": 8.223974151857837e-06,
|
240562 |
+
"loss": 0.9649,
|
240563 |
+
"step": 110980
|
240564 |
+
},
|
240565 |
+
{
|
240566 |
+
"epoch": 894.84,
|
240567 |
+
"learning_rate": 8.223893376413571e-06,
|
240568 |
+
"loss": 0.7675,
|
240569 |
+
"step": 110985
|
240570 |
+
},
|
240571 |
+
{
|
240572 |
+
"epoch": 894.88,
|
240573 |
+
"learning_rate": 8.223812600969307e-06,
|
240574 |
+
"loss": 0.3051,
|
240575 |
+
"step": 110990
|
240576 |
+
},
|
240577 |
+
{
|
240578 |
+
"epoch": 894.92,
|
240579 |
+
"learning_rate": 8.223731825525041e-06,
|
240580 |
+
"loss": 0.2799,
|
240581 |
+
"step": 110995
|
240582 |
+
},
|
240583 |
+
{
|
240584 |
+
"epoch": 894.96,
|
240585 |
+
"learning_rate": 8.223651050080777e-06,
|
240586 |
+
"loss": 0.411,
|
240587 |
+
"step": 111000
|
240588 |
+
},
|
240589 |
+
{
|
240590 |
+
"epoch": 895.0,
|
240591 |
+
"learning_rate": 8.223570274636511e-06,
|
240592 |
+
"loss": 1.2268,
|
240593 |
+
"step": 111005
|
240594 |
+
},
|
240595 |
+
{
|
240596 |
+
"epoch": 895.0,
|
240597 |
+
"eval_loss": 0.4037691652774811,
|
240598 |
+
"eval_runtime": 44.1709,
|
240599 |
+
"eval_samples_per_second": 18.949,
|
240600 |
+
"eval_steps_per_second": 0.611,
|
240601 |
+
"eval_wer": 0.1897884018198888,
|
240602 |
+
"step": 111005
|
240603 |
+
},
|
240604 |
+
{
|
240605 |
+
"epoch": 888.04,
|
240606 |
+
"learning_rate": 8.223489499192247e-06,
|
240607 |
+
"loss": 0.2999,
|
240608 |
+
"step": 111010
|
240609 |
+
},
|
240610 |
+
{
|
240611 |
+
"epoch": 888.08,
|
240612 |
+
"learning_rate": 8.223408723747981e-06,
|
240613 |
+
"loss": 0.3209,
|
240614 |
+
"step": 111015
|
240615 |
+
},
|
240616 |
+
{
|
240617 |
+
"epoch": 888.12,
|
240618 |
+
"learning_rate": 8.223327948303717e-06,
|
240619 |
+
"loss": 0.398,
|
240620 |
+
"step": 111020
|
240621 |
+
},
|
240622 |
+
{
|
240623 |
+
"epoch": 888.16,
|
240624 |
+
"learning_rate": 8.223247172859451e-06,
|
240625 |
+
"loss": 0.4157,
|
240626 |
+
"step": 111025
|
240627 |
+
},
|
240628 |
+
{
|
240629 |
+
"epoch": 888.2,
|
240630 |
+
"learning_rate": 8.223166397415187e-06,
|
240631 |
+
"loss": 1.1022,
|
240632 |
+
"step": 111030
|
240633 |
+
},
|
240634 |
+
{
|
240635 |
+
"epoch": 888.24,
|
240636 |
+
"learning_rate": 8.223085621970921e-06,
|
240637 |
+
"loss": 0.4123,
|
240638 |
+
"step": 111035
|
240639 |
+
},
|
240640 |
+
{
|
240641 |
+
"epoch": 888.28,
|
240642 |
+
"learning_rate": 8.223004846526657e-06,
|
240643 |
+
"loss": 0.2387,
|
240644 |
+
"step": 111040
|
240645 |
+
},
|
240646 |
+
{
|
240647 |
+
"epoch": 888.32,
|
240648 |
+
"learning_rate": 8.222924071082393e-06,
|
240649 |
+
"loss": 0.3189,
|
240650 |
+
"step": 111045
|
240651 |
+
},
|
240652 |
+
{
|
240653 |
+
"epoch": 888.36,
|
240654 |
+
"learning_rate": 8.222843295638127e-06,
|
240655 |
+
"loss": 0.5308,
|
240656 |
+
"step": 111050
|
240657 |
+
},
|
240658 |
+
{
|
240659 |
+
"epoch": 888.4,
|
240660 |
+
"learning_rate": 8.222762520193863e-06,
|
240661 |
+
"loss": 1.1537,
|
240662 |
+
"step": 111055
|
240663 |
+
},
|
240664 |
+
{
|
240665 |
+
"epoch": 888.44,
|
240666 |
+
"learning_rate": 8.222681744749597e-06,
|
240667 |
+
"loss": 0.3102,
|
240668 |
+
"step": 111060
|
240669 |
+
},
|
240670 |
+
{
|
240671 |
+
"epoch": 888.48,
|
240672 |
+
"learning_rate": 8.222600969305333e-06,
|
240673 |
+
"loss": 0.2456,
|
240674 |
+
"step": 111065
|
240675 |
+
},
|
240676 |
+
{
|
240677 |
+
"epoch": 888.52,
|
240678 |
+
"learning_rate": 8.222520193861067e-06,
|
240679 |
+
"loss": 0.2704,
|
240680 |
+
"step": 111070
|
240681 |
+
},
|
240682 |
+
{
|
240683 |
+
"epoch": 888.56,
|
240684 |
+
"learning_rate": 8.222439418416803e-06,
|
240685 |
+
"loss": 0.3987,
|
240686 |
+
"step": 111075
|
240687 |
+
},
|
240688 |
+
{
|
240689 |
+
"epoch": 888.6,
|
240690 |
+
"learning_rate": 8.222358642972537e-06,
|
240691 |
+
"loss": 1.1706,
|
240692 |
+
"step": 111080
|
240693 |
+
},
|
240694 |
+
{
|
240695 |
+
"epoch": 888.64,
|
240696 |
+
"learning_rate": 8.222277867528273e-06,
|
240697 |
+
"loss": 0.3518,
|
240698 |
+
"step": 111085
|
240699 |
+
},
|
240700 |
+
{
|
240701 |
+
"epoch": 888.68,
|
240702 |
+
"learning_rate": 8.222197092084007e-06,
|
240703 |
+
"loss": 0.2963,
|
240704 |
+
"step": 111090
|
240705 |
+
},
|
240706 |
+
{
|
240707 |
+
"epoch": 888.72,
|
240708 |
+
"learning_rate": 8.222116316639743e-06,
|
240709 |
+
"loss": 0.3069,
|
240710 |
+
"step": 111095
|
240711 |
+
},
|
240712 |
+
{
|
240713 |
+
"epoch": 888.76,
|
240714 |
+
"learning_rate": 8.222035541195479e-06,
|
240715 |
+
"loss": 0.5053,
|
240716 |
+
"step": 111100
|
240717 |
+
},
|
240718 |
+
{
|
240719 |
+
"epoch": 888.8,
|
240720 |
+
"learning_rate": 8.221954765751213e-06,
|
240721 |
+
"loss": 1.1867,
|
240722 |
+
"step": 111105
|
240723 |
+
},
|
240724 |
+
{
|
240725 |
+
"epoch": 888.84,
|
240726 |
+
"learning_rate": 8.221873990306949e-06,
|
240727 |
+
"loss": 0.3344,
|
240728 |
+
"step": 111110
|
240729 |
+
},
|
240730 |
+
{
|
240731 |
+
"epoch": 888.88,
|
240732 |
+
"learning_rate": 8.221793214862683e-06,
|
240733 |
+
"loss": 0.2685,
|
240734 |
+
"step": 111115
|
240735 |
+
},
|
240736 |
+
{
|
240737 |
+
"epoch": 888.92,
|
240738 |
+
"learning_rate": 8.221712439418419e-06,
|
240739 |
+
"loss": 0.346,
|
240740 |
+
"step": 111120
|
240741 |
+
},
|
240742 |
+
{
|
240743 |
+
"epoch": 888.96,
|
240744 |
+
"learning_rate": 8.221631663974153e-06,
|
240745 |
+
"loss": 0.413,
|
240746 |
+
"step": 111125
|
240747 |
+
},
|
240748 |
+
{
|
240749 |
+
"epoch": 889.0,
|
240750 |
+
"learning_rate": 8.221550888529888e-06,
|
240751 |
+
"loss": 1.43,
|
240752 |
+
"step": 111130
|
240753 |
+
},
|
240754 |
+
{
|
240755 |
+
"epoch": 889.0,
|
240756 |
+
"eval_loss": 0.4563208818435669,
|
240757 |
+
"eval_runtime": 45.6982,
|
240758 |
+
"eval_samples_per_second": 18.316,
|
240759 |
+
"eval_steps_per_second": 0.591,
|
240760 |
+
"eval_wer": 0.17906810035842294,
|
240761 |
+
"step": 111130
|
240762 |
+
},
|
240763 |
+
{
|
240764 |
+
"epoch": 889.04,
|
240765 |
+
"learning_rate": 8.221470113085623e-06,
|
240766 |
+
"loss": 0.3338,
|
240767 |
+
"step": 111135
|
240768 |
+
},
|
240769 |
+
{
|
240770 |
+
"epoch": 889.08,
|
240771 |
+
"learning_rate": 8.221389337641358e-06,
|
240772 |
+
"loss": 0.2645,
|
240773 |
+
"step": 111140
|
240774 |
+
},
|
240775 |
+
{
|
240776 |
+
"epoch": 889.12,
|
240777 |
+
"learning_rate": 8.221308562197093e-06,
|
240778 |
+
"loss": 0.3281,
|
240779 |
+
"step": 111145
|
240780 |
+
},
|
240781 |
+
{
|
240782 |
+
"epoch": 889.16,
|
240783 |
+
"learning_rate": 8.221227786752828e-06,
|
240784 |
+
"loss": 0.4546,
|
240785 |
+
"step": 111150
|
240786 |
+
},
|
240787 |
+
{
|
240788 |
+
"epoch": 889.2,
|
240789 |
+
"learning_rate": 8.221147011308563e-06,
|
240790 |
+
"loss": 1.1137,
|
240791 |
+
"step": 111155
|
240792 |
+
},
|
240793 |
+
{
|
240794 |
+
"epoch": 889.24,
|
240795 |
+
"learning_rate": 8.221066235864298e-06,
|
240796 |
+
"loss": 0.3458,
|
240797 |
+
"step": 111160
|
240798 |
+
},
|
240799 |
+
{
|
240800 |
+
"epoch": 889.28,
|
240801 |
+
"learning_rate": 8.220985460420034e-06,
|
240802 |
+
"loss": 0.2908,
|
240803 |
+
"step": 111165
|
240804 |
+
},
|
240805 |
+
{
|
240806 |
+
"epoch": 889.32,
|
240807 |
+
"learning_rate": 8.220904684975768e-06,
|
240808 |
+
"loss": 0.3467,
|
240809 |
+
"step": 111170
|
240810 |
+
},
|
240811 |
+
{
|
240812 |
+
"epoch": 889.36,
|
240813 |
+
"learning_rate": 8.220823909531504e-06,
|
240814 |
+
"loss": 0.4733,
|
240815 |
+
"step": 111175
|
240816 |
+
},
|
240817 |
+
{
|
240818 |
+
"epoch": 889.4,
|
240819 |
+
"learning_rate": 8.220743134087238e-06,
|
240820 |
+
"loss": 1.119,
|
240821 |
+
"step": 111180
|
240822 |
+
},
|
240823 |
+
{
|
240824 |
+
"epoch": 889.44,
|
240825 |
+
"learning_rate": 8.220662358642974e-06,
|
240826 |
+
"loss": 0.3293,
|
240827 |
+
"step": 111185
|
240828 |
+
},
|
240829 |
+
{
|
240830 |
+
"epoch": 889.48,
|
240831 |
+
"learning_rate": 8.220581583198708e-06,
|
240832 |
+
"loss": 0.2844,
|
240833 |
+
"step": 111190
|
240834 |
+
},
|
240835 |
+
{
|
240836 |
+
"epoch": 889.52,
|
240837 |
+
"learning_rate": 8.220500807754444e-06,
|
240838 |
+
"loss": 0.294,
|
240839 |
+
"step": 111195
|
240840 |
+
},
|
240841 |
+
{
|
240842 |
+
"epoch": 889.56,
|
240843 |
+
"learning_rate": 8.220420032310178e-06,
|
240844 |
+
"loss": 0.4952,
|
240845 |
+
"step": 111200
|
240846 |
+
},
|
240847 |
+
{
|
240848 |
+
"epoch": 889.6,
|
240849 |
+
"learning_rate": 8.220339256865914e-06,
|
240850 |
+
"loss": 1.2199,
|
240851 |
+
"step": 111205
|
240852 |
+
},
|
240853 |
+
{
|
240854 |
+
"epoch": 889.64,
|
240855 |
+
"learning_rate": 8.220258481421648e-06,
|
240856 |
+
"loss": 0.329,
|
240857 |
+
"step": 111210
|
240858 |
+
},
|
240859 |
+
{
|
240860 |
+
"epoch": 889.68,
|
240861 |
+
"learning_rate": 8.220177705977384e-06,
|
240862 |
+
"loss": 0.3095,
|
240863 |
+
"step": 111215
|
240864 |
+
},
|
240865 |
+
{
|
240866 |
+
"epoch": 889.72,
|
240867 |
+
"learning_rate": 8.220096930533118e-06,
|
240868 |
+
"loss": 0.3582,
|
240869 |
+
"step": 111220
|
240870 |
+
},
|
240871 |
+
{
|
240872 |
+
"epoch": 889.76,
|
240873 |
+
"learning_rate": 8.220016155088854e-06,
|
240874 |
+
"loss": 0.4658,
|
240875 |
+
"step": 111225
|
240876 |
+
},
|
240877 |
+
{
|
240878 |
+
"epoch": 889.8,
|
240879 |
+
"learning_rate": 8.21993537964459e-06,
|
240880 |
+
"loss": 1.2943,
|
240881 |
+
"step": 111230
|
240882 |
+
},
|
240883 |
+
{
|
240884 |
+
"epoch": 889.84,
|
240885 |
+
"learning_rate": 8.219854604200324e-06,
|
240886 |
+
"loss": 0.3334,
|
240887 |
+
"step": 111235
|
240888 |
+
},
|
240889 |
+
{
|
240890 |
+
"epoch": 889.88,
|
240891 |
+
"learning_rate": 8.21977382875606e-06,
|
240892 |
+
"loss": 0.2945,
|
240893 |
+
"step": 111240
|
240894 |
+
},
|
240895 |
+
{
|
240896 |
+
"epoch": 889.92,
|
240897 |
+
"learning_rate": 8.219693053311794e-06,
|
240898 |
+
"loss": 0.2881,
|
240899 |
+
"step": 111245
|
240900 |
+
},
|
240901 |
+
{
|
240902 |
+
"epoch": 889.96,
|
240903 |
+
"learning_rate": 8.21961227786753e-06,
|
240904 |
+
"loss": 0.4068,
|
240905 |
+
"step": 111250
|
240906 |
+
},
|
240907 |
+
{
|
240908 |
+
"epoch": 890.0,
|
240909 |
+
"learning_rate": 8.219531502423264e-06,
|
240910 |
+
"loss": 1.2953,
|
240911 |
+
"step": 111255
|
240912 |
+
},
|
240913 |
+
{
|
240914 |
+
"epoch": 890.0,
|
240915 |
+
"eval_loss": 0.39579546451568604,
|
240916 |
+
"eval_runtime": 41.6212,
|
240917 |
+
"eval_samples_per_second": 20.11,
|
240918 |
+
"eval_steps_per_second": 0.649,
|
240919 |
+
"eval_wer": 0.18293677266325697,
|
240920 |
+
"step": 111255
|
240921 |
+
},
|
240922 |
+
{
|
240923 |
+
"epoch": 897.04,
|
240924 |
+
"learning_rate": 8.219450726979e-06,
|
240925 |
+
"loss": 0.3568,
|
240926 |
+
"step": 111260
|
240927 |
+
},
|
240928 |
+
{
|
240929 |
+
"epoch": 897.08,
|
240930 |
+
"learning_rate": 8.219369951534734e-06,
|
240931 |
+
"loss": 0.2539,
|
240932 |
+
"step": 111265
|
240933 |
+
},
|
240934 |
+
{
|
240935 |
+
"epoch": 897.12,
|
240936 |
+
"learning_rate": 8.21928917609047e-06,
|
240937 |
+
"loss": 0.3165,
|
240938 |
+
"step": 111270
|
240939 |
+
},
|
240940 |
+
{
|
240941 |
+
"epoch": 897.16,
|
240942 |
+
"learning_rate": 8.219208400646204e-06,
|
240943 |
+
"loss": 0.4533,
|
240944 |
+
"step": 111275
|
240945 |
+
},
|
240946 |
+
{
|
240947 |
+
"epoch": 897.2,
|
240948 |
+
"learning_rate": 8.21912762520194e-06,
|
240949 |
+
"loss": 1.0925,
|
240950 |
+
"step": 111280
|
240951 |
+
},
|
240952 |
+
{
|
240953 |
+
"epoch": 897.24,
|
240954 |
+
"learning_rate": 8.219046849757674e-06,
|
240955 |
+
"loss": 0.3357,
|
240956 |
+
"step": 111285
|
240957 |
+
},
|
240958 |
+
{
|
240959 |
+
"epoch": 897.28,
|
240960 |
+
"learning_rate": 8.21896607431341e-06,
|
240961 |
+
"loss": 0.2692,
|
240962 |
+
"step": 111290
|
240963 |
+
},
|
240964 |
+
{
|
240965 |
+
"epoch": 897.32,
|
240966 |
+
"learning_rate": 8.218885298869146e-06,
|
240967 |
+
"loss": 0.3194,
|
240968 |
+
"step": 111295
|
240969 |
+
},
|
240970 |
+
{
|
240971 |
+
"epoch": 897.36,
|
240972 |
+
"learning_rate": 8.21880452342488e-06,
|
240973 |
+
"loss": 0.555,
|
240974 |
+
"step": 111300
|
240975 |
+
},
|
240976 |
+
{
|
240977 |
+
"epoch": 897.4,
|
240978 |
+
"learning_rate": 8.218723747980616e-06,
|
240979 |
+
"loss": 1.1612,
|
240980 |
+
"step": 111305
|
240981 |
+
},
|
240982 |
+
{
|
240983 |
+
"epoch": 897.44,
|
240984 |
+
"learning_rate": 8.21864297253635e-06,
|
240985 |
+
"loss": 0.3347,
|
240986 |
+
"step": 111310
|
240987 |
+
},
|
240988 |
+
{
|
240989 |
+
"epoch": 897.48,
|
240990 |
+
"learning_rate": 8.218562197092086e-06,
|
240991 |
+
"loss": 0.2972,
|
240992 |
+
"step": 111315
|
240993 |
+
},
|
240994 |
+
{
|
240995 |
+
"epoch": 897.52,
|
240996 |
+
"learning_rate": 8.21848142164782e-06,
|
240997 |
+
"loss": 0.3082,
|
240998 |
+
"step": 111320
|
240999 |
+
},
|
241000 |
+
{
|
241001 |
+
"epoch": 897.56,
|
241002 |
+
"learning_rate": 8.218400646203556e-06,
|
241003 |
+
"loss": 0.4182,
|
241004 |
+
"step": 111325
|
241005 |
+
},
|
241006 |
+
{
|
241007 |
+
"epoch": 897.6,
|
241008 |
+
"learning_rate": 8.21831987075929e-06,
|
241009 |
+
"loss": 1.192,
|
241010 |
+
"step": 111330
|
241011 |
+
},
|
241012 |
+
{
|
241013 |
+
"epoch": 897.65,
|
241014 |
+
"learning_rate": 8.218239095315026e-06,
|
241015 |
+
"loss": 0.3286,
|
241016 |
+
"step": 111335
|
241017 |
+
},
|
241018 |
+
{
|
241019 |
+
"epoch": 897.69,
|
241020 |
+
"learning_rate": 8.21815831987076e-06,
|
241021 |
+
"loss": 0.2918,
|
241022 |
+
"step": 111340
|
241023 |
+
},
|
241024 |
+
{
|
241025 |
+
"epoch": 897.73,
|
241026 |
+
"learning_rate": 8.218077544426495e-06,
|
241027 |
+
"loss": 0.3403,
|
241028 |
+
"step": 111345
|
241029 |
+
},
|
241030 |
+
{
|
241031 |
+
"epoch": 897.77,
|
241032 |
+
"learning_rate": 8.21799676898223e-06,
|
241033 |
+
"loss": 0.3825,
|
241034 |
+
"step": 111350
|
241035 |
+
},
|
241036 |
+
{
|
241037 |
+
"epoch": 897.81,
|
241038 |
+
"learning_rate": 8.217915993537965e-06,
|
241039 |
+
"loss": 1.0865,
|
241040 |
+
"step": 111355
|
241041 |
+
},
|
241042 |
+
{
|
241043 |
+
"epoch": 897.85,
|
241044 |
+
"learning_rate": 8.2178352180937e-06,
|
241045 |
+
"loss": 0.3358,
|
241046 |
+
"step": 111360
|
241047 |
+
},
|
241048 |
+
{
|
241049 |
+
"epoch": 897.89,
|
241050 |
+
"learning_rate": 8.217754442649435e-06,
|
241051 |
+
"loss": 0.3336,
|
241052 |
+
"step": 111365
|
241053 |
+
},
|
241054 |
+
{
|
241055 |
+
"epoch": 897.93,
|
241056 |
+
"learning_rate": 8.217673667205171e-06,
|
241057 |
+
"loss": 0.3484,
|
241058 |
+
"step": 111370
|
241059 |
+
},
|
241060 |
+
{
|
241061 |
+
"epoch": 897.97,
|
241062 |
+
"learning_rate": 8.217592891760905e-06,
|
241063 |
+
"loss": 0.4983,
|
241064 |
+
"step": 111375
|
241065 |
+
},
|
241066 |
+
{
|
241067 |
+
"epoch": 898.0,
|
241068 |
+
"eval_loss": 0.3549356162548065,
|
241069 |
+
"eval_runtime": 42.074,
|
241070 |
+
"eval_samples_per_second": 19.894,
|
241071 |
+
"eval_steps_per_second": 0.642,
|
241072 |
+
"eval_wer": 0.18447376051029285,
|
241073 |
+
"step": 111379
|
241074 |
}
|
241075 |
],
|
241076 |
"max_steps": 620000,
|
241077 |
"num_train_epochs": 5000,
|
241078 |
+
"total_flos": 3.134467323546747e+20,
|
241079 |
"trial_name": null,
|
241080 |
"trial_params": null
|
241081 |
}
|
model-bin/finetune/base/{checkpoint-110758 β checkpoint-111379}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630045620.5655968/events.out.tfevents.1630045620.52f5c7e305a3.886.71
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:789077e8a431041de909256226a12d9fd795a843e189968fcbd73a7d940c1bb0
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630046118.6803894/events.out.tfevents.1630046118.52f5c7e305a3.886.73
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d4196b10fc377834886dc0d987e6817a8ea8f93225248f6d0c19c43600641b2
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630046681.8092928/events.out.tfevents.1630046681.52f5c7e305a3.886.75
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:abdd13375aec82b2ef9f339c6eb574e37141831c3243c001d07c32e4ab0cf46f
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630047172.2078328/events.out.tfevents.1630047172.52f5c7e305a3.886.77
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:232390e2831e3b474ff01088741dca913da5c1a45a34e06d55f1e7af7f05c211
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630047639.1210005/events.out.tfevents.1630047639.52f5c7e305a3.886.79
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:63970938dab9c44f0940f5ad235a4c8e66651ae8540d2f66744893ea940f0ac2
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630045620.52f5c7e305a3.886.70
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b8314966b5c7b8b998ed76cc6f32d38a8a2fd310b0f566a0ba90b2151c0e7b1f
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630046118.52f5c7e305a3.886.72
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2fca3e324b891d951182c535e721c6721fa0e5b2649c8e079560f44cf322a59
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630046681.52f5c7e305a3.886.74
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f79dfe9c06a70a4a2dc57702c07e9552251702809224aa774e01719c63bd929a
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630047172.52f5c7e305a3.886.76
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86cc4daafbb29ca0d18efe9c264f9c5feb271b6fcd28a84322fbf266ffb68542
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630047639.52f5c7e305a3.886.78
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f6abc8f2020bcd704cade369621eba3e292108e363a1184c6f8b5a5f47ca085
|
3 |
+
size 8462
|