"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630246352.8300264/events.out.tfevents.1630246352.cc93b136ebf5.1086.189 +3 -0
- model-bin/finetune/base/log/1630246785.4372547/events.out.tfevents.1630246785.cc93b136ebf5.1086.191 +3 -0
- model-bin/finetune/base/log/1630247219.004471/events.out.tfevents.1630247219.cc93b136ebf5.1086.193 +3 -0
- model-bin/finetune/base/log/1630247655.9469938/events.out.tfevents.1630247655.cc93b136ebf5.1086.195 +3 -0
- model-bin/finetune/base/log/1630248088.0553622/events.out.tfevents.1630248088.cc93b136ebf5.1086.197 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630246352.cc93b136ebf5.1086.188 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630246785.cc93b136ebf5.1086.190 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630247218.cc93b136ebf5.1086.192 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630247655.cc93b136ebf5.1086.194 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630248088.cc93b136ebf5.1086.196 +3 -0
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e9eede5251c3f12c63b46d66a179f1dbd2b6d9f12ab83956ee5715976561602
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a6bd337b8f0e2f3b46fc1c4ecb5f3bafbfc576e1e6bde2071ce0f670336e2879
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a1d6790720c3684b5fab1b80fa056d808968462ad941960785ea5dec6ae7387
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:884cb1f5ec43d0bdfcfad8769f8573a767f703fd8562f4c240659e0fdd8266d6
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be40c7eaf300a7ee9457562db278d81160ffb7d793ceebd4250c4e629248949f
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -291264,11 +291264,800 @@
|
|
291264 |
"eval_steps_per_second": 0.671,
|
291265 |
"eval_wer": 0.182829413960433,
|
291266 |
"step": 150827
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
291267 |
}
|
291268 |
],
|
291269 |
-
"max_steps":
|
291270 |
"num_train_epochs": 5000,
|
291271 |
-
"total_flos": 4.
|
291272 |
"trial_name": null,
|
291273 |
"trial_params": null
|
291274 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1220.995983935743,
|
5 |
+
"global_step": 151448,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
291264 |
"eval_steps_per_second": 0.671,
|
291265 |
"eval_wer": 0.182829413960433,
|
291266 |
"step": 150827
|
291267 |
+
},
|
291268 |
+
{
|
291269 |
+
"epoch": 1216.02,
|
291270 |
+
"learning_rate": 7.580436187399032e-06,
|
291271 |
+
"loss": 0.2669,
|
291272 |
+
"step": 150830
|
291273 |
+
},
|
291274 |
+
{
|
291275 |
+
"epoch": 1216.06,
|
291276 |
+
"learning_rate": 7.580355411954766e-06,
|
291277 |
+
"loss": 0.2552,
|
291278 |
+
"step": 150835
|
291279 |
+
},
|
291280 |
+
{
|
291281 |
+
"epoch": 1216.1,
|
291282 |
+
"learning_rate": 7.580274636510502e-06,
|
291283 |
+
"loss": 0.2963,
|
291284 |
+
"step": 150840
|
291285 |
+
},
|
291286 |
+
{
|
291287 |
+
"epoch": 1216.14,
|
291288 |
+
"learning_rate": 7.580193861066237e-06,
|
291289 |
+
"loss": 0.3486,
|
291290 |
+
"step": 150845
|
291291 |
+
},
|
291292 |
+
{
|
291293 |
+
"epoch": 1216.18,
|
291294 |
+
"learning_rate": 7.580113085621972e-06,
|
291295 |
+
"loss": 0.7076,
|
291296 |
+
"step": 150850
|
291297 |
+
},
|
291298 |
+
{
|
291299 |
+
"epoch": 1216.22,
|
291300 |
+
"learning_rate": 7.580032310177707e-06,
|
291301 |
+
"loss": 0.7936,
|
291302 |
+
"step": 150855
|
291303 |
+
},
|
291304 |
+
{
|
291305 |
+
"epoch": 1216.27,
|
291306 |
+
"learning_rate": 7.579951534733442e-06,
|
291307 |
+
"loss": 0.2483,
|
291308 |
+
"step": 150860
|
291309 |
+
},
|
291310 |
+
{
|
291311 |
+
"epoch": 1216.31,
|
291312 |
+
"learning_rate": 7.579870759289177e-06,
|
291313 |
+
"loss": 0.3237,
|
291314 |
+
"step": 150865
|
291315 |
+
},
|
291316 |
+
{
|
291317 |
+
"epoch": 1216.35,
|
291318 |
+
"learning_rate": 7.579789983844912e-06,
|
291319 |
+
"loss": 0.3846,
|
291320 |
+
"step": 150870
|
291321 |
+
},
|
291322 |
+
{
|
291323 |
+
"epoch": 1216.39,
|
291324 |
+
"learning_rate": 7.5797092084006466e-06,
|
291325 |
+
"loss": 0.673,
|
291326 |
+
"step": 150875
|
291327 |
+
},
|
291328 |
+
{
|
291329 |
+
"epoch": 1216.43,
|
291330 |
+
"learning_rate": 7.5796284329563816e-06,
|
291331 |
+
"loss": 0.897,
|
291332 |
+
"step": 150880
|
291333 |
+
},
|
291334 |
+
{
|
291335 |
+
"epoch": 1216.47,
|
291336 |
+
"learning_rate": 7.579547657512117e-06,
|
291337 |
+
"loss": 0.2648,
|
291338 |
+
"step": 150885
|
291339 |
+
},
|
291340 |
+
{
|
291341 |
+
"epoch": 1216.51,
|
291342 |
+
"learning_rate": 7.5794668820678515e-06,
|
291343 |
+
"loss": 0.3112,
|
291344 |
+
"step": 150890
|
291345 |
+
},
|
291346 |
+
{
|
291347 |
+
"epoch": 1216.55,
|
291348 |
+
"learning_rate": 7.579386106623587e-06,
|
291349 |
+
"loss": 0.367,
|
291350 |
+
"step": 150895
|
291351 |
+
},
|
291352 |
+
{
|
291353 |
+
"epoch": 1216.59,
|
291354 |
+
"learning_rate": 7.5793053311793215e-06,
|
291355 |
+
"loss": 0.7596,
|
291356 |
+
"step": 150900
|
291357 |
+
},
|
291358 |
+
{
|
291359 |
+
"epoch": 1216.63,
|
291360 |
+
"learning_rate": 7.579224555735057e-06,
|
291361 |
+
"loss": 0.8846,
|
291362 |
+
"step": 150905
|
291363 |
+
},
|
291364 |
+
{
|
291365 |
+
"epoch": 1216.67,
|
291366 |
+
"learning_rate": 7.5791437802907915e-06,
|
291367 |
+
"loss": 0.2878,
|
291368 |
+
"step": 150910
|
291369 |
+
},
|
291370 |
+
{
|
291371 |
+
"epoch": 1216.71,
|
291372 |
+
"learning_rate": 7.579063004846527e-06,
|
291373 |
+
"loss": 0.2911,
|
291374 |
+
"step": 150915
|
291375 |
+
},
|
291376 |
+
{
|
291377 |
+
"epoch": 1216.75,
|
291378 |
+
"learning_rate": 7.578982229402262e-06,
|
291379 |
+
"loss": 0.4213,
|
291380 |
+
"step": 150920
|
291381 |
+
},
|
291382 |
+
{
|
291383 |
+
"epoch": 1216.79,
|
291384 |
+
"learning_rate": 7.578901453957997e-06,
|
291385 |
+
"loss": 0.6366,
|
291386 |
+
"step": 150925
|
291387 |
+
},
|
291388 |
+
{
|
291389 |
+
"epoch": 1216.83,
|
291390 |
+
"learning_rate": 7.578820678513732e-06,
|
291391 |
+
"loss": 0.886,
|
291392 |
+
"step": 150930
|
291393 |
+
},
|
291394 |
+
{
|
291395 |
+
"epoch": 1216.87,
|
291396 |
+
"learning_rate": 7.578739903069467e-06,
|
291397 |
+
"loss": 0.2832,
|
291398 |
+
"step": 150935
|
291399 |
+
},
|
291400 |
+
{
|
291401 |
+
"epoch": 1216.91,
|
291402 |
+
"learning_rate": 7.578659127625203e-06,
|
291403 |
+
"loss": 0.2781,
|
291404 |
+
"step": 150940
|
291405 |
+
},
|
291406 |
+
{
|
291407 |
+
"epoch": 1216.95,
|
291408 |
+
"learning_rate": 7.578578352180937e-06,
|
291409 |
+
"loss": 0.3466,
|
291410 |
+
"step": 150945
|
291411 |
+
},
|
291412 |
+
{
|
291413 |
+
"epoch": 1216.99,
|
291414 |
+
"learning_rate": 7.578497576736673e-06,
|
291415 |
+
"loss": 0.7638,
|
291416 |
+
"step": 150950
|
291417 |
+
},
|
291418 |
+
{
|
291419 |
+
"epoch": 1217.0,
|
291420 |
+
"eval_loss": 0.3567521274089813,
|
291421 |
+
"eval_runtime": 40.4597,
|
291422 |
+
"eval_samples_per_second": 20.687,
|
291423 |
+
"eval_steps_per_second": 0.667,
|
291424 |
+
"eval_wer": 0.17730812013348166,
|
291425 |
+
"step": 150951
|
291426 |
+
},
|
291427 |
+
{
|
291428 |
+
"epoch": 1207.03,
|
291429 |
+
"learning_rate": 7.578416801292407e-06,
|
291430 |
+
"loss": 0.2737,
|
291431 |
+
"step": 150955
|
291432 |
+
},
|
291433 |
+
{
|
291434 |
+
"epoch": 1207.07,
|
291435 |
+
"learning_rate": 7.578336025848143e-06,
|
291436 |
+
"loss": 0.261,
|
291437 |
+
"step": 150960
|
291438 |
+
},
|
291439 |
+
{
|
291440 |
+
"epoch": 1207.11,
|
291441 |
+
"learning_rate": 7.578255250403877e-06,
|
291442 |
+
"loss": 0.2653,
|
291443 |
+
"step": 150965
|
291444 |
+
},
|
291445 |
+
{
|
291446 |
+
"epoch": 1207.15,
|
291447 |
+
"learning_rate": 7.578174474959613e-06,
|
291448 |
+
"loss": 0.4096,
|
291449 |
+
"step": 150970
|
291450 |
+
},
|
291451 |
+
{
|
291452 |
+
"epoch": 1207.19,
|
291453 |
+
"learning_rate": 7.578093699515347e-06,
|
291454 |
+
"loss": 0.8663,
|
291455 |
+
"step": 150975
|
291456 |
+
},
|
291457 |
+
{
|
291458 |
+
"epoch": 1207.23,
|
291459 |
+
"learning_rate": 7.578012924071083e-06,
|
291460 |
+
"loss": 0.5652,
|
291461 |
+
"step": 150980
|
291462 |
+
},
|
291463 |
+
{
|
291464 |
+
"epoch": 1207.27,
|
291465 |
+
"learning_rate": 7.577932148626818e-06,
|
291466 |
+
"loss": 0.2107,
|
291467 |
+
"step": 150985
|
291468 |
+
},
|
291469 |
+
{
|
291470 |
+
"epoch": 1207.31,
|
291471 |
+
"learning_rate": 7.577851373182553e-06,
|
291472 |
+
"loss": 0.2782,
|
291473 |
+
"step": 150990
|
291474 |
+
},
|
291475 |
+
{
|
291476 |
+
"epoch": 1207.35,
|
291477 |
+
"learning_rate": 7.577770597738289e-06,
|
291478 |
+
"loss": 0.4066,
|
291479 |
+
"step": 150995
|
291480 |
+
},
|
291481 |
+
{
|
291482 |
+
"epoch": 1207.39,
|
291483 |
+
"learning_rate": 7.577689822294023e-06,
|
291484 |
+
"loss": 0.9056,
|
291485 |
+
"step": 151000
|
291486 |
+
},
|
291487 |
+
{
|
291488 |
+
"epoch": 1207.43,
|
291489 |
+
"learning_rate": 7.577609046849759e-06,
|
291490 |
+
"loss": 0.6806,
|
291491 |
+
"step": 151005
|
291492 |
+
},
|
291493 |
+
{
|
291494 |
+
"epoch": 1207.47,
|
291495 |
+
"learning_rate": 7.577528271405493e-06,
|
291496 |
+
"loss": 0.3028,
|
291497 |
+
"step": 151010
|
291498 |
+
},
|
291499 |
+
{
|
291500 |
+
"epoch": 1207.51,
|
291501 |
+
"learning_rate": 7.577447495961229e-06,
|
291502 |
+
"loss": 0.3177,
|
291503 |
+
"step": 151015
|
291504 |
+
},
|
291505 |
+
{
|
291506 |
+
"epoch": 1207.55,
|
291507 |
+
"learning_rate": 7.577366720516963e-06,
|
291508 |
+
"loss": 0.37,
|
291509 |
+
"step": 151020
|
291510 |
+
},
|
291511 |
+
{
|
291512 |
+
"epoch": 1207.59,
|
291513 |
+
"learning_rate": 7.577285945072699e-06,
|
291514 |
+
"loss": 0.8868,
|
291515 |
+
"step": 151025
|
291516 |
+
},
|
291517 |
+
{
|
291518 |
+
"epoch": 1207.63,
|
291519 |
+
"learning_rate": 7.577205169628433e-06,
|
291520 |
+
"loss": 0.588,
|
291521 |
+
"step": 151030
|
291522 |
+
},
|
291523 |
+
{
|
291524 |
+
"epoch": 1207.67,
|
291525 |
+
"learning_rate": 7.577124394184169e-06,
|
291526 |
+
"loss": 0.277,
|
291527 |
+
"step": 151035
|
291528 |
+
},
|
291529 |
+
{
|
291530 |
+
"epoch": 1207.71,
|
291531 |
+
"learning_rate": 7.577043618739903e-06,
|
291532 |
+
"loss": 0.2769,
|
291533 |
+
"step": 151040
|
291534 |
+
},
|
291535 |
+
{
|
291536 |
+
"epoch": 1207.75,
|
291537 |
+
"learning_rate": 7.576962843295639e-06,
|
291538 |
+
"loss": 0.3973,
|
291539 |
+
"step": 151045
|
291540 |
+
},
|
291541 |
+
{
|
291542 |
+
"epoch": 1207.79,
|
291543 |
+
"learning_rate": 7.576882067851374e-06,
|
291544 |
+
"loss": 0.9731,
|
291545 |
+
"step": 151050
|
291546 |
+
},
|
291547 |
+
{
|
291548 |
+
"epoch": 1207.83,
|
291549 |
+
"learning_rate": 7.576801292407109e-06,
|
291550 |
+
"loss": 0.6487,
|
291551 |
+
"step": 151055
|
291552 |
+
},
|
291553 |
+
{
|
291554 |
+
"epoch": 1207.87,
|
291555 |
+
"learning_rate": 7.5767205169628445e-06,
|
291556 |
+
"loss": 0.2917,
|
291557 |
+
"step": 151060
|
291558 |
+
},
|
291559 |
+
{
|
291560 |
+
"epoch": 1207.91,
|
291561 |
+
"learning_rate": 7.576639741518579e-06,
|
291562 |
+
"loss": 0.2741,
|
291563 |
+
"step": 151065
|
291564 |
+
},
|
291565 |
+
{
|
291566 |
+
"epoch": 1207.95,
|
291567 |
+
"learning_rate": 7.5765589660743145e-06,
|
291568 |
+
"loss": 0.3645,
|
291569 |
+
"step": 151070
|
291570 |
+
},
|
291571 |
+
{
|
291572 |
+
"epoch": 1207.99,
|
291573 |
+
"learning_rate": 7.576478190630049e-06,
|
291574 |
+
"loss": 0.825,
|
291575 |
+
"step": 151075
|
291576 |
+
},
|
291577 |
+
{
|
291578 |
+
"epoch": 1208.0,
|
291579 |
+
"eval_loss": 0.3887929916381836,
|
291580 |
+
"eval_runtime": 39.8424,
|
291581 |
+
"eval_samples_per_second": 21.008,
|
291582 |
+
"eval_steps_per_second": 0.678,
|
291583 |
+
"eval_wer": 0.17792285757166443,
|
291584 |
+
"step": 151076
|
291585 |
+
},
|
291586 |
+
{
|
291587 |
+
"epoch": 1218.03,
|
291588 |
+
"learning_rate": 7.5763974151857844e-06,
|
291589 |
+
"loss": 0.347,
|
291590 |
+
"step": 151080
|
291591 |
+
},
|
291592 |
+
{
|
291593 |
+
"epoch": 1218.07,
|
291594 |
+
"learning_rate": 7.576316639741519e-06,
|
291595 |
+
"loss": 0.2568,
|
291596 |
+
"step": 151085
|
291597 |
+
},
|
291598 |
+
{
|
291599 |
+
"epoch": 1218.11,
|
291600 |
+
"learning_rate": 7.576235864297254e-06,
|
291601 |
+
"loss": 0.3279,
|
291602 |
+
"step": 151090
|
291603 |
+
},
|
291604 |
+
{
|
291605 |
+
"epoch": 1218.15,
|
291606 |
+
"learning_rate": 7.5761550888529886e-06,
|
291607 |
+
"loss": 0.3822,
|
291608 |
+
"step": 151095
|
291609 |
+
},
|
291610 |
+
{
|
291611 |
+
"epoch": 1218.19,
|
291612 |
+
"learning_rate": 7.576074313408724e-06,
|
291613 |
+
"loss": 0.8134,
|
291614 |
+
"step": 151100
|
291615 |
+
},
|
291616 |
+
{
|
291617 |
+
"epoch": 1218.23,
|
291618 |
+
"learning_rate": 7.5759935379644585e-06,
|
291619 |
+
"loss": 0.6241,
|
291620 |
+
"step": 151105
|
291621 |
+
},
|
291622 |
+
{
|
291623 |
+
"epoch": 1218.27,
|
291624 |
+
"learning_rate": 7.575912762520194e-06,
|
291625 |
+
"loss": 0.2843,
|
291626 |
+
"step": 151110
|
291627 |
+
},
|
291628 |
+
{
|
291629 |
+
"epoch": 1218.31,
|
291630 |
+
"learning_rate": 7.57583198707593e-06,
|
291631 |
+
"loss": 0.2978,
|
291632 |
+
"step": 151115
|
291633 |
+
},
|
291634 |
+
{
|
291635 |
+
"epoch": 1218.35,
|
291636 |
+
"learning_rate": 7.575751211631664e-06,
|
291637 |
+
"loss": 0.3903,
|
291638 |
+
"step": 151120
|
291639 |
+
},
|
291640 |
+
{
|
291641 |
+
"epoch": 1218.39,
|
291642 |
+
"learning_rate": 7.5756704361874e-06,
|
291643 |
+
"loss": 0.8314,
|
291644 |
+
"step": 151125
|
291645 |
+
},
|
291646 |
+
{
|
291647 |
+
"epoch": 1218.43,
|
291648 |
+
"learning_rate": 7.575589660743134e-06,
|
291649 |
+
"loss": 0.6191,
|
291650 |
+
"step": 151130
|
291651 |
+
},
|
291652 |
+
{
|
291653 |
+
"epoch": 1218.47,
|
291654 |
+
"learning_rate": 7.57550888529887e-06,
|
291655 |
+
"loss": 0.2565,
|
291656 |
+
"step": 151135
|
291657 |
+
},
|
291658 |
+
{
|
291659 |
+
"epoch": 1218.51,
|
291660 |
+
"learning_rate": 7.575428109854604e-06,
|
291661 |
+
"loss": 0.2411,
|
291662 |
+
"step": 151140
|
291663 |
+
},
|
291664 |
+
{
|
291665 |
+
"epoch": 1218.55,
|
291666 |
+
"learning_rate": 7.57534733441034e-06,
|
291667 |
+
"loss": 0.3642,
|
291668 |
+
"step": 151145
|
291669 |
+
},
|
291670 |
+
{
|
291671 |
+
"epoch": 1218.59,
|
291672 |
+
"learning_rate": 7.575266558966074e-06,
|
291673 |
+
"loss": 0.9774,
|
291674 |
+
"step": 151150
|
291675 |
+
},
|
291676 |
+
{
|
291677 |
+
"epoch": 1218.63,
|
291678 |
+
"learning_rate": 7.57518578352181e-06,
|
291679 |
+
"loss": 0.6026,
|
291680 |
+
"step": 151155
|
291681 |
+
},
|
291682 |
+
{
|
291683 |
+
"epoch": 1218.67,
|
291684 |
+
"learning_rate": 7.575105008077544e-06,
|
291685 |
+
"loss": 0.254,
|
291686 |
+
"step": 151160
|
291687 |
+
},
|
291688 |
+
{
|
291689 |
+
"epoch": 1218.71,
|
291690 |
+
"learning_rate": 7.57502423263328e-06,
|
291691 |
+
"loss": 0.3303,
|
291692 |
+
"step": 151165
|
291693 |
+
},
|
291694 |
+
{
|
291695 |
+
"epoch": 1218.76,
|
291696 |
+
"learning_rate": 7.574943457189016e-06,
|
291697 |
+
"loss": 0.3565,
|
291698 |
+
"step": 151170
|
291699 |
+
},
|
291700 |
+
{
|
291701 |
+
"epoch": 1218.8,
|
291702 |
+
"learning_rate": 7.57486268174475e-06,
|
291703 |
+
"loss": 0.7702,
|
291704 |
+
"step": 151175
|
291705 |
+
},
|
291706 |
+
{
|
291707 |
+
"epoch": 1218.84,
|
291708 |
+
"learning_rate": 7.574781906300486e-06,
|
291709 |
+
"loss": 0.5369,
|
291710 |
+
"step": 151180
|
291711 |
+
},
|
291712 |
+
{
|
291713 |
+
"epoch": 1218.88,
|
291714 |
+
"learning_rate": 7.57470113085622e-06,
|
291715 |
+
"loss": 0.2883,
|
291716 |
+
"step": 151185
|
291717 |
+
},
|
291718 |
+
{
|
291719 |
+
"epoch": 1218.92,
|
291720 |
+
"learning_rate": 7.574620355411956e-06,
|
291721 |
+
"loss": 0.2753,
|
291722 |
+
"step": 151190
|
291723 |
+
},
|
291724 |
+
{
|
291725 |
+
"epoch": 1218.96,
|
291726 |
+
"learning_rate": 7.57453957996769e-06,
|
291727 |
+
"loss": 0.4346,
|
291728 |
+
"step": 151195
|
291729 |
+
},
|
291730 |
+
{
|
291731 |
+
"epoch": 1219.0,
|
291732 |
+
"learning_rate": 7.574458804523426e-06,
|
291733 |
+
"loss": 0.9678,
|
291734 |
+
"step": 151200
|
291735 |
+
},
|
291736 |
+
{
|
291737 |
+
"epoch": 1219.0,
|
291738 |
+
"eval_loss": 0.40293607115745544,
|
291739 |
+
"eval_runtime": 40.725,
|
291740 |
+
"eval_samples_per_second": 20.528,
|
291741 |
+
"eval_steps_per_second": 0.663,
|
291742 |
+
"eval_wer": 0.17951517798645994,
|
291743 |
+
"step": 151200
|
291744 |
+
},
|
291745 |
+
{
|
291746 |
+
"epoch": 1219.04,
|
291747 |
+
"learning_rate": 7.57437802907916e-06,
|
291748 |
+
"loss": 0.3041,
|
291749 |
+
"step": 151205
|
291750 |
+
},
|
291751 |
+
{
|
291752 |
+
"epoch": 1219.08,
|
291753 |
+
"learning_rate": 7.574297253634896e-06,
|
291754 |
+
"loss": 0.2819,
|
291755 |
+
"step": 151210
|
291756 |
+
},
|
291757 |
+
{
|
291758 |
+
"epoch": 1219.12,
|
291759 |
+
"learning_rate": 7.57421647819063e-06,
|
291760 |
+
"loss": 0.2779,
|
291761 |
+
"step": 151215
|
291762 |
+
},
|
291763 |
+
{
|
291764 |
+
"epoch": 1219.16,
|
291765 |
+
"learning_rate": 7.574135702746366e-06,
|
291766 |
+
"loss": 0.4908,
|
291767 |
+
"step": 151220
|
291768 |
+
},
|
291769 |
+
{
|
291770 |
+
"epoch": 1219.2,
|
291771 |
+
"learning_rate": 7.5740549273021e-06,
|
291772 |
+
"loss": 1.049,
|
291773 |
+
"step": 151225
|
291774 |
+
},
|
291775 |
+
{
|
291776 |
+
"epoch": 1219.24,
|
291777 |
+
"learning_rate": 7.573974151857836e-06,
|
291778 |
+
"loss": 0.3181,
|
291779 |
+
"step": 151230
|
291780 |
+
},
|
291781 |
+
{
|
291782 |
+
"epoch": 1219.28,
|
291783 |
+
"learning_rate": 7.573893376413572e-06,
|
291784 |
+
"loss": 0.2837,
|
291785 |
+
"step": 151235
|
291786 |
+
},
|
291787 |
+
{
|
291788 |
+
"epoch": 1219.32,
|
291789 |
+
"learning_rate": 7.573812600969306e-06,
|
291790 |
+
"loss": 0.3345,
|
291791 |
+
"step": 151240
|
291792 |
+
},
|
291793 |
+
{
|
291794 |
+
"epoch": 1219.36,
|
291795 |
+
"learning_rate": 7.5737318255250416e-06,
|
291796 |
+
"loss": 0.4296,
|
291797 |
+
"step": 151245
|
291798 |
+
},
|
291799 |
+
{
|
291800 |
+
"epoch": 1219.4,
|
291801 |
+
"learning_rate": 7.573651050080776e-06,
|
291802 |
+
"loss": 1.2842,
|
291803 |
+
"step": 151250
|
291804 |
+
},
|
291805 |
+
{
|
291806 |
+
"epoch": 1219.44,
|
291807 |
+
"learning_rate": 7.5735702746365115e-06,
|
291808 |
+
"loss": 0.3158,
|
291809 |
+
"step": 151255
|
291810 |
+
},
|
291811 |
+
{
|
291812 |
+
"epoch": 1219.48,
|
291813 |
+
"learning_rate": 7.573489499192246e-06,
|
291814 |
+
"loss": 0.2343,
|
291815 |
+
"step": 151260
|
291816 |
+
},
|
291817 |
+
{
|
291818 |
+
"epoch": 1219.52,
|
291819 |
+
"learning_rate": 7.5734087237479815e-06,
|
291820 |
+
"loss": 0.2876,
|
291821 |
+
"step": 151265
|
291822 |
+
},
|
291823 |
+
{
|
291824 |
+
"epoch": 1219.56,
|
291825 |
+
"learning_rate": 7.573327948303716e-06,
|
291826 |
+
"loss": 0.4508,
|
291827 |
+
"step": 151270
|
291828 |
+
},
|
291829 |
+
{
|
291830 |
+
"epoch": 1219.6,
|
291831 |
+
"learning_rate": 7.5732471728594515e-06,
|
291832 |
+
"loss": 1.1773,
|
291833 |
+
"step": 151275
|
291834 |
+
},
|
291835 |
+
{
|
291836 |
+
"epoch": 1219.64,
|
291837 |
+
"learning_rate": 7.573166397415186e-06,
|
291838 |
+
"loss": 0.27,
|
291839 |
+
"step": 151280
|
291840 |
+
},
|
291841 |
+
{
|
291842 |
+
"epoch": 1219.68,
|
291843 |
+
"learning_rate": 7.5730856219709215e-06,
|
291844 |
+
"loss": 0.2492,
|
291845 |
+
"step": 151285
|
291846 |
+
},
|
291847 |
+
{
|
291848 |
+
"epoch": 1219.72,
|
291849 |
+
"learning_rate": 7.573004846526657e-06,
|
291850 |
+
"loss": 0.3389,
|
291851 |
+
"step": 151290
|
291852 |
+
},
|
291853 |
+
{
|
291854 |
+
"epoch": 1219.76,
|
291855 |
+
"learning_rate": 7.5729240710823914e-06,
|
291856 |
+
"loss": 0.39,
|
291857 |
+
"step": 151295
|
291858 |
+
},
|
291859 |
+
{
|
291860 |
+
"epoch": 1219.8,
|
291861 |
+
"learning_rate": 7.572843295638127e-06,
|
291862 |
+
"loss": 1.2323,
|
291863 |
+
"step": 151300
|
291864 |
+
},
|
291865 |
+
{
|
291866 |
+
"epoch": 1219.84,
|
291867 |
+
"learning_rate": 7.572762520193861e-06,
|
291868 |
+
"loss": 0.2676,
|
291869 |
+
"step": 151305
|
291870 |
+
},
|
291871 |
+
{
|
291872 |
+
"epoch": 1219.88,
|
291873 |
+
"learning_rate": 7.572681744749597e-06,
|
291874 |
+
"loss": 0.2777,
|
291875 |
+
"step": 151310
|
291876 |
+
},
|
291877 |
+
{
|
291878 |
+
"epoch": 1219.92,
|
291879 |
+
"learning_rate": 7.572600969305331e-06,
|
291880 |
+
"loss": 0.3009,
|
291881 |
+
"step": 151315
|
291882 |
+
},
|
291883 |
+
{
|
291884 |
+
"epoch": 1219.96,
|
291885 |
+
"learning_rate": 7.572520193861067e-06,
|
291886 |
+
"loss": 0.4905,
|
291887 |
+
"step": 151320
|
291888 |
+
},
|
291889 |
+
{
|
291890 |
+
"epoch": 1220.0,
|
291891 |
+
"eval_loss": 0.40317270159721375,
|
291892 |
+
"eval_runtime": 40.9503,
|
291893 |
+
"eval_samples_per_second": 20.415,
|
291894 |
+
"eval_steps_per_second": 0.659,
|
291895 |
+
"eval_wer": 0.18318384015168088,
|
291896 |
+
"step": 151324
|
291897 |
+
},
|
291898 |
+
{
|
291899 |
+
"epoch": 1220.01,
|
291900 |
+
"learning_rate": 7.572439418416801e-06,
|
291901 |
+
"loss": 0.2942,
|
291902 |
+
"step": 151325
|
291903 |
+
},
|
291904 |
+
{
|
291905 |
+
"epoch": 1220.05,
|
291906 |
+
"learning_rate": 7.572358642972537e-06,
|
291907 |
+
"loss": 0.2593,
|
291908 |
+
"step": 151330
|
291909 |
+
},
|
291910 |
+
{
|
291911 |
+
"epoch": 1220.09,
|
291912 |
+
"learning_rate": 7.572277867528271e-06,
|
291913 |
+
"loss": 0.2962,
|
291914 |
+
"step": 151335
|
291915 |
+
},
|
291916 |
+
{
|
291917 |
+
"epoch": 1220.13,
|
291918 |
+
"learning_rate": 7.572197092084007e-06,
|
291919 |
+
"loss": 0.3793,
|
291920 |
+
"step": 151340
|
291921 |
+
},
|
291922 |
+
{
|
291923 |
+
"epoch": 1220.17,
|
291924 |
+
"learning_rate": 7.572116316639743e-06,
|
291925 |
+
"loss": 0.473,
|
291926 |
+
"step": 151345
|
291927 |
+
},
|
291928 |
+
{
|
291929 |
+
"epoch": 1220.21,
|
291930 |
+
"learning_rate": 7.572035541195477e-06,
|
291931 |
+
"loss": 1.116,
|
291932 |
+
"step": 151350
|
291933 |
+
},
|
291934 |
+
{
|
291935 |
+
"epoch": 1220.25,
|
291936 |
+
"learning_rate": 7.571954765751213e-06,
|
291937 |
+
"loss": 0.2653,
|
291938 |
+
"step": 151355
|
291939 |
+
},
|
291940 |
+
{
|
291941 |
+
"epoch": 1220.29,
|
291942 |
+
"learning_rate": 7.571873990306947e-06,
|
291943 |
+
"loss": 0.2706,
|
291944 |
+
"step": 151360
|
291945 |
+
},
|
291946 |
+
{
|
291947 |
+
"epoch": 1220.33,
|
291948 |
+
"learning_rate": 7.571793214862683e-06,
|
291949 |
+
"loss": 0.2761,
|
291950 |
+
"step": 151365
|
291951 |
+
},
|
291952 |
+
{
|
291953 |
+
"epoch": 1220.37,
|
291954 |
+
"learning_rate": 7.571712439418417e-06,
|
291955 |
+
"loss": 0.5144,
|
291956 |
+
"step": 151370
|
291957 |
+
},
|
291958 |
+
{
|
291959 |
+
"epoch": 1220.41,
|
291960 |
+
"learning_rate": 7.571631663974153e-06,
|
291961 |
+
"loss": 1.0997,
|
291962 |
+
"step": 151375
|
291963 |
+
},
|
291964 |
+
{
|
291965 |
+
"epoch": 1220.45,
|
291966 |
+
"learning_rate": 7.571550888529887e-06,
|
291967 |
+
"loss": 0.2801,
|
291968 |
+
"step": 151380
|
291969 |
+
},
|
291970 |
+
{
|
291971 |
+
"epoch": 1220.49,
|
291972 |
+
"learning_rate": 7.571470113085623e-06,
|
291973 |
+
"loss": 0.3037,
|
291974 |
+
"step": 151385
|
291975 |
+
},
|
291976 |
+
{
|
291977 |
+
"epoch": 1220.53,
|
291978 |
+
"learning_rate": 7.571389337641357e-06,
|
291979 |
+
"loss": 0.3857,
|
291980 |
+
"step": 151390
|
291981 |
+
},
|
291982 |
+
{
|
291983 |
+
"epoch": 1220.57,
|
291984 |
+
"learning_rate": 7.571308562197093e-06,
|
291985 |
+
"loss": 0.554,
|
291986 |
+
"step": 151395
|
291987 |
+
},
|
291988 |
+
{
|
291989 |
+
"epoch": 1220.61,
|
291990 |
+
"learning_rate": 7.571227786752827e-06,
|
291991 |
+
"loss": 1.0511,
|
291992 |
+
"step": 151400
|
291993 |
+
},
|
291994 |
+
{
|
291995 |
+
"epoch": 1220.65,
|
291996 |
+
"learning_rate": 7.571147011308563e-06,
|
291997 |
+
"loss": 0.2562,
|
291998 |
+
"step": 151405
|
291999 |
+
},
|
292000 |
+
{
|
292001 |
+
"epoch": 1220.69,
|
292002 |
+
"learning_rate": 7.571066235864299e-06,
|
292003 |
+
"loss": 0.2393,
|
292004 |
+
"step": 151410
|
292005 |
+
},
|
292006 |
+
{
|
292007 |
+
"epoch": 1220.73,
|
292008 |
+
"learning_rate": 7.570985460420033e-06,
|
292009 |
+
"loss": 0.3112,
|
292010 |
+
"step": 151415
|
292011 |
+
},
|
292012 |
+
{
|
292013 |
+
"epoch": 1220.77,
|
292014 |
+
"learning_rate": 7.570904684975769e-06,
|
292015 |
+
"loss": 0.5436,
|
292016 |
+
"step": 151420
|
292017 |
+
},
|
292018 |
+
{
|
292019 |
+
"epoch": 1220.81,
|
292020 |
+
"learning_rate": 7.570823909531503e-06,
|
292021 |
+
"loss": 0.9745,
|
292022 |
+
"step": 151425
|
292023 |
+
},
|
292024 |
+
{
|
292025 |
+
"epoch": 1220.85,
|
292026 |
+
"learning_rate": 7.570743134087239e-06,
|
292027 |
+
"loss": 0.2797,
|
292028 |
+
"step": 151430
|
292029 |
+
},
|
292030 |
+
{
|
292031 |
+
"epoch": 1220.89,
|
292032 |
+
"learning_rate": 7.570662358642973e-06,
|
292033 |
+
"loss": 0.2683,
|
292034 |
+
"step": 151435
|
292035 |
+
},
|
292036 |
+
{
|
292037 |
+
"epoch": 1220.93,
|
292038 |
+
"learning_rate": 7.570581583198709e-06,
|
292039 |
+
"loss": 0.3428,
|
292040 |
+
"step": 151440
|
292041 |
+
},
|
292042 |
+
{
|
292043 |
+
"epoch": 1220.97,
|
292044 |
+
"learning_rate": 7.570500807754443e-06,
|
292045 |
+
"loss": 0.4689,
|
292046 |
+
"step": 151445
|
292047 |
+
},
|
292048 |
+
{
|
292049 |
+
"epoch": 1221.0,
|
292050 |
+
"eval_loss": 0.3577888607978821,
|
292051 |
+
"eval_runtime": 42.2225,
|
292052 |
+
"eval_samples_per_second": 19.8,
|
292053 |
+
"eval_steps_per_second": 0.639,
|
292054 |
+
"eval_wer": 0.17625899280575538,
|
292055 |
+
"step": 151448
|
292056 |
}
|
292057 |
],
|
292058 |
+
"max_steps": 620000,
|
292059 |
"num_train_epochs": 5000,
|
292060 |
+
"total_flos": 4.2618930913164073e+20,
|
292061 |
"trial_name": null,
|
292062 |
"trial_params": null
|
292063 |
}
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630246352.8300264/events.out.tfevents.1630246352.cc93b136ebf5.1086.189
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1d2e006a76411ca66fef0f0601b536f525ad6d1bf56071220648ca0968c5ce39
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630246785.4372547/events.out.tfevents.1630246785.cc93b136ebf5.1086.191
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88c677975dcb5799d881c0a06fd4f929de0324135c035871197a803122b7e182
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630247219.004471/events.out.tfevents.1630247219.cc93b136ebf5.1086.193
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba94dba6d83ae89638f94a50b5ed007dd6fa5fff4a9ef3c1d84e599ce33866a2
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630247655.9469938/events.out.tfevents.1630247655.cc93b136ebf5.1086.195
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b8bb97898f6233ed672cf04852addd9a7e44ee8bfe0ec7376dc5ec8454e6735
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630248088.0553622/events.out.tfevents.1630248088.cc93b136ebf5.1086.197
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b42f4a6bdb41edd1853e3d18e435253edac9dffde47c25b93a644f5ef02d2e20
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630246352.cc93b136ebf5.1086.188
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d09b4ca59b66bc0c0684dee2fe769e86548e33094e265be58139e3f0a4d37830
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630246785.cc93b136ebf5.1086.190
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0056a189953ac9b863ab9640ac16ddada116751dc2c5670f09d12244155c7710
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630247218.cc93b136ebf5.1086.192
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b7cc40664c5864ee6af5d3191ef612920626590925ef149aacf562c7b185192
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630247655.cc93b136ebf5.1086.194
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b5e892ecadb74ebdc66ca0fcae41c145b395b336ec9b9b70d9c3394abfcfe60f
|
3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630248088.cc93b136ebf5.1086.196
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:222a21717f9e387720acc83eac9a3017b439135e5baedab94209e89526ba9fa0
|
3 |
+
size 8622
|