Training in progress, step 2320
Browse files- {checkpoint-2000 β checkpoint-2320}/config.json +0 -0
- {checkpoint-2000 β checkpoint-2320}/optimizer.pt +1 -1
- {checkpoint-2000 β checkpoint-2320}/pytorch_model.bin +1 -1
- {checkpoint-2000 β checkpoint-2320}/rng_state.pth +1 -1
- {checkpoint-2000 β checkpoint-2320}/scheduler.pt +1 -1
- {checkpoint-2000 β checkpoint-2320}/trainer_state.json +2227 -3
- {checkpoint-2000 β checkpoint-2320}/training_args.bin +0 -0
- pytorch_model.bin +1 -1
- runs/Feb29_21-23-49_robolidar/events.out.tfevents.1709263483.robolidar.58688.0 +2 -2
{checkpoint-2000 β checkpoint-2320}/config.json
RENAMED
File without changes
|
{checkpoint-2000 β checkpoint-2320}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 677462463
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b8604b9c6e315d290f86b28be2c59f0c89129b2bbbb45ff341adb2ec2ca83c83
|
3 |
size 677462463
|
{checkpoint-2000 β checkpoint-2320}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338799561
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30f2cfebe23afbdeec3f33b5f0aee22d89a51a708ecf56af5d97a012fafcea25
|
3 |
size 338799561
|
{checkpoint-2000 β checkpoint-2320}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a43682d5348164d7f251a244eede93a644c7241a8515371677d90533574ca419
|
3 |
size 14575
|
{checkpoint-2000 β checkpoint-2320}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:70a709ba6f89236232051c65150c56c4aa02e3bf4304fad9df772c066c8edbe4
|
3 |
size 627
|
{checkpoint-2000 β checkpoint-2320}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.06833437085151672,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/parking-terrain_marks/checkpoint-240",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -13906,11 +13906,2235 @@
|
|
13906 |
"eval_samples_per_second": 1.165,
|
13907 |
"eval_steps_per_second": 1.165,
|
13908 |
"step": 2000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13909 |
}
|
13910 |
],
|
13911 |
"max_steps": 2400,
|
13912 |
"num_train_epochs": 600,
|
13913 |
-
"total_flos": 1.
|
13914 |
"trial_name": null,
|
13915 |
"trial_params": null
|
13916 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.06833437085151672,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/parking-terrain_marks/checkpoint-240",
|
4 |
+
"epoch": 580.0,
|
5 |
+
"global_step": 2320,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
13906 |
"eval_samples_per_second": 1.165,
|
13907 |
"eval_steps_per_second": 1.165,
|
13908 |
"step": 2000
|
13909 |
+
},
|
13910 |
+
{
|
13911 |
+
"epoch": 500.25,
|
13912 |
+
"learning_rate": 8.75e-06,
|
13913 |
+
"loss": 0.0049,
|
13914 |
+
"step": 2001
|
13915 |
+
},
|
13916 |
+
{
|
13917 |
+
"epoch": 500.5,
|
13918 |
+
"learning_rate": 8.728070175438596e-06,
|
13919 |
+
"loss": 0.0044,
|
13920 |
+
"step": 2002
|
13921 |
+
},
|
13922 |
+
{
|
13923 |
+
"epoch": 500.75,
|
13924 |
+
"learning_rate": 8.706140350877192e-06,
|
13925 |
+
"loss": 0.0045,
|
13926 |
+
"step": 2003
|
13927 |
+
},
|
13928 |
+
{
|
13929 |
+
"epoch": 501.0,
|
13930 |
+
"learning_rate": 8.68421052631579e-06,
|
13931 |
+
"loss": 0.004,
|
13932 |
+
"step": 2004
|
13933 |
+
},
|
13934 |
+
{
|
13935 |
+
"epoch": 501.25,
|
13936 |
+
"learning_rate": 8.662280701754386e-06,
|
13937 |
+
"loss": 0.0059,
|
13938 |
+
"step": 2005
|
13939 |
+
},
|
13940 |
+
{
|
13941 |
+
"epoch": 501.5,
|
13942 |
+
"learning_rate": 8.640350877192983e-06,
|
13943 |
+
"loss": 0.0035,
|
13944 |
+
"step": 2006
|
13945 |
+
},
|
13946 |
+
{
|
13947 |
+
"epoch": 501.75,
|
13948 |
+
"learning_rate": 8.61842105263158e-06,
|
13949 |
+
"loss": 0.0046,
|
13950 |
+
"step": 2007
|
13951 |
+
},
|
13952 |
+
{
|
13953 |
+
"epoch": 502.0,
|
13954 |
+
"learning_rate": 8.596491228070176e-06,
|
13955 |
+
"loss": 0.0035,
|
13956 |
+
"step": 2008
|
13957 |
+
},
|
13958 |
+
{
|
13959 |
+
"epoch": 502.25,
|
13960 |
+
"learning_rate": 8.574561403508772e-06,
|
13961 |
+
"loss": 0.0061,
|
13962 |
+
"step": 2009
|
13963 |
+
},
|
13964 |
+
{
|
13965 |
+
"epoch": 502.5,
|
13966 |
+
"learning_rate": 8.552631578947368e-06,
|
13967 |
+
"loss": 0.0038,
|
13968 |
+
"step": 2010
|
13969 |
+
},
|
13970 |
+
{
|
13971 |
+
"epoch": 502.75,
|
13972 |
+
"learning_rate": 8.530701754385965e-06,
|
13973 |
+
"loss": 0.0036,
|
13974 |
+
"step": 2011
|
13975 |
+
},
|
13976 |
+
{
|
13977 |
+
"epoch": 503.0,
|
13978 |
+
"learning_rate": 8.508771929824563e-06,
|
13979 |
+
"loss": 0.004,
|
13980 |
+
"step": 2012
|
13981 |
+
},
|
13982 |
+
{
|
13983 |
+
"epoch": 503.25,
|
13984 |
+
"learning_rate": 8.486842105263159e-06,
|
13985 |
+
"loss": 0.0044,
|
13986 |
+
"step": 2013
|
13987 |
+
},
|
13988 |
+
{
|
13989 |
+
"epoch": 503.5,
|
13990 |
+
"learning_rate": 8.464912280701755e-06,
|
13991 |
+
"loss": 0.0042,
|
13992 |
+
"step": 2014
|
13993 |
+
},
|
13994 |
+
{
|
13995 |
+
"epoch": 503.75,
|
13996 |
+
"learning_rate": 8.442982456140352e-06,
|
13997 |
+
"loss": 0.0049,
|
13998 |
+
"step": 2015
|
13999 |
+
},
|
14000 |
+
{
|
14001 |
+
"epoch": 504.0,
|
14002 |
+
"learning_rate": 8.421052631578948e-06,
|
14003 |
+
"loss": 0.0037,
|
14004 |
+
"step": 2016
|
14005 |
+
},
|
14006 |
+
{
|
14007 |
+
"epoch": 504.25,
|
14008 |
+
"learning_rate": 8.399122807017545e-06,
|
14009 |
+
"loss": 0.0053,
|
14010 |
+
"step": 2017
|
14011 |
+
},
|
14012 |
+
{
|
14013 |
+
"epoch": 504.5,
|
14014 |
+
"learning_rate": 8.377192982456141e-06,
|
14015 |
+
"loss": 0.0056,
|
14016 |
+
"step": 2018
|
14017 |
+
},
|
14018 |
+
{
|
14019 |
+
"epoch": 504.75,
|
14020 |
+
"learning_rate": 8.355263157894737e-06,
|
14021 |
+
"loss": 0.0034,
|
14022 |
+
"step": 2019
|
14023 |
+
},
|
14024 |
+
{
|
14025 |
+
"epoch": 505.0,
|
14026 |
+
"learning_rate": 8.333333333333334e-06,
|
14027 |
+
"loss": 0.0044,
|
14028 |
+
"step": 2020
|
14029 |
+
},
|
14030 |
+
{
|
14031 |
+
"epoch": 505.0,
|
14032 |
+
"eval_accuracy_ELSE": 1.0,
|
14033 |
+
"eval_accuracy_X": NaN,
|
14034 |
+
"eval_accuracy_parking_lines": 0.0,
|
14035 |
+
"eval_accuracy_unlabeled": NaN,
|
14036 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14037 |
+
"eval_iou_X": NaN,
|
14038 |
+
"eval_iou_parking_lines": 0.0,
|
14039 |
+
"eval_iou_unlabeled": NaN,
|
14040 |
+
"eval_loss": 0.08571472764015198,
|
14041 |
+
"eval_mean_accuracy": 0.5,
|
14042 |
+
"eval_mean_iou": 0.49420738220214844,
|
14043 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14044 |
+
"eval_runtime": 0.8632,
|
14045 |
+
"eval_samples_per_second": 1.158,
|
14046 |
+
"eval_steps_per_second": 1.158,
|
14047 |
+
"step": 2020
|
14048 |
+
},
|
14049 |
+
{
|
14050 |
+
"epoch": 505.25,
|
14051 |
+
"learning_rate": 8.31140350877193e-06,
|
14052 |
+
"loss": 0.0061,
|
14053 |
+
"step": 2021
|
14054 |
+
},
|
14055 |
+
{
|
14056 |
+
"epoch": 505.5,
|
14057 |
+
"learning_rate": 8.289473684210526e-06,
|
14058 |
+
"loss": 0.0041,
|
14059 |
+
"step": 2022
|
14060 |
+
},
|
14061 |
+
{
|
14062 |
+
"epoch": 505.75,
|
14063 |
+
"learning_rate": 8.267543859649123e-06,
|
14064 |
+
"loss": 0.0041,
|
14065 |
+
"step": 2023
|
14066 |
+
},
|
14067 |
+
{
|
14068 |
+
"epoch": 506.0,
|
14069 |
+
"learning_rate": 8.245614035087721e-06,
|
14070 |
+
"loss": 0.0036,
|
14071 |
+
"step": 2024
|
14072 |
+
},
|
14073 |
+
{
|
14074 |
+
"epoch": 506.25,
|
14075 |
+
"learning_rate": 8.223684210526317e-06,
|
14076 |
+
"loss": 0.0035,
|
14077 |
+
"step": 2025
|
14078 |
+
},
|
14079 |
+
{
|
14080 |
+
"epoch": 506.5,
|
14081 |
+
"learning_rate": 8.201754385964914e-06,
|
14082 |
+
"loss": 0.0049,
|
14083 |
+
"step": 2026
|
14084 |
+
},
|
14085 |
+
{
|
14086 |
+
"epoch": 506.75,
|
14087 |
+
"learning_rate": 8.17982456140351e-06,
|
14088 |
+
"loss": 0.0049,
|
14089 |
+
"step": 2027
|
14090 |
+
},
|
14091 |
+
{
|
14092 |
+
"epoch": 507.0,
|
14093 |
+
"learning_rate": 8.157894736842106e-06,
|
14094 |
+
"loss": 0.0045,
|
14095 |
+
"step": 2028
|
14096 |
+
},
|
14097 |
+
{
|
14098 |
+
"epoch": 507.25,
|
14099 |
+
"learning_rate": 8.135964912280703e-06,
|
14100 |
+
"loss": 0.0042,
|
14101 |
+
"step": 2029
|
14102 |
+
},
|
14103 |
+
{
|
14104 |
+
"epoch": 507.5,
|
14105 |
+
"learning_rate": 8.1140350877193e-06,
|
14106 |
+
"loss": 0.0033,
|
14107 |
+
"step": 2030
|
14108 |
+
},
|
14109 |
+
{
|
14110 |
+
"epoch": 507.75,
|
14111 |
+
"learning_rate": 8.092105263157896e-06,
|
14112 |
+
"loss": 0.0056,
|
14113 |
+
"step": 2031
|
14114 |
+
},
|
14115 |
+
{
|
14116 |
+
"epoch": 508.0,
|
14117 |
+
"learning_rate": 8.070175438596492e-06,
|
14118 |
+
"loss": 0.0045,
|
14119 |
+
"step": 2032
|
14120 |
+
},
|
14121 |
+
{
|
14122 |
+
"epoch": 508.25,
|
14123 |
+
"learning_rate": 8.048245614035088e-06,
|
14124 |
+
"loss": 0.0047,
|
14125 |
+
"step": 2033
|
14126 |
+
},
|
14127 |
+
{
|
14128 |
+
"epoch": 508.5,
|
14129 |
+
"learning_rate": 8.026315789473685e-06,
|
14130 |
+
"loss": 0.0048,
|
14131 |
+
"step": 2034
|
14132 |
+
},
|
14133 |
+
{
|
14134 |
+
"epoch": 508.75,
|
14135 |
+
"learning_rate": 8.004385964912281e-06,
|
14136 |
+
"loss": 0.003,
|
14137 |
+
"step": 2035
|
14138 |
+
},
|
14139 |
+
{
|
14140 |
+
"epoch": 509.0,
|
14141 |
+
"learning_rate": 7.982456140350877e-06,
|
14142 |
+
"loss": 0.0043,
|
14143 |
+
"step": 2036
|
14144 |
+
},
|
14145 |
+
{
|
14146 |
+
"epoch": 509.25,
|
14147 |
+
"learning_rate": 7.960526315789474e-06,
|
14148 |
+
"loss": 0.0057,
|
14149 |
+
"step": 2037
|
14150 |
+
},
|
14151 |
+
{
|
14152 |
+
"epoch": 509.5,
|
14153 |
+
"learning_rate": 7.93859649122807e-06,
|
14154 |
+
"loss": 0.0044,
|
14155 |
+
"step": 2038
|
14156 |
+
},
|
14157 |
+
{
|
14158 |
+
"epoch": 509.75,
|
14159 |
+
"learning_rate": 7.916666666666667e-06,
|
14160 |
+
"loss": 0.0038,
|
14161 |
+
"step": 2039
|
14162 |
+
},
|
14163 |
+
{
|
14164 |
+
"epoch": 510.0,
|
14165 |
+
"learning_rate": 7.894736842105263e-06,
|
14166 |
+
"loss": 0.0036,
|
14167 |
+
"step": 2040
|
14168 |
+
},
|
14169 |
+
{
|
14170 |
+
"epoch": 510.0,
|
14171 |
+
"eval_accuracy_ELSE": 1.0,
|
14172 |
+
"eval_accuracy_X": NaN,
|
14173 |
+
"eval_accuracy_parking_lines": 0.0,
|
14174 |
+
"eval_accuracy_unlabeled": NaN,
|
14175 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14176 |
+
"eval_iou_X": NaN,
|
14177 |
+
"eval_iou_parking_lines": 0.0,
|
14178 |
+
"eval_iou_unlabeled": NaN,
|
14179 |
+
"eval_loss": 0.08490405231714249,
|
14180 |
+
"eval_mean_accuracy": 0.5,
|
14181 |
+
"eval_mean_iou": 0.49420738220214844,
|
14182 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14183 |
+
"eval_runtime": 0.8468,
|
14184 |
+
"eval_samples_per_second": 1.181,
|
14185 |
+
"eval_steps_per_second": 1.181,
|
14186 |
+
"step": 2040
|
14187 |
+
},
|
14188 |
+
{
|
14189 |
+
"epoch": 510.25,
|
14190 |
+
"learning_rate": 7.87280701754386e-06,
|
14191 |
+
"loss": 0.0041,
|
14192 |
+
"step": 2041
|
14193 |
+
},
|
14194 |
+
{
|
14195 |
+
"epoch": 510.5,
|
14196 |
+
"learning_rate": 7.850877192982456e-06,
|
14197 |
+
"loss": 0.0039,
|
14198 |
+
"step": 2042
|
14199 |
+
},
|
14200 |
+
{
|
14201 |
+
"epoch": 510.75,
|
14202 |
+
"learning_rate": 7.828947368421052e-06,
|
14203 |
+
"loss": 0.0047,
|
14204 |
+
"step": 2043
|
14205 |
+
},
|
14206 |
+
{
|
14207 |
+
"epoch": 511.0,
|
14208 |
+
"learning_rate": 7.80701754385965e-06,
|
14209 |
+
"loss": 0.0047,
|
14210 |
+
"step": 2044
|
14211 |
+
},
|
14212 |
+
{
|
14213 |
+
"epoch": 511.25,
|
14214 |
+
"learning_rate": 7.785087719298246e-06,
|
14215 |
+
"loss": 0.004,
|
14216 |
+
"step": 2045
|
14217 |
+
},
|
14218 |
+
{
|
14219 |
+
"epoch": 511.5,
|
14220 |
+
"learning_rate": 7.763157894736843e-06,
|
14221 |
+
"loss": 0.0048,
|
14222 |
+
"step": 2046
|
14223 |
+
},
|
14224 |
+
{
|
14225 |
+
"epoch": 511.75,
|
14226 |
+
"learning_rate": 7.74122807017544e-06,
|
14227 |
+
"loss": 0.005,
|
14228 |
+
"step": 2047
|
14229 |
+
},
|
14230 |
+
{
|
14231 |
+
"epoch": 512.0,
|
14232 |
+
"learning_rate": 7.719298245614036e-06,
|
14233 |
+
"loss": 0.0047,
|
14234 |
+
"step": 2048
|
14235 |
+
},
|
14236 |
+
{
|
14237 |
+
"epoch": 512.25,
|
14238 |
+
"learning_rate": 7.697368421052632e-06,
|
14239 |
+
"loss": 0.0045,
|
14240 |
+
"step": 2049
|
14241 |
+
},
|
14242 |
+
{
|
14243 |
+
"epoch": 512.5,
|
14244 |
+
"learning_rate": 7.675438596491228e-06,
|
14245 |
+
"loss": 0.005,
|
14246 |
+
"step": 2050
|
14247 |
+
},
|
14248 |
+
{
|
14249 |
+
"epoch": 512.75,
|
14250 |
+
"learning_rate": 7.653508771929825e-06,
|
14251 |
+
"loss": 0.0048,
|
14252 |
+
"step": 2051
|
14253 |
+
},
|
14254 |
+
{
|
14255 |
+
"epoch": 513.0,
|
14256 |
+
"learning_rate": 7.631578947368421e-06,
|
14257 |
+
"loss": 0.003,
|
14258 |
+
"step": 2052
|
14259 |
+
},
|
14260 |
+
{
|
14261 |
+
"epoch": 513.25,
|
14262 |
+
"learning_rate": 7.609649122807018e-06,
|
14263 |
+
"loss": 0.0038,
|
14264 |
+
"step": 2053
|
14265 |
+
},
|
14266 |
+
{
|
14267 |
+
"epoch": 513.5,
|
14268 |
+
"learning_rate": 7.587719298245615e-06,
|
14269 |
+
"loss": 0.0049,
|
14270 |
+
"step": 2054
|
14271 |
+
},
|
14272 |
+
{
|
14273 |
+
"epoch": 513.75,
|
14274 |
+
"learning_rate": 7.565789473684211e-06,
|
14275 |
+
"loss": 0.0044,
|
14276 |
+
"step": 2055
|
14277 |
+
},
|
14278 |
+
{
|
14279 |
+
"epoch": 514.0,
|
14280 |
+
"learning_rate": 7.5438596491228074e-06,
|
14281 |
+
"loss": 0.0037,
|
14282 |
+
"step": 2056
|
14283 |
+
},
|
14284 |
+
{
|
14285 |
+
"epoch": 514.25,
|
14286 |
+
"learning_rate": 7.521929824561404e-06,
|
14287 |
+
"loss": 0.0046,
|
14288 |
+
"step": 2057
|
14289 |
+
},
|
14290 |
+
{
|
14291 |
+
"epoch": 514.5,
|
14292 |
+
"learning_rate": 7.5e-06,
|
14293 |
+
"loss": 0.0049,
|
14294 |
+
"step": 2058
|
14295 |
+
},
|
14296 |
+
{
|
14297 |
+
"epoch": 514.75,
|
14298 |
+
"learning_rate": 7.4780701754385966e-06,
|
14299 |
+
"loss": 0.0045,
|
14300 |
+
"step": 2059
|
14301 |
+
},
|
14302 |
+
{
|
14303 |
+
"epoch": 515.0,
|
14304 |
+
"learning_rate": 7.456140350877193e-06,
|
14305 |
+
"loss": 0.0038,
|
14306 |
+
"step": 2060
|
14307 |
+
},
|
14308 |
+
{
|
14309 |
+
"epoch": 515.0,
|
14310 |
+
"eval_accuracy_ELSE": 1.0,
|
14311 |
+
"eval_accuracy_X": NaN,
|
14312 |
+
"eval_accuracy_parking_lines": 0.0,
|
14313 |
+
"eval_accuracy_unlabeled": NaN,
|
14314 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14315 |
+
"eval_iou_X": NaN,
|
14316 |
+
"eval_iou_parking_lines": 0.0,
|
14317 |
+
"eval_iou_unlabeled": NaN,
|
14318 |
+
"eval_loss": 0.08451391011476517,
|
14319 |
+
"eval_mean_accuracy": 0.5,
|
14320 |
+
"eval_mean_iou": 0.49420738220214844,
|
14321 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14322 |
+
"eval_runtime": 0.7684,
|
14323 |
+
"eval_samples_per_second": 1.301,
|
14324 |
+
"eval_steps_per_second": 1.301,
|
14325 |
+
"step": 2060
|
14326 |
+
},
|
14327 |
+
{
|
14328 |
+
"epoch": 515.25,
|
14329 |
+
"learning_rate": 7.434210526315789e-06,
|
14330 |
+
"loss": 0.0043,
|
14331 |
+
"step": 2061
|
14332 |
+
},
|
14333 |
+
{
|
14334 |
+
"epoch": 515.5,
|
14335 |
+
"learning_rate": 7.412280701754386e-06,
|
14336 |
+
"loss": 0.0033,
|
14337 |
+
"step": 2062
|
14338 |
+
},
|
14339 |
+
{
|
14340 |
+
"epoch": 515.75,
|
14341 |
+
"learning_rate": 7.390350877192982e-06,
|
14342 |
+
"loss": 0.0062,
|
14343 |
+
"step": 2063
|
14344 |
+
},
|
14345 |
+
{
|
14346 |
+
"epoch": 516.0,
|
14347 |
+
"learning_rate": 7.3684210526315784e-06,
|
14348 |
+
"loss": 0.0031,
|
14349 |
+
"step": 2064
|
14350 |
+
},
|
14351 |
+
{
|
14352 |
+
"epoch": 516.25,
|
14353 |
+
"learning_rate": 7.3464912280701765e-06,
|
14354 |
+
"loss": 0.0037,
|
14355 |
+
"step": 2065
|
14356 |
+
},
|
14357 |
+
{
|
14358 |
+
"epoch": 516.5,
|
14359 |
+
"learning_rate": 7.324561403508773e-06,
|
14360 |
+
"loss": 0.0036,
|
14361 |
+
"step": 2066
|
14362 |
+
},
|
14363 |
+
{
|
14364 |
+
"epoch": 516.75,
|
14365 |
+
"learning_rate": 7.302631578947369e-06,
|
14366 |
+
"loss": 0.0044,
|
14367 |
+
"step": 2067
|
14368 |
+
},
|
14369 |
+
{
|
14370 |
+
"epoch": 517.0,
|
14371 |
+
"learning_rate": 7.280701754385966e-06,
|
14372 |
+
"loss": 0.0059,
|
14373 |
+
"step": 2068
|
14374 |
+
},
|
14375 |
+
{
|
14376 |
+
"epoch": 517.25,
|
14377 |
+
"learning_rate": 7.258771929824562e-06,
|
14378 |
+
"loss": 0.004,
|
14379 |
+
"step": 2069
|
14380 |
+
},
|
14381 |
+
{
|
14382 |
+
"epoch": 517.5,
|
14383 |
+
"learning_rate": 7.236842105263158e-06,
|
14384 |
+
"loss": 0.0051,
|
14385 |
+
"step": 2070
|
14386 |
+
},
|
14387 |
+
{
|
14388 |
+
"epoch": 517.75,
|
14389 |
+
"learning_rate": 7.214912280701755e-06,
|
14390 |
+
"loss": 0.0034,
|
14391 |
+
"step": 2071
|
14392 |
+
},
|
14393 |
+
{
|
14394 |
+
"epoch": 518.0,
|
14395 |
+
"learning_rate": 7.192982456140351e-06,
|
14396 |
+
"loss": 0.0044,
|
14397 |
+
"step": 2072
|
14398 |
+
},
|
14399 |
+
{
|
14400 |
+
"epoch": 518.25,
|
14401 |
+
"learning_rate": 7.1710526315789475e-06,
|
14402 |
+
"loss": 0.0046,
|
14403 |
+
"step": 2073
|
14404 |
+
},
|
14405 |
+
{
|
14406 |
+
"epoch": 518.5,
|
14407 |
+
"learning_rate": 7.149122807017544e-06,
|
14408 |
+
"loss": 0.0047,
|
14409 |
+
"step": 2074
|
14410 |
+
},
|
14411 |
+
{
|
14412 |
+
"epoch": 518.75,
|
14413 |
+
"learning_rate": 7.12719298245614e-06,
|
14414 |
+
"loss": 0.0045,
|
14415 |
+
"step": 2075
|
14416 |
+
},
|
14417 |
+
{
|
14418 |
+
"epoch": 519.0,
|
14419 |
+
"learning_rate": 7.1052631578947375e-06,
|
14420 |
+
"loss": 0.003,
|
14421 |
+
"step": 2076
|
14422 |
+
},
|
14423 |
+
{
|
14424 |
+
"epoch": 519.25,
|
14425 |
+
"learning_rate": 7.083333333333334e-06,
|
14426 |
+
"loss": 0.0039,
|
14427 |
+
"step": 2077
|
14428 |
+
},
|
14429 |
+
{
|
14430 |
+
"epoch": 519.5,
|
14431 |
+
"learning_rate": 7.06140350877193e-06,
|
14432 |
+
"loss": 0.0037,
|
14433 |
+
"step": 2078
|
14434 |
+
},
|
14435 |
+
{
|
14436 |
+
"epoch": 519.75,
|
14437 |
+
"learning_rate": 7.039473684210527e-06,
|
14438 |
+
"loss": 0.0055,
|
14439 |
+
"step": 2079
|
14440 |
+
},
|
14441 |
+
{
|
14442 |
+
"epoch": 520.0,
|
14443 |
+
"learning_rate": 7.017543859649123e-06,
|
14444 |
+
"loss": 0.0047,
|
14445 |
+
"step": 2080
|
14446 |
+
},
|
14447 |
+
{
|
14448 |
+
"epoch": 520.0,
|
14449 |
+
"eval_accuracy_ELSE": 1.0,
|
14450 |
+
"eval_accuracy_X": NaN,
|
14451 |
+
"eval_accuracy_parking_lines": 0.0,
|
14452 |
+
"eval_accuracy_unlabeled": NaN,
|
14453 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14454 |
+
"eval_iou_X": NaN,
|
14455 |
+
"eval_iou_parking_lines": 0.0,
|
14456 |
+
"eval_iou_unlabeled": NaN,
|
14457 |
+
"eval_loss": 0.08511849492788315,
|
14458 |
+
"eval_mean_accuracy": 0.5,
|
14459 |
+
"eval_mean_iou": 0.49420738220214844,
|
14460 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14461 |
+
"eval_runtime": 0.786,
|
14462 |
+
"eval_samples_per_second": 1.272,
|
14463 |
+
"eval_steps_per_second": 1.272,
|
14464 |
+
"step": 2080
|
14465 |
+
},
|
14466 |
+
{
|
14467 |
+
"epoch": 520.25,
|
14468 |
+
"learning_rate": 6.995614035087719e-06,
|
14469 |
+
"loss": 0.0033,
|
14470 |
+
"step": 2081
|
14471 |
+
},
|
14472 |
+
{
|
14473 |
+
"epoch": 520.5,
|
14474 |
+
"learning_rate": 6.973684210526316e-06,
|
14475 |
+
"loss": 0.0044,
|
14476 |
+
"step": 2082
|
14477 |
+
},
|
14478 |
+
{
|
14479 |
+
"epoch": 520.75,
|
14480 |
+
"learning_rate": 6.951754385964912e-06,
|
14481 |
+
"loss": 0.0049,
|
14482 |
+
"step": 2083
|
14483 |
+
},
|
14484 |
+
{
|
14485 |
+
"epoch": 521.0,
|
14486 |
+
"learning_rate": 6.9298245614035085e-06,
|
14487 |
+
"loss": 0.0043,
|
14488 |
+
"step": 2084
|
14489 |
+
},
|
14490 |
+
{
|
14491 |
+
"epoch": 521.25,
|
14492 |
+
"learning_rate": 6.9078947368421065e-06,
|
14493 |
+
"loss": 0.0043,
|
14494 |
+
"step": 2085
|
14495 |
+
},
|
14496 |
+
{
|
14497 |
+
"epoch": 521.5,
|
14498 |
+
"learning_rate": 6.885964912280703e-06,
|
14499 |
+
"loss": 0.0056,
|
14500 |
+
"step": 2086
|
14501 |
+
},
|
14502 |
+
{
|
14503 |
+
"epoch": 521.75,
|
14504 |
+
"learning_rate": 6.864035087719299e-06,
|
14505 |
+
"loss": 0.0036,
|
14506 |
+
"step": 2087
|
14507 |
+
},
|
14508 |
+
{
|
14509 |
+
"epoch": 522.0,
|
14510 |
+
"learning_rate": 6.842105263157896e-06,
|
14511 |
+
"loss": 0.0036,
|
14512 |
+
"step": 2088
|
14513 |
+
},
|
14514 |
+
{
|
14515 |
+
"epoch": 522.25,
|
14516 |
+
"learning_rate": 6.820175438596492e-06,
|
14517 |
+
"loss": 0.0055,
|
14518 |
+
"step": 2089
|
14519 |
+
},
|
14520 |
+
{
|
14521 |
+
"epoch": 522.5,
|
14522 |
+
"learning_rate": 6.798245614035088e-06,
|
14523 |
+
"loss": 0.0035,
|
14524 |
+
"step": 2090
|
14525 |
+
},
|
14526 |
+
{
|
14527 |
+
"epoch": 522.75,
|
14528 |
+
"learning_rate": 6.776315789473685e-06,
|
14529 |
+
"loss": 0.0036,
|
14530 |
+
"step": 2091
|
14531 |
+
},
|
14532 |
+
{
|
14533 |
+
"epoch": 523.0,
|
14534 |
+
"learning_rate": 6.754385964912281e-06,
|
14535 |
+
"loss": 0.005,
|
14536 |
+
"step": 2092
|
14537 |
+
},
|
14538 |
+
{
|
14539 |
+
"epoch": 523.25,
|
14540 |
+
"learning_rate": 6.7324561403508775e-06,
|
14541 |
+
"loss": 0.0053,
|
14542 |
+
"step": 2093
|
14543 |
+
},
|
14544 |
+
{
|
14545 |
+
"epoch": 523.5,
|
14546 |
+
"learning_rate": 6.710526315789474e-06,
|
14547 |
+
"loss": 0.0041,
|
14548 |
+
"step": 2094
|
14549 |
+
},
|
14550 |
+
{
|
14551 |
+
"epoch": 523.75,
|
14552 |
+
"learning_rate": 6.68859649122807e-06,
|
14553 |
+
"loss": 0.004,
|
14554 |
+
"step": 2095
|
14555 |
+
},
|
14556 |
+
{
|
14557 |
+
"epoch": 524.0,
|
14558 |
+
"learning_rate": 6.666666666666667e-06,
|
14559 |
+
"loss": 0.0031,
|
14560 |
+
"step": 2096
|
14561 |
+
},
|
14562 |
+
{
|
14563 |
+
"epoch": 524.25,
|
14564 |
+
"learning_rate": 6.644736842105263e-06,
|
14565 |
+
"loss": 0.0034,
|
14566 |
+
"step": 2097
|
14567 |
+
},
|
14568 |
+
{
|
14569 |
+
"epoch": 524.5,
|
14570 |
+
"learning_rate": 6.622807017543859e-06,
|
14571 |
+
"loss": 0.0049,
|
14572 |
+
"step": 2098
|
14573 |
+
},
|
14574 |
+
{
|
14575 |
+
"epoch": 524.75,
|
14576 |
+
"learning_rate": 6.600877192982456e-06,
|
14577 |
+
"loss": 0.0048,
|
14578 |
+
"step": 2099
|
14579 |
+
},
|
14580 |
+
{
|
14581 |
+
"epoch": 525.0,
|
14582 |
+
"learning_rate": 6.578947368421053e-06,
|
14583 |
+
"loss": 0.0045,
|
14584 |
+
"step": 2100
|
14585 |
+
},
|
14586 |
+
{
|
14587 |
+
"epoch": 525.0,
|
14588 |
+
"eval_accuracy_ELSE": 1.0,
|
14589 |
+
"eval_accuracy_X": NaN,
|
14590 |
+
"eval_accuracy_parking_lines": 0.0,
|
14591 |
+
"eval_accuracy_unlabeled": NaN,
|
14592 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14593 |
+
"eval_iou_X": NaN,
|
14594 |
+
"eval_iou_parking_lines": 0.0,
|
14595 |
+
"eval_iou_unlabeled": NaN,
|
14596 |
+
"eval_loss": 0.0851300060749054,
|
14597 |
+
"eval_mean_accuracy": 0.5,
|
14598 |
+
"eval_mean_iou": 0.49420738220214844,
|
14599 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14600 |
+
"eval_runtime": 0.8298,
|
14601 |
+
"eval_samples_per_second": 1.205,
|
14602 |
+
"eval_steps_per_second": 1.205,
|
14603 |
+
"step": 2100
|
14604 |
+
},
|
14605 |
+
{
|
14606 |
+
"epoch": 525.25,
|
14607 |
+
"learning_rate": 6.557017543859649e-06,
|
14608 |
+
"loss": 0.0042,
|
14609 |
+
"step": 2101
|
14610 |
+
},
|
14611 |
+
{
|
14612 |
+
"epoch": 525.5,
|
14613 |
+
"learning_rate": 6.535087719298246e-06,
|
14614 |
+
"loss": 0.0038,
|
14615 |
+
"step": 2102
|
14616 |
+
},
|
14617 |
+
{
|
14618 |
+
"epoch": 525.75,
|
14619 |
+
"learning_rate": 6.513157894736842e-06,
|
14620 |
+
"loss": 0.0039,
|
14621 |
+
"step": 2103
|
14622 |
+
},
|
14623 |
+
{
|
14624 |
+
"epoch": 526.0,
|
14625 |
+
"learning_rate": 6.4912280701754385e-06,
|
14626 |
+
"loss": 0.0059,
|
14627 |
+
"step": 2104
|
14628 |
+
},
|
14629 |
+
{
|
14630 |
+
"epoch": 526.25,
|
14631 |
+
"learning_rate": 6.469298245614036e-06,
|
14632 |
+
"loss": 0.0045,
|
14633 |
+
"step": 2105
|
14634 |
+
},
|
14635 |
+
{
|
14636 |
+
"epoch": 526.5,
|
14637 |
+
"learning_rate": 6.447368421052632e-06,
|
14638 |
+
"loss": 0.0046,
|
14639 |
+
"step": 2106
|
14640 |
+
},
|
14641 |
+
{
|
14642 |
+
"epoch": 526.75,
|
14643 |
+
"learning_rate": 6.425438596491229e-06,
|
14644 |
+
"loss": 0.0036,
|
14645 |
+
"step": 2107
|
14646 |
+
},
|
14647 |
+
{
|
14648 |
+
"epoch": 527.0,
|
14649 |
+
"learning_rate": 6.403508771929826e-06,
|
14650 |
+
"loss": 0.0043,
|
14651 |
+
"step": 2108
|
14652 |
+
},
|
14653 |
+
{
|
14654 |
+
"epoch": 527.25,
|
14655 |
+
"learning_rate": 6.381578947368422e-06,
|
14656 |
+
"loss": 0.0043,
|
14657 |
+
"step": 2109
|
14658 |
+
},
|
14659 |
+
{
|
14660 |
+
"epoch": 527.5,
|
14661 |
+
"learning_rate": 6.3596491228070184e-06,
|
14662 |
+
"loss": 0.0045,
|
14663 |
+
"step": 2110
|
14664 |
+
},
|
14665 |
+
{
|
14666 |
+
"epoch": 527.75,
|
14667 |
+
"learning_rate": 6.337719298245615e-06,
|
14668 |
+
"loss": 0.0037,
|
14669 |
+
"step": 2111
|
14670 |
+
},
|
14671 |
+
{
|
14672 |
+
"epoch": 528.0,
|
14673 |
+
"learning_rate": 6.315789473684211e-06,
|
14674 |
+
"loss": 0.0044,
|
14675 |
+
"step": 2112
|
14676 |
+
},
|
14677 |
+
{
|
14678 |
+
"epoch": 528.25,
|
14679 |
+
"learning_rate": 6.2938596491228076e-06,
|
14680 |
+
"loss": 0.004,
|
14681 |
+
"step": 2113
|
14682 |
+
},
|
14683 |
+
{
|
14684 |
+
"epoch": 528.5,
|
14685 |
+
"learning_rate": 6.271929824561404e-06,
|
14686 |
+
"loss": 0.0044,
|
14687 |
+
"step": 2114
|
14688 |
+
},
|
14689 |
+
{
|
14690 |
+
"epoch": 528.75,
|
14691 |
+
"learning_rate": 6.25e-06,
|
14692 |
+
"loss": 0.005,
|
14693 |
+
"step": 2115
|
14694 |
+
},
|
14695 |
+
{
|
14696 |
+
"epoch": 529.0,
|
14697 |
+
"learning_rate": 6.228070175438597e-06,
|
14698 |
+
"loss": 0.0036,
|
14699 |
+
"step": 2116
|
14700 |
+
},
|
14701 |
+
{
|
14702 |
+
"epoch": 529.25,
|
14703 |
+
"learning_rate": 6.206140350877193e-06,
|
14704 |
+
"loss": 0.0038,
|
14705 |
+
"step": 2117
|
14706 |
+
},
|
14707 |
+
{
|
14708 |
+
"epoch": 529.5,
|
14709 |
+
"learning_rate": 6.184210526315789e-06,
|
14710 |
+
"loss": 0.0032,
|
14711 |
+
"step": 2118
|
14712 |
+
},
|
14713 |
+
{
|
14714 |
+
"epoch": 529.75,
|
14715 |
+
"learning_rate": 6.162280701754386e-06,
|
14716 |
+
"loss": 0.0054,
|
14717 |
+
"step": 2119
|
14718 |
+
},
|
14719 |
+
{
|
14720 |
+
"epoch": 530.0,
|
14721 |
+
"learning_rate": 6.140350877192982e-06,
|
14722 |
+
"loss": 0.0056,
|
14723 |
+
"step": 2120
|
14724 |
+
},
|
14725 |
+
{
|
14726 |
+
"epoch": 530.0,
|
14727 |
+
"eval_accuracy_ELSE": 1.0,
|
14728 |
+
"eval_accuracy_X": NaN,
|
14729 |
+
"eval_accuracy_parking_lines": 0.0,
|
14730 |
+
"eval_accuracy_unlabeled": NaN,
|
14731 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14732 |
+
"eval_iou_X": NaN,
|
14733 |
+
"eval_iou_parking_lines": 0.0,
|
14734 |
+
"eval_iou_unlabeled": NaN,
|
14735 |
+
"eval_loss": 0.08506093919277191,
|
14736 |
+
"eval_mean_accuracy": 0.5,
|
14737 |
+
"eval_mean_iou": 0.49420738220214844,
|
14738 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14739 |
+
"eval_runtime": 0.9918,
|
14740 |
+
"eval_samples_per_second": 1.008,
|
14741 |
+
"eval_steps_per_second": 1.008,
|
14742 |
+
"step": 2120
|
14743 |
+
},
|
14744 |
+
{
|
14745 |
+
"epoch": 530.25,
|
14746 |
+
"learning_rate": 6.118421052631579e-06,
|
14747 |
+
"loss": 0.0043,
|
14748 |
+
"step": 2121
|
14749 |
+
},
|
14750 |
+
{
|
14751 |
+
"epoch": 530.5,
|
14752 |
+
"learning_rate": 6.096491228070176e-06,
|
14753 |
+
"loss": 0.0049,
|
14754 |
+
"step": 2122
|
14755 |
+
},
|
14756 |
+
{
|
14757 |
+
"epoch": 530.75,
|
14758 |
+
"learning_rate": 6.074561403508772e-06,
|
14759 |
+
"loss": 0.0036,
|
14760 |
+
"step": 2123
|
14761 |
+
},
|
14762 |
+
{
|
14763 |
+
"epoch": 531.0,
|
14764 |
+
"learning_rate": 6.0526315789473685e-06,
|
14765 |
+
"loss": 0.0042,
|
14766 |
+
"step": 2124
|
14767 |
+
},
|
14768 |
+
{
|
14769 |
+
"epoch": 531.25,
|
14770 |
+
"learning_rate": 6.030701754385965e-06,
|
14771 |
+
"loss": 0.0056,
|
14772 |
+
"step": 2125
|
14773 |
+
},
|
14774 |
+
{
|
14775 |
+
"epoch": 531.5,
|
14776 |
+
"learning_rate": 6.008771929824561e-06,
|
14777 |
+
"loss": 0.003,
|
14778 |
+
"step": 2126
|
14779 |
+
},
|
14780 |
+
{
|
14781 |
+
"epoch": 531.75,
|
14782 |
+
"learning_rate": 5.986842105263158e-06,
|
14783 |
+
"loss": 0.0048,
|
14784 |
+
"step": 2127
|
14785 |
+
},
|
14786 |
+
{
|
14787 |
+
"epoch": 532.0,
|
14788 |
+
"learning_rate": 5.964912280701755e-06,
|
14789 |
+
"loss": 0.004,
|
14790 |
+
"step": 2128
|
14791 |
+
},
|
14792 |
+
{
|
14793 |
+
"epoch": 532.25,
|
14794 |
+
"learning_rate": 5.942982456140351e-06,
|
14795 |
+
"loss": 0.0043,
|
14796 |
+
"step": 2129
|
14797 |
+
},
|
14798 |
+
{
|
14799 |
+
"epoch": 532.5,
|
14800 |
+
"learning_rate": 5.921052631578948e-06,
|
14801 |
+
"loss": 0.0031,
|
14802 |
+
"step": 2130
|
14803 |
+
},
|
14804 |
+
{
|
14805 |
+
"epoch": 532.75,
|
14806 |
+
"learning_rate": 5.899122807017545e-06,
|
14807 |
+
"loss": 0.005,
|
14808 |
+
"step": 2131
|
14809 |
+
},
|
14810 |
+
{
|
14811 |
+
"epoch": 533.0,
|
14812 |
+
"learning_rate": 5.877192982456141e-06,
|
14813 |
+
"loss": 0.0048,
|
14814 |
+
"step": 2132
|
14815 |
+
},
|
14816 |
+
{
|
14817 |
+
"epoch": 533.25,
|
14818 |
+
"learning_rate": 5.855263157894738e-06,
|
14819 |
+
"loss": 0.0033,
|
14820 |
+
"step": 2133
|
14821 |
+
},
|
14822 |
+
{
|
14823 |
+
"epoch": 533.5,
|
14824 |
+
"learning_rate": 5.833333333333334e-06,
|
14825 |
+
"loss": 0.0049,
|
14826 |
+
"step": 2134
|
14827 |
+
},
|
14828 |
+
{
|
14829 |
+
"epoch": 533.75,
|
14830 |
+
"learning_rate": 5.81140350877193e-06,
|
14831 |
+
"loss": 0.0044,
|
14832 |
+
"step": 2135
|
14833 |
+
},
|
14834 |
+
{
|
14835 |
+
"epoch": 534.0,
|
14836 |
+
"learning_rate": 5.789473684210527e-06,
|
14837 |
+
"loss": 0.0046,
|
14838 |
+
"step": 2136
|
14839 |
+
},
|
14840 |
+
{
|
14841 |
+
"epoch": 534.25,
|
14842 |
+
"learning_rate": 5.767543859649123e-06,
|
14843 |
+
"loss": 0.0046,
|
14844 |
+
"step": 2137
|
14845 |
+
},
|
14846 |
+
{
|
14847 |
+
"epoch": 534.5,
|
14848 |
+
"learning_rate": 5.7456140350877194e-06,
|
14849 |
+
"loss": 0.004,
|
14850 |
+
"step": 2138
|
14851 |
+
},
|
14852 |
+
{
|
14853 |
+
"epoch": 534.75,
|
14854 |
+
"learning_rate": 5.723684210526316e-06,
|
14855 |
+
"loss": 0.004,
|
14856 |
+
"step": 2139
|
14857 |
+
},
|
14858 |
+
{
|
14859 |
+
"epoch": 535.0,
|
14860 |
+
"learning_rate": 5.701754385964912e-06,
|
14861 |
+
"loss": 0.0041,
|
14862 |
+
"step": 2140
|
14863 |
+
},
|
14864 |
+
{
|
14865 |
+
"epoch": 535.0,
|
14866 |
+
"eval_accuracy_ELSE": 1.0,
|
14867 |
+
"eval_accuracy_X": NaN,
|
14868 |
+
"eval_accuracy_parking_lines": 0.0,
|
14869 |
+
"eval_accuracy_unlabeled": NaN,
|
14870 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
14871 |
+
"eval_iou_X": NaN,
|
14872 |
+
"eval_iou_parking_lines": 0.0,
|
14873 |
+
"eval_iou_unlabeled": NaN,
|
14874 |
+
"eval_loss": 0.08593860268592834,
|
14875 |
+
"eval_mean_accuracy": 0.5,
|
14876 |
+
"eval_mean_iou": 0.49420738220214844,
|
14877 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
14878 |
+
"eval_runtime": 1.0246,
|
14879 |
+
"eval_samples_per_second": 0.976,
|
14880 |
+
"eval_steps_per_second": 0.976,
|
14881 |
+
"step": 2140
|
14882 |
+
},
|
14883 |
+
{
|
14884 |
+
"epoch": 535.25,
|
14885 |
+
"learning_rate": 5.679824561403509e-06,
|
14886 |
+
"loss": 0.0045,
|
14887 |
+
"step": 2141
|
14888 |
+
},
|
14889 |
+
{
|
14890 |
+
"epoch": 535.5,
|
14891 |
+
"learning_rate": 5.657894736842106e-06,
|
14892 |
+
"loss": 0.0042,
|
14893 |
+
"step": 2142
|
14894 |
+
},
|
14895 |
+
{
|
14896 |
+
"epoch": 535.75,
|
14897 |
+
"learning_rate": 5.635964912280702e-06,
|
14898 |
+
"loss": 0.0046,
|
14899 |
+
"step": 2143
|
14900 |
+
},
|
14901 |
+
{
|
14902 |
+
"epoch": 536.0,
|
14903 |
+
"learning_rate": 5.6140350877192985e-06,
|
14904 |
+
"loss": 0.005,
|
14905 |
+
"step": 2144
|
14906 |
+
},
|
14907 |
+
{
|
14908 |
+
"epoch": 536.25,
|
14909 |
+
"learning_rate": 5.592105263157895e-06,
|
14910 |
+
"loss": 0.0044,
|
14911 |
+
"step": 2145
|
14912 |
+
},
|
14913 |
+
{
|
14914 |
+
"epoch": 536.5,
|
14915 |
+
"learning_rate": 5.570175438596491e-06,
|
14916 |
+
"loss": 0.0044,
|
14917 |
+
"step": 2146
|
14918 |
+
},
|
14919 |
+
{
|
14920 |
+
"epoch": 536.75,
|
14921 |
+
"learning_rate": 5.548245614035088e-06,
|
14922 |
+
"loss": 0.0049,
|
14923 |
+
"step": 2147
|
14924 |
+
},
|
14925 |
+
{
|
14926 |
+
"epoch": 537.0,
|
14927 |
+
"learning_rate": 5.526315789473684e-06,
|
14928 |
+
"loss": 0.0037,
|
14929 |
+
"step": 2148
|
14930 |
+
},
|
14931 |
+
{
|
14932 |
+
"epoch": 537.25,
|
14933 |
+
"learning_rate": 5.50438596491228e-06,
|
14934 |
+
"loss": 0.004,
|
14935 |
+
"step": 2149
|
14936 |
+
},
|
14937 |
+
{
|
14938 |
+
"epoch": 537.5,
|
14939 |
+
"learning_rate": 5.482456140350877e-06,
|
14940 |
+
"loss": 0.0042,
|
14941 |
+
"step": 2150
|
14942 |
+
},
|
14943 |
+
{
|
14944 |
+
"epoch": 537.75,
|
14945 |
+
"learning_rate": 5.460526315789474e-06,
|
14946 |
+
"loss": 0.0032,
|
14947 |
+
"step": 2151
|
14948 |
+
},
|
14949 |
+
{
|
14950 |
+
"epoch": 538.0,
|
14951 |
+
"learning_rate": 5.43859649122807e-06,
|
14952 |
+
"loss": 0.006,
|
14953 |
+
"step": 2152
|
14954 |
+
},
|
14955 |
+
{
|
14956 |
+
"epoch": 538.25,
|
14957 |
+
"learning_rate": 5.416666666666667e-06,
|
14958 |
+
"loss": 0.0045,
|
14959 |
+
"step": 2153
|
14960 |
+
},
|
14961 |
+
{
|
14962 |
+
"epoch": 538.5,
|
14963 |
+
"learning_rate": 5.394736842105263e-06,
|
14964 |
+
"loss": 0.0056,
|
14965 |
+
"step": 2154
|
14966 |
+
},
|
14967 |
+
{
|
14968 |
+
"epoch": 538.75,
|
14969 |
+
"learning_rate": 5.37280701754386e-06,
|
14970 |
+
"loss": 0.0037,
|
14971 |
+
"step": 2155
|
14972 |
+
},
|
14973 |
+
{
|
14974 |
+
"epoch": 539.0,
|
14975 |
+
"learning_rate": 5.350877192982457e-06,
|
14976 |
+
"loss": 0.0042,
|
14977 |
+
"step": 2156
|
14978 |
+
},
|
14979 |
+
{
|
14980 |
+
"epoch": 539.25,
|
14981 |
+
"learning_rate": 5.328947368421053e-06,
|
14982 |
+
"loss": 0.0047,
|
14983 |
+
"step": 2157
|
14984 |
+
},
|
14985 |
+
{
|
14986 |
+
"epoch": 539.5,
|
14987 |
+
"learning_rate": 5.3070175438596495e-06,
|
14988 |
+
"loss": 0.0042,
|
14989 |
+
"step": 2158
|
14990 |
+
},
|
14991 |
+
{
|
14992 |
+
"epoch": 539.75,
|
14993 |
+
"learning_rate": 5.285087719298246e-06,
|
14994 |
+
"loss": 0.0045,
|
14995 |
+
"step": 2159
|
14996 |
+
},
|
14997 |
+
{
|
14998 |
+
"epoch": 540.0,
|
14999 |
+
"learning_rate": 5.263157894736842e-06,
|
15000 |
+
"loss": 0.004,
|
15001 |
+
"step": 2160
|
15002 |
+
},
|
15003 |
+
{
|
15004 |
+
"epoch": 540.0,
|
15005 |
+
"eval_accuracy_ELSE": 1.0,
|
15006 |
+
"eval_accuracy_X": NaN,
|
15007 |
+
"eval_accuracy_parking_lines": 0.0,
|
15008 |
+
"eval_accuracy_unlabeled": NaN,
|
15009 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15010 |
+
"eval_iou_X": NaN,
|
15011 |
+
"eval_iou_parking_lines": 0.0,
|
15012 |
+
"eval_iou_unlabeled": NaN,
|
15013 |
+
"eval_loss": 0.08665323257446289,
|
15014 |
+
"eval_mean_accuracy": 0.5,
|
15015 |
+
"eval_mean_iou": 0.49420738220214844,
|
15016 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15017 |
+
"eval_runtime": 0.8938,
|
15018 |
+
"eval_samples_per_second": 1.119,
|
15019 |
+
"eval_steps_per_second": 1.119,
|
15020 |
+
"step": 2160
|
15021 |
+
},
|
15022 |
+
{
|
15023 |
+
"epoch": 540.25,
|
15024 |
+
"learning_rate": 5.2412280701754394e-06,
|
15025 |
+
"loss": 0.0045,
|
15026 |
+
"step": 2161
|
15027 |
+
},
|
15028 |
+
{
|
15029 |
+
"epoch": 540.5,
|
15030 |
+
"learning_rate": 5.219298245614036e-06,
|
15031 |
+
"loss": 0.0039,
|
15032 |
+
"step": 2162
|
15033 |
+
},
|
15034 |
+
{
|
15035 |
+
"epoch": 540.75,
|
15036 |
+
"learning_rate": 5.197368421052632e-06,
|
15037 |
+
"loss": 0.0047,
|
15038 |
+
"step": 2163
|
15039 |
+
},
|
15040 |
+
{
|
15041 |
+
"epoch": 541.0,
|
15042 |
+
"learning_rate": 5.1754385964912286e-06,
|
15043 |
+
"loss": 0.0047,
|
15044 |
+
"step": 2164
|
15045 |
+
},
|
15046 |
+
{
|
15047 |
+
"epoch": 541.25,
|
15048 |
+
"learning_rate": 5.153508771929825e-06,
|
15049 |
+
"loss": 0.0041,
|
15050 |
+
"step": 2165
|
15051 |
+
},
|
15052 |
+
{
|
15053 |
+
"epoch": 541.5,
|
15054 |
+
"learning_rate": 5.131578947368421e-06,
|
15055 |
+
"loss": 0.0039,
|
15056 |
+
"step": 2166
|
15057 |
+
},
|
15058 |
+
{
|
15059 |
+
"epoch": 541.75,
|
15060 |
+
"learning_rate": 5.109649122807018e-06,
|
15061 |
+
"loss": 0.0047,
|
15062 |
+
"step": 2167
|
15063 |
+
},
|
15064 |
+
{
|
15065 |
+
"epoch": 542.0,
|
15066 |
+
"learning_rate": 5.087719298245614e-06,
|
15067 |
+
"loss": 0.0044,
|
15068 |
+
"step": 2168
|
15069 |
+
},
|
15070 |
+
{
|
15071 |
+
"epoch": 542.25,
|
15072 |
+
"learning_rate": 5.0657894736842104e-06,
|
15073 |
+
"loss": 0.0056,
|
15074 |
+
"step": 2169
|
15075 |
+
},
|
15076 |
+
{
|
15077 |
+
"epoch": 542.5,
|
15078 |
+
"learning_rate": 5.043859649122807e-06,
|
15079 |
+
"loss": 0.0035,
|
15080 |
+
"step": 2170
|
15081 |
+
},
|
15082 |
+
{
|
15083 |
+
"epoch": 542.75,
|
15084 |
+
"learning_rate": 5.021929824561404e-06,
|
15085 |
+
"loss": 0.004,
|
15086 |
+
"step": 2171
|
15087 |
+
},
|
15088 |
+
{
|
15089 |
+
"epoch": 543.0,
|
15090 |
+
"learning_rate": 5e-06,
|
15091 |
+
"loss": 0.0045,
|
15092 |
+
"step": 2172
|
15093 |
+
},
|
15094 |
+
{
|
15095 |
+
"epoch": 543.25,
|
15096 |
+
"learning_rate": 4.978070175438597e-06,
|
15097 |
+
"loss": 0.0034,
|
15098 |
+
"step": 2173
|
15099 |
+
},
|
15100 |
+
{
|
15101 |
+
"epoch": 543.5,
|
15102 |
+
"learning_rate": 4.956140350877193e-06,
|
15103 |
+
"loss": 0.0044,
|
15104 |
+
"step": 2174
|
15105 |
+
},
|
15106 |
+
{
|
15107 |
+
"epoch": 543.75,
|
15108 |
+
"learning_rate": 4.9342105263157895e-06,
|
15109 |
+
"loss": 0.0051,
|
15110 |
+
"step": 2175
|
15111 |
+
},
|
15112 |
+
{
|
15113 |
+
"epoch": 544.0,
|
15114 |
+
"learning_rate": 4.912280701754386e-06,
|
15115 |
+
"loss": 0.0042,
|
15116 |
+
"step": 2176
|
15117 |
+
},
|
15118 |
+
{
|
15119 |
+
"epoch": 544.25,
|
15120 |
+
"learning_rate": 4.890350877192982e-06,
|
15121 |
+
"loss": 0.0033,
|
15122 |
+
"step": 2177
|
15123 |
+
},
|
15124 |
+
{
|
15125 |
+
"epoch": 544.5,
|
15126 |
+
"learning_rate": 4.868421052631579e-06,
|
15127 |
+
"loss": 0.0043,
|
15128 |
+
"step": 2178
|
15129 |
+
},
|
15130 |
+
{
|
15131 |
+
"epoch": 544.75,
|
15132 |
+
"learning_rate": 4.846491228070176e-06,
|
15133 |
+
"loss": 0.0044,
|
15134 |
+
"step": 2179
|
15135 |
+
},
|
15136 |
+
{
|
15137 |
+
"epoch": 545.0,
|
15138 |
+
"learning_rate": 4.824561403508772e-06,
|
15139 |
+
"loss": 0.0054,
|
15140 |
+
"step": 2180
|
15141 |
+
},
|
15142 |
+
{
|
15143 |
+
"epoch": 545.0,
|
15144 |
+
"eval_accuracy_ELSE": 1.0,
|
15145 |
+
"eval_accuracy_X": NaN,
|
15146 |
+
"eval_accuracy_parking_lines": 0.0,
|
15147 |
+
"eval_accuracy_unlabeled": NaN,
|
15148 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15149 |
+
"eval_iou_X": NaN,
|
15150 |
+
"eval_iou_parking_lines": 0.0,
|
15151 |
+
"eval_iou_unlabeled": NaN,
|
15152 |
+
"eval_loss": 0.08619242161512375,
|
15153 |
+
"eval_mean_accuracy": 0.5,
|
15154 |
+
"eval_mean_iou": 0.49420738220214844,
|
15155 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15156 |
+
"eval_runtime": 0.8752,
|
15157 |
+
"eval_samples_per_second": 1.143,
|
15158 |
+
"eval_steps_per_second": 1.143,
|
15159 |
+
"step": 2180
|
15160 |
+
},
|
15161 |
+
{
|
15162 |
+
"epoch": 545.25,
|
15163 |
+
"learning_rate": 4.802631578947369e-06,
|
15164 |
+
"loss": 0.0037,
|
15165 |
+
"step": 2181
|
15166 |
+
},
|
15167 |
+
{
|
15168 |
+
"epoch": 545.5,
|
15169 |
+
"learning_rate": 4.780701754385966e-06,
|
15170 |
+
"loss": 0.0044,
|
15171 |
+
"step": 2182
|
15172 |
+
},
|
15173 |
+
{
|
15174 |
+
"epoch": 545.75,
|
15175 |
+
"learning_rate": 4.758771929824562e-06,
|
15176 |
+
"loss": 0.0041,
|
15177 |
+
"step": 2183
|
15178 |
+
},
|
15179 |
+
{
|
15180 |
+
"epoch": 546.0,
|
15181 |
+
"learning_rate": 4.736842105263159e-06,
|
15182 |
+
"loss": 0.0049,
|
15183 |
+
"step": 2184
|
15184 |
+
},
|
15185 |
+
{
|
15186 |
+
"epoch": 546.25,
|
15187 |
+
"learning_rate": 4.714912280701755e-06,
|
15188 |
+
"loss": 0.0045,
|
15189 |
+
"step": 2185
|
15190 |
+
},
|
15191 |
+
{
|
15192 |
+
"epoch": 546.5,
|
15193 |
+
"learning_rate": 4.692982456140351e-06,
|
15194 |
+
"loss": 0.0048,
|
15195 |
+
"step": 2186
|
15196 |
+
},
|
15197 |
+
{
|
15198 |
+
"epoch": 546.75,
|
15199 |
+
"learning_rate": 4.671052631578948e-06,
|
15200 |
+
"loss": 0.0044,
|
15201 |
+
"step": 2187
|
15202 |
+
},
|
15203 |
+
{
|
15204 |
+
"epoch": 547.0,
|
15205 |
+
"learning_rate": 4.649122807017544e-06,
|
15206 |
+
"loss": 0.0039,
|
15207 |
+
"step": 2188
|
15208 |
+
},
|
15209 |
+
{
|
15210 |
+
"epoch": 547.25,
|
15211 |
+
"learning_rate": 4.6271929824561405e-06,
|
15212 |
+
"loss": 0.0039,
|
15213 |
+
"step": 2189
|
15214 |
+
},
|
15215 |
+
{
|
15216 |
+
"epoch": 547.5,
|
15217 |
+
"learning_rate": 4.605263157894737e-06,
|
15218 |
+
"loss": 0.0048,
|
15219 |
+
"step": 2190
|
15220 |
+
},
|
15221 |
+
{
|
15222 |
+
"epoch": 547.75,
|
15223 |
+
"learning_rate": 4.583333333333333e-06,
|
15224 |
+
"loss": 0.0043,
|
15225 |
+
"step": 2191
|
15226 |
+
},
|
15227 |
+
{
|
15228 |
+
"epoch": 548.0,
|
15229 |
+
"learning_rate": 4.5614035087719304e-06,
|
15230 |
+
"loss": 0.0042,
|
15231 |
+
"step": 2192
|
15232 |
+
},
|
15233 |
+
{
|
15234 |
+
"epoch": 548.25,
|
15235 |
+
"learning_rate": 4.539473684210527e-06,
|
15236 |
+
"loss": 0.0046,
|
15237 |
+
"step": 2193
|
15238 |
+
},
|
15239 |
+
{
|
15240 |
+
"epoch": 548.5,
|
15241 |
+
"learning_rate": 4.517543859649123e-06,
|
15242 |
+
"loss": 0.0038,
|
15243 |
+
"step": 2194
|
15244 |
+
},
|
15245 |
+
{
|
15246 |
+
"epoch": 548.75,
|
15247 |
+
"learning_rate": 4.4956140350877196e-06,
|
15248 |
+
"loss": 0.0038,
|
15249 |
+
"step": 2195
|
15250 |
+
},
|
15251 |
+
{
|
15252 |
+
"epoch": 549.0,
|
15253 |
+
"learning_rate": 4.473684210526316e-06,
|
15254 |
+
"loss": 0.0048,
|
15255 |
+
"step": 2196
|
15256 |
+
},
|
15257 |
+
{
|
15258 |
+
"epoch": 549.25,
|
15259 |
+
"learning_rate": 4.451754385964912e-06,
|
15260 |
+
"loss": 0.0045,
|
15261 |
+
"step": 2197
|
15262 |
+
},
|
15263 |
+
{
|
15264 |
+
"epoch": 549.5,
|
15265 |
+
"learning_rate": 4.429824561403509e-06,
|
15266 |
+
"loss": 0.0036,
|
15267 |
+
"step": 2198
|
15268 |
+
},
|
15269 |
+
{
|
15270 |
+
"epoch": 549.75,
|
15271 |
+
"learning_rate": 4.407894736842105e-06,
|
15272 |
+
"loss": 0.0042,
|
15273 |
+
"step": 2199
|
15274 |
+
},
|
15275 |
+
{
|
15276 |
+
"epoch": 550.0,
|
15277 |
+
"learning_rate": 4.3859649122807014e-06,
|
15278 |
+
"loss": 0.0047,
|
15279 |
+
"step": 2200
|
15280 |
+
},
|
15281 |
+
{
|
15282 |
+
"epoch": 550.0,
|
15283 |
+
"eval_accuracy_ELSE": 1.0,
|
15284 |
+
"eval_accuracy_X": NaN,
|
15285 |
+
"eval_accuracy_parking_lines": 0.0,
|
15286 |
+
"eval_accuracy_unlabeled": NaN,
|
15287 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15288 |
+
"eval_iou_X": NaN,
|
15289 |
+
"eval_iou_parking_lines": 0.0,
|
15290 |
+
"eval_iou_unlabeled": NaN,
|
15291 |
+
"eval_loss": 0.08726246654987335,
|
15292 |
+
"eval_mean_accuracy": 0.5,
|
15293 |
+
"eval_mean_iou": 0.49420738220214844,
|
15294 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15295 |
+
"eval_runtime": 0.8385,
|
15296 |
+
"eval_samples_per_second": 1.193,
|
15297 |
+
"eval_steps_per_second": 1.193,
|
15298 |
+
"step": 2200
|
15299 |
+
},
|
15300 |
+
{
|
15301 |
+
"epoch": 550.25,
|
15302 |
+
"learning_rate": 4.364035087719298e-06,
|
15303 |
+
"loss": 0.0062,
|
15304 |
+
"step": 2201
|
15305 |
+
},
|
15306 |
+
{
|
15307 |
+
"epoch": 550.5,
|
15308 |
+
"learning_rate": 4.342105263157895e-06,
|
15309 |
+
"loss": 0.003,
|
15310 |
+
"step": 2202
|
15311 |
+
},
|
15312 |
+
{
|
15313 |
+
"epoch": 550.75,
|
15314 |
+
"learning_rate": 4.320175438596491e-06,
|
15315 |
+
"loss": 0.0034,
|
15316 |
+
"step": 2203
|
15317 |
+
},
|
15318 |
+
{
|
15319 |
+
"epoch": 551.0,
|
15320 |
+
"learning_rate": 4.298245614035088e-06,
|
15321 |
+
"loss": 0.0046,
|
15322 |
+
"step": 2204
|
15323 |
+
},
|
15324 |
+
{
|
15325 |
+
"epoch": 551.25,
|
15326 |
+
"learning_rate": 4.276315789473684e-06,
|
15327 |
+
"loss": 0.0043,
|
15328 |
+
"step": 2205
|
15329 |
+
},
|
15330 |
+
{
|
15331 |
+
"epoch": 551.5,
|
15332 |
+
"learning_rate": 4.254385964912281e-06,
|
15333 |
+
"loss": 0.005,
|
15334 |
+
"step": 2206
|
15335 |
+
},
|
15336 |
+
{
|
15337 |
+
"epoch": 551.75,
|
15338 |
+
"learning_rate": 4.232456140350878e-06,
|
15339 |
+
"loss": 0.004,
|
15340 |
+
"step": 2207
|
15341 |
+
},
|
15342 |
+
{
|
15343 |
+
"epoch": 552.0,
|
15344 |
+
"learning_rate": 4.210526315789474e-06,
|
15345 |
+
"loss": 0.0036,
|
15346 |
+
"step": 2208
|
15347 |
+
},
|
15348 |
+
{
|
15349 |
+
"epoch": 552.25,
|
15350 |
+
"learning_rate": 4.1885964912280705e-06,
|
15351 |
+
"loss": 0.0043,
|
15352 |
+
"step": 2209
|
15353 |
+
},
|
15354 |
+
{
|
15355 |
+
"epoch": 552.5,
|
15356 |
+
"learning_rate": 4.166666666666667e-06,
|
15357 |
+
"loss": 0.0056,
|
15358 |
+
"step": 2210
|
15359 |
+
},
|
15360 |
+
{
|
15361 |
+
"epoch": 552.75,
|
15362 |
+
"learning_rate": 4.144736842105263e-06,
|
15363 |
+
"loss": 0.0036,
|
15364 |
+
"step": 2211
|
15365 |
+
},
|
15366 |
+
{
|
15367 |
+
"epoch": 553.0,
|
15368 |
+
"learning_rate": 4.1228070175438605e-06,
|
15369 |
+
"loss": 0.0045,
|
15370 |
+
"step": 2212
|
15371 |
+
},
|
15372 |
+
{
|
15373 |
+
"epoch": 553.25,
|
15374 |
+
"learning_rate": 4.100877192982457e-06,
|
15375 |
+
"loss": 0.0044,
|
15376 |
+
"step": 2213
|
15377 |
+
},
|
15378 |
+
{
|
15379 |
+
"epoch": 553.5,
|
15380 |
+
"learning_rate": 4.078947368421053e-06,
|
15381 |
+
"loss": 0.0041,
|
15382 |
+
"step": 2214
|
15383 |
+
},
|
15384 |
+
{
|
15385 |
+
"epoch": 553.75,
|
15386 |
+
"learning_rate": 4.05701754385965e-06,
|
15387 |
+
"loss": 0.0053,
|
15388 |
+
"step": 2215
|
15389 |
+
},
|
15390 |
+
{
|
15391 |
+
"epoch": 554.0,
|
15392 |
+
"learning_rate": 4.035087719298246e-06,
|
15393 |
+
"loss": 0.0035,
|
15394 |
+
"step": 2216
|
15395 |
+
},
|
15396 |
+
{
|
15397 |
+
"epoch": 554.25,
|
15398 |
+
"learning_rate": 4.013157894736842e-06,
|
15399 |
+
"loss": 0.0037,
|
15400 |
+
"step": 2217
|
15401 |
+
},
|
15402 |
+
{
|
15403 |
+
"epoch": 554.5,
|
15404 |
+
"learning_rate": 3.991228070175439e-06,
|
15405 |
+
"loss": 0.0046,
|
15406 |
+
"step": 2218
|
15407 |
+
},
|
15408 |
+
{
|
15409 |
+
"epoch": 554.75,
|
15410 |
+
"learning_rate": 3.969298245614035e-06,
|
15411 |
+
"loss": 0.0042,
|
15412 |
+
"step": 2219
|
15413 |
+
},
|
15414 |
+
{
|
15415 |
+
"epoch": 555.0,
|
15416 |
+
"learning_rate": 3.9473684210526315e-06,
|
15417 |
+
"loss": 0.0048,
|
15418 |
+
"step": 2220
|
15419 |
+
},
|
15420 |
+
{
|
15421 |
+
"epoch": 555.0,
|
15422 |
+
"eval_accuracy_ELSE": 1.0,
|
15423 |
+
"eval_accuracy_X": NaN,
|
15424 |
+
"eval_accuracy_parking_lines": 0.0,
|
15425 |
+
"eval_accuracy_unlabeled": NaN,
|
15426 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15427 |
+
"eval_iou_X": NaN,
|
15428 |
+
"eval_iou_parking_lines": 0.0,
|
15429 |
+
"eval_iou_unlabeled": NaN,
|
15430 |
+
"eval_loss": 0.08687347173690796,
|
15431 |
+
"eval_mean_accuracy": 0.5,
|
15432 |
+
"eval_mean_iou": 0.49420738220214844,
|
15433 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15434 |
+
"eval_runtime": 1.005,
|
15435 |
+
"eval_samples_per_second": 0.995,
|
15436 |
+
"eval_steps_per_second": 0.995,
|
15437 |
+
"step": 2220
|
15438 |
+
},
|
15439 |
+
{
|
15440 |
+
"epoch": 555.25,
|
15441 |
+
"learning_rate": 3.925438596491228e-06,
|
15442 |
+
"loss": 0.0045,
|
15443 |
+
"step": 2221
|
15444 |
+
},
|
15445 |
+
{
|
15446 |
+
"epoch": 555.5,
|
15447 |
+
"learning_rate": 3.903508771929825e-06,
|
15448 |
+
"loss": 0.0042,
|
15449 |
+
"step": 2222
|
15450 |
+
},
|
15451 |
+
{
|
15452 |
+
"epoch": 555.75,
|
15453 |
+
"learning_rate": 3.8815789473684214e-06,
|
15454 |
+
"loss": 0.0037,
|
15455 |
+
"step": 2223
|
15456 |
+
},
|
15457 |
+
{
|
15458 |
+
"epoch": 556.0,
|
15459 |
+
"learning_rate": 3.859649122807018e-06,
|
15460 |
+
"loss": 0.005,
|
15461 |
+
"step": 2224
|
15462 |
+
},
|
15463 |
+
{
|
15464 |
+
"epoch": 556.25,
|
15465 |
+
"learning_rate": 3.837719298245614e-06,
|
15466 |
+
"loss": 0.0035,
|
15467 |
+
"step": 2225
|
15468 |
+
},
|
15469 |
+
{
|
15470 |
+
"epoch": 556.5,
|
15471 |
+
"learning_rate": 3.8157894736842105e-06,
|
15472 |
+
"loss": 0.0051,
|
15473 |
+
"step": 2226
|
15474 |
+
},
|
15475 |
+
{
|
15476 |
+
"epoch": 556.75,
|
15477 |
+
"learning_rate": 3.7938596491228073e-06,
|
15478 |
+
"loss": 0.0053,
|
15479 |
+
"step": 2227
|
15480 |
+
},
|
15481 |
+
{
|
15482 |
+
"epoch": 557.0,
|
15483 |
+
"learning_rate": 3.7719298245614037e-06,
|
15484 |
+
"loss": 0.0037,
|
15485 |
+
"step": 2228
|
15486 |
+
},
|
15487 |
+
{
|
15488 |
+
"epoch": 557.25,
|
15489 |
+
"learning_rate": 3.75e-06,
|
15490 |
+
"loss": 0.004,
|
15491 |
+
"step": 2229
|
15492 |
+
},
|
15493 |
+
{
|
15494 |
+
"epoch": 557.5,
|
15495 |
+
"learning_rate": 3.7280701754385965e-06,
|
15496 |
+
"loss": 0.0032,
|
15497 |
+
"step": 2230
|
15498 |
+
},
|
15499 |
+
{
|
15500 |
+
"epoch": 557.75,
|
15501 |
+
"learning_rate": 3.706140350877193e-06,
|
15502 |
+
"loss": 0.0038,
|
15503 |
+
"step": 2231
|
15504 |
+
},
|
15505 |
+
{
|
15506 |
+
"epoch": 558.0,
|
15507 |
+
"learning_rate": 3.6842105263157892e-06,
|
15508 |
+
"loss": 0.0066,
|
15509 |
+
"step": 2232
|
15510 |
+
},
|
15511 |
+
{
|
15512 |
+
"epoch": 558.25,
|
15513 |
+
"learning_rate": 3.6622807017543864e-06,
|
15514 |
+
"loss": 0.0056,
|
15515 |
+
"step": 2233
|
15516 |
+
},
|
15517 |
+
{
|
15518 |
+
"epoch": 558.5,
|
15519 |
+
"learning_rate": 3.640350877192983e-06,
|
15520 |
+
"loss": 0.0047,
|
15521 |
+
"step": 2234
|
15522 |
+
},
|
15523 |
+
{
|
15524 |
+
"epoch": 558.75,
|
15525 |
+
"learning_rate": 3.618421052631579e-06,
|
15526 |
+
"loss": 0.0036,
|
15527 |
+
"step": 2235
|
15528 |
+
},
|
15529 |
+
{
|
15530 |
+
"epoch": 559.0,
|
15531 |
+
"learning_rate": 3.5964912280701756e-06,
|
15532 |
+
"loss": 0.0033,
|
15533 |
+
"step": 2236
|
15534 |
+
},
|
15535 |
+
{
|
15536 |
+
"epoch": 559.25,
|
15537 |
+
"learning_rate": 3.574561403508772e-06,
|
15538 |
+
"loss": 0.0035,
|
15539 |
+
"step": 2237
|
15540 |
+
},
|
15541 |
+
{
|
15542 |
+
"epoch": 559.5,
|
15543 |
+
"learning_rate": 3.5526315789473687e-06,
|
15544 |
+
"loss": 0.0052,
|
15545 |
+
"step": 2238
|
15546 |
+
},
|
15547 |
+
{
|
15548 |
+
"epoch": 559.75,
|
15549 |
+
"learning_rate": 3.530701754385965e-06,
|
15550 |
+
"loss": 0.0036,
|
15551 |
+
"step": 2239
|
15552 |
+
},
|
15553 |
+
{
|
15554 |
+
"epoch": 560.0,
|
15555 |
+
"learning_rate": 3.5087719298245615e-06,
|
15556 |
+
"loss": 0.0045,
|
15557 |
+
"step": 2240
|
15558 |
+
},
|
15559 |
+
{
|
15560 |
+
"epoch": 560.0,
|
15561 |
+
"eval_accuracy_ELSE": 1.0,
|
15562 |
+
"eval_accuracy_X": NaN,
|
15563 |
+
"eval_accuracy_parking_lines": 0.0,
|
15564 |
+
"eval_accuracy_unlabeled": NaN,
|
15565 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15566 |
+
"eval_iou_X": NaN,
|
15567 |
+
"eval_iou_parking_lines": 0.0,
|
15568 |
+
"eval_iou_unlabeled": NaN,
|
15569 |
+
"eval_loss": 0.08646020293235779,
|
15570 |
+
"eval_mean_accuracy": 0.5,
|
15571 |
+
"eval_mean_iou": 0.49420738220214844,
|
15572 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15573 |
+
"eval_runtime": 0.8637,
|
15574 |
+
"eval_samples_per_second": 1.158,
|
15575 |
+
"eval_steps_per_second": 1.158,
|
15576 |
+
"step": 2240
|
15577 |
+
},
|
15578 |
+
{
|
15579 |
+
"epoch": 560.25,
|
15580 |
+
"learning_rate": 3.486842105263158e-06,
|
15581 |
+
"loss": 0.005,
|
15582 |
+
"step": 2241
|
15583 |
+
},
|
15584 |
+
{
|
15585 |
+
"epoch": 560.5,
|
15586 |
+
"learning_rate": 3.4649122807017542e-06,
|
15587 |
+
"loss": 0.0038,
|
15588 |
+
"step": 2242
|
15589 |
+
},
|
15590 |
+
{
|
15591 |
+
"epoch": 560.75,
|
15592 |
+
"learning_rate": 3.4429824561403515e-06,
|
15593 |
+
"loss": 0.0045,
|
15594 |
+
"step": 2243
|
15595 |
+
},
|
15596 |
+
{
|
15597 |
+
"epoch": 561.0,
|
15598 |
+
"learning_rate": 3.421052631578948e-06,
|
15599 |
+
"loss": 0.0046,
|
15600 |
+
"step": 2244
|
15601 |
+
},
|
15602 |
+
{
|
15603 |
+
"epoch": 561.25,
|
15604 |
+
"learning_rate": 3.399122807017544e-06,
|
15605 |
+
"loss": 0.0029,
|
15606 |
+
"step": 2245
|
15607 |
+
},
|
15608 |
+
{
|
15609 |
+
"epoch": 561.5,
|
15610 |
+
"learning_rate": 3.3771929824561406e-06,
|
15611 |
+
"loss": 0.0045,
|
15612 |
+
"step": 2246
|
15613 |
+
},
|
15614 |
+
{
|
15615 |
+
"epoch": 561.75,
|
15616 |
+
"learning_rate": 3.355263157894737e-06,
|
15617 |
+
"loss": 0.0042,
|
15618 |
+
"step": 2247
|
15619 |
+
},
|
15620 |
+
{
|
15621 |
+
"epoch": 562.0,
|
15622 |
+
"learning_rate": 3.3333333333333333e-06,
|
15623 |
+
"loss": 0.0057,
|
15624 |
+
"step": 2248
|
15625 |
+
},
|
15626 |
+
{
|
15627 |
+
"epoch": 562.25,
|
15628 |
+
"learning_rate": 3.3114035087719297e-06,
|
15629 |
+
"loss": 0.0044,
|
15630 |
+
"step": 2249
|
15631 |
+
},
|
15632 |
+
{
|
15633 |
+
"epoch": 562.5,
|
15634 |
+
"learning_rate": 3.2894736842105265e-06,
|
15635 |
+
"loss": 0.0049,
|
15636 |
+
"step": 2250
|
15637 |
+
},
|
15638 |
+
{
|
15639 |
+
"epoch": 562.75,
|
15640 |
+
"learning_rate": 3.267543859649123e-06,
|
15641 |
+
"loss": 0.0035,
|
15642 |
+
"step": 2251
|
15643 |
+
},
|
15644 |
+
{
|
15645 |
+
"epoch": 563.0,
|
15646 |
+
"learning_rate": 3.2456140350877192e-06,
|
15647 |
+
"loss": 0.0042,
|
15648 |
+
"step": 2252
|
15649 |
+
},
|
15650 |
+
{
|
15651 |
+
"epoch": 563.25,
|
15652 |
+
"learning_rate": 3.223684210526316e-06,
|
15653 |
+
"loss": 0.0045,
|
15654 |
+
"step": 2253
|
15655 |
+
},
|
15656 |
+
{
|
15657 |
+
"epoch": 563.5,
|
15658 |
+
"learning_rate": 3.201754385964913e-06,
|
15659 |
+
"loss": 0.0036,
|
15660 |
+
"step": 2254
|
15661 |
+
},
|
15662 |
+
{
|
15663 |
+
"epoch": 563.75,
|
15664 |
+
"learning_rate": 3.1798245614035092e-06,
|
15665 |
+
"loss": 0.0053,
|
15666 |
+
"step": 2255
|
15667 |
+
},
|
15668 |
+
{
|
15669 |
+
"epoch": 564.0,
|
15670 |
+
"learning_rate": 3.1578947368421056e-06,
|
15671 |
+
"loss": 0.0036,
|
15672 |
+
"step": 2256
|
15673 |
+
},
|
15674 |
+
{
|
15675 |
+
"epoch": 564.25,
|
15676 |
+
"learning_rate": 3.135964912280702e-06,
|
15677 |
+
"loss": 0.0028,
|
15678 |
+
"step": 2257
|
15679 |
+
},
|
15680 |
+
{
|
15681 |
+
"epoch": 564.5,
|
15682 |
+
"learning_rate": 3.1140350877192983e-06,
|
15683 |
+
"loss": 0.0066,
|
15684 |
+
"step": 2258
|
15685 |
+
},
|
15686 |
+
{
|
15687 |
+
"epoch": 564.75,
|
15688 |
+
"learning_rate": 3.0921052631578947e-06,
|
15689 |
+
"loss": 0.0034,
|
15690 |
+
"step": 2259
|
15691 |
+
},
|
15692 |
+
{
|
15693 |
+
"epoch": 565.0,
|
15694 |
+
"learning_rate": 3.070175438596491e-06,
|
15695 |
+
"loss": 0.0049,
|
15696 |
+
"step": 2260
|
15697 |
+
},
|
15698 |
+
{
|
15699 |
+
"epoch": 565.0,
|
15700 |
+
"eval_accuracy_ELSE": 1.0,
|
15701 |
+
"eval_accuracy_X": NaN,
|
15702 |
+
"eval_accuracy_parking_lines": 0.0,
|
15703 |
+
"eval_accuracy_unlabeled": NaN,
|
15704 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15705 |
+
"eval_iou_X": NaN,
|
15706 |
+
"eval_iou_parking_lines": 0.0,
|
15707 |
+
"eval_iou_unlabeled": NaN,
|
15708 |
+
"eval_loss": 0.08613073825836182,
|
15709 |
+
"eval_mean_accuracy": 0.5,
|
15710 |
+
"eval_mean_iou": 0.49420738220214844,
|
15711 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15712 |
+
"eval_runtime": 0.8878,
|
15713 |
+
"eval_samples_per_second": 1.126,
|
15714 |
+
"eval_steps_per_second": 1.126,
|
15715 |
+
"step": 2260
|
15716 |
+
},
|
15717 |
+
{
|
15718 |
+
"epoch": 565.25,
|
15719 |
+
"learning_rate": 3.048245614035088e-06,
|
15720 |
+
"loss": 0.0035,
|
15721 |
+
"step": 2261
|
15722 |
+
},
|
15723 |
+
{
|
15724 |
+
"epoch": 565.5,
|
15725 |
+
"learning_rate": 3.0263157894736843e-06,
|
15726 |
+
"loss": 0.004,
|
15727 |
+
"step": 2262
|
15728 |
+
},
|
15729 |
+
{
|
15730 |
+
"epoch": 565.75,
|
15731 |
+
"learning_rate": 3.0043859649122806e-06,
|
15732 |
+
"loss": 0.0048,
|
15733 |
+
"step": 2263
|
15734 |
+
},
|
15735 |
+
{
|
15736 |
+
"epoch": 566.0,
|
15737 |
+
"learning_rate": 2.9824561403508774e-06,
|
15738 |
+
"loss": 0.0045,
|
15739 |
+
"step": 2264
|
15740 |
+
},
|
15741 |
+
{
|
15742 |
+
"epoch": 566.25,
|
15743 |
+
"learning_rate": 2.960526315789474e-06,
|
15744 |
+
"loss": 0.0046,
|
15745 |
+
"step": 2265
|
15746 |
+
},
|
15747 |
+
{
|
15748 |
+
"epoch": 566.5,
|
15749 |
+
"learning_rate": 2.9385964912280706e-06,
|
15750 |
+
"loss": 0.0036,
|
15751 |
+
"step": 2266
|
15752 |
+
},
|
15753 |
+
{
|
15754 |
+
"epoch": 566.75,
|
15755 |
+
"learning_rate": 2.916666666666667e-06,
|
15756 |
+
"loss": 0.0051,
|
15757 |
+
"step": 2267
|
15758 |
+
},
|
15759 |
+
{
|
15760 |
+
"epoch": 567.0,
|
15761 |
+
"learning_rate": 2.8947368421052634e-06,
|
15762 |
+
"loss": 0.0039,
|
15763 |
+
"step": 2268
|
15764 |
+
},
|
15765 |
+
{
|
15766 |
+
"epoch": 567.25,
|
15767 |
+
"learning_rate": 2.8728070175438597e-06,
|
15768 |
+
"loss": 0.0057,
|
15769 |
+
"step": 2269
|
15770 |
+
},
|
15771 |
+
{
|
15772 |
+
"epoch": 567.5,
|
15773 |
+
"learning_rate": 2.850877192982456e-06,
|
15774 |
+
"loss": 0.0033,
|
15775 |
+
"step": 2270
|
15776 |
+
},
|
15777 |
+
{
|
15778 |
+
"epoch": 567.75,
|
15779 |
+
"learning_rate": 2.828947368421053e-06,
|
15780 |
+
"loss": 0.0044,
|
15781 |
+
"step": 2271
|
15782 |
+
},
|
15783 |
+
{
|
15784 |
+
"epoch": 568.0,
|
15785 |
+
"learning_rate": 2.8070175438596493e-06,
|
15786 |
+
"loss": 0.0044,
|
15787 |
+
"step": 2272
|
15788 |
+
},
|
15789 |
+
{
|
15790 |
+
"epoch": 568.25,
|
15791 |
+
"learning_rate": 2.7850877192982456e-06,
|
15792 |
+
"loss": 0.0027,
|
15793 |
+
"step": 2273
|
15794 |
+
},
|
15795 |
+
{
|
15796 |
+
"epoch": 568.5,
|
15797 |
+
"learning_rate": 2.763157894736842e-06,
|
15798 |
+
"loss": 0.0051,
|
15799 |
+
"step": 2274
|
15800 |
+
},
|
15801 |
+
{
|
15802 |
+
"epoch": 568.75,
|
15803 |
+
"learning_rate": 2.7412280701754384e-06,
|
15804 |
+
"loss": 0.0041,
|
15805 |
+
"step": 2275
|
15806 |
+
},
|
15807 |
+
{
|
15808 |
+
"epoch": 569.0,
|
15809 |
+
"learning_rate": 2.719298245614035e-06,
|
15810 |
+
"loss": 0.0048,
|
15811 |
+
"step": 2276
|
15812 |
+
},
|
15813 |
+
{
|
15814 |
+
"epoch": 569.25,
|
15815 |
+
"learning_rate": 2.6973684210526316e-06,
|
15816 |
+
"loss": 0.0043,
|
15817 |
+
"step": 2277
|
15818 |
+
},
|
15819 |
+
{
|
15820 |
+
"epoch": 569.5,
|
15821 |
+
"learning_rate": 2.6754385964912284e-06,
|
15822 |
+
"loss": 0.0038,
|
15823 |
+
"step": 2278
|
15824 |
+
},
|
15825 |
+
{
|
15826 |
+
"epoch": 569.75,
|
15827 |
+
"learning_rate": 2.6535087719298247e-06,
|
15828 |
+
"loss": 0.004,
|
15829 |
+
"step": 2279
|
15830 |
+
},
|
15831 |
+
{
|
15832 |
+
"epoch": 570.0,
|
15833 |
+
"learning_rate": 2.631578947368421e-06,
|
15834 |
+
"loss": 0.005,
|
15835 |
+
"step": 2280
|
15836 |
+
},
|
15837 |
+
{
|
15838 |
+
"epoch": 570.0,
|
15839 |
+
"eval_accuracy_ELSE": 1.0,
|
15840 |
+
"eval_accuracy_X": NaN,
|
15841 |
+
"eval_accuracy_parking_lines": 0.0,
|
15842 |
+
"eval_accuracy_unlabeled": NaN,
|
15843 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15844 |
+
"eval_iou_X": NaN,
|
15845 |
+
"eval_iou_parking_lines": 0.0,
|
15846 |
+
"eval_iou_unlabeled": NaN,
|
15847 |
+
"eval_loss": 0.08558374643325806,
|
15848 |
+
"eval_mean_accuracy": 0.5,
|
15849 |
+
"eval_mean_iou": 0.49420738220214844,
|
15850 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15851 |
+
"eval_runtime": 0.9203,
|
15852 |
+
"eval_samples_per_second": 1.087,
|
15853 |
+
"eval_steps_per_second": 1.087,
|
15854 |
+
"step": 2280
|
15855 |
+
},
|
15856 |
+
{
|
15857 |
+
"epoch": 570.25,
|
15858 |
+
"learning_rate": 2.609649122807018e-06,
|
15859 |
+
"loss": 0.0052,
|
15860 |
+
"step": 2281
|
15861 |
+
},
|
15862 |
+
{
|
15863 |
+
"epoch": 570.5,
|
15864 |
+
"learning_rate": 2.5877192982456143e-06,
|
15865 |
+
"loss": 0.0051,
|
15866 |
+
"step": 2282
|
15867 |
+
},
|
15868 |
+
{
|
15869 |
+
"epoch": 570.75,
|
15870 |
+
"learning_rate": 2.5657894736842107e-06,
|
15871 |
+
"loss": 0.0035,
|
15872 |
+
"step": 2283
|
15873 |
+
},
|
15874 |
+
{
|
15875 |
+
"epoch": 571.0,
|
15876 |
+
"learning_rate": 2.543859649122807e-06,
|
15877 |
+
"loss": 0.0036,
|
15878 |
+
"step": 2284
|
15879 |
+
},
|
15880 |
+
{
|
15881 |
+
"epoch": 571.25,
|
15882 |
+
"learning_rate": 2.5219298245614034e-06,
|
15883 |
+
"loss": 0.0036,
|
15884 |
+
"step": 2285
|
15885 |
+
},
|
15886 |
+
{
|
15887 |
+
"epoch": 571.5,
|
15888 |
+
"learning_rate": 2.5e-06,
|
15889 |
+
"loss": 0.0056,
|
15890 |
+
"step": 2286
|
15891 |
+
},
|
15892 |
+
{
|
15893 |
+
"epoch": 571.75,
|
15894 |
+
"learning_rate": 2.4780701754385966e-06,
|
15895 |
+
"loss": 0.0034,
|
15896 |
+
"step": 2287
|
15897 |
+
},
|
15898 |
+
{
|
15899 |
+
"epoch": 572.0,
|
15900 |
+
"learning_rate": 2.456140350877193e-06,
|
15901 |
+
"loss": 0.0042,
|
15902 |
+
"step": 2288
|
15903 |
+
},
|
15904 |
+
{
|
15905 |
+
"epoch": 572.25,
|
15906 |
+
"learning_rate": 2.4342105263157893e-06,
|
15907 |
+
"loss": 0.0051,
|
15908 |
+
"step": 2289
|
15909 |
+
},
|
15910 |
+
{
|
15911 |
+
"epoch": 572.5,
|
15912 |
+
"learning_rate": 2.412280701754386e-06,
|
15913 |
+
"loss": 0.0042,
|
15914 |
+
"step": 2290
|
15915 |
+
},
|
15916 |
+
{
|
15917 |
+
"epoch": 572.75,
|
15918 |
+
"learning_rate": 2.390350877192983e-06,
|
15919 |
+
"loss": 0.004,
|
15920 |
+
"step": 2291
|
15921 |
+
},
|
15922 |
+
{
|
15923 |
+
"epoch": 573.0,
|
15924 |
+
"learning_rate": 2.3684210526315793e-06,
|
15925 |
+
"loss": 0.0047,
|
15926 |
+
"step": 2292
|
15927 |
+
},
|
15928 |
+
{
|
15929 |
+
"epoch": 573.25,
|
15930 |
+
"learning_rate": 2.3464912280701757e-06,
|
15931 |
+
"loss": 0.0034,
|
15932 |
+
"step": 2293
|
15933 |
+
},
|
15934 |
+
{
|
15935 |
+
"epoch": 573.5,
|
15936 |
+
"learning_rate": 2.324561403508772e-06,
|
15937 |
+
"loss": 0.0044,
|
15938 |
+
"step": 2294
|
15939 |
+
},
|
15940 |
+
{
|
15941 |
+
"epoch": 573.75,
|
15942 |
+
"learning_rate": 2.3026315789473684e-06,
|
15943 |
+
"loss": 0.0052,
|
15944 |
+
"step": 2295
|
15945 |
+
},
|
15946 |
+
{
|
15947 |
+
"epoch": 574.0,
|
15948 |
+
"learning_rate": 2.2807017543859652e-06,
|
15949 |
+
"loss": 0.0042,
|
15950 |
+
"step": 2296
|
15951 |
+
},
|
15952 |
+
{
|
15953 |
+
"epoch": 574.25,
|
15954 |
+
"learning_rate": 2.2587719298245616e-06,
|
15955 |
+
"loss": 0.0044,
|
15956 |
+
"step": 2297
|
15957 |
+
},
|
15958 |
+
{
|
15959 |
+
"epoch": 574.5,
|
15960 |
+
"learning_rate": 2.236842105263158e-06,
|
15961 |
+
"loss": 0.0036,
|
15962 |
+
"step": 2298
|
15963 |
+
},
|
15964 |
+
{
|
15965 |
+
"epoch": 574.75,
|
15966 |
+
"learning_rate": 2.2149122807017543e-06,
|
15967 |
+
"loss": 0.0048,
|
15968 |
+
"step": 2299
|
15969 |
+
},
|
15970 |
+
{
|
15971 |
+
"epoch": 575.0,
|
15972 |
+
"learning_rate": 2.1929824561403507e-06,
|
15973 |
+
"loss": 0.0043,
|
15974 |
+
"step": 2300
|
15975 |
+
},
|
15976 |
+
{
|
15977 |
+
"epoch": 575.0,
|
15978 |
+
"eval_accuracy_ELSE": 1.0,
|
15979 |
+
"eval_accuracy_X": NaN,
|
15980 |
+
"eval_accuracy_parking_lines": 0.0,
|
15981 |
+
"eval_accuracy_unlabeled": NaN,
|
15982 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
15983 |
+
"eval_iou_X": NaN,
|
15984 |
+
"eval_iou_parking_lines": 0.0,
|
15985 |
+
"eval_iou_unlabeled": NaN,
|
15986 |
+
"eval_loss": 0.08545289933681488,
|
15987 |
+
"eval_mean_accuracy": 0.5,
|
15988 |
+
"eval_mean_iou": 0.49420738220214844,
|
15989 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
15990 |
+
"eval_runtime": 0.9167,
|
15991 |
+
"eval_samples_per_second": 1.091,
|
15992 |
+
"eval_steps_per_second": 1.091,
|
15993 |
+
"step": 2300
|
15994 |
+
},
|
15995 |
+
{
|
15996 |
+
"epoch": 575.25,
|
15997 |
+
"learning_rate": 2.1710526315789475e-06,
|
15998 |
+
"loss": 0.0034,
|
15999 |
+
"step": 2301
|
16000 |
+
},
|
16001 |
+
{
|
16002 |
+
"epoch": 575.5,
|
16003 |
+
"learning_rate": 2.149122807017544e-06,
|
16004 |
+
"loss": 0.0051,
|
16005 |
+
"step": 2302
|
16006 |
+
},
|
16007 |
+
{
|
16008 |
+
"epoch": 575.75,
|
16009 |
+
"learning_rate": 2.1271929824561407e-06,
|
16010 |
+
"loss": 0.0042,
|
16011 |
+
"step": 2303
|
16012 |
+
},
|
16013 |
+
{
|
16014 |
+
"epoch": 576.0,
|
16015 |
+
"learning_rate": 2.105263157894737e-06,
|
16016 |
+
"loss": 0.0052,
|
16017 |
+
"step": 2304
|
16018 |
+
},
|
16019 |
+
{
|
16020 |
+
"epoch": 576.25,
|
16021 |
+
"learning_rate": 2.0833333333333334e-06,
|
16022 |
+
"loss": 0.0042,
|
16023 |
+
"step": 2305
|
16024 |
+
},
|
16025 |
+
{
|
16026 |
+
"epoch": 576.5,
|
16027 |
+
"learning_rate": 2.0614035087719302e-06,
|
16028 |
+
"loss": 0.0048,
|
16029 |
+
"step": 2306
|
16030 |
+
},
|
16031 |
+
{
|
16032 |
+
"epoch": 576.75,
|
16033 |
+
"learning_rate": 2.0394736842105266e-06,
|
16034 |
+
"loss": 0.0044,
|
16035 |
+
"step": 2307
|
16036 |
+
},
|
16037 |
+
{
|
16038 |
+
"epoch": 577.0,
|
16039 |
+
"learning_rate": 2.017543859649123e-06,
|
16040 |
+
"loss": 0.0033,
|
16041 |
+
"step": 2308
|
16042 |
+
},
|
16043 |
+
{
|
16044 |
+
"epoch": 577.25,
|
16045 |
+
"learning_rate": 1.9956140350877194e-06,
|
16046 |
+
"loss": 0.0041,
|
16047 |
+
"step": 2309
|
16048 |
+
},
|
16049 |
+
{
|
16050 |
+
"epoch": 577.5,
|
16051 |
+
"learning_rate": 1.9736842105263157e-06,
|
16052 |
+
"loss": 0.0045,
|
16053 |
+
"step": 2310
|
16054 |
+
},
|
16055 |
+
{
|
16056 |
+
"epoch": 577.75,
|
16057 |
+
"learning_rate": 1.9517543859649125e-06,
|
16058 |
+
"loss": 0.0039,
|
16059 |
+
"step": 2311
|
16060 |
+
},
|
16061 |
+
{
|
16062 |
+
"epoch": 578.0,
|
16063 |
+
"learning_rate": 1.929824561403509e-06,
|
16064 |
+
"loss": 0.0042,
|
16065 |
+
"step": 2312
|
16066 |
+
},
|
16067 |
+
{
|
16068 |
+
"epoch": 578.25,
|
16069 |
+
"learning_rate": 1.9078947368421053e-06,
|
16070 |
+
"loss": 0.0045,
|
16071 |
+
"step": 2313
|
16072 |
+
},
|
16073 |
+
{
|
16074 |
+
"epoch": 578.5,
|
16075 |
+
"learning_rate": 1.8859649122807019e-06,
|
16076 |
+
"loss": 0.0037,
|
16077 |
+
"step": 2314
|
16078 |
+
},
|
16079 |
+
{
|
16080 |
+
"epoch": 578.75,
|
16081 |
+
"learning_rate": 1.8640350877192982e-06,
|
16082 |
+
"loss": 0.0034,
|
16083 |
+
"step": 2315
|
16084 |
+
},
|
16085 |
+
{
|
16086 |
+
"epoch": 579.0,
|
16087 |
+
"learning_rate": 1.8421052631578946e-06,
|
16088 |
+
"loss": 0.0071,
|
16089 |
+
"step": 2316
|
16090 |
+
},
|
16091 |
+
{
|
16092 |
+
"epoch": 579.25,
|
16093 |
+
"learning_rate": 1.8201754385964914e-06,
|
16094 |
+
"loss": 0.0046,
|
16095 |
+
"step": 2317
|
16096 |
+
},
|
16097 |
+
{
|
16098 |
+
"epoch": 579.5,
|
16099 |
+
"learning_rate": 1.7982456140350878e-06,
|
16100 |
+
"loss": 0.0044,
|
16101 |
+
"step": 2318
|
16102 |
+
},
|
16103 |
+
{
|
16104 |
+
"epoch": 579.75,
|
16105 |
+
"learning_rate": 1.7763157894736844e-06,
|
16106 |
+
"loss": 0.0042,
|
16107 |
+
"step": 2319
|
16108 |
+
},
|
16109 |
+
{
|
16110 |
+
"epoch": 580.0,
|
16111 |
+
"learning_rate": 1.7543859649122807e-06,
|
16112 |
+
"loss": 0.0031,
|
16113 |
+
"step": 2320
|
16114 |
+
},
|
16115 |
+
{
|
16116 |
+
"epoch": 580.0,
|
16117 |
+
"eval_accuracy_ELSE": 1.0,
|
16118 |
+
"eval_accuracy_X": NaN,
|
16119 |
+
"eval_accuracy_parking_lines": 0.0,
|
16120 |
+
"eval_accuracy_unlabeled": NaN,
|
16121 |
+
"eval_iou_ELSE": 0.9884147644042969,
|
16122 |
+
"eval_iou_X": NaN,
|
16123 |
+
"eval_iou_parking_lines": 0.0,
|
16124 |
+
"eval_iou_unlabeled": NaN,
|
16125 |
+
"eval_loss": 0.08664220571517944,
|
16126 |
+
"eval_mean_accuracy": 0.5,
|
16127 |
+
"eval_mean_iou": 0.49420738220214844,
|
16128 |
+
"eval_overall_accuracy": 0.9884147644042969,
|
16129 |
+
"eval_runtime": 0.8636,
|
16130 |
+
"eval_samples_per_second": 1.158,
|
16131 |
+
"eval_steps_per_second": 1.158,
|
16132 |
+
"step": 2320
|
16133 |
}
|
16134 |
],
|
16135 |
"max_steps": 2400,
|
16136 |
"num_train_epochs": 600,
|
16137 |
+
"total_flos": 1.4585890698008986e+19,
|
16138 |
"trial_name": null,
|
16139 |
"trial_params": null
|
16140 |
}
|
{checkpoint-2000 β checkpoint-2320}/training_args.bin
RENAMED
File without changes
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338799561
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30f2cfebe23afbdeec3f33b5f0aee22d89a51a708ecf56af5d97a012fafcea25
|
3 |
size 338799561
|
runs/Feb29_21-23-49_robolidar/events.out.tfevents.1709263483.robolidar.58688.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e70cb421cec333781c7824ef4b6809fe486993688200de073d63248ffe06d6a
|
3 |
+
size 472665
|