Upload checkpoints and training logs
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +24 -0
- downstream-shadow-removal/ckpt/step039999.pt +3 -0
- downstream-shadow-removal/config-2023-02-12-13-23-18.yaml +70 -0
- downstream-shadow-removal/config-2023-02-13-02-44-14.yaml +70 -0
- downstream-shadow-removal/output-2023-02-12-13-23-18.log +224 -0
- downstream-shadow-removal/output-2023-02-13-02-44-14.log +680 -0
- downstream-shadow-removal/samples/step000999.png +3 -0
- downstream-shadow-removal/samples/step001999.png +3 -0
- downstream-shadow-removal/samples/step002999.png +3 -0
- downstream-shadow-removal/samples/step003999.png +3 -0
- downstream-shadow-removal/samples/step004999.png +3 -0
- downstream-shadow-removal/samples/step005999.png +3 -0
- downstream-shadow-removal/samples/step006999.png +3 -0
- downstream-shadow-removal/samples/step007999.png +3 -0
- downstream-shadow-removal/samples/step008999.png +3 -0
- downstream-shadow-removal/samples/step009999.png +3 -0
- downstream-shadow-removal/tensorboard/events.out.tfevents.1676208201.jason-system.49917.0 +3 -0
- downstream-shadow-removal/tensorboard/events.out.tfevents.1676256254.jason-system.54160.0 +3 -0
- downstream-watermark-removal/ckpt/step099999.pt +3 -0
- downstream-watermark-removal/config-2023-02-12-14-28-38.yaml +70 -0
- downstream-watermark-removal/config-2023-02-12-15-41-28.yaml +70 -0
- downstream-watermark-removal/output-2023-02-12-14-28-38.log +223 -0
- downstream-watermark-removal/output-2023-02-12-15-41-28.log +0 -0
- downstream-watermark-removal/samples/step009999.png +3 -0
- downstream-watermark-removal/samples/step099999.png +3 -0
- downstream-watermark-removal/tensorboard/events.out.tfevents.1676212121.jason-system.50864.0 +3 -0
- downstream-watermark-removal/tensorboard/events.out.tfevents.1676216490.jason-system.51652.0 +3 -0
- joint_ffhq_brush_realnoise/ckpt/step149999/meta.pt +3 -0
- joint_ffhq_brush_realnoise/ckpt/step149999/model.pt +3 -0
- joint_ffhq_brush_realnoise/ckpt/step149999/optimizer.pt +3 -0
- joint_ffhq_brush_realnoise/config-2023-03-20-11-52-44.yaml +57 -0
- joint_ffhq_brush_realnoise/output-2023-03-20-11-52-44.log +0 -0
- joint_ffhq_brush_realnoise/samples/step149999.png +3 -0
- joint_ffhq_brush_realnoise/tensorboard/events.out.tfevents.1679284370.admin.cluster.local.35162.0 +3 -0
- joint_ffhq_maskdir_realnoise/ckpt/step149999/meta.pt +3 -0
- joint_ffhq_maskdir_realnoise/ckpt/step149999/model.pt +3 -0
- joint_ffhq_maskdir_realnoise/ckpt/step149999/optimizer.pt +3 -0
- joint_ffhq_maskdir_realnoise/config-2023-03-20-11-55-31.yaml +61 -0
- joint_ffhq_maskdir_realnoise/output-2023-03-20-11-55-31.log +0 -0
- joint_ffhq_maskdir_realnoise/samples/step149999.png +3 -0
- joint_ffhq_maskdir_realnoise/tensorboard/events.out.tfevents.1679284533.admin.cluster.local.727.0 +3 -0
- joint_imagenet_brush_realnoise/ckpt/step149999/meta.pt +3 -0
- joint_imagenet_brush_realnoise/ckpt/step149999/model.pt +3 -0
- joint_imagenet_brush_realnoise/ckpt/step149999/optimizer.pt +3 -0
- joint_imagenet_brush_realnoise/config-2023-03-22-15-20-53.yaml +54 -0
- joint_imagenet_brush_realnoise/output-2023-03-22-15-20-53.log +0 -0
- joint_imagenet_brush_realnoise/samples/step149999.png +3 -0
- joint_imagenet_brush_realnoise/tensorboard/events.out.tfevents.1679469675.admin.cluster.local.8096.0 +3 -0
- joint_imagenet_maskdir_realnoise/ckpt/step149999/meta.pt +3 -0
- joint_imagenet_maskdir_realnoise/ckpt/step149999/model.pt +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,27 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
downstream-shadow-removal/samples/step000999.png filter=lfs diff=lfs merge=lfs -text
|
37 |
+
downstream-shadow-removal/samples/step001999.png filter=lfs diff=lfs merge=lfs -text
|
38 |
+
downstream-shadow-removal/samples/step002999.png filter=lfs diff=lfs merge=lfs -text
|
39 |
+
downstream-shadow-removal/samples/step003999.png filter=lfs diff=lfs merge=lfs -text
|
40 |
+
downstream-shadow-removal/samples/step004999.png filter=lfs diff=lfs merge=lfs -text
|
41 |
+
downstream-shadow-removal/samples/step005999.png filter=lfs diff=lfs merge=lfs -text
|
42 |
+
downstream-shadow-removal/samples/step006999.png filter=lfs diff=lfs merge=lfs -text
|
43 |
+
downstream-shadow-removal/samples/step007999.png filter=lfs diff=lfs merge=lfs -text
|
44 |
+
downstream-shadow-removal/samples/step008999.png filter=lfs diff=lfs merge=lfs -text
|
45 |
+
downstream-shadow-removal/samples/step009999.png filter=lfs diff=lfs merge=lfs -text
|
46 |
+
downstream-watermark-removal/samples/step009999.png filter=lfs diff=lfs merge=lfs -text
|
47 |
+
downstream-watermark-removal/samples/step099999.png filter=lfs diff=lfs merge=lfs -text
|
48 |
+
joint_ffhq_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
49 |
+
joint_ffhq_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
50 |
+
joint_imagenet_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
51 |
+
joint_imagenet_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
52 |
+
joint_places_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
53 |
+
joint_places_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
54 |
+
separate_ffhq_brush_realnoise/samples/step079999.png filter=lfs diff=lfs merge=lfs -text
|
55 |
+
separate_ffhq_maskdir_realnoise/samples/step079999.png filter=lfs diff=lfs merge=lfs -text
|
56 |
+
separate_imagenet_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
57 |
+
separate_imagenet_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
58 |
+
separate_places_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
59 |
+
separate_places_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
|
downstream-shadow-removal/ckpt/step039999.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:550ed527f61244a228d83cbf3c8d3afbbb59238b61eb01aef029a3d73c4f10e6
|
3 |
+
size 160487047
|
downstream-shadow-removal/config-2023-02-12-13-23-18.yaml
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
DATA:
|
2 |
+
DATAROOT: /data/ISTD/
|
3 |
+
IMG_SIZE: 256
|
4 |
+
NAME: ISTD
|
5 |
+
DATALOADER:
|
6 |
+
BATCH_SIZE: 4
|
7 |
+
MICRO_BATCH: 0
|
8 |
+
NUM_WORKERS: 4
|
9 |
+
PIN_MEMORY: true
|
10 |
+
PREFETCH_FACTOR: 2
|
11 |
+
EVALUATE:
|
12 |
+
N_EVAL: 10000
|
13 |
+
MASK:
|
14 |
+
BRUSH_LENGTH_RATIO:
|
15 |
+
- 0.1
|
16 |
+
- 0.25
|
17 |
+
BRUSH_NUM: &id001
|
18 |
+
- 0
|
19 |
+
- 0
|
20 |
+
BRUSH_TURNS:
|
21 |
+
- 4
|
22 |
+
- 18
|
23 |
+
BRUSH_WIDTH_RATIO:
|
24 |
+
- 0.02
|
25 |
+
- 0.1
|
26 |
+
MASK_DIR: null
|
27 |
+
NOISE_DATASETS: []
|
28 |
+
RECT_LENGTH_RATIO:
|
29 |
+
- 0.2
|
30 |
+
- 0.8
|
31 |
+
RECT_NUM: *id001
|
32 |
+
SMOOTH_ITERS: 4
|
33 |
+
SMOOTH_KERNEL_SIZE: 15
|
34 |
+
SMOOTH_SIGMA: 4
|
35 |
+
MODEL:
|
36 |
+
MPN:
|
37 |
+
BASE_N_CHANNELS: 64
|
38 |
+
NECK_N_CHANNELS: 128
|
39 |
+
RIN:
|
40 |
+
BASE_N_CHANNELS: 32
|
41 |
+
NECK_N_CHANNELS: 128
|
42 |
+
WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
|
43 |
+
SAMPLE:
|
44 |
+
N_SAMPLES: 100
|
45 |
+
RANDOM: false
|
46 |
+
SAVE_DIR: ./samples/
|
47 |
+
SEED: 1234
|
48 |
+
TRAIN:
|
49 |
+
COEF_ADV: 0.001
|
50 |
+
COEF_IDMRF: 0.001
|
51 |
+
COEF_MPN: 2.0
|
52 |
+
COEF_REC: 1.4
|
53 |
+
COEF_SEMANTIC: 0.0001
|
54 |
+
EVAL_FREQ: 1000
|
55 |
+
OPTIM_DISC:
|
56 |
+
BETAS: &id002
|
57 |
+
- 0.5
|
58 |
+
- 0.9
|
59 |
+
LR: 1.0e-05
|
60 |
+
OPTIM_MPN:
|
61 |
+
BETAS: *id002
|
62 |
+
LR: 1.0e-05
|
63 |
+
OPTIM_RIN:
|
64 |
+
BETAS: *id002
|
65 |
+
LR: 1.0e-05
|
66 |
+
PRINT_FREQ: 100
|
67 |
+
RESUME: null
|
68 |
+
SAMPLE_FREQ: 1000
|
69 |
+
SAVE_FREQ: 5000
|
70 |
+
TRAIN_STEPS: 10000
|
downstream-shadow-removal/config-2023-02-13-02-44-14.yaml
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
DATA:
|
2 |
+
DATAROOT: /data/ISTD/
|
3 |
+
IMG_SIZE: 256
|
4 |
+
NAME: ISTD
|
5 |
+
DATALOADER:
|
6 |
+
BATCH_SIZE: 4
|
7 |
+
MICRO_BATCH: 0
|
8 |
+
NUM_WORKERS: 4
|
9 |
+
PIN_MEMORY: true
|
10 |
+
PREFETCH_FACTOR: 2
|
11 |
+
EVALUATE:
|
12 |
+
N_EVAL: 10000
|
13 |
+
MASK:
|
14 |
+
BRUSH_LENGTH_RATIO:
|
15 |
+
- 0.1
|
16 |
+
- 0.25
|
17 |
+
BRUSH_NUM: &id001
|
18 |
+
- 0
|
19 |
+
- 0
|
20 |
+
BRUSH_TURNS:
|
21 |
+
- 4
|
22 |
+
- 18
|
23 |
+
BRUSH_WIDTH_RATIO:
|
24 |
+
- 0.02
|
25 |
+
- 0.1
|
26 |
+
MASK_DIR: null
|
27 |
+
NOISE_DATASETS: []
|
28 |
+
RECT_LENGTH_RATIO:
|
29 |
+
- 0.2
|
30 |
+
- 0.8
|
31 |
+
RECT_NUM: *id001
|
32 |
+
SMOOTH_ITERS: 4
|
33 |
+
SMOOTH_KERNEL_SIZE: 15
|
34 |
+
SMOOTH_SIGMA: 4
|
35 |
+
MODEL:
|
36 |
+
MPN:
|
37 |
+
BASE_N_CHANNELS: 64
|
38 |
+
NECK_N_CHANNELS: 128
|
39 |
+
RIN:
|
40 |
+
BASE_N_CHANNELS: 32
|
41 |
+
NECK_N_CHANNELS: 128
|
42 |
+
WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
|
43 |
+
SAMPLE:
|
44 |
+
N_SAMPLES: 100
|
45 |
+
RANDOM: false
|
46 |
+
SAVE_DIR: ./samples/
|
47 |
+
SEED: 1234
|
48 |
+
TRAIN:
|
49 |
+
COEF_ADV: 0.001
|
50 |
+
COEF_IDMRF: 0.001
|
51 |
+
COEF_MPN: 2.0
|
52 |
+
COEF_REC: 1.4
|
53 |
+
COEF_SEMANTIC: 0.0001
|
54 |
+
EVAL_FREQ: 1000
|
55 |
+
OPTIM_DISC:
|
56 |
+
BETAS: &id002
|
57 |
+
- 0.5
|
58 |
+
- 0.9
|
59 |
+
LR: 1.0e-05
|
60 |
+
OPTIM_MPN:
|
61 |
+
BETAS: *id002
|
62 |
+
LR: 1.0e-05
|
63 |
+
OPTIM_RIN:
|
64 |
+
BETAS: *id002
|
65 |
+
LR: 1.0e-05
|
66 |
+
PRINT_FREQ: 100
|
67 |
+
RESUME: latest
|
68 |
+
SAMPLE_FREQ: 1000
|
69 |
+
SAVE_FREQ: 5000
|
70 |
+
TRAIN_STEPS: 100000
|
downstream-shadow-removal/output-2023-02-12-13-23-18.log
ADDED
@@ -0,0 +1,224 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-02-12 13:23:21,129 - INFO - Experiment directory: runs/downstream-shadow-removal
|
2 |
+
2023-02-12 13:23:21,129 - INFO - Device: cuda
|
3 |
+
2023-02-12 13:23:21,129 - INFO - Number of devices: 1
|
4 |
+
2023-02-12 13:23:21,135 - WARNING - Replace split `valid` with split `test`
|
5 |
+
2023-02-12 13:23:21,137 - INFO - Size of training set: 1330
|
6 |
+
2023-02-12 13:23:21,137 - INFO - Size of validation set: 540
|
7 |
+
2023-02-12 13:23:21,137 - INFO - Batch size per device: 4
|
8 |
+
2023-02-12 13:23:21,137 - INFO - Effective batch size: 4
|
9 |
+
2023-02-12 13:23:22,639 - INFO - Successfully load mpn from ./runs/places-joint/ckpt/step149999.pt
|
10 |
+
2023-02-12 13:23:22,643 - INFO - Successfully load rin from ./runs/places-joint/ckpt/step149999.pt
|
11 |
+
2023-02-12 13:23:22,645 - INFO - Successfully load disc from ./runs/places-joint/ckpt/step149999.pt
|
12 |
+
2023-02-12 13:23:22,647 - INFO - Successfully load pdisc from ./runs/places-joint/ckpt/step149999.pt
|
13 |
+
2023-02-12 13:23:22,650 - INFO - Start training...
|
14 |
+
2023-02-12 13:23:54,134 - INFO - [Train] step: 99, loss_adv_disc: -1.281875
|
15 |
+
2023-02-12 13:23:54,342 - INFO - [Train] step: 99, loss_mpn: 0.162302, loss_rec: 0.088649, loss_semantic: 0.554951, loss_idmrf: 7.990071, loss_adv_gen: -119.046860
|
16 |
+
2023-02-12 13:24:23,357 - INFO - [Train] step: 199, loss_adv_disc: -4.481179
|
17 |
+
2023-02-12 13:24:23,565 - INFO - [Train] step: 199, loss_mpn: 0.074521, loss_rec: 0.074638, loss_semantic: 0.563851, loss_idmrf: 7.788015, loss_adv_gen: -152.929199
|
18 |
+
2023-02-12 13:24:52,586 - INFO - [Train] step: 299, loss_adv_disc: 10.212236
|
19 |
+
2023-02-12 13:24:52,793 - INFO - [Train] step: 299, loss_mpn: 0.058758, loss_rec: 0.090663, loss_semantic: 0.657425, loss_idmrf: 8.309335, loss_adv_gen: -154.783676
|
20 |
+
2023-02-12 13:25:21,980 - INFO - [Train] step: 399, loss_adv_disc: -1.535188
|
21 |
+
2023-02-12 13:25:22,187 - INFO - [Train] step: 399, loss_mpn: 0.035261, loss_rec: 0.079162, loss_semantic: 0.496730, loss_idmrf: 4.807649, loss_adv_gen: -116.015060
|
22 |
+
2023-02-12 13:25:51,186 - INFO - [Train] step: 499, loss_adv_disc: -12.597924
|
23 |
+
2023-02-12 13:25:51,393 - INFO - [Train] step: 499, loss_mpn: 0.090085, loss_rec: 0.132387, loss_semantic: 0.493011, loss_idmrf: 6.727902, loss_adv_gen: -152.829514
|
24 |
+
2023-02-12 13:26:20,392 - INFO - [Train] step: 599, loss_adv_disc: -1.783383
|
25 |
+
2023-02-12 13:26:20,599 - INFO - [Train] step: 599, loss_mpn: 0.072984, loss_rec: 0.063185, loss_semantic: 0.514825, loss_idmrf: 5.854929, loss_adv_gen: -98.764038
|
26 |
+
2023-02-12 13:26:49,784 - INFO - [Train] step: 699, loss_adv_disc: -0.208335
|
27 |
+
2023-02-12 13:26:49,991 - INFO - [Train] step: 699, loss_mpn: 0.039042, loss_rec: 0.099020, loss_semantic: 0.674136, loss_idmrf: 6.369217, loss_adv_gen: -154.383133
|
28 |
+
2023-02-12 13:27:18,971 - INFO - [Train] step: 799, loss_adv_disc: 0.360612
|
29 |
+
2023-02-12 13:27:19,179 - INFO - [Train] step: 799, loss_mpn: 0.025015, loss_rec: 0.058072, loss_semantic: 0.417860, loss_idmrf: 3.865017, loss_adv_gen: -109.052032
|
30 |
+
2023-02-12 13:27:48,162 - INFO - [Train] step: 899, loss_adv_disc: -1.592671
|
31 |
+
2023-02-12 13:27:48,370 - INFO - [Train] step: 899, loss_mpn: 0.020442, loss_rec: 0.082523, loss_semantic: 0.448917, loss_idmrf: 5.112060, loss_adv_gen: -113.044441
|
32 |
+
2023-02-12 13:28:17,555 - INFO - [Train] step: 999, loss_adv_disc: -3.951293
|
33 |
+
2023-02-12 13:28:17,763 - INFO - [Train] step: 999, loss_mpn: 0.089411, loss_rec: 0.061914, loss_semantic: 0.476440, loss_idmrf: 5.286564, loss_adv_gen: -118.216484
|
34 |
+
2023-02-12 13:28:22,927 - INFO - [Eval] step: 999, bce: 0.387083, psnr: 22.987576, ssim: 0.910526
|
35 |
+
2023-02-12 13:28:53,546 - INFO - [Train] step: 1099, loss_adv_disc: -1.074106
|
36 |
+
2023-02-12 13:28:53,753 - INFO - [Train] step: 1099, loss_mpn: 0.014869, loss_rec: 0.055306, loss_semantic: 0.405501, loss_idmrf: 3.033794, loss_adv_gen: -105.430122
|
37 |
+
2023-02-12 13:29:22,756 - INFO - [Train] step: 1199, loss_adv_disc: -9.270275
|
38 |
+
2023-02-12 13:29:22,963 - INFO - [Train] step: 1199, loss_mpn: 0.025306, loss_rec: 0.071859, loss_semantic: 0.442767, loss_idmrf: 8.446411, loss_adv_gen: -92.132133
|
39 |
+
2023-02-12 13:29:51,972 - INFO - [Train] step: 1299, loss_adv_disc: -14.562277
|
40 |
+
2023-02-12 13:29:52,179 - INFO - [Train] step: 1299, loss_mpn: 0.049613, loss_rec: 0.083995, loss_semantic: 0.550962, loss_idmrf: 12.367452, loss_adv_gen: -142.591858
|
41 |
+
2023-02-12 13:30:21,389 - INFO - [Train] step: 1399, loss_adv_disc: 1.822066
|
42 |
+
2023-02-12 13:30:21,596 - INFO - [Train] step: 1399, loss_mpn: 0.029735, loss_rec: 0.060577, loss_semantic: 0.492058, loss_idmrf: 9.837570, loss_adv_gen: -199.041153
|
43 |
+
2023-02-12 13:30:50,607 - INFO - [Train] step: 1499, loss_adv_disc: -6.002779
|
44 |
+
2023-02-12 13:30:50,814 - INFO - [Train] step: 1499, loss_mpn: 0.030221, loss_rec: 0.072016, loss_semantic: 0.597045, loss_idmrf: 4.388852, loss_adv_gen: -157.418549
|
45 |
+
2023-02-12 13:31:19,828 - INFO - [Train] step: 1599, loss_adv_disc: -7.131116
|
46 |
+
2023-02-12 13:31:20,035 - INFO - [Train] step: 1599, loss_mpn: 0.026604, loss_rec: 0.073666, loss_semantic: 0.473443, loss_idmrf: 6.644470, loss_adv_gen: -159.523529
|
47 |
+
2023-02-12 13:31:49,247 - INFO - [Train] step: 1699, loss_adv_disc: 6.895418
|
48 |
+
2023-02-12 13:31:49,455 - INFO - [Train] step: 1699, loss_mpn: 0.018450, loss_rec: 0.074060, loss_semantic: 0.419208, loss_idmrf: 5.020046, loss_adv_gen: -115.342003
|
49 |
+
2023-02-12 13:32:18,483 - INFO - [Train] step: 1799, loss_adv_disc: -5.128633
|
50 |
+
2023-02-12 13:32:18,690 - INFO - [Train] step: 1799, loss_mpn: 0.025845, loss_rec: 0.077207, loss_semantic: 0.518926, loss_idmrf: 7.905535, loss_adv_gen: -180.514923
|
51 |
+
2023-02-12 13:32:47,721 - INFO - [Train] step: 1899, loss_adv_disc: -7.722089
|
52 |
+
2023-02-12 13:32:47,928 - INFO - [Train] step: 1899, loss_mpn: 0.019188, loss_rec: 0.056849, loss_semantic: 0.458317, loss_idmrf: 2.534275, loss_adv_gen: -159.524200
|
53 |
+
2023-02-12 13:33:17,162 - INFO - [Train] step: 1999, loss_adv_disc: -13.602573
|
54 |
+
2023-02-12 13:33:17,369 - INFO - [Train] step: 1999, loss_mpn: 0.048309, loss_rec: 0.095544, loss_semantic: 0.487569, loss_idmrf: 6.279430, loss_adv_gen: -165.249130
|
55 |
+
2023-02-12 13:33:22,523 - INFO - [Eval] step: 1999, bce: 0.361587, psnr: 23.048809, ssim: 0.913292
|
56 |
+
2023-02-12 13:33:52,822 - INFO - [Train] step: 2099, loss_adv_disc: 4.722687
|
57 |
+
2023-02-12 13:33:53,029 - INFO - [Train] step: 2099, loss_mpn: 0.020883, loss_rec: 0.064225, loss_semantic: 0.439803, loss_idmrf: 5.173006, loss_adv_gen: -127.306946
|
58 |
+
2023-02-12 13:34:22,033 - INFO - [Train] step: 2199, loss_adv_disc: -7.286263
|
59 |
+
2023-02-12 13:34:22,241 - INFO - [Train] step: 2199, loss_mpn: 0.025944, loss_rec: 0.060299, loss_semantic: 0.436272, loss_idmrf: 3.550529, loss_adv_gen: -123.120049
|
60 |
+
2023-02-12 13:34:51,247 - INFO - [Train] step: 2299, loss_adv_disc: -0.899125
|
61 |
+
2023-02-12 13:34:51,454 - INFO - [Train] step: 2299, loss_mpn: 0.015718, loss_rec: 0.044088, loss_semantic: 0.435449, loss_idmrf: 6.796015, loss_adv_gen: -150.392487
|
62 |
+
2023-02-12 13:35:20,660 - INFO - [Train] step: 2399, loss_adv_disc: 1.281568
|
63 |
+
2023-02-12 13:35:20,868 - INFO - [Train] step: 2399, loss_mpn: 0.036071, loss_rec: 0.046558, loss_semantic: 0.442822, loss_idmrf: 7.079724, loss_adv_gen: -145.621643
|
64 |
+
2023-02-12 13:35:49,875 - INFO - [Train] step: 2499, loss_adv_disc: -0.852216
|
65 |
+
2023-02-12 13:35:50,082 - INFO - [Train] step: 2499, loss_mpn: 0.025623, loss_rec: 0.052791, loss_semantic: 0.480111, loss_idmrf: 6.043885, loss_adv_gen: -119.099579
|
66 |
+
2023-02-12 13:36:19,086 - INFO - [Train] step: 2599, loss_adv_disc: 0.180318
|
67 |
+
2023-02-12 13:36:19,293 - INFO - [Train] step: 2599, loss_mpn: 0.016104, loss_rec: 0.048713, loss_semantic: 0.442174, loss_idmrf: 9.205061, loss_adv_gen: -186.506836
|
68 |
+
2023-02-12 13:36:48,498 - INFO - [Train] step: 2699, loss_adv_disc: 1.399606
|
69 |
+
2023-02-12 13:36:48,705 - INFO - [Train] step: 2699, loss_mpn: 0.009911, loss_rec: 0.038589, loss_semantic: 0.336942, loss_idmrf: 2.168548, loss_adv_gen: -112.174011
|
70 |
+
2023-02-12 13:37:17,709 - INFO - [Train] step: 2799, loss_adv_disc: -2.826076
|
71 |
+
2023-02-12 13:37:17,916 - INFO - [Train] step: 2799, loss_mpn: 0.019236, loss_rec: 0.056004, loss_semantic: 0.425428, loss_idmrf: 4.864495, loss_adv_gen: -128.027161
|
72 |
+
2023-02-12 13:37:46,925 - INFO - [Train] step: 2899, loss_adv_disc: -8.600943
|
73 |
+
2023-02-12 13:37:47,133 - INFO - [Train] step: 2899, loss_mpn: 0.035397, loss_rec: 0.070796, loss_semantic: 0.379691, loss_idmrf: 3.269684, loss_adv_gen: -134.056183
|
74 |
+
2023-02-12 13:38:16,347 - INFO - [Train] step: 2999, loss_adv_disc: -4.140905
|
75 |
+
2023-02-12 13:38:16,555 - INFO - [Train] step: 2999, loss_mpn: 0.015209, loss_rec: 0.055310, loss_semantic: 0.407628, loss_idmrf: 9.434046, loss_adv_gen: -126.835373
|
76 |
+
2023-02-12 13:38:21,704 - INFO - [Eval] step: 2999, bce: 0.289332, psnr: 23.335571, ssim: 0.917456
|
77 |
+
2023-02-12 13:38:51,972 - INFO - [Train] step: 3099, loss_adv_disc: 0.191693
|
78 |
+
2023-02-12 13:38:52,179 - INFO - [Train] step: 3099, loss_mpn: 0.010881, loss_rec: 0.039351, loss_semantic: 0.360058, loss_idmrf: 5.650393, loss_adv_gen: -82.992958
|
79 |
+
2023-02-12 13:39:21,195 - INFO - [Train] step: 3199, loss_adv_disc: 0.658779
|
80 |
+
2023-02-12 13:39:21,402 - INFO - [Train] step: 3199, loss_mpn: 0.014635, loss_rec: 0.053971, loss_semantic: 0.422246, loss_idmrf: 3.753058, loss_adv_gen: -134.257843
|
81 |
+
2023-02-12 13:39:50,413 - INFO - [Train] step: 3299, loss_adv_disc: 3.345398
|
82 |
+
2023-02-12 13:39:50,621 - INFO - [Train] step: 3299, loss_mpn: 0.011842, loss_rec: 0.055801, loss_semantic: 0.412139, loss_idmrf: 3.589517, loss_adv_gen: -142.031662
|
83 |
+
2023-02-12 13:40:19,828 - INFO - [Train] step: 3399, loss_adv_disc: 0.072930
|
84 |
+
2023-02-12 13:40:20,035 - INFO - [Train] step: 3399, loss_mpn: 0.011662, loss_rec: 0.046599, loss_semantic: 0.414424, loss_idmrf: 6.025554, loss_adv_gen: -136.393219
|
85 |
+
2023-02-12 13:40:49,051 - INFO - [Train] step: 3499, loss_adv_disc: -2.746962
|
86 |
+
2023-02-12 13:40:49,258 - INFO - [Train] step: 3499, loss_mpn: 0.011284, loss_rec: 0.072484, loss_semantic: 0.463066, loss_idmrf: 4.362463, loss_adv_gen: -117.525566
|
87 |
+
2023-02-12 13:41:18,282 - INFO - [Train] step: 3599, loss_adv_disc: -5.855626
|
88 |
+
2023-02-12 13:41:18,489 - INFO - [Train] step: 3599, loss_mpn: 0.036952, loss_rec: 0.063538, loss_semantic: 0.474658, loss_idmrf: 6.167106, loss_adv_gen: -145.618546
|
89 |
+
2023-02-12 13:41:47,702 - INFO - [Train] step: 3699, loss_adv_disc: -1.438740
|
90 |
+
2023-02-12 13:41:47,910 - INFO - [Train] step: 3699, loss_mpn: 0.021134, loss_rec: 0.054315, loss_semantic: 0.490676, loss_idmrf: 3.553780, loss_adv_gen: -86.800690
|
91 |
+
2023-02-12 13:42:16,919 - INFO - [Train] step: 3799, loss_adv_disc: -6.214058
|
92 |
+
2023-02-12 13:42:17,126 - INFO - [Train] step: 3799, loss_mpn: 0.073864, loss_rec: 0.111120, loss_semantic: 0.516712, loss_idmrf: 7.677469, loss_adv_gen: -118.615211
|
93 |
+
2023-02-12 13:42:46,136 - INFO - [Train] step: 3899, loss_adv_disc: 0.043025
|
94 |
+
2023-02-12 13:42:46,343 - INFO - [Train] step: 3899, loss_mpn: 0.018508, loss_rec: 0.053039, loss_semantic: 0.664957, loss_idmrf: 3.672100, loss_adv_gen: -142.985031
|
95 |
+
2023-02-12 13:43:15,559 - INFO - [Train] step: 3999, loss_adv_disc: 0.784531
|
96 |
+
2023-02-12 13:43:15,767 - INFO - [Train] step: 3999, loss_mpn: 0.012252, loss_rec: 0.043319, loss_semantic: 0.343232, loss_idmrf: 3.316679, loss_adv_gen: -110.461723
|
97 |
+
2023-02-12 13:43:20,880 - INFO - [Eval] step: 3999, bce: 0.354661, psnr: 22.951654, ssim: 0.917451
|
98 |
+
2023-02-12 13:43:50,951 - INFO - [Train] step: 4099, loss_adv_disc: -0.960966
|
99 |
+
2023-02-12 13:43:51,157 - INFO - [Train] step: 4099, loss_mpn: 0.031528, loss_rec: 0.049657, loss_semantic: 0.404885, loss_idmrf: 4.565471, loss_adv_gen: -120.507851
|
100 |
+
2023-02-12 13:44:20,173 - INFO - [Train] step: 4199, loss_adv_disc: 2.489911
|
101 |
+
2023-02-12 13:44:20,381 - INFO - [Train] step: 4199, loss_mpn: 0.011535, loss_rec: 0.054218, loss_semantic: 0.445787, loss_idmrf: 3.776466, loss_adv_gen: -103.148987
|
102 |
+
2023-02-12 13:44:49,396 - INFO - [Train] step: 4299, loss_adv_disc: -9.166822
|
103 |
+
2023-02-12 13:44:49,603 - INFO - [Train] step: 4299, loss_mpn: 0.024727, loss_rec: 0.083794, loss_semantic: 0.552959, loss_idmrf: 8.386415, loss_adv_gen: -101.256668
|
104 |
+
2023-02-12 13:45:18,813 - INFO - [Train] step: 4399, loss_adv_disc: -9.202044
|
105 |
+
2023-02-12 13:45:19,022 - INFO - [Train] step: 4399, loss_mpn: 0.011078, loss_rec: 0.059956, loss_semantic: 0.367525, loss_idmrf: 4.487533, loss_adv_gen: -93.162827
|
106 |
+
2023-02-12 13:45:48,031 - INFO - [Train] step: 4499, loss_adv_disc: -2.651873
|
107 |
+
2023-02-12 13:45:48,239 - INFO - [Train] step: 4499, loss_mpn: 0.017707, loss_rec: 0.047575, loss_semantic: 0.423964, loss_idmrf: 7.904148, loss_adv_gen: -92.877556
|
108 |
+
2023-02-12 13:46:17,249 - INFO - [Train] step: 4599, loss_adv_disc: -1.531205
|
109 |
+
2023-02-12 13:46:17,456 - INFO - [Train] step: 4599, loss_mpn: 0.020947, loss_rec: 0.068436, loss_semantic: 0.545828, loss_idmrf: 4.993389, loss_adv_gen: -90.154297
|
110 |
+
2023-02-12 13:46:46,710 - INFO - [Train] step: 4699, loss_adv_disc: 4.040330
|
111 |
+
2023-02-12 13:46:46,918 - INFO - [Train] step: 4699, loss_mpn: 0.014810, loss_rec: 0.070468, loss_semantic: 0.525938, loss_idmrf: 5.143922, loss_adv_gen: -120.824249
|
112 |
+
2023-02-12 13:47:15,956 - INFO - [Train] step: 4799, loss_adv_disc: -1.453163
|
113 |
+
2023-02-12 13:47:16,164 - INFO - [Train] step: 4799, loss_mpn: 0.015021, loss_rec: 0.053999, loss_semantic: 0.496558, loss_idmrf: 3.925681, loss_adv_gen: -118.246658
|
114 |
+
2023-02-12 13:47:45,201 - INFO - [Train] step: 4899, loss_adv_disc: -0.842131
|
115 |
+
2023-02-12 13:47:45,408 - INFO - [Train] step: 4899, loss_mpn: 0.013811, loss_rec: 0.038493, loss_semantic: 0.386875, loss_idmrf: 6.477338, loss_adv_gen: -102.733505
|
116 |
+
2023-02-12 13:48:14,643 - INFO - [Train] step: 4999, loss_adv_disc: 1.860435
|
117 |
+
2023-02-12 13:48:14,851 - INFO - [Train] step: 4999, loss_mpn: 0.026306, loss_rec: 0.045528, loss_semantic: 0.423802, loss_idmrf: 4.491714, loss_adv_gen: -81.664734
|
118 |
+
2023-02-12 13:48:20,010 - INFO - [Eval] step: 4999, bce: 0.390663, psnr: 23.292404, ssim: 0.920074
|
119 |
+
2023-02-12 13:48:50,229 - INFO - [Train] step: 5099, loss_adv_disc: 4.181933
|
120 |
+
2023-02-12 13:48:50,436 - INFO - [Train] step: 5099, loss_mpn: 0.018531, loss_rec: 0.056876, loss_semantic: 0.372196, loss_idmrf: 5.690077, loss_adv_gen: -91.499268
|
121 |
+
2023-02-12 13:49:19,440 - INFO - [Train] step: 5199, loss_adv_disc: 3.452145
|
122 |
+
2023-02-12 13:49:19,648 - INFO - [Train] step: 5199, loss_mpn: 0.012540, loss_rec: 0.050392, loss_semantic: 0.367254, loss_idmrf: 5.612234, loss_adv_gen: -104.571259
|
123 |
+
2023-02-12 13:49:48,664 - INFO - [Train] step: 5299, loss_adv_disc: 0.072335
|
124 |
+
2023-02-12 13:49:48,871 - INFO - [Train] step: 5299, loss_mpn: 0.009281, loss_rec: 0.034018, loss_semantic: 0.373055, loss_idmrf: 9.307406, loss_adv_gen: -86.459152
|
125 |
+
2023-02-12 13:50:18,082 - INFO - [Train] step: 5399, loss_adv_disc: -1.546246
|
126 |
+
2023-02-12 13:50:18,289 - INFO - [Train] step: 5399, loss_mpn: 0.015896, loss_rec: 0.049707, loss_semantic: 0.560504, loss_idmrf: 4.826586, loss_adv_gen: -86.684578
|
127 |
+
2023-02-12 13:50:47,301 - INFO - [Train] step: 5499, loss_adv_disc: -3.722799
|
128 |
+
2023-02-12 13:50:47,508 - INFO - [Train] step: 5499, loss_mpn: 0.012618, loss_rec: 0.040985, loss_semantic: 0.321694, loss_idmrf: 3.553346, loss_adv_gen: -95.202797
|
129 |
+
2023-02-12 13:51:16,520 - INFO - [Train] step: 5599, loss_adv_disc: -1.551634
|
130 |
+
2023-02-12 13:51:16,727 - INFO - [Train] step: 5599, loss_mpn: 0.010603, loss_rec: 0.041937, loss_semantic: 0.368314, loss_idmrf: 6.840730, loss_adv_gen: -122.285248
|
131 |
+
2023-02-12 13:51:45,929 - INFO - [Train] step: 5699, loss_adv_disc: 1.480345
|
132 |
+
2023-02-12 13:51:46,137 - INFO - [Train] step: 5699, loss_mpn: 0.022207, loss_rec: 0.033643, loss_semantic: 0.418103, loss_idmrf: 3.735212, loss_adv_gen: -70.136971
|
133 |
+
2023-02-12 13:52:15,152 - INFO - [Train] step: 5799, loss_adv_disc: -5.609029
|
134 |
+
2023-02-12 13:52:15,358 - INFO - [Train] step: 5799, loss_mpn: 0.018745, loss_rec: 0.051710, loss_semantic: 0.431293, loss_idmrf: 6.607909, loss_adv_gen: -109.463600
|
135 |
+
2023-02-12 13:52:44,366 - INFO - [Train] step: 5899, loss_adv_disc: 1.087741
|
136 |
+
2023-02-12 13:52:44,574 - INFO - [Train] step: 5899, loss_mpn: 0.019051, loss_rec: 0.037814, loss_semantic: 0.404567, loss_idmrf: 2.744005, loss_adv_gen: -85.043411
|
137 |
+
2023-02-12 13:53:13,796 - INFO - [Train] step: 5999, loss_adv_disc: 1.313997
|
138 |
+
2023-02-12 13:53:14,004 - INFO - [Train] step: 5999, loss_mpn: 0.007717, loss_rec: 0.039593, loss_semantic: 0.355299, loss_idmrf: 3.153148, loss_adv_gen: -92.488480
|
139 |
+
2023-02-12 13:53:19,152 - INFO - [Eval] step: 5999, bce: 0.360973, psnr: 23.440897, ssim: 0.921864
|
140 |
+
2023-02-12 13:53:49,448 - INFO - [Train] step: 6099, loss_adv_disc: -6.527198
|
141 |
+
2023-02-12 13:53:49,655 - INFO - [Train] step: 6099, loss_mpn: 0.012591, loss_rec: 0.060579, loss_semantic: 0.470930, loss_idmrf: 5.310204, loss_adv_gen: -102.355423
|
142 |
+
2023-02-12 13:54:18,664 - INFO - [Train] step: 6199, loss_adv_disc: -1.580616
|
143 |
+
2023-02-12 13:54:18,872 - INFO - [Train] step: 6199, loss_mpn: 0.014308, loss_rec: 0.048241, loss_semantic: 0.396192, loss_idmrf: 6.307591, loss_adv_gen: -101.506081
|
144 |
+
2023-02-12 13:54:47,889 - INFO - [Train] step: 6299, loss_adv_disc: -0.838180
|
145 |
+
2023-02-12 13:54:48,096 - INFO - [Train] step: 6299, loss_mpn: 0.014393, loss_rec: 0.046224, loss_semantic: 0.443932, loss_idmrf: 4.024370, loss_adv_gen: -59.003387
|
146 |
+
2023-02-12 13:55:17,334 - INFO - [Train] step: 6399, loss_adv_disc: -1.536690
|
147 |
+
2023-02-12 13:55:17,541 - INFO - [Train] step: 6399, loss_mpn: 0.017005, loss_rec: 0.041010, loss_semantic: 0.443276, loss_idmrf: 5.697750, loss_adv_gen: -82.131157
|
148 |
+
2023-02-12 13:55:46,587 - INFO - [Train] step: 6499, loss_adv_disc: -2.835433
|
149 |
+
2023-02-12 13:55:46,795 - INFO - [Train] step: 6499, loss_mpn: 0.008315, loss_rec: 0.038857, loss_semantic: 0.353050, loss_idmrf: 3.457996, loss_adv_gen: -78.603981
|
150 |
+
2023-02-12 13:56:15,842 - INFO - [Train] step: 6599, loss_adv_disc: -2.492158
|
151 |
+
2023-02-12 13:56:16,049 - INFO - [Train] step: 6599, loss_mpn: 0.008221, loss_rec: 0.046404, loss_semantic: 0.431954, loss_idmrf: 4.713380, loss_adv_gen: -86.217865
|
152 |
+
2023-02-12 13:56:45,275 - INFO - [Train] step: 6699, loss_adv_disc: -0.614049
|
153 |
+
2023-02-12 13:56:45,482 - INFO - [Train] step: 6699, loss_mpn: 0.010530, loss_rec: 0.046490, loss_semantic: 0.398438, loss_idmrf: 2.225234, loss_adv_gen: -66.221626
|
154 |
+
2023-02-12 13:57:14,500 - INFO - [Train] step: 6799, loss_adv_disc: -2.154045
|
155 |
+
2023-02-12 13:57:14,707 - INFO - [Train] step: 6799, loss_mpn: 0.014369, loss_rec: 0.050858, loss_semantic: 0.448220, loss_idmrf: 3.556105, loss_adv_gen: -105.186676
|
156 |
+
2023-02-12 13:57:43,719 - INFO - [Train] step: 6899, loss_adv_disc: -1.190847
|
157 |
+
2023-02-12 13:57:43,927 - INFO - [Train] step: 6899, loss_mpn: 0.028931, loss_rec: 0.051065, loss_semantic: 0.450494, loss_idmrf: 6.116441, loss_adv_gen: -104.228516
|
158 |
+
2023-02-12 13:58:13,150 - INFO - [Train] step: 6999, loss_adv_disc: -4.266356
|
159 |
+
2023-02-12 13:58:13,357 - INFO - [Train] step: 6999, loss_mpn: 0.013678, loss_rec: 0.041581, loss_semantic: 0.424674, loss_idmrf: 4.031384, loss_adv_gen: -84.939545
|
160 |
+
2023-02-12 13:58:18,504 - INFO - [Eval] step: 6999, bce: 0.392871, psnr: 23.324829, ssim: 0.921624
|
161 |
+
2023-02-12 13:58:48,580 - INFO - [Train] step: 7099, loss_adv_disc: -4.838001
|
162 |
+
2023-02-12 13:58:48,788 - INFO - [Train] step: 7099, loss_mpn: 0.006589, loss_rec: 0.039819, loss_semantic: 0.366608, loss_idmrf: 3.555822, loss_adv_gen: -76.142303
|
163 |
+
2023-02-12 13:59:17,801 - INFO - [Train] step: 7199, loss_adv_disc: -1.795486
|
164 |
+
2023-02-12 13:59:18,008 - INFO - [Train] step: 7199, loss_mpn: 0.007781, loss_rec: 0.037948, loss_semantic: 0.380204, loss_idmrf: 3.421008, loss_adv_gen: -98.430969
|
165 |
+
2023-02-12 13:59:47,024 - INFO - [Train] step: 7299, loss_adv_disc: -5.008701
|
166 |
+
2023-02-12 13:59:47,231 - INFO - [Train] step: 7299, loss_mpn: 0.006212, loss_rec: 0.050326, loss_semantic: 0.343246, loss_idmrf: 3.782480, loss_adv_gen: -89.840332
|
167 |
+
2023-02-12 14:00:16,439 - INFO - [Train] step: 7399, loss_adv_disc: 0.952745
|
168 |
+
2023-02-12 14:00:16,647 - INFO - [Train] step: 7399, loss_mpn: 0.011856, loss_rec: 0.037551, loss_semantic: 0.361128, loss_idmrf: 4.399923, loss_adv_gen: -75.590004
|
169 |
+
2023-02-12 14:00:45,662 - INFO - [Train] step: 7499, loss_adv_disc: -0.309092
|
170 |
+
2023-02-12 14:00:45,869 - INFO - [Train] step: 7499, loss_mpn: 0.032277, loss_rec: 0.040263, loss_semantic: 0.608975, loss_idmrf: 3.067678, loss_adv_gen: -71.062065
|
171 |
+
2023-02-12 14:01:14,880 - INFO - [Train] step: 7599, loss_adv_disc: -3.749333
|
172 |
+
2023-02-12 14:01:15,087 - INFO - [Train] step: 7599, loss_mpn: 0.025242, loss_rec: 0.047163, loss_semantic: 0.448062, loss_idmrf: 3.615059, loss_adv_gen: -102.538254
|
173 |
+
2023-02-12 14:01:44,291 - INFO - [Train] step: 7699, loss_adv_disc: -1.246231
|
174 |
+
2023-02-12 14:01:44,499 - INFO - [Train] step: 7699, loss_mpn: 0.011135, loss_rec: 0.053402, loss_semantic: 0.458806, loss_idmrf: 2.834428, loss_adv_gen: -89.317276
|
175 |
+
2023-02-12 14:02:13,491 - INFO - [Train] step: 7799, loss_adv_disc: -0.799770
|
176 |
+
2023-02-12 14:02:13,698 - INFO - [Train] step: 7799, loss_mpn: 0.024141, loss_rec: 0.038572, loss_semantic: 0.450097, loss_idmrf: 4.533717, loss_adv_gen: -91.607513
|
177 |
+
2023-02-12 14:02:42,691 - INFO - [Train] step: 7899, loss_adv_disc: -1.562794
|
178 |
+
2023-02-12 14:02:42,898 - INFO - [Train] step: 7899, loss_mpn: 0.014830, loss_rec: 0.039014, loss_semantic: 0.443213, loss_idmrf: 7.125534, loss_adv_gen: -99.803688
|
179 |
+
2023-02-12 14:03:12,101 - INFO - [Train] step: 7999, loss_adv_disc: -1.292525
|
180 |
+
2023-02-12 14:03:12,308 - INFO - [Train] step: 7999, loss_mpn: 0.009162, loss_rec: 0.035810, loss_semantic: 0.339180, loss_idmrf: 4.937554, loss_adv_gen: -66.376938
|
181 |
+
2023-02-12 14:03:17,463 - INFO - [Eval] step: 7999, bce: 0.410713, psnr: 23.380175, ssim: 0.922875
|
182 |
+
2023-02-12 14:03:47,540 - INFO - [Train] step: 8099, loss_adv_disc: -1.598471
|
183 |
+
2023-02-12 14:03:47,747 - INFO - [Train] step: 8099, loss_mpn: 0.008017, loss_rec: 0.038835, loss_semantic: 0.342724, loss_idmrf: 3.858693, loss_adv_gen: -72.846497
|
184 |
+
2023-02-12 14:04:16,766 - INFO - [Train] step: 8199, loss_adv_disc: -1.526141
|
185 |
+
2023-02-12 14:04:16,974 - INFO - [Train] step: 8199, loss_mpn: 0.007769, loss_rec: 0.048700, loss_semantic: 0.370412, loss_idmrf: 4.246994, loss_adv_gen: -87.642441
|
186 |
+
2023-02-12 14:04:45,993 - INFO - [Train] step: 8299, loss_adv_disc: -4.967025
|
187 |
+
2023-02-12 14:04:46,200 - INFO - [Train] step: 8299, loss_mpn: 0.017493, loss_rec: 0.051434, loss_semantic: 0.451939, loss_idmrf: 4.851607, loss_adv_gen: -92.471909
|
188 |
+
2023-02-12 14:05:15,432 - INFO - [Train] step: 8399, loss_adv_disc: -0.423972
|
189 |
+
2023-02-12 14:05:15,639 - INFO - [Train] step: 8399, loss_mpn: 0.009851, loss_rec: 0.035539, loss_semantic: 0.387267, loss_idmrf: 4.063365, loss_adv_gen: -88.120407
|
190 |
+
2023-02-12 14:05:44,657 - INFO - [Train] step: 8499, loss_adv_disc: 0.920736
|
191 |
+
2023-02-12 14:05:44,864 - INFO - [Train] step: 8499, loss_mpn: 0.008419, loss_rec: 0.030417, loss_semantic: 0.338675, loss_idmrf: 3.181174, loss_adv_gen: -59.173306
|
192 |
+
2023-02-12 14:06:13,877 - INFO - [Train] step: 8599, loss_adv_disc: -0.524680
|
193 |
+
2023-02-12 14:06:14,084 - INFO - [Train] step: 8599, loss_mpn: 0.009725, loss_rec: 0.048215, loss_semantic: 0.435598, loss_idmrf: 2.502195, loss_adv_gen: -71.777512
|
194 |
+
2023-02-12 14:06:43,310 - INFO - [Train] step: 8699, loss_adv_disc: -1.867347
|
195 |
+
2023-02-12 14:06:43,517 - INFO - [Train] step: 8699, loss_mpn: 0.005162, loss_rec: 0.028492, loss_semantic: 0.312991, loss_idmrf: 4.448667, loss_adv_gen: -58.639214
|
196 |
+
2023-02-12 14:07:12,536 - INFO - [Train] step: 8799, loss_adv_disc: 0.215890
|
197 |
+
2023-02-12 14:07:12,743 - INFO - [Train] step: 8799, loss_mpn: 0.007085, loss_rec: 0.027526, loss_semantic: 0.321581, loss_idmrf: 2.571146, loss_adv_gen: -79.483131
|
198 |
+
2023-02-12 14:07:41,757 - INFO - [Train] step: 8899, loss_adv_disc: -5.541975
|
199 |
+
2023-02-12 14:07:41,964 - INFO - [Train] step: 8899, loss_mpn: 0.014692, loss_rec: 0.058386, loss_semantic: 0.391197, loss_idmrf: 6.332783, loss_adv_gen: -77.094063
|
200 |
+
2023-02-12 14:08:11,177 - INFO - [Train] step: 8999, loss_adv_disc: -4.298098
|
201 |
+
2023-02-12 14:08:11,385 - INFO - [Train] step: 8999, loss_mpn: 0.009263, loss_rec: 0.041659, loss_semantic: 0.368018, loss_idmrf: 3.096043, loss_adv_gen: -94.624344
|
202 |
+
2023-02-12 14:08:16,542 - INFO - [Eval] step: 8999, bce: 0.350487, psnr: 23.695768, ssim: 0.925011
|
203 |
+
2023-02-12 14:08:46,831 - INFO - [Train] step: 9099, loss_adv_disc: -0.096787
|
204 |
+
2023-02-12 14:08:47,038 - INFO - [Train] step: 9099, loss_mpn: 0.007456, loss_rec: 0.050012, loss_semantic: 0.391861, loss_idmrf: 4.834039, loss_adv_gen: -34.803146
|
205 |
+
2023-02-12 14:09:16,052 - INFO - [Train] step: 9199, loss_adv_disc: -2.421654
|
206 |
+
2023-02-12 14:09:16,260 - INFO - [Train] step: 9199, loss_mpn: 0.010739, loss_rec: 0.043619, loss_semantic: 0.375820, loss_idmrf: 2.879037, loss_adv_gen: -95.356018
|
207 |
+
2023-02-12 14:09:45,468 - INFO - [Train] step: 9299, loss_adv_disc: -0.639228
|
208 |
+
2023-02-12 14:09:45,675 - INFO - [Train] step: 9299, loss_mpn: 0.009492, loss_rec: 0.041271, loss_semantic: 0.385236, loss_idmrf: 3.942233, loss_adv_gen: -81.359604
|
209 |
+
2023-02-12 14:10:14,686 - INFO - [Train] step: 9399, loss_adv_disc: -0.824157
|
210 |
+
2023-02-12 14:10:14,893 - INFO - [Train] step: 9399, loss_mpn: 0.007119, loss_rec: 0.035177, loss_semantic: 0.354077, loss_idmrf: 4.972061, loss_adv_gen: -74.898308
|
211 |
+
2023-02-12 14:10:43,902 - INFO - [Train] step: 9499, loss_adv_disc: -2.232660
|
212 |
+
2023-02-12 14:10:44,110 - INFO - [Train] step: 9499, loss_mpn: 0.014257, loss_rec: 0.035584, loss_semantic: 0.429310, loss_idmrf: 5.182566, loss_adv_gen: -64.690308
|
213 |
+
2023-02-12 14:11:13,120 - INFO - [Train] step: 9599, loss_adv_disc: -1.639446
|
214 |
+
2023-02-12 14:11:13,327 - INFO - [Train] step: 9599, loss_mpn: 0.019186, loss_rec: 0.041353, loss_semantic: 0.504692, loss_idmrf: 3.800310, loss_adv_gen: -68.249176
|
215 |
+
2023-02-12 14:11:42,542 - INFO - [Train] step: 9699, loss_adv_disc: -3.789857
|
216 |
+
2023-02-12 14:11:42,750 - INFO - [Train] step: 9699, loss_mpn: 0.021227, loss_rec: 0.046244, loss_semantic: 0.464764, loss_idmrf: 4.286100, loss_adv_gen: -80.508911
|
217 |
+
2023-02-12 14:12:11,774 - INFO - [Train] step: 9799, loss_adv_disc: -1.276845
|
218 |
+
2023-02-12 14:12:11,981 - INFO - [Train] step: 9799, loss_mpn: 0.008188, loss_rec: 0.036653, loss_semantic: 0.358453, loss_idmrf: 2.212314, loss_adv_gen: -55.353634
|
219 |
+
2023-02-12 14:12:41,001 - INFO - [Train] step: 9899, loss_adv_disc: -1.013932
|
220 |
+
2023-02-12 14:12:41,209 - INFO - [Train] step: 9899, loss_mpn: 0.008457, loss_rec: 0.037062, loss_semantic: 0.404350, loss_idmrf: 2.998356, loss_adv_gen: -77.622223
|
221 |
+
2023-02-12 14:13:10,420 - INFO - [Train] step: 9999, loss_adv_disc: -2.420930
|
222 |
+
2023-02-12 14:13:10,628 - INFO - [Train] step: 9999, loss_mpn: 0.009475, loss_rec: 0.047726, loss_semantic: 0.396993, loss_idmrf: 5.336214, loss_adv_gen: -73.804047
|
223 |
+
2023-02-12 14:13:15,773 - INFO - [Eval] step: 9999, bce: 0.441264, psnr: 23.645063, ssim: 0.924997
|
224 |
+
2023-02-12 14:13:16,954 - INFO - End of training
|
downstream-shadow-removal/output-2023-02-13-02-44-14.log
ADDED
@@ -0,0 +1,680 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-02-13 02:44:14,898 - INFO - Experiment directory: runs/downstream-shadow-removal
|
2 |
+
2023-02-13 02:44:14,898 - INFO - Device: cuda
|
3 |
+
2023-02-13 02:44:14,898 - INFO - Number of devices: 1
|
4 |
+
2023-02-13 02:44:14,904 - WARNING - Replace split `valid` with split `test`
|
5 |
+
2023-02-13 02:44:14,907 - INFO - Size of training set: 1330
|
6 |
+
2023-02-13 02:44:14,907 - INFO - Size of validation set: 540
|
7 |
+
2023-02-13 02:44:14,907 - INFO - Batch size per device: 4
|
8 |
+
2023-02-13 02:44:14,907 - INFO - Effective batch size: 4
|
9 |
+
2023-02-13 02:44:16,374 - INFO - Successfully load mpn from ./runs/places-joint/ckpt/step149999.pt
|
10 |
+
2023-02-13 02:44:16,378 - INFO - Successfully load rin from ./runs/places-joint/ckpt/step149999.pt
|
11 |
+
2023-02-13 02:44:16,384 - INFO - Successfully load disc from ./runs/places-joint/ckpt/step149999.pt
|
12 |
+
2023-02-13 02:44:16,385 - INFO - Successfully load pdisc from ./runs/places-joint/ckpt/step149999.pt
|
13 |
+
2023-02-13 02:44:16,388 - INFO - Resume from runs/downstream-shadow-removal/ckpt/step009999.pt
|
14 |
+
2023-02-13 02:44:16,472 - INFO - Successfully load mpn from runs/downstream-shadow-removal/ckpt/step009999.pt
|
15 |
+
2023-02-13 02:44:16,476 - INFO - Successfully load rin from runs/downstream-shadow-removal/ckpt/step009999.pt
|
16 |
+
2023-02-13 02:44:16,478 - INFO - Successfully load disc from runs/downstream-shadow-removal/ckpt/step009999.pt
|
17 |
+
2023-02-13 02:44:16,480 - INFO - Successfully load pdisc from runs/downstream-shadow-removal/ckpt/step009999.pt
|
18 |
+
2023-02-13 02:44:16,515 - INFO - Successfully load optimizers from runs/downstream-shadow-removal/ckpt/step009999.pt
|
19 |
+
2023-02-13 02:44:16,515 - INFO - Restart training at step 10000
|
20 |
+
2023-02-13 02:44:16,515 - INFO - Best psnr so far: 23.695768356323242
|
21 |
+
2023-02-13 02:44:16,518 - INFO - Start training...
|
22 |
+
2023-02-13 02:44:48,197 - INFO - [Train] step: 10099, loss_adv_disc: 0.078588
|
23 |
+
2023-02-13 02:44:48,404 - INFO - [Train] step: 10099, loss_mpn: 0.011509, loss_rec: 0.039201, loss_semantic: 0.377556, loss_idmrf: 3.615375, loss_adv_gen: -48.026169
|
24 |
+
2023-02-13 02:45:17,488 - INFO - [Train] step: 10199, loss_adv_disc: -1.947309
|
25 |
+
2023-02-13 02:45:17,697 - INFO - [Train] step: 10199, loss_mpn: 0.012715, loss_rec: 0.030864, loss_semantic: 0.360930, loss_idmrf: 2.600466, loss_adv_gen: -85.842705
|
26 |
+
2023-02-13 02:45:46,859 - INFO - [Train] step: 10299, loss_adv_disc: -0.551243
|
27 |
+
2023-02-13 02:45:47,067 - INFO - [Train] step: 10299, loss_mpn: 0.014206, loss_rec: 0.043966, loss_semantic: 0.523468, loss_idmrf: 5.591407, loss_adv_gen: -92.931442
|
28 |
+
2023-02-13 02:46:16,444 - INFO - [Train] step: 10399, loss_adv_disc: -4.300162
|
29 |
+
2023-02-13 02:46:16,652 - INFO - [Train] step: 10399, loss_mpn: 0.008325, loss_rec: 0.037018, loss_semantic: 0.390521, loss_idmrf: 2.957632, loss_adv_gen: -63.995998
|
30 |
+
2023-02-13 02:46:45,872 - INFO - [Train] step: 10499, loss_adv_disc: -0.984805
|
31 |
+
2023-02-13 02:46:46,081 - INFO - [Train] step: 10499, loss_mpn: 0.025511, loss_rec: 0.042500, loss_semantic: 0.398691, loss_idmrf: 4.002370, loss_adv_gen: -72.055634
|
32 |
+
2023-02-13 02:47:15,306 - INFO - [Train] step: 10599, loss_adv_disc: -0.168864
|
33 |
+
2023-02-13 02:47:15,514 - INFO - [Train] step: 10599, loss_mpn: 0.010539, loss_rec: 0.032968, loss_semantic: 0.400717, loss_idmrf: 3.446010, loss_adv_gen: -75.776711
|
34 |
+
2023-02-13 02:47:44,915 - INFO - [Train] step: 10699, loss_adv_disc: -0.191629
|
35 |
+
2023-02-13 02:47:45,123 - INFO - [Train] step: 10699, loss_mpn: 0.010557, loss_rec: 0.038703, loss_semantic: 0.471336, loss_idmrf: 2.717856, loss_adv_gen: -81.432404
|
36 |
+
2023-02-13 02:48:14,342 - INFO - [Train] step: 10799, loss_adv_disc: 1.988065
|
37 |
+
2023-02-13 02:48:14,551 - INFO - [Train] step: 10799, loss_mpn: 0.008534, loss_rec: 0.033567, loss_semantic: 0.337367, loss_idmrf: 2.358154, loss_adv_gen: -62.045174
|
38 |
+
2023-02-13 02:48:43,779 - INFO - [Train] step: 10899, loss_adv_disc: -0.890945
|
39 |
+
2023-02-13 02:48:43,988 - INFO - [Train] step: 10899, loss_mpn: 0.009466, loss_rec: 0.031909, loss_semantic: 0.363756, loss_idmrf: 2.512322, loss_adv_gen: -55.988701
|
40 |
+
2023-02-13 02:49:13,412 - INFO - [Train] step: 10999, loss_adv_disc: -0.436340
|
41 |
+
2023-02-13 02:49:13,620 - INFO - [Train] step: 10999, loss_mpn: 0.013038, loss_rec: 0.034377, loss_semantic: 0.406465, loss_idmrf: 3.428352, loss_adv_gen: -76.155876
|
42 |
+
2023-02-13 02:49:18,810 - INFO - [Eval] step: 10999, bce: 0.356415, psnr: 23.687820, ssim: 0.926956
|
43 |
+
2023-02-13 02:49:49,575 - INFO - [Train] step: 11099, loss_adv_disc: -1.143975
|
44 |
+
2023-02-13 02:49:49,784 - INFO - [Train] step: 11099, loss_mpn: 0.006747, loss_rec: 0.029099, loss_semantic: 0.350319, loss_idmrf: 2.055681, loss_adv_gen: -41.675575
|
45 |
+
2023-02-13 02:50:19,039 - INFO - [Train] step: 11199, loss_adv_disc: -2.175141
|
46 |
+
2023-02-13 02:50:19,248 - INFO - [Train] step: 11199, loss_mpn: 0.014661, loss_rec: 0.030048, loss_semantic: 0.366945, loss_idmrf: 5.886166, loss_adv_gen: -57.670628
|
47 |
+
2023-02-13 02:50:48,495 - INFO - [Train] step: 11299, loss_adv_disc: -2.454169
|
48 |
+
2023-02-13 02:50:48,704 - INFO - [Train] step: 11299, loss_mpn: 0.011991, loss_rec: 0.038149, loss_semantic: 0.407933, loss_idmrf: 8.903237, loss_adv_gen: -93.871811
|
49 |
+
2023-02-13 02:51:18,146 - INFO - [Train] step: 11399, loss_adv_disc: -2.127088
|
50 |
+
2023-02-13 02:51:18,355 - INFO - [Train] step: 11399, loss_mpn: 0.012583, loss_rec: 0.028059, loss_semantic: 0.370067, loss_idmrf: 5.352232, loss_adv_gen: -75.710938
|
51 |
+
2023-02-13 02:51:47,616 - INFO - [Train] step: 11499, loss_adv_disc: -3.866915
|
52 |
+
2023-02-13 02:51:47,825 - INFO - [Train] step: 11499, loss_mpn: 0.010172, loss_rec: 0.039636, loss_semantic: 0.481377, loss_idmrf: 2.577618, loss_adv_gen: -50.608372
|
53 |
+
2023-02-13 02:52:17,063 - INFO - [Train] step: 11599, loss_adv_disc: -0.510452
|
54 |
+
2023-02-13 02:52:17,272 - INFO - [Train] step: 11599, loss_mpn: 0.012594, loss_rec: 0.036184, loss_semantic: 0.380372, loss_idmrf: 4.289717, loss_adv_gen: -74.201294
|
55 |
+
2023-02-13 02:52:46,702 - INFO - [Train] step: 11699, loss_adv_disc: -3.221022
|
56 |
+
2023-02-13 02:52:46,911 - INFO - [Train] step: 11699, loss_mpn: 0.007079, loss_rec: 0.033923, loss_semantic: 0.347901, loss_idmrf: 3.232406, loss_adv_gen: -49.549866
|
57 |
+
2023-02-13 02:53:16,156 - INFO - [Train] step: 11799, loss_adv_disc: -2.184320
|
58 |
+
2023-02-13 02:53:16,365 - INFO - [Train] step: 11799, loss_mpn: 0.009117, loss_rec: 0.041747, loss_semantic: 0.414465, loss_idmrf: 4.049892, loss_adv_gen: -93.858597
|
59 |
+
2023-02-13 02:53:45,611 - INFO - [Train] step: 11899, loss_adv_disc: -0.988897
|
60 |
+
2023-02-13 02:53:45,819 - INFO - [Train] step: 11899, loss_mpn: 0.005853, loss_rec: 0.026617, loss_semantic: 0.351113, loss_idmrf: 1.215564, loss_adv_gen: -80.830864
|
61 |
+
2023-02-13 02:54:15,241 - INFO - [Train] step: 11999, loss_adv_disc: 0.980708
|
62 |
+
2023-02-13 02:54:15,449 - INFO - [Train] step: 11999, loss_mpn: 0.021106, loss_rec: 0.040558, loss_semantic: 0.430699, loss_idmrf: 4.337394, loss_adv_gen: -98.472916
|
63 |
+
2023-02-13 02:54:20,591 - INFO - [Eval] step: 11999, bce: 0.448820, psnr: 23.549391, ssim: 0.926096
|
64 |
+
2023-02-13 02:54:50,910 - INFO - [Train] step: 12099, loss_adv_disc: -0.077401
|
65 |
+
2023-02-13 02:54:51,118 - INFO - [Train] step: 12099, loss_mpn: 0.010991, loss_rec: 0.039699, loss_semantic: 0.377598, loss_idmrf: 3.547094, loss_adv_gen: -59.509033
|
66 |
+
2023-02-13 02:55:20,352 - INFO - [Train] step: 12199, loss_adv_disc: -1.588708
|
67 |
+
2023-02-13 02:55:20,561 - INFO - [Train] step: 12199, loss_mpn: 0.012134, loss_rec: 0.032258, loss_semantic: 0.373746, loss_idmrf: 2.459065, loss_adv_gen: -80.642197
|
68 |
+
2023-02-13 02:55:49,796 - INFO - [Train] step: 12299, loss_adv_disc: -0.440839
|
69 |
+
2023-02-13 02:55:50,005 - INFO - [Train] step: 12299, loss_mpn: 0.007193, loss_rec: 0.028371, loss_semantic: 0.365545, loss_idmrf: 4.296700, loss_adv_gen: -86.054810
|
70 |
+
2023-02-13 02:56:19,442 - INFO - [Train] step: 12399, loss_adv_disc: -1.179346
|
71 |
+
2023-02-13 02:56:19,651 - INFO - [Train] step: 12399, loss_mpn: 0.014047, loss_rec: 0.025998, loss_semantic: 0.354992, loss_idmrf: 3.493587, loss_adv_gen: -57.490608
|
72 |
+
2023-02-13 02:56:48,891 - INFO - [Train] step: 12499, loss_adv_disc: -1.112655
|
73 |
+
2023-02-13 02:56:49,099 - INFO - [Train] step: 12499, loss_mpn: 0.013486, loss_rec: 0.032641, loss_semantic: 0.408824, loss_idmrf: 4.551135, loss_adv_gen: -69.951340
|
74 |
+
2023-02-13 02:57:18,345 - INFO - [Train] step: 12599, loss_adv_disc: 0.841728
|
75 |
+
2023-02-13 02:57:18,554 - INFO - [Train] step: 12599, loss_mpn: 0.008797, loss_rec: 0.028766, loss_semantic: 0.358703, loss_idmrf: 6.030992, loss_adv_gen: -90.953262
|
76 |
+
2023-02-13 02:57:47,993 - INFO - [Train] step: 12699, loss_adv_disc: 0.809182
|
77 |
+
2023-02-13 02:57:48,201 - INFO - [Train] step: 12699, loss_mpn: 0.004429, loss_rec: 0.026492, loss_semantic: 0.300694, loss_idmrf: 1.625929, loss_adv_gen: -77.897903
|
78 |
+
2023-02-13 02:58:17,440 - INFO - [Train] step: 12799, loss_adv_disc: -0.801807
|
79 |
+
2023-02-13 02:58:17,649 - INFO - [Train] step: 12799, loss_mpn: 0.014015, loss_rec: 0.034848, loss_semantic: 0.370866, loss_idmrf: 2.979873, loss_adv_gen: -68.310692
|
80 |
+
2023-02-13 02:58:46,886 - INFO - [Train] step: 12899, loss_adv_disc: -5.047458
|
81 |
+
2023-02-13 02:58:47,095 - INFO - [Train] step: 12899, loss_mpn: 0.010504, loss_rec: 0.033995, loss_semantic: 0.352904, loss_idmrf: 2.596021, loss_adv_gen: -68.085014
|
82 |
+
2023-02-13 02:59:16,521 - INFO - [Train] step: 12999, loss_adv_disc: -6.217038
|
83 |
+
2023-02-13 02:59:16,729 - INFO - [Train] step: 12999, loss_mpn: 0.009560, loss_rec: 0.042248, loss_semantic: 0.351271, loss_idmrf: 7.592016, loss_adv_gen: -72.404373
|
84 |
+
2023-02-13 02:59:21,915 - INFO - [Eval] step: 12999, bce: 0.403072, psnr: 23.723335, ssim: 0.927467
|
85 |
+
2023-02-13 02:59:52,416 - INFO - [Train] step: 13099, loss_adv_disc: -0.344683
|
86 |
+
2023-02-13 02:59:52,625 - INFO - [Train] step: 13099, loss_mpn: 0.008536, loss_rec: 0.027696, loss_semantic: 0.320824, loss_idmrf: 4.384482, loss_adv_gen: -36.807159
|
87 |
+
2023-02-13 03:00:21,858 - INFO - [Train] step: 13199, loss_adv_disc: -0.633883
|
88 |
+
2023-02-13 03:00:22,067 - INFO - [Train] step: 13199, loss_mpn: 0.011513, loss_rec: 0.032611, loss_semantic: 0.386401, loss_idmrf: 3.015574, loss_adv_gen: -84.160339
|
89 |
+
2023-02-13 03:00:51,326 - INFO - [Train] step: 13299, loss_adv_disc: -1.454661
|
90 |
+
2023-02-13 03:00:51,535 - INFO - [Train] step: 13299, loss_mpn: 0.010252, loss_rec: 0.031422, loss_semantic: 0.363536, loss_idmrf: 2.590449, loss_adv_gen: -77.080185
|
91 |
+
2023-02-13 03:01:20,963 - INFO - [Train] step: 13399, loss_adv_disc: -1.952087
|
92 |
+
2023-02-13 03:01:21,172 - INFO - [Train] step: 13399, loss_mpn: 0.006985, loss_rec: 0.028874, loss_semantic: 0.363051, loss_idmrf: 4.615744, loss_adv_gen: -86.813782
|
93 |
+
2023-02-13 03:01:50,420 - INFO - [Train] step: 13499, loss_adv_disc: -2.931163
|
94 |
+
2023-02-13 03:01:50,629 - INFO - [Train] step: 13499, loss_mpn: 0.007509, loss_rec: 0.041209, loss_semantic: 0.391465, loss_idmrf: 2.554505, loss_adv_gen: -65.555023
|
95 |
+
2023-02-13 03:02:19,886 - INFO - [Train] step: 13599, loss_adv_disc: -1.157901
|
96 |
+
2023-02-13 03:02:20,095 - INFO - [Train] step: 13599, loss_mpn: 0.024059, loss_rec: 0.033903, loss_semantic: 0.395873, loss_idmrf: 3.517569, loss_adv_gen: -80.516960
|
97 |
+
2023-02-13 03:02:49,519 - INFO - [Train] step: 13699, loss_adv_disc: -0.466357
|
98 |
+
2023-02-13 03:02:49,728 - INFO - [Train] step: 13699, loss_mpn: 0.009441, loss_rec: 0.038954, loss_semantic: 0.422468, loss_idmrf: 2.220746, loss_adv_gen: -57.919315
|
99 |
+
2023-02-13 03:03:18,965 - INFO - [Train] step: 13799, loss_adv_disc: -1.499573
|
100 |
+
2023-02-13 03:03:19,174 - INFO - [Train] step: 13799, loss_mpn: 0.021366, loss_rec: 0.046169, loss_semantic: 0.447270, loss_idmrf: 5.458885, loss_adv_gen: -85.748489
|
101 |
+
2023-02-13 03:03:48,422 - INFO - [Train] step: 13899, loss_adv_disc: -1.536951
|
102 |
+
2023-02-13 03:03:48,630 - INFO - [Train] step: 13899, loss_mpn: 0.009894, loss_rec: 0.037270, loss_semantic: 0.543508, loss_idmrf: 1.969022, loss_adv_gen: -76.695808
|
103 |
+
2023-02-13 03:04:18,050 - INFO - [Train] step: 13999, loss_adv_disc: -4.334954
|
104 |
+
2023-02-13 03:04:18,258 - INFO - [Train] step: 13999, loss_mpn: 0.008074, loss_rec: 0.032596, loss_semantic: 0.314528, loss_idmrf: 2.552278, loss_adv_gen: -43.856140
|
105 |
+
2023-02-13 03:04:23,391 - INFO - [Eval] step: 13999, bce: 0.459337, psnr: 23.695547, ssim: 0.927995
|
106 |
+
2023-02-13 03:04:53,694 - INFO - [Train] step: 14099, loss_adv_disc: -1.992434
|
107 |
+
2023-02-13 03:04:53,903 - INFO - [Train] step: 14099, loss_mpn: 0.022011, loss_rec: 0.037382, loss_semantic: 0.366072, loss_idmrf: 3.695474, loss_adv_gen: -59.897938
|
108 |
+
2023-02-13 03:05:23,137 - INFO - [Train] step: 14199, loss_adv_disc: -1.008330
|
109 |
+
2023-02-13 03:05:23,346 - INFO - [Train] step: 14199, loss_mpn: 0.006269, loss_rec: 0.036688, loss_semantic: 0.391129, loss_idmrf: 2.103096, loss_adv_gen: -51.805092
|
110 |
+
2023-02-13 03:05:52,601 - INFO - [Train] step: 14299, loss_adv_disc: -4.214204
|
111 |
+
2023-02-13 03:05:52,809 - INFO - [Train] step: 14299, loss_mpn: 0.014387, loss_rec: 0.059094, loss_semantic: 0.497419, loss_idmrf: 6.891026, loss_adv_gen: -68.543396
|
112 |
+
2023-02-13 03:06:22,244 - INFO - [Train] step: 14399, loss_adv_disc: -2.768673
|
113 |
+
2023-02-13 03:06:22,453 - INFO - [Train] step: 14399, loss_mpn: 0.006042, loss_rec: 0.028931, loss_semantic: 0.317719, loss_idmrf: 2.839618, loss_adv_gen: -49.560658
|
114 |
+
2023-02-13 03:06:51,708 - INFO - [Train] step: 14499, loss_adv_disc: -1.587522
|
115 |
+
2023-02-13 03:06:51,916 - INFO - [Train] step: 14499, loss_mpn: 0.013344, loss_rec: 0.031495, loss_semantic: 0.366431, loss_idmrf: 5.531614, loss_adv_gen: -43.433792
|
116 |
+
2023-02-13 03:07:21,159 - INFO - [Train] step: 14599, loss_adv_disc: -3.142626
|
117 |
+
2023-02-13 03:07:21,368 - INFO - [Train] step: 14599, loss_mpn: 0.015624, loss_rec: 0.041417, loss_semantic: 0.470021, loss_idmrf: 3.430955, loss_adv_gen: -44.935692
|
118 |
+
2023-02-13 03:07:50,822 - INFO - [Train] step: 14699, loss_adv_disc: -2.612098
|
119 |
+
2023-02-13 03:07:51,031 - INFO - [Train] step: 14699, loss_mpn: 0.011433, loss_rec: 0.045211, loss_semantic: 0.458431, loss_idmrf: 3.423684, loss_adv_gen: -78.737206
|
120 |
+
2023-02-13 03:08:20,280 - INFO - [Train] step: 14799, loss_adv_disc: -1.572463
|
121 |
+
2023-02-13 03:08:20,489 - INFO - [Train] step: 14799, loss_mpn: 0.010487, loss_rec: 0.042552, loss_semantic: 0.449359, loss_idmrf: 2.908170, loss_adv_gen: -67.777924
|
122 |
+
2023-02-13 03:08:49,729 - INFO - [Train] step: 14899, loss_adv_disc: -0.717800
|
123 |
+
2023-02-13 03:08:49,938 - INFO - [Train] step: 14899, loss_mpn: 0.010333, loss_rec: 0.027488, loss_semantic: 0.339489, loss_idmrf: 5.085710, loss_adv_gen: -69.917725
|
124 |
+
2023-02-13 03:09:19,376 - INFO - [Train] step: 14999, loss_adv_disc: -1.933161
|
125 |
+
2023-02-13 03:09:19,585 - INFO - [Train] step: 14999, loss_mpn: 0.018537, loss_rec: 0.033588, loss_semantic: 0.374348, loss_idmrf: 3.241080, loss_adv_gen: -35.936356
|
126 |
+
2023-02-13 03:09:24,728 - INFO - [Eval] step: 14999, bce: 0.398218, psnr: 23.808777, ssim: 0.929202
|
127 |
+
2023-02-13 03:09:55,338 - INFO - [Train] step: 15099, loss_adv_disc: -1.123843
|
128 |
+
2023-02-13 03:09:55,547 - INFO - [Train] step: 15099, loss_mpn: 0.010467, loss_rec: 0.036722, loss_semantic: 0.335451, loss_idmrf: 3.921709, loss_adv_gen: -65.612755
|
129 |
+
2023-02-13 03:10:24,781 - INFO - [Train] step: 15199, loss_adv_disc: -0.055525
|
130 |
+
2023-02-13 03:10:24,989 - INFO - [Train] step: 15199, loss_mpn: 0.008737, loss_rec: 0.030392, loss_semantic: 0.323074, loss_idmrf: 4.016256, loss_adv_gen: -63.581787
|
131 |
+
2023-02-13 03:10:54,249 - INFO - [Train] step: 15299, loss_adv_disc: -0.904782
|
132 |
+
2023-02-13 03:10:54,457 - INFO - [Train] step: 15299, loss_mpn: 0.007227, loss_rec: 0.025108, loss_semantic: 0.326624, loss_idmrf: 7.887135, loss_adv_gen: -52.554401
|
133 |
+
2023-02-13 03:11:23,871 - INFO - [Train] step: 15399, loss_adv_disc: -2.399804
|
134 |
+
2023-02-13 03:11:24,080 - INFO - [Train] step: 15399, loss_mpn: 0.011453, loss_rec: 0.036506, loss_semantic: 0.499180, loss_idmrf: 3.387949, loss_adv_gen: -48.605957
|
135 |
+
2023-02-13 03:11:53,310 - INFO - [Train] step: 15499, loss_adv_disc: -1.687280
|
136 |
+
2023-02-13 03:11:53,519 - INFO - [Train] step: 15499, loss_mpn: 0.006589, loss_rec: 0.029240, loss_semantic: 0.297095, loss_idmrf: 2.843406, loss_adv_gen: -62.710510
|
137 |
+
2023-02-13 03:12:22,762 - INFO - [Train] step: 15599, loss_adv_disc: -0.255836
|
138 |
+
2023-02-13 03:12:22,970 - INFO - [Train] step: 15599, loss_mpn: 0.005989, loss_rec: 0.031084, loss_semantic: 0.322900, loss_idmrf: 5.402665, loss_adv_gen: -63.991249
|
139 |
+
2023-02-13 03:12:52,388 - INFO - [Train] step: 15699, loss_adv_disc: -0.566187
|
140 |
+
2023-02-13 03:12:52,597 - INFO - [Train] step: 15699, loss_mpn: 0.013880, loss_rec: 0.024043, loss_semantic: 0.370826, loss_idmrf: 2.768170, loss_adv_gen: -44.828705
|
141 |
+
2023-02-13 03:13:21,837 - INFO - [Train] step: 15799, loss_adv_disc: -3.458609
|
142 |
+
2023-02-13 03:13:22,046 - INFO - [Train] step: 15799, loss_mpn: 0.011703, loss_rec: 0.035102, loss_semantic: 0.382987, loss_idmrf: 4.488460, loss_adv_gen: -65.005875
|
143 |
+
2023-02-13 03:13:51,286 - INFO - [Train] step: 15899, loss_adv_disc: -0.599114
|
144 |
+
2023-02-13 03:13:51,495 - INFO - [Train] step: 15899, loss_mpn: 0.012524, loss_rec: 0.025607, loss_semantic: 0.360270, loss_idmrf: 1.651819, loss_adv_gen: -54.153267
|
145 |
+
2023-02-13 03:14:20,919 - INFO - [Train] step: 15999, loss_adv_disc: -0.632309
|
146 |
+
2023-02-13 03:14:21,128 - INFO - [Train] step: 15999, loss_mpn: 0.005059, loss_rec: 0.024989, loss_semantic: 0.313581, loss_idmrf: 2.090605, loss_adv_gen: -50.700420
|
147 |
+
2023-02-13 03:14:26,263 - INFO - [Eval] step: 15999, bce: 0.407530, psnr: 23.910851, ssim: 0.929588
|
148 |
+
2023-02-13 03:14:56,787 - INFO - [Train] step: 16099, loss_adv_disc: -2.858291
|
149 |
+
2023-02-13 03:14:56,996 - INFO - [Train] step: 16099, loss_mpn: 0.009401, loss_rec: 0.040451, loss_semantic: 0.429576, loss_idmrf: 3.903901, loss_adv_gen: -67.689865
|
150 |
+
2023-02-13 03:15:26,239 - INFO - [Train] step: 16199, loss_adv_disc: -4.446222
|
151 |
+
2023-02-13 03:15:26,448 - INFO - [Train] step: 16199, loss_mpn: 0.010294, loss_rec: 0.029341, loss_semantic: 0.350636, loss_idmrf: 4.611250, loss_adv_gen: -58.366482
|
152 |
+
2023-02-13 03:15:55,697 - INFO - [Train] step: 16299, loss_adv_disc: -1.042235
|
153 |
+
2023-02-13 03:15:55,905 - INFO - [Train] step: 16299, loss_mpn: 0.008295, loss_rec: 0.034843, loss_semantic: 0.395846, loss_idmrf: 2.474969, loss_adv_gen: -40.464108
|
154 |
+
2023-02-13 03:16:25,337 - INFO - [Train] step: 16399, loss_adv_disc: -0.666275
|
155 |
+
2023-02-13 03:16:25,546 - INFO - [Train] step: 16399, loss_mpn: 0.008661, loss_rec: 0.032371, loss_semantic: 0.398836, loss_idmrf: 4.959391, loss_adv_gen: -58.231514
|
156 |
+
2023-02-13 03:16:54,787 - INFO - [Train] step: 16499, loss_adv_disc: -3.111873
|
157 |
+
2023-02-13 03:16:54,996 - INFO - [Train] step: 16499, loss_mpn: 0.006355, loss_rec: 0.030116, loss_semantic: 0.322977, loss_idmrf: 2.656320, loss_adv_gen: -59.407661
|
158 |
+
2023-02-13 03:17:24,246 - INFO - [Train] step: 16599, loss_adv_disc: -0.969188
|
159 |
+
2023-02-13 03:17:24,454 - INFO - [Train] step: 16599, loss_mpn: 0.006812, loss_rec: 0.030045, loss_semantic: 0.388530, loss_idmrf: 3.429312, loss_adv_gen: -55.191212
|
160 |
+
2023-02-13 03:17:53,885 - INFO - [Train] step: 16699, loss_adv_disc: -1.384235
|
161 |
+
2023-02-13 03:17:54,094 - INFO - [Train] step: 16699, loss_mpn: 0.007724, loss_rec: 0.036217, loss_semantic: 0.368885, loss_idmrf: 1.676530, loss_adv_gen: -57.781860
|
162 |
+
2023-02-13 03:18:23,332 - INFO - [Train] step: 16799, loss_adv_disc: -1.833328
|
163 |
+
2023-02-13 03:18:23,541 - INFO - [Train] step: 16799, loss_mpn: 0.011292, loss_rec: 0.038726, loss_semantic: 0.407536, loss_idmrf: 2.409158, loss_adv_gen: -69.407043
|
164 |
+
2023-02-13 03:18:52,791 - INFO - [Train] step: 16899, loss_adv_disc: -2.642775
|
165 |
+
2023-02-13 03:18:53,000 - INFO - [Train] step: 16899, loss_mpn: 0.017453, loss_rec: 0.039096, loss_semantic: 0.409500, loss_idmrf: 4.307015, loss_adv_gen: -69.141121
|
166 |
+
2023-02-13 03:19:22,424 - INFO - [Train] step: 16999, loss_adv_disc: -2.500937
|
167 |
+
2023-02-13 03:19:22,633 - INFO - [Train] step: 16999, loss_mpn: 0.011810, loss_rec: 0.031087, loss_semantic: 0.387880, loss_idmrf: 3.024927, loss_adv_gen: -49.219688
|
168 |
+
2023-02-13 03:19:27,776 - INFO - [Eval] step: 16999, bce: 0.381526, psnr: 23.834940, ssim: 0.929512
|
169 |
+
2023-02-13 03:19:58,077 - INFO - [Train] step: 17099, loss_adv_disc: -4.640062
|
170 |
+
2023-02-13 03:19:58,286 - INFO - [Train] step: 17099, loss_mpn: 0.005222, loss_rec: 0.031922, loss_semantic: 0.341579, loss_idmrf: 2.995645, loss_adv_gen: -48.879826
|
171 |
+
2023-02-13 03:20:27,517 - INFO - [Train] step: 17199, loss_adv_disc: -1.268022
|
172 |
+
2023-02-13 03:20:27,726 - INFO - [Train] step: 17199, loss_mpn: 0.006362, loss_rec: 0.028963, loss_semantic: 0.348143, loss_idmrf: 2.592186, loss_adv_gen: -58.502922
|
173 |
+
2023-02-13 03:20:56,971 - INFO - [Train] step: 17299, loss_adv_disc: -3.076265
|
174 |
+
2023-02-13 03:20:57,179 - INFO - [Train] step: 17299, loss_mpn: 0.005794, loss_rec: 0.032290, loss_semantic: 0.320258, loss_idmrf: 3.073877, loss_adv_gen: -75.206711
|
175 |
+
2023-02-13 03:21:26,600 - INFO - [Train] step: 17399, loss_adv_disc: -0.722616
|
176 |
+
2023-02-13 03:21:26,809 - INFO - [Train] step: 17399, loss_mpn: 0.008862, loss_rec: 0.027422, loss_semantic: 0.329777, loss_idmrf: 3.391365, loss_adv_gen: -45.755348
|
177 |
+
2023-02-13 03:21:56,062 - INFO - [Train] step: 17499, loss_adv_disc: -1.951949
|
178 |
+
2023-02-13 03:21:56,271 - INFO - [Train] step: 17499, loss_mpn: 0.022730, loss_rec: 0.036895, loss_semantic: 0.536559, loss_idmrf: 2.266129, loss_adv_gen: -33.774021
|
179 |
+
2023-02-13 03:22:25,513 - INFO - [Train] step: 17599, loss_adv_disc: -2.635029
|
180 |
+
2023-02-13 03:22:25,721 - INFO - [Train] step: 17599, loss_mpn: 0.020863, loss_rec: 0.034489, loss_semantic: 0.406302, loss_idmrf: 2.749728, loss_adv_gen: -68.144875
|
181 |
+
2023-02-13 03:22:55,157 - INFO - [Train] step: 17699, loss_adv_disc: -1.602936
|
182 |
+
2023-02-13 03:22:55,366 - INFO - [Train] step: 17699, loss_mpn: 0.009050, loss_rec: 0.035981, loss_semantic: 0.420545, loss_idmrf: 2.019552, loss_adv_gen: -59.465828
|
183 |
+
2023-02-13 03:23:24,609 - INFO - [Train] step: 17799, loss_adv_disc: -1.937828
|
184 |
+
2023-02-13 03:23:24,817 - INFO - [Train] step: 17799, loss_mpn: 0.016830, loss_rec: 0.032039, loss_semantic: 0.409238, loss_idmrf: 3.658442, loss_adv_gen: -62.794762
|
185 |
+
2023-02-13 03:23:54,047 - INFO - [Train] step: 17899, loss_adv_disc: -1.917367
|
186 |
+
2023-02-13 03:23:54,256 - INFO - [Train] step: 17899, loss_mpn: 0.011632, loss_rec: 0.032111, loss_semantic: 0.397115, loss_idmrf: 5.596826, loss_adv_gen: -54.250221
|
187 |
+
2023-02-13 03:24:23,692 - INFO - [Train] step: 17999, loss_adv_disc: -1.522330
|
188 |
+
2023-02-13 03:24:23,900 - INFO - [Train] step: 17999, loss_mpn: 0.008025, loss_rec: 0.027907, loss_semantic: 0.317568, loss_idmrf: 4.100923, loss_adv_gen: -41.896065
|
189 |
+
2023-02-13 03:24:29,031 - INFO - [Eval] step: 17999, bce: 0.442065, psnr: 23.760124, ssim: 0.929170
|
190 |
+
2023-02-13 03:24:59,321 - INFO - [Train] step: 18099, loss_adv_disc: -0.926828
|
191 |
+
2023-02-13 03:24:59,530 - INFO - [Train] step: 18099, loss_mpn: 0.005604, loss_rec: 0.026934, loss_semantic: 0.317587, loss_idmrf: 3.018780, loss_adv_gen: -43.219757
|
192 |
+
2023-02-13 03:25:28,761 - INFO - [Train] step: 18199, loss_adv_disc: -1.252304
|
193 |
+
2023-02-13 03:25:28,970 - INFO - [Train] step: 18199, loss_mpn: 0.006583, loss_rec: 0.036220, loss_semantic: 0.343108, loss_idmrf: 3.407384, loss_adv_gen: -59.127075
|
194 |
+
2023-02-13 03:25:58,215 - INFO - [Train] step: 18299, loss_adv_disc: -4.361234
|
195 |
+
2023-02-13 03:25:58,423 - INFO - [Train] step: 18299, loss_mpn: 0.015157, loss_rec: 0.037108, loss_semantic: 0.418717, loss_idmrf: 3.749311, loss_adv_gen: -60.900532
|
196 |
+
2023-02-13 03:26:27,850 - INFO - [Train] step: 18399, loss_adv_disc: -2.134433
|
197 |
+
2023-02-13 03:26:28,059 - INFO - [Train] step: 18399, loss_mpn: 0.008116, loss_rec: 0.027292, loss_semantic: 0.342987, loss_idmrf: 2.560256, loss_adv_gen: -48.888443
|
198 |
+
2023-02-13 03:26:57,292 - INFO - [Train] step: 18499, loss_adv_disc: 0.527846
|
199 |
+
2023-02-13 03:26:57,501 - INFO - [Train] step: 18499, loss_mpn: 0.006677, loss_rec: 0.026621, loss_semantic: 0.318599, loss_idmrf: 2.754797, loss_adv_gen: -46.165695
|
200 |
+
2023-02-13 03:27:26,725 - INFO - [Train] step: 18599, loss_adv_disc: -1.308938
|
201 |
+
2023-02-13 03:27:26,934 - INFO - [Train] step: 18599, loss_mpn: 0.007462, loss_rec: 0.039733, loss_semantic: 0.407243, loss_idmrf: 1.858337, loss_adv_gen: -37.666771
|
202 |
+
2023-02-13 03:27:56,360 - INFO - [Train] step: 18699, loss_adv_disc: -2.993707
|
203 |
+
2023-02-13 03:27:56,569 - INFO - [Train] step: 18699, loss_mpn: 0.004185, loss_rec: 0.024976, loss_semantic: 0.292228, loss_idmrf: 3.635601, loss_adv_gen: -41.624527
|
204 |
+
2023-02-13 03:28:25,807 - INFO - [Train] step: 18799, loss_adv_disc: -0.925315
|
205 |
+
2023-02-13 03:28:26,016 - INFO - [Train] step: 18799, loss_mpn: 0.005924, loss_rec: 0.023125, loss_semantic: 0.296790, loss_idmrf: 2.006775, loss_adv_gen: -57.868706
|
206 |
+
2023-02-13 03:28:55,264 - INFO - [Train] step: 18899, loss_adv_disc: -4.718306
|
207 |
+
2023-02-13 03:28:55,472 - INFO - [Train] step: 18899, loss_mpn: 0.010640, loss_rec: 0.041306, loss_semantic: 0.368742, loss_idmrf: 5.275894, loss_adv_gen: -66.048042
|
208 |
+
2023-02-13 03:29:24,901 - INFO - [Train] step: 18999, loss_adv_disc: -3.436166
|
209 |
+
2023-02-13 03:29:25,109 - INFO - [Train] step: 18999, loss_mpn: 0.008324, loss_rec: 0.031876, loss_semantic: 0.347499, loss_idmrf: 2.720975, loss_adv_gen: -78.016502
|
210 |
+
2023-02-13 03:29:30,252 - INFO - [Eval] step: 18999, bce: 0.449807, psnr: 23.802105, ssim: 0.928289
|
211 |
+
2023-02-13 03:30:00,570 - INFO - [Train] step: 19099, loss_adv_disc: -1.636162
|
212 |
+
2023-02-13 03:30:00,779 - INFO - [Train] step: 19099, loss_mpn: 0.006203, loss_rec: 0.037541, loss_semantic: 0.359517, loss_idmrf: 3.860636, loss_adv_gen: -16.672791
|
213 |
+
2023-02-13 03:30:30,016 - INFO - [Train] step: 19199, loss_adv_disc: -3.129387
|
214 |
+
2023-02-13 03:30:30,225 - INFO - [Train] step: 19199, loss_mpn: 0.007494, loss_rec: 0.033616, loss_semantic: 0.351235, loss_idmrf: 2.313427, loss_adv_gen: -65.895172
|
215 |
+
2023-02-13 03:30:59,645 - INFO - [Train] step: 19299, loss_adv_disc: -1.372097
|
216 |
+
2023-02-13 03:30:59,853 - INFO - [Train] step: 19299, loss_mpn: 0.007141, loss_rec: 0.030316, loss_semantic: 0.351007, loss_idmrf: 2.720200, loss_adv_gen: -61.847340
|
217 |
+
2023-02-13 03:31:29,092 - INFO - [Train] step: 19399, loss_adv_disc: -1.721190
|
218 |
+
2023-02-13 03:31:29,301 - INFO - [Train] step: 19399, loss_mpn: 0.006103, loss_rec: 0.029222, loss_semantic: 0.334996, loss_idmrf: 4.253203, loss_adv_gen: -60.822281
|
219 |
+
2023-02-13 03:31:58,551 - INFO - [Train] step: 19499, loss_adv_disc: -1.837099
|
220 |
+
2023-02-13 03:31:58,760 - INFO - [Train] step: 19499, loss_mpn: 0.010932, loss_rec: 0.026487, loss_semantic: 0.384942, loss_idmrf: 3.565819, loss_adv_gen: -45.608253
|
221 |
+
2023-02-13 03:32:28,006 - INFO - [Train] step: 19599, loss_adv_disc: -2.718598
|
222 |
+
2023-02-13 03:32:28,214 - INFO - [Train] step: 19599, loss_mpn: 0.015166, loss_rec: 0.035973, loss_semantic: 0.467589, loss_idmrf: 2.883530, loss_adv_gen: -50.148312
|
223 |
+
2023-02-13 03:32:57,640 - INFO - [Train] step: 19699, loss_adv_disc: -2.816616
|
224 |
+
2023-02-13 03:32:57,848 - INFO - [Train] step: 19699, loss_mpn: 0.016170, loss_rec: 0.035296, loss_semantic: 0.428836, loss_idmrf: 3.616713, loss_adv_gen: -58.801323
|
225 |
+
2023-02-13 03:33:27,079 - INFO - [Train] step: 19799, loss_adv_disc: -0.808167
|
226 |
+
2023-02-13 03:33:27,288 - INFO - [Train] step: 19799, loss_mpn: 0.006861, loss_rec: 0.030523, loss_semantic: 0.333927, loss_idmrf: 1.760766, loss_adv_gen: -48.423855
|
227 |
+
2023-02-13 03:33:56,540 - INFO - [Train] step: 19899, loss_adv_disc: -1.257056
|
228 |
+
2023-02-13 03:33:56,748 - INFO - [Train] step: 19899, loss_mpn: 0.007679, loss_rec: 0.029557, loss_semantic: 0.375780, loss_idmrf: 2.455881, loss_adv_gen: -60.748562
|
229 |
+
2023-02-13 03:34:26,168 - INFO - [Train] step: 19999, loss_adv_disc: -2.131047
|
230 |
+
2023-02-13 03:34:26,376 - INFO - [Train] step: 19999, loss_mpn: 0.007155, loss_rec: 0.033969, loss_semantic: 0.365665, loss_idmrf: 4.215178, loss_adv_gen: -42.286987
|
231 |
+
2023-02-13 03:34:31,535 - INFO - [Eval] step: 19999, bce: 0.532807, psnr: 23.844255, ssim: 0.928977
|
232 |
+
2023-02-13 03:35:01,994 - INFO - [Train] step: 20099, loss_adv_disc: -0.821020
|
233 |
+
2023-02-13 03:35:02,203 - INFO - [Train] step: 20099, loss_mpn: 0.002918, loss_rec: 0.019707, loss_semantic: 0.295061, loss_idmrf: 1.051597, loss_adv_gen: -54.710880
|
234 |
+
2023-02-13 03:35:31,436 - INFO - [Train] step: 20199, loss_adv_disc: -1.382209
|
235 |
+
2023-02-13 03:35:31,645 - INFO - [Train] step: 20199, loss_mpn: 0.008895, loss_rec: 0.024759, loss_semantic: 0.342038, loss_idmrf: 4.306022, loss_adv_gen: -59.216305
|
236 |
+
2023-02-13 03:36:01,086 - INFO - [Train] step: 20299, loss_adv_disc: -1.843528
|
237 |
+
2023-02-13 03:36:01,295 - INFO - [Train] step: 20299, loss_mpn: 0.021462, loss_rec: 0.036182, loss_semantic: 0.377382, loss_idmrf: 4.597513, loss_adv_gen: -64.265610
|
238 |
+
2023-02-13 03:36:30,533 - INFO - [Train] step: 20399, loss_adv_disc: -1.680128
|
239 |
+
2023-02-13 03:36:30,741 - INFO - [Train] step: 20399, loss_mpn: 0.006874, loss_rec: 0.026922, loss_semantic: 0.314610, loss_idmrf: 2.805595, loss_adv_gen: -47.953209
|
240 |
+
2023-02-13 03:36:59,988 - INFO - [Train] step: 20499, loss_adv_disc: -0.908575
|
241 |
+
2023-02-13 03:37:00,197 - INFO - [Train] step: 20499, loss_mpn: 0.012316, loss_rec: 0.027403, loss_semantic: 0.369375, loss_idmrf: 2.440661, loss_adv_gen: -50.821514
|
242 |
+
2023-02-13 03:37:29,441 - INFO - [Train] step: 20599, loss_adv_disc: -0.791389
|
243 |
+
2023-02-13 03:37:29,650 - INFO - [Train] step: 20599, loss_mpn: 0.005344, loss_rec: 0.021288, loss_semantic: 0.310707, loss_idmrf: 5.183637, loss_adv_gen: -62.245640
|
244 |
+
2023-02-13 03:37:59,069 - INFO - [Train] step: 20699, loss_adv_disc: -1.344117
|
245 |
+
2023-02-13 03:37:59,278 - INFO - [Train] step: 20699, loss_mpn: 0.007965, loss_rec: 0.025558, loss_semantic: 0.339915, loss_idmrf: 1.723730, loss_adv_gen: -37.892769
|
246 |
+
2023-02-13 03:38:28,512 - INFO - [Train] step: 20799, loss_adv_disc: -4.988375
|
247 |
+
2023-02-13 03:38:28,720 - INFO - [Train] step: 20799, loss_mpn: 0.012833, loss_rec: 0.037847, loss_semantic: 0.357962, loss_idmrf: 3.132791, loss_adv_gen: -62.692024
|
248 |
+
2023-02-13 03:38:57,955 - INFO - [Train] step: 20899, loss_adv_disc: -1.214686
|
249 |
+
2023-02-13 03:38:58,164 - INFO - [Train] step: 20899, loss_mpn: 0.010181, loss_rec: 0.043168, loss_semantic: 0.396244, loss_idmrf: 3.857690, loss_adv_gen: -67.181183
|
250 |
+
2023-02-13 03:39:27,591 - INFO - [Train] step: 20999, loss_adv_disc: -1.825191
|
251 |
+
2023-02-13 03:39:27,800 - INFO - [Train] step: 20999, loss_mpn: 0.010744, loss_rec: 0.033796, loss_semantic: 0.396960, loss_idmrf: 3.169326, loss_adv_gen: -53.664833
|
252 |
+
2023-02-13 03:39:32,938 - INFO - [Eval] step: 20999, bce: 0.412189, psnr: 23.995857, ssim: 0.931168
|
253 |
+
2023-02-13 03:40:03,449 - INFO - [Train] step: 21099, loss_adv_disc: -2.373983
|
254 |
+
2023-02-13 03:40:03,657 - INFO - [Train] step: 21099, loss_mpn: 0.016065, loss_rec: 0.031946, loss_semantic: 0.375466, loss_idmrf: 3.098600, loss_adv_gen: -42.354683
|
255 |
+
2023-02-13 03:40:32,892 - INFO - [Train] step: 21199, loss_adv_disc: -4.281846
|
256 |
+
2023-02-13 03:40:33,101 - INFO - [Train] step: 21199, loss_mpn: 0.008469, loss_rec: 0.030433, loss_semantic: 0.343839, loss_idmrf: 2.670589, loss_adv_gen: -42.038010
|
257 |
+
2023-02-13 03:41:02,534 - INFO - [Train] step: 21299, loss_adv_disc: -2.047416
|
258 |
+
2023-02-13 03:41:02,743 - INFO - [Train] step: 21299, loss_mpn: 0.007427, loss_rec: 0.032674, loss_semantic: 0.365108, loss_idmrf: 2.909791, loss_adv_gen: -58.266251
|
259 |
+
2023-02-13 03:41:31,982 - INFO - [Train] step: 21399, loss_adv_disc: -0.787121
|
260 |
+
2023-02-13 03:41:32,191 - INFO - [Train] step: 21399, loss_mpn: 0.007601, loss_rec: 0.032305, loss_semantic: 0.341065, loss_idmrf: 1.518853, loss_adv_gen: -49.361404
|
261 |
+
2023-02-13 03:42:01,440 - INFO - [Train] step: 21499, loss_adv_disc: -2.960994
|
262 |
+
2023-02-13 03:42:01,649 - INFO - [Train] step: 21499, loss_mpn: 0.010133, loss_rec: 0.033218, loss_semantic: 0.454507, loss_idmrf: 3.152321, loss_adv_gen: -43.761852
|
263 |
+
2023-02-13 03:42:30,902 - INFO - [Train] step: 21599, loss_adv_disc: -0.531810
|
264 |
+
2023-02-13 03:42:31,111 - INFO - [Train] step: 21599, loss_mpn: 0.006240, loss_rec: 0.026150, loss_semantic: 0.333196, loss_idmrf: 3.342569, loss_adv_gen: -49.895046
|
265 |
+
2023-02-13 03:43:00,552 - INFO - [Train] step: 21699, loss_adv_disc: -0.934840
|
266 |
+
2023-02-13 03:43:00,760 - INFO - [Train] step: 21699, loss_mpn: 0.006143, loss_rec: 0.025827, loss_semantic: 0.347423, loss_idmrf: 5.185640, loss_adv_gen: -61.407597
|
267 |
+
2023-02-13 03:43:30,007 - INFO - [Train] step: 21799, loss_adv_disc: -1.708362
|
268 |
+
2023-02-13 03:43:30,215 - INFO - [Train] step: 21799, loss_mpn: 0.007157, loss_rec: 0.033153, loss_semantic: 0.366003, loss_idmrf: 2.537788, loss_adv_gen: -56.243927
|
269 |
+
2023-02-13 03:43:59,453 - INFO - [Train] step: 21899, loss_adv_disc: -2.662016
|
270 |
+
2023-02-13 03:43:59,662 - INFO - [Train] step: 21899, loss_mpn: 0.011560, loss_rec: 0.034256, loss_semantic: 0.414814, loss_idmrf: 2.540663, loss_adv_gen: -47.157578
|
271 |
+
2023-02-13 03:44:29,088 - INFO - [Train] step: 21999, loss_adv_disc: -1.665895
|
272 |
+
2023-02-13 03:44:29,297 - INFO - [Train] step: 21999, loss_mpn: 0.012481, loss_rec: 0.041681, loss_semantic: 0.410288, loss_idmrf: 1.742691, loss_adv_gen: -62.708092
|
273 |
+
2023-02-13 03:44:34,479 - INFO - [Eval] step: 21999, bce: 0.464714, psnr: 23.809441, ssim: 0.930178
|
274 |
+
2023-02-13 03:45:04,803 - INFO - [Train] step: 22099, loss_adv_disc: -2.987253
|
275 |
+
2023-02-13 03:45:05,012 - INFO - [Train] step: 22099, loss_mpn: 0.007824, loss_rec: 0.031368, loss_semantic: 0.363590, loss_idmrf: 2.717836, loss_adv_gen: -58.619572
|
276 |
+
2023-02-13 03:45:34,241 - INFO - [Train] step: 22199, loss_adv_disc: -0.920993
|
277 |
+
2023-02-13 03:45:34,449 - INFO - [Train] step: 22199, loss_mpn: 0.009114, loss_rec: 0.032073, loss_semantic: 0.363013, loss_idmrf: 3.228037, loss_adv_gen: -51.014500
|
278 |
+
2023-02-13 03:46:03,893 - INFO - [Train] step: 22299, loss_adv_disc: -1.332216
|
279 |
+
2023-02-13 03:46:04,102 - INFO - [Train] step: 22299, loss_mpn: 0.005149, loss_rec: 0.023768, loss_semantic: 0.309608, loss_idmrf: 3.535915, loss_adv_gen: -58.052727
|
280 |
+
2023-02-13 03:46:33,337 - INFO - [Train] step: 22399, loss_adv_disc: -2.954251
|
281 |
+
2023-02-13 03:46:33,546 - INFO - [Train] step: 22399, loss_mpn: 0.008305, loss_rec: 0.042678, loss_semantic: 0.423436, loss_idmrf: 3.085995, loss_adv_gen: -57.386398
|
282 |
+
2023-02-13 03:47:02,796 - INFO - [Train] step: 22499, loss_adv_disc: -2.537665
|
283 |
+
2023-02-13 03:47:03,005 - INFO - [Train] step: 22499, loss_mpn: 0.010203, loss_rec: 0.033067, loss_semantic: 0.427873, loss_idmrf: 3.364084, loss_adv_gen: -31.977356
|
284 |
+
2023-02-13 03:47:32,266 - INFO - [Train] step: 22599, loss_adv_disc: -2.119593
|
285 |
+
2023-02-13 03:47:32,475 - INFO - [Train] step: 22599, loss_mpn: 0.009034, loss_rec: 0.026896, loss_semantic: 0.351944, loss_idmrf: 2.590875, loss_adv_gen: -39.202179
|
286 |
+
2023-02-13 03:48:01,904 - INFO - [Train] step: 22699, loss_adv_disc: -2.218577
|
287 |
+
2023-02-13 03:48:02,113 - INFO - [Train] step: 22699, loss_mpn: 0.005697, loss_rec: 0.031317, loss_semantic: 0.377514, loss_idmrf: 1.533009, loss_adv_gen: -43.550716
|
288 |
+
2023-02-13 03:48:31,363 - INFO - [Train] step: 22799, loss_adv_disc: -3.975915
|
289 |
+
2023-02-13 03:48:31,571 - INFO - [Train] step: 22799, loss_mpn: 0.004963, loss_rec: 0.030778, loss_semantic: 0.357679, loss_idmrf: 1.428168, loss_adv_gen: -64.493042
|
290 |
+
2023-02-13 03:49:00,815 - INFO - [Train] step: 22899, loss_adv_disc: -2.386568
|
291 |
+
2023-02-13 03:49:01,024 - INFO - [Train] step: 22899, loss_mpn: 0.008505, loss_rec: 0.033951, loss_semantic: 0.400746, loss_idmrf: 2.240148, loss_adv_gen: -60.405560
|
292 |
+
2023-02-13 03:49:30,445 - INFO - [Train] step: 22999, loss_adv_disc: -1.425534
|
293 |
+
2023-02-13 03:49:30,654 - INFO - [Train] step: 22999, loss_mpn: 0.004612, loss_rec: 0.022062, loss_semantic: 0.292219, loss_idmrf: 2.723917, loss_adv_gen: -54.489628
|
294 |
+
2023-02-13 03:49:35,816 - INFO - [Eval] step: 22999, bce: 0.496233, psnr: 23.705549, ssim: 0.930028
|
295 |
+
2023-02-13 03:50:06,130 - INFO - [Train] step: 23099, loss_adv_disc: -3.629030
|
296 |
+
2023-02-13 03:50:06,339 - INFO - [Train] step: 23099, loss_mpn: 0.017113, loss_rec: 0.039963, loss_semantic: 0.420710, loss_idmrf: 3.228497, loss_adv_gen: -66.667511
|
297 |
+
2023-02-13 03:50:35,579 - INFO - [Train] step: 23199, loss_adv_disc: -2.341859
|
298 |
+
2023-02-13 03:50:35,787 - INFO - [Train] step: 23199, loss_mpn: 0.007523, loss_rec: 0.030686, loss_semantic: 0.342419, loss_idmrf: 2.144747, loss_adv_gen: -73.500671
|
299 |
+
2023-02-13 03:51:05,245 - INFO - [Train] step: 23299, loss_adv_disc: -1.051663
|
300 |
+
2023-02-13 03:51:05,454 - INFO - [Train] step: 23299, loss_mpn: 0.005567, loss_rec: 0.023323, loss_semantic: 0.330779, loss_idmrf: 2.536284, loss_adv_gen: -54.168274
|
301 |
+
2023-02-13 03:51:34,704 - INFO - [Train] step: 23399, loss_adv_disc: -2.064084
|
302 |
+
2023-02-13 03:51:34,914 - INFO - [Train] step: 23399, loss_mpn: 0.005655, loss_rec: 0.025069, loss_semantic: 0.338070, loss_idmrf: 2.127142, loss_adv_gen: -64.372055
|
303 |
+
2023-02-13 03:52:04,164 - INFO - [Train] step: 23499, loss_adv_disc: -2.628796
|
304 |
+
2023-02-13 03:52:04,373 - INFO - [Train] step: 23499, loss_mpn: 0.007882, loss_rec: 0.035950, loss_semantic: 0.411128, loss_idmrf: 1.505162, loss_adv_gen: -54.485676
|
305 |
+
2023-02-13 03:52:33,627 - INFO - [Train] step: 23599, loss_adv_disc: -2.026811
|
306 |
+
2023-02-13 03:52:33,836 - INFO - [Train] step: 23599, loss_mpn: 0.008700, loss_rec: 0.021994, loss_semantic: 0.314727, loss_idmrf: 2.674738, loss_adv_gen: -50.828640
|
307 |
+
2023-02-13 03:53:03,268 - INFO - [Train] step: 23699, loss_adv_disc: -2.408108
|
308 |
+
2023-02-13 03:53:03,477 - INFO - [Train] step: 23699, loss_mpn: 0.007119, loss_rec: 0.029397, loss_semantic: 0.370820, loss_idmrf: 2.113973, loss_adv_gen: -64.023994
|
309 |
+
2023-02-13 03:53:32,736 - INFO - [Train] step: 23799, loss_adv_disc: -2.448371
|
310 |
+
2023-02-13 03:53:32,945 - INFO - [Train] step: 23799, loss_mpn: 0.009236, loss_rec: 0.038028, loss_semantic: 0.389971, loss_idmrf: 2.496542, loss_adv_gen: -68.355583
|
311 |
+
2023-02-13 03:54:02,191 - INFO - [Train] step: 23899, loss_adv_disc: -2.646384
|
312 |
+
2023-02-13 03:54:02,400 - INFO - [Train] step: 23899, loss_mpn: 0.005686, loss_rec: 0.026619, loss_semantic: 0.305794, loss_idmrf: 2.764585, loss_adv_gen: -58.164246
|
313 |
+
2023-02-13 03:54:31,844 - INFO - [Train] step: 23999, loss_adv_disc: -2.244050
|
314 |
+
2023-02-13 03:54:32,053 - INFO - [Train] step: 23999, loss_mpn: 0.019072, loss_rec: 0.036695, loss_semantic: 0.419800, loss_idmrf: 4.794349, loss_adv_gen: -56.691669
|
315 |
+
2023-02-13 03:54:37,188 - INFO - [Eval] step: 23999, bce: 0.423072, psnr: 23.929384, ssim: 0.931687
|
316 |
+
2023-02-13 03:55:07,474 - INFO - [Train] step: 24099, loss_adv_disc: -1.824091
|
317 |
+
2023-02-13 03:55:07,682 - INFO - [Train] step: 24099, loss_mpn: 0.007323, loss_rec: 0.025409, loss_semantic: 0.367337, loss_idmrf: 1.783256, loss_adv_gen: -41.997749
|
318 |
+
2023-02-13 03:55:36,922 - INFO - [Train] step: 24199, loss_adv_disc: -1.705926
|
319 |
+
2023-02-13 03:55:37,131 - INFO - [Train] step: 24199, loss_mpn: 0.006608, loss_rec: 0.024922, loss_semantic: 0.325300, loss_idmrf: 1.772946, loss_adv_gen: -55.280464
|
320 |
+
2023-02-13 03:56:06,571 - INFO - [Train] step: 24299, loss_adv_disc: -1.387813
|
321 |
+
2023-02-13 03:56:06,780 - INFO - [Train] step: 24299, loss_mpn: 0.008532, loss_rec: 0.024819, loss_semantic: 0.333036, loss_idmrf: 3.660525, loss_adv_gen: -42.061749
|
322 |
+
2023-02-13 03:56:36,009 - INFO - [Train] step: 24399, loss_adv_disc: -3.398212
|
323 |
+
2023-02-13 03:56:36,218 - INFO - [Train] step: 24399, loss_mpn: 0.004310, loss_rec: 0.025555, loss_semantic: 0.320220, loss_idmrf: 0.711168, loss_adv_gen: -52.639893
|
324 |
+
2023-02-13 03:57:05,452 - INFO - [Train] step: 24499, loss_adv_disc: -1.182582
|
325 |
+
2023-02-13 03:57:05,661 - INFO - [Train] step: 24499, loss_mpn: 0.004868, loss_rec: 0.026463, loss_semantic: 0.320026, loss_idmrf: 2.703800, loss_adv_gen: -44.305458
|
326 |
+
2023-02-13 03:57:34,890 - INFO - [Train] step: 24599, loss_adv_disc: -2.179983
|
327 |
+
2023-02-13 03:57:35,098 - INFO - [Train] step: 24599, loss_mpn: 0.008763, loss_rec: 0.021728, loss_semantic: 0.337773, loss_idmrf: 2.351331, loss_adv_gen: -39.223248
|
328 |
+
2023-02-13 03:58:04,537 - INFO - [Train] step: 24699, loss_adv_disc: -1.959560
|
329 |
+
2023-02-13 03:58:04,745 - INFO - [Train] step: 24699, loss_mpn: 0.007004, loss_rec: 0.029935, loss_semantic: 0.346946, loss_idmrf: 2.790915, loss_adv_gen: -56.569611
|
330 |
+
2023-02-13 03:58:33,993 - INFO - [Train] step: 24799, loss_adv_disc: -1.558927
|
331 |
+
2023-02-13 03:58:34,202 - INFO - [Train] step: 24799, loss_mpn: 0.013352, loss_rec: 0.030124, loss_semantic: 0.413617, loss_idmrf: 2.552111, loss_adv_gen: -56.299240
|
332 |
+
2023-02-13 03:59:03,447 - INFO - [Train] step: 24899, loss_adv_disc: -2.618589
|
333 |
+
2023-02-13 03:59:03,656 - INFO - [Train] step: 24899, loss_mpn: 0.012856, loss_rec: 0.033626, loss_semantic: 0.396584, loss_idmrf: 5.144225, loss_adv_gen: -46.024830
|
334 |
+
2023-02-13 03:59:33,088 - INFO - [Train] step: 24999, loss_adv_disc: -2.875488
|
335 |
+
2023-02-13 03:59:33,296 - INFO - [Train] step: 24999, loss_mpn: 0.006294, loss_rec: 0.032758, loss_semantic: 0.382921, loss_idmrf: 4.101409, loss_adv_gen: -38.091202
|
336 |
+
2023-02-13 03:59:38,456 - INFO - [Eval] step: 24999, bce: 0.494293, psnr: 23.910641, ssim: 0.931275
|
337 |
+
2023-02-13 04:00:08,901 - INFO - [Train] step: 25099, loss_adv_disc: -3.002494
|
338 |
+
2023-02-13 04:00:09,110 - INFO - [Train] step: 25099, loss_mpn: 0.012923, loss_rec: 0.042998, loss_semantic: 0.389094, loss_idmrf: 3.376410, loss_adv_gen: -64.067093
|
339 |
+
2023-02-13 04:00:38,346 - INFO - [Train] step: 25199, loss_adv_disc: -5.016815
|
340 |
+
2023-02-13 04:00:38,555 - INFO - [Train] step: 25199, loss_mpn: 0.014368, loss_rec: 0.053279, loss_semantic: 0.451728, loss_idmrf: 3.427337, loss_adv_gen: -61.353497
|
341 |
+
2023-02-13 04:01:07,993 - INFO - [Train] step: 25299, loss_adv_disc: -2.128964
|
342 |
+
2023-02-13 04:01:08,202 - INFO - [Train] step: 25299, loss_mpn: 0.007653, loss_rec: 0.031010, loss_semantic: 0.372408, loss_idmrf: 3.696013, loss_adv_gen: -37.674156
|
343 |
+
2023-02-13 04:01:37,446 - INFO - [Train] step: 25399, loss_adv_disc: -1.170438
|
344 |
+
2023-02-13 04:01:37,655 - INFO - [Train] step: 25399, loss_mpn: 0.007121, loss_rec: 0.023605, loss_semantic: 0.337867, loss_idmrf: 2.371561, loss_adv_gen: -44.922745
|
345 |
+
2023-02-13 04:02:06,886 - INFO - [Train] step: 25499, loss_adv_disc: -2.289455
|
346 |
+
2023-02-13 04:02:07,095 - INFO - [Train] step: 25499, loss_mpn: 0.009439, loss_rec: 0.030207, loss_semantic: 0.341227, loss_idmrf: 3.571287, loss_adv_gen: -52.630623
|
347 |
+
2023-02-13 04:02:36,331 - INFO - [Train] step: 25599, loss_adv_disc: -2.275971
|
348 |
+
2023-02-13 04:02:36,540 - INFO - [Train] step: 25599, loss_mpn: 0.010324, loss_rec: 0.033716, loss_semantic: 0.369033, loss_idmrf: 2.983631, loss_adv_gen: -56.771912
|
349 |
+
2023-02-13 04:03:05,956 - INFO - [Train] step: 25699, loss_adv_disc: -3.929438
|
350 |
+
2023-02-13 04:03:06,165 - INFO - [Train] step: 25699, loss_mpn: 0.005107, loss_rec: 0.023621, loss_semantic: 0.317293, loss_idmrf: 1.977758, loss_adv_gen: -38.194695
|
351 |
+
2023-02-13 04:03:35,407 - INFO - [Train] step: 25799, loss_adv_disc: -2.947953
|
352 |
+
2023-02-13 04:03:35,616 - INFO - [Train] step: 25799, loss_mpn: 0.007025, loss_rec: 0.026580, loss_semantic: 0.322663, loss_idmrf: 1.570272, loss_adv_gen: -59.964291
|
353 |
+
2023-02-13 04:04:04,848 - INFO - [Train] step: 25899, loss_adv_disc: -1.598645
|
354 |
+
2023-02-13 04:04:05,057 - INFO - [Train] step: 25899, loss_mpn: 0.005981, loss_rec: 0.039875, loss_semantic: 0.387087, loss_idmrf: 2.259921, loss_adv_gen: -31.552460
|
355 |
+
2023-02-13 04:04:34,480 - INFO - [Train] step: 25999, loss_adv_disc: -1.000123
|
356 |
+
2023-02-13 04:04:34,688 - INFO - [Train] step: 25999, loss_mpn: 0.011854, loss_rec: 0.030783, loss_semantic: 0.366924, loss_idmrf: 2.419249, loss_adv_gen: -42.398209
|
357 |
+
2023-02-13 04:04:39,819 - INFO - [Eval] step: 25999, bce: 0.379053, psnr: 23.972345, ssim: 0.932233
|
358 |
+
2023-02-13 04:05:10,126 - INFO - [Train] step: 26099, loss_adv_disc: -1.131559
|
359 |
+
2023-02-13 04:05:10,335 - INFO - [Train] step: 26099, loss_mpn: 0.007134, loss_rec: 0.021641, loss_semantic: 0.318042, loss_idmrf: 3.112147, loss_adv_gen: -53.450821
|
360 |
+
2023-02-13 04:05:39,581 - INFO - [Train] step: 26199, loss_adv_disc: -1.814842
|
361 |
+
2023-02-13 04:05:39,789 - INFO - [Train] step: 26199, loss_mpn: 0.006957, loss_rec: 0.031351, loss_semantic: 0.359431, loss_idmrf: 3.939166, loss_adv_gen: -49.258316
|
362 |
+
2023-02-13 04:06:09,217 - INFO - [Train] step: 26299, loss_adv_disc: -1.674930
|
363 |
+
2023-02-13 04:06:09,426 - INFO - [Train] step: 26299, loss_mpn: 0.005781, loss_rec: 0.029166, loss_semantic: 0.291324, loss_idmrf: 3.769212, loss_adv_gen: -36.555950
|
364 |
+
2023-02-13 04:06:38,659 - INFO - [Train] step: 26399, loss_adv_disc: -1.669163
|
365 |
+
2023-02-13 04:06:38,867 - INFO - [Train] step: 26399, loss_mpn: 0.009312, loss_rec: 0.021297, loss_semantic: 0.302380, loss_idmrf: 3.968942, loss_adv_gen: -35.490517
|
366 |
+
2023-02-13 04:07:08,106 - INFO - [Train] step: 26499, loss_adv_disc: -3.611588
|
367 |
+
2023-02-13 04:07:08,314 - INFO - [Train] step: 26499, loss_mpn: 0.012238, loss_rec: 0.030930, loss_semantic: 0.397214, loss_idmrf: 4.898618, loss_adv_gen: -69.384735
|
368 |
+
2023-02-13 04:07:37,547 - INFO - [Train] step: 26599, loss_adv_disc: -4.188798
|
369 |
+
2023-02-13 04:07:37,755 - INFO - [Train] step: 26599, loss_mpn: 0.014178, loss_rec: 0.041108, loss_semantic: 0.457722, loss_idmrf: 3.513246, loss_adv_gen: -88.662796
|
370 |
+
2023-02-13 04:08:07,192 - INFO - [Train] step: 26699, loss_adv_disc: -2.032719
|
371 |
+
2023-02-13 04:08:07,400 - INFO - [Train] step: 26699, loss_mpn: 0.005614, loss_rec: 0.028942, loss_semantic: 0.317826, loss_idmrf: 2.767789, loss_adv_gen: -53.963402
|
372 |
+
2023-02-13 04:08:36,641 - INFO - [Train] step: 26799, loss_adv_disc: -1.779737
|
373 |
+
2023-02-13 04:08:36,850 - INFO - [Train] step: 26799, loss_mpn: 0.005052, loss_rec: 0.023449, loss_semantic: 0.296901, loss_idmrf: 1.286634, loss_adv_gen: -62.788879
|
374 |
+
2023-02-13 04:09:06,077 - INFO - [Train] step: 26899, loss_adv_disc: -3.180129
|
375 |
+
2023-02-13 04:09:06,286 - INFO - [Train] step: 26899, loss_mpn: 0.010187, loss_rec: 0.027996, loss_semantic: 0.343080, loss_idmrf: 2.296761, loss_adv_gen: -52.778481
|
376 |
+
2023-02-13 04:09:35,690 - INFO - [Train] step: 26999, loss_adv_disc: -0.776358
|
377 |
+
2023-02-13 04:09:35,903 - INFO - [Train] step: 26999, loss_mpn: 0.003122, loss_rec: 0.022430, loss_semantic: 0.296052, loss_idmrf: 1.410728, loss_adv_gen: -56.631275
|
378 |
+
2023-02-13 04:09:41,048 - INFO - [Eval] step: 26999, bce: 0.467820, psnr: 23.892040, ssim: 0.931154
|
379 |
+
2023-02-13 04:10:11,354 - INFO - [Train] step: 27099, loss_adv_disc: -3.549163
|
380 |
+
2023-02-13 04:10:11,563 - INFO - [Train] step: 27099, loss_mpn: 0.012723, loss_rec: 0.045948, loss_semantic: 0.476646, loss_idmrf: 3.937117, loss_adv_gen: -45.297981
|
381 |
+
2023-02-13 04:10:40,799 - INFO - [Train] step: 27199, loss_adv_disc: -1.461819
|
382 |
+
2023-02-13 04:10:41,008 - INFO - [Train] step: 27199, loss_mpn: 0.006455, loss_rec: 0.025119, loss_semantic: 0.319218, loss_idmrf: 1.629803, loss_adv_gen: -49.119114
|
383 |
+
2023-02-13 04:11:10,427 - INFO - [Train] step: 27299, loss_adv_disc: -1.982147
|
384 |
+
2023-02-13 04:11:10,636 - INFO - [Train] step: 27299, loss_mpn: 0.007175, loss_rec: 0.028984, loss_semantic: 0.362375, loss_idmrf: 1.994591, loss_adv_gen: -35.928234
|
385 |
+
2023-02-13 04:11:39,875 - INFO - [Train] step: 27399, loss_adv_disc: -1.590581
|
386 |
+
2023-02-13 04:11:40,084 - INFO - [Train] step: 27399, loss_mpn: 0.008258, loss_rec: 0.030676, loss_semantic: 0.359340, loss_idmrf: 5.009028, loss_adv_gen: -66.535286
|
387 |
+
2023-02-13 04:12:09,315 - INFO - [Train] step: 27499, loss_adv_disc: -0.402845
|
388 |
+
2023-02-13 04:12:09,524 - INFO - [Train] step: 27499, loss_mpn: 0.005017, loss_rec: 0.018790, loss_semantic: 0.267146, loss_idmrf: 3.032572, loss_adv_gen: -31.063499
|
389 |
+
2023-02-13 04:12:38,945 - INFO - [Train] step: 27599, loss_adv_disc: -2.334307
|
390 |
+
2023-02-13 04:12:39,153 - INFO - [Train] step: 27599, loss_mpn: 0.009680, loss_rec: 0.031130, loss_semantic: 0.403578, loss_idmrf: 2.780985, loss_adv_gen: -43.041565
|
391 |
+
2023-02-13 04:13:08,383 - INFO - [Train] step: 27699, loss_adv_disc: -3.015743
|
392 |
+
2023-02-13 04:13:08,591 - INFO - [Train] step: 27699, loss_mpn: 0.018975, loss_rec: 0.047783, loss_semantic: 0.425497, loss_idmrf: 3.275642, loss_adv_gen: -46.561310
|
393 |
+
2023-02-13 04:13:37,834 - INFO - [Train] step: 27799, loss_adv_disc: -2.397944
|
394 |
+
2023-02-13 04:13:38,042 - INFO - [Train] step: 27799, loss_mpn: 0.008875, loss_rec: 0.022515, loss_semantic: 0.318298, loss_idmrf: 3.172273, loss_adv_gen: -50.527237
|
395 |
+
2023-02-13 04:14:07,284 - INFO - [Train] step: 27899, loss_adv_disc: -3.253990
|
396 |
+
2023-02-13 04:14:07,492 - INFO - [Train] step: 27899, loss_mpn: 0.008475, loss_rec: 0.030102, loss_semantic: 0.385625, loss_idmrf: 3.359437, loss_adv_gen: -54.444069
|
397 |
+
2023-02-13 04:14:36,918 - INFO - [Train] step: 27999, loss_adv_disc: -1.056982
|
398 |
+
2023-02-13 04:14:37,126 - INFO - [Train] step: 27999, loss_mpn: 0.007983, loss_rec: 0.024990, loss_semantic: 0.312865, loss_idmrf: 1.289099, loss_adv_gen: -40.702690
|
399 |
+
2023-02-13 04:14:42,289 - INFO - [Eval] step: 27999, bce: 0.407889, psnr: 24.050518, ssim: 0.932894
|
400 |
+
2023-02-13 04:15:12,757 - INFO - [Train] step: 28099, loss_adv_disc: -3.248437
|
401 |
+
2023-02-13 04:15:12,965 - INFO - [Train] step: 28099, loss_mpn: 0.006869, loss_rec: 0.028881, loss_semantic: 0.362487, loss_idmrf: 1.182486, loss_adv_gen: -54.308258
|
402 |
+
2023-02-13 04:15:42,196 - INFO - [Train] step: 28199, loss_adv_disc: -2.695652
|
403 |
+
2023-02-13 04:15:42,405 - INFO - [Train] step: 28199, loss_mpn: 0.006013, loss_rec: 0.027336, loss_semantic: 0.359971, loss_idmrf: 1.179807, loss_adv_gen: -31.996471
|
404 |
+
2023-02-13 04:16:11,831 - INFO - [Train] step: 28299, loss_adv_disc: -2.555561
|
405 |
+
2023-02-13 04:16:12,040 - INFO - [Train] step: 28299, loss_mpn: 0.006889, loss_rec: 0.026413, loss_semantic: 0.323360, loss_idmrf: 2.315033, loss_adv_gen: -45.354450
|
406 |
+
2023-02-13 04:16:41,278 - INFO - [Train] step: 28399, loss_adv_disc: -0.930417
|
407 |
+
2023-02-13 04:16:41,486 - INFO - [Train] step: 28399, loss_mpn: 0.003901, loss_rec: 0.020817, loss_semantic: 0.298452, loss_idmrf: 2.879560, loss_adv_gen: -52.044228
|
408 |
+
2023-02-13 04:17:10,724 - INFO - [Train] step: 28499, loss_adv_disc: -1.263307
|
409 |
+
2023-02-13 04:17:10,933 - INFO - [Train] step: 28499, loss_mpn: 0.008175, loss_rec: 0.027284, loss_semantic: 0.342604, loss_idmrf: 1.002726, loss_adv_gen: -47.339424
|
410 |
+
2023-02-13 04:17:40,360 - INFO - [Train] step: 28599, loss_adv_disc: -3.608934
|
411 |
+
2023-02-13 04:17:40,568 - INFO - [Train] step: 28599, loss_mpn: 0.009251, loss_rec: 0.031929, loss_semantic: 0.374587, loss_idmrf: 1.810358, loss_adv_gen: -42.233673
|
412 |
+
2023-02-13 04:18:09,808 - INFO - [Train] step: 28699, loss_adv_disc: -1.356530
|
413 |
+
2023-02-13 04:18:10,017 - INFO - [Train] step: 28699, loss_mpn: 0.006944, loss_rec: 0.023396, loss_semantic: 0.342738, loss_idmrf: 1.533525, loss_adv_gen: -38.768314
|
414 |
+
2023-02-13 04:18:39,252 - INFO - [Train] step: 28799, loss_adv_disc: -1.913511
|
415 |
+
2023-02-13 04:18:39,461 - INFO - [Train] step: 28799, loss_mpn: 0.005483, loss_rec: 0.022142, loss_semantic: 0.296328, loss_idmrf: 1.239522, loss_adv_gen: -41.124989
|
416 |
+
2023-02-13 04:19:08,710 - INFO - [Train] step: 28899, loss_adv_disc: -1.888143
|
417 |
+
2023-02-13 04:19:08,918 - INFO - [Train] step: 28899, loss_mpn: 0.006542, loss_rec: 0.025403, loss_semantic: 0.326764, loss_idmrf: 4.936932, loss_adv_gen: -59.378212
|
418 |
+
2023-02-13 04:19:38,344 - INFO - [Train] step: 28999, loss_adv_disc: -1.588324
|
419 |
+
2023-02-13 04:19:38,553 - INFO - [Train] step: 28999, loss_mpn: 0.006560, loss_rec: 0.025576, loss_semantic: 0.327476, loss_idmrf: 3.260027, loss_adv_gen: -48.005432
|
420 |
+
2023-02-13 04:19:43,698 - INFO - [Eval] step: 28999, bce: 0.465755, psnr: 23.979197, ssim: 0.932822
|
421 |
+
2023-02-13 04:20:13,994 - INFO - [Train] step: 29099, loss_adv_disc: -2.685792
|
422 |
+
2023-02-13 04:20:14,203 - INFO - [Train] step: 29099, loss_mpn: 0.012146, loss_rec: 0.029994, loss_semantic: 0.365296, loss_idmrf: 2.682022, loss_adv_gen: -36.540726
|
423 |
+
2023-02-13 04:20:43,452 - INFO - [Train] step: 29199, loss_adv_disc: -1.354690
|
424 |
+
2023-02-13 04:20:43,661 - INFO - [Train] step: 29199, loss_mpn: 0.007733, loss_rec: 0.025824, loss_semantic: 0.352530, loss_idmrf: 1.492018, loss_adv_gen: -59.247963
|
425 |
+
2023-02-13 04:21:13,086 - INFO - [Train] step: 29299, loss_adv_disc: -0.308071
|
426 |
+
2023-02-13 04:21:13,295 - INFO - [Train] step: 29299, loss_mpn: 0.005732, loss_rec: 0.027066, loss_semantic: 0.429230, loss_idmrf: 5.198031, loss_adv_gen: -36.152596
|
427 |
+
2023-02-13 04:21:42,523 - INFO - [Train] step: 29399, loss_adv_disc: -5.704453
|
428 |
+
2023-02-13 04:21:42,733 - INFO - [Train] step: 29399, loss_mpn: 0.009742, loss_rec: 0.036058, loss_semantic: 0.422115, loss_idmrf: 3.425453, loss_adv_gen: -64.740868
|
429 |
+
2023-02-13 04:22:11,979 - INFO - [Train] step: 29499, loss_adv_disc: -3.116785
|
430 |
+
2023-02-13 04:22:12,187 - INFO - [Train] step: 29499, loss_mpn: 0.009215, loss_rec: 0.025835, loss_semantic: 0.381380, loss_idmrf: 4.609172, loss_adv_gen: -35.275311
|
431 |
+
2023-02-13 04:22:41,608 - INFO - [Train] step: 29599, loss_adv_disc: -3.281649
|
432 |
+
2023-02-13 04:22:41,817 - INFO - [Train] step: 29599, loss_mpn: 0.008651, loss_rec: 0.032746, loss_semantic: 0.376462, loss_idmrf: 1.623248, loss_adv_gen: -55.076706
|
433 |
+
2023-02-13 04:23:11,065 - INFO - [Train] step: 29699, loss_adv_disc: -1.218207
|
434 |
+
2023-02-13 04:23:11,274 - INFO - [Train] step: 29699, loss_mpn: 0.003264, loss_rec: 0.019132, loss_semantic: 0.286781, loss_idmrf: 2.779821, loss_adv_gen: -47.997040
|
435 |
+
2023-02-13 04:23:40,521 - INFO - [Train] step: 29799, loss_adv_disc: -4.113495
|
436 |
+
2023-02-13 04:23:40,730 - INFO - [Train] step: 29799, loss_mpn: 0.014174, loss_rec: 0.033425, loss_semantic: 0.392298, loss_idmrf: 2.298213, loss_adv_gen: -53.244530
|
437 |
+
2023-02-13 04:24:09,972 - INFO - [Train] step: 29899, loss_adv_disc: -2.777640
|
438 |
+
2023-02-13 04:24:10,181 - INFO - [Train] step: 29899, loss_mpn: 0.010283, loss_rec: 0.027365, loss_semantic: 0.370754, loss_idmrf: 2.673162, loss_adv_gen: -44.858421
|
439 |
+
2023-02-13 04:24:39,617 - INFO - [Train] step: 29999, loss_adv_disc: -1.686746
|
440 |
+
2023-02-13 04:24:39,826 - INFO - [Train] step: 29999, loss_mpn: 0.006264, loss_rec: 0.021137, loss_semantic: 0.292509, loss_idmrf: 2.118923, loss_adv_gen: -39.865761
|
441 |
+
2023-02-13 04:24:44,947 - INFO - [Eval] step: 29999, bce: 0.492185, psnr: 23.945683, ssim: 0.932370
|
442 |
+
2023-02-13 04:25:15,393 - INFO - [Train] step: 30099, loss_adv_disc: -2.197589
|
443 |
+
2023-02-13 04:25:15,602 - INFO - [Train] step: 30099, loss_mpn: 0.014454, loss_rec: 0.025383, loss_semantic: 0.374786, loss_idmrf: 2.624727, loss_adv_gen: -40.343857
|
444 |
+
2023-02-13 04:25:44,882 - INFO - [Train] step: 30199, loss_adv_disc: -1.674134
|
445 |
+
2023-02-13 04:25:45,091 - INFO - [Train] step: 30199, loss_mpn: 0.006576, loss_rec: 0.023323, loss_semantic: 0.302643, loss_idmrf: 1.844227, loss_adv_gen: -39.347649
|
446 |
+
2023-02-13 04:26:14,522 - INFO - [Train] step: 30299, loss_adv_disc: -2.467939
|
447 |
+
2023-02-13 04:26:14,730 - INFO - [Train] step: 30299, loss_mpn: 0.004610, loss_rec: 0.019071, loss_semantic: 0.272741, loss_idmrf: 2.557956, loss_adv_gen: -35.477177
|
448 |
+
2023-02-13 04:26:43,963 - INFO - [Train] step: 30399, loss_adv_disc: -3.106873
|
449 |
+
2023-02-13 04:26:44,171 - INFO - [Train] step: 30399, loss_mpn: 0.011434, loss_rec: 0.034334, loss_semantic: 0.442513, loss_idmrf: 2.949084, loss_adv_gen: -48.490547
|
450 |
+
2023-02-13 04:27:13,406 - INFO - [Train] step: 30499, loss_adv_disc: -3.356421
|
451 |
+
2023-02-13 04:27:13,614 - INFO - [Train] step: 30499, loss_mpn: 0.006068, loss_rec: 0.027413, loss_semantic: 0.352242, loss_idmrf: 2.848948, loss_adv_gen: -59.479187
|
452 |
+
2023-02-13 04:27:43,044 - INFO - [Train] step: 30599, loss_adv_disc: -1.141179
|
453 |
+
2023-02-13 04:27:43,253 - INFO - [Train] step: 30599, loss_mpn: 0.004855, loss_rec: 0.017252, loss_semantic: 0.252036, loss_idmrf: 1.422145, loss_adv_gen: -30.667336
|
454 |
+
2023-02-13 04:28:12,498 - INFO - [Train] step: 30699, loss_adv_disc: -2.391990
|
455 |
+
2023-02-13 04:28:12,707 - INFO - [Train] step: 30699, loss_mpn: 0.008719, loss_rec: 0.027900, loss_semantic: 0.338170, loss_idmrf: 3.741982, loss_adv_gen: -48.991299
|
456 |
+
2023-02-13 04:28:41,943 - INFO - [Train] step: 30799, loss_adv_disc: -2.743029
|
457 |
+
2023-02-13 04:28:42,152 - INFO - [Train] step: 30799, loss_mpn: 0.007150, loss_rec: 0.029743, loss_semantic: 0.353363, loss_idmrf: 1.883271, loss_adv_gen: -57.676750
|
458 |
+
2023-02-13 04:29:11,394 - INFO - [Train] step: 30899, loss_adv_disc: -1.434591
|
459 |
+
2023-02-13 04:29:11,603 - INFO - [Train] step: 30899, loss_mpn: 0.004940, loss_rec: 0.021217, loss_semantic: 0.291477, loss_idmrf: 1.621642, loss_adv_gen: -61.386360
|
460 |
+
2023-02-13 04:29:41,027 - INFO - [Train] step: 30999, loss_adv_disc: -3.122560
|
461 |
+
2023-02-13 04:29:41,235 - INFO - [Train] step: 30999, loss_mpn: 0.007449, loss_rec: 0.035113, loss_semantic: 0.435334, loss_idmrf: 3.818401, loss_adv_gen: -52.961311
|
462 |
+
2023-02-13 04:29:46,393 - INFO - [Eval] step: 30999, bce: 0.490164, psnr: 23.965448, ssim: 0.932622
|
463 |
+
2023-02-13 04:30:16,716 - INFO - [Train] step: 31099, loss_adv_disc: -2.185221
|
464 |
+
2023-02-13 04:30:16,924 - INFO - [Train] step: 31099, loss_mpn: 0.004554, loss_rec: 0.020886, loss_semantic: 0.288429, loss_idmrf: 1.497186, loss_adv_gen: -52.509830
|
465 |
+
2023-02-13 04:30:46,170 - INFO - [Train] step: 31199, loss_adv_disc: -4.675382
|
466 |
+
2023-02-13 04:30:46,378 - INFO - [Train] step: 31199, loss_mpn: 0.011436, loss_rec: 0.037177, loss_semantic: 0.422422, loss_idmrf: 3.885715, loss_adv_gen: -55.271225
|
467 |
+
2023-02-13 04:31:15,811 - INFO - [Train] step: 31299, loss_adv_disc: -1.755943
|
468 |
+
2023-02-13 04:31:16,020 - INFO - [Train] step: 31299, loss_mpn: 0.009999, loss_rec: 0.023714, loss_semantic: 0.357692, loss_idmrf: 2.204484, loss_adv_gen: -36.626755
|
469 |
+
2023-02-13 04:31:45,265 - INFO - [Train] step: 31399, loss_adv_disc: -1.392416
|
470 |
+
2023-02-13 04:31:45,474 - INFO - [Train] step: 31399, loss_mpn: 0.007405, loss_rec: 0.023420, loss_semantic: 0.318830, loss_idmrf: 2.317692, loss_adv_gen: -43.350834
|
471 |
+
2023-02-13 04:32:14,705 - INFO - [Train] step: 31499, loss_adv_disc: -2.211383
|
472 |
+
2023-02-13 04:32:14,914 - INFO - [Train] step: 31499, loss_mpn: 0.006516, loss_rec: 0.028322, loss_semantic: 0.381882, loss_idmrf: 1.678529, loss_adv_gen: -37.920712
|
473 |
+
2023-02-13 04:32:44,356 - INFO - [Train] step: 31599, loss_adv_disc: -1.601540
|
474 |
+
2023-02-13 04:32:44,564 - INFO - [Train] step: 31599, loss_mpn: 0.009201, loss_rec: 0.031845, loss_semantic: 0.375913, loss_idmrf: 1.841415, loss_adv_gen: -47.110973
|
475 |
+
2023-02-13 04:33:13,811 - INFO - [Train] step: 31699, loss_adv_disc: -1.545524
|
476 |
+
2023-02-13 04:33:14,019 - INFO - [Train] step: 31699, loss_mpn: 0.008448, loss_rec: 0.037653, loss_semantic: 0.368679, loss_idmrf: 2.739496, loss_adv_gen: -25.949142
|
477 |
+
2023-02-13 04:33:43,266 - INFO - [Train] step: 31799, loss_adv_disc: -3.020606
|
478 |
+
2023-02-13 04:33:43,475 - INFO - [Train] step: 31799, loss_mpn: 0.015079, loss_rec: 0.038613, loss_semantic: 0.437632, loss_idmrf: 2.923442, loss_adv_gen: -38.652390
|
479 |
+
2023-02-13 04:34:12,732 - INFO - [Train] step: 31899, loss_adv_disc: -2.436349
|
480 |
+
2023-02-13 04:34:12,940 - INFO - [Train] step: 31899, loss_mpn: 0.007766, loss_rec: 0.024789, loss_semantic: 0.337582, loss_idmrf: 2.340315, loss_adv_gen: -43.865879
|
481 |
+
2023-02-13 04:34:42,373 - INFO - [Train] step: 31999, loss_adv_disc: -2.298278
|
482 |
+
2023-02-13 04:34:42,582 - INFO - [Train] step: 31999, loss_mpn: 0.008034, loss_rec: 0.031308, loss_semantic: 0.375757, loss_idmrf: 2.158618, loss_adv_gen: -53.958466
|
483 |
+
2023-02-13 04:34:47,725 - INFO - [Eval] step: 31999, bce: 0.511144, psnr: 23.987217, ssim: 0.932225
|
484 |
+
2023-02-13 04:35:18,021 - INFO - [Train] step: 32099, loss_adv_disc: -2.084677
|
485 |
+
2023-02-13 04:35:18,229 - INFO - [Train] step: 32099, loss_mpn: 0.009502, loss_rec: 0.025188, loss_semantic: 0.323963, loss_idmrf: 3.507452, loss_adv_gen: -46.396492
|
486 |
+
2023-02-13 04:35:47,464 - INFO - [Train] step: 32199, loss_adv_disc: -4.015974
|
487 |
+
2023-02-13 04:35:47,673 - INFO - [Train] step: 32199, loss_mpn: 0.009766, loss_rec: 0.030390, loss_semantic: 0.407730, loss_idmrf: 5.195652, loss_adv_gen: -59.783085
|
488 |
+
2023-02-13 04:36:17,108 - INFO - [Train] step: 32299, loss_adv_disc: -3.174493
|
489 |
+
2023-02-13 04:36:17,317 - INFO - [Train] step: 32299, loss_mpn: 0.009587, loss_rec: 0.022740, loss_semantic: 0.340636, loss_idmrf: 1.810525, loss_adv_gen: -43.734375
|
490 |
+
2023-02-13 04:36:46,557 - INFO - [Train] step: 32399, loss_adv_disc: -2.961869
|
491 |
+
2023-02-13 04:36:46,766 - INFO - [Train] step: 32399, loss_mpn: 0.006834, loss_rec: 0.026820, loss_semantic: 0.367303, loss_idmrf: 3.318485, loss_adv_gen: -48.750732
|
492 |
+
2023-02-13 04:37:16,014 - INFO - [Train] step: 32499, loss_adv_disc: -1.361956
|
493 |
+
2023-02-13 04:37:16,223 - INFO - [Train] step: 32499, loss_mpn: 0.003526, loss_rec: 0.021565, loss_semantic: 0.330520, loss_idmrf: 0.975429, loss_adv_gen: -30.753990
|
494 |
+
2023-02-13 04:37:45,659 - INFO - [Train] step: 32599, loss_adv_disc: 0.066008
|
495 |
+
2023-02-13 04:37:45,868 - INFO - [Train] step: 32599, loss_mpn: 0.011720, loss_rec: 0.032479, loss_semantic: 0.369994, loss_idmrf: 3.079811, loss_adv_gen: -62.844032
|
496 |
+
2023-02-13 04:38:15,115 - INFO - [Train] step: 32699, loss_adv_disc: -2.147907
|
497 |
+
2023-02-13 04:38:15,323 - INFO - [Train] step: 32699, loss_mpn: 0.010529, loss_rec: 0.030653, loss_semantic: 0.365616, loss_idmrf: 2.815894, loss_adv_gen: -38.179665
|
498 |
+
2023-02-13 04:38:44,576 - INFO - [Train] step: 32799, loss_adv_disc: -3.022984
|
499 |
+
2023-02-13 04:38:44,784 - INFO - [Train] step: 32799, loss_mpn: 0.007264, loss_rec: 0.024473, loss_semantic: 0.354572, loss_idmrf: 2.879258, loss_adv_gen: -47.123955
|
500 |
+
2023-02-13 04:39:14,018 - INFO - [Train] step: 32899, loss_adv_disc: -2.134367
|
501 |
+
2023-02-13 04:39:14,226 - INFO - [Train] step: 32899, loss_mpn: 0.010902, loss_rec: 0.024346, loss_semantic: 0.349439, loss_idmrf: 2.405934, loss_adv_gen: -49.264977
|
502 |
+
2023-02-13 04:39:43,683 - INFO - [Train] step: 32999, loss_adv_disc: -1.867232
|
503 |
+
2023-02-13 04:39:43,891 - INFO - [Train] step: 32999, loss_mpn: 0.007945, loss_rec: 0.030071, loss_semantic: 0.401522, loss_idmrf: 1.317394, loss_adv_gen: -43.877331
|
504 |
+
2023-02-13 04:39:49,055 - INFO - [Eval] step: 32999, bce: 0.449431, psnr: 23.902346, ssim: 0.932984
|
505 |
+
2023-02-13 04:40:19,349 - INFO - [Train] step: 33099, loss_adv_disc: -2.116894
|
506 |
+
2023-02-13 04:40:19,558 - INFO - [Train] step: 33099, loss_mpn: 0.010758, loss_rec: 0.024110, loss_semantic: 0.369346, loss_idmrf: 1.759542, loss_adv_gen: -47.994667
|
507 |
+
2023-02-13 04:40:48,792 - INFO - [Train] step: 33199, loss_adv_disc: -1.882688
|
508 |
+
2023-02-13 04:40:49,001 - INFO - [Train] step: 33199, loss_mpn: 0.008150, loss_rec: 0.024921, loss_semantic: 0.324246, loss_idmrf: 3.030573, loss_adv_gen: -52.880379
|
509 |
+
2023-02-13 04:41:18,442 - INFO - [Train] step: 33299, loss_adv_disc: -1.602988
|
510 |
+
2023-02-13 04:41:18,651 - INFO - [Train] step: 33299, loss_mpn: 0.005471, loss_rec: 0.018395, loss_semantic: 0.298615, loss_idmrf: 5.933259, loss_adv_gen: -33.227077
|
511 |
+
2023-02-13 04:41:47,905 - INFO - [Train] step: 33399, loss_adv_disc: -1.262914
|
512 |
+
2023-02-13 04:41:48,113 - INFO - [Train] step: 33399, loss_mpn: 0.004389, loss_rec: 0.023651, loss_semantic: 0.292535, loss_idmrf: 1.633751, loss_adv_gen: -47.819061
|
513 |
+
2023-02-13 04:42:17,374 - INFO - [Train] step: 33499, loss_adv_disc: -4.915746
|
514 |
+
2023-02-13 04:42:17,583 - INFO - [Train] step: 33499, loss_mpn: 0.006723, loss_rec: 0.039050, loss_semantic: 0.405457, loss_idmrf: 2.034822, loss_adv_gen: -73.422531
|
515 |
+
2023-02-13 04:42:47,032 - INFO - [Train] step: 33599, loss_adv_disc: -3.144320
|
516 |
+
2023-02-13 04:42:47,241 - INFO - [Train] step: 33599, loss_mpn: 0.009663, loss_rec: 0.031112, loss_semantic: 0.353883, loss_idmrf: 2.664726, loss_adv_gen: -30.228935
|
517 |
+
2023-02-13 04:43:16,501 - INFO - [Train] step: 33699, loss_adv_disc: -4.212373
|
518 |
+
2023-02-13 04:43:16,710 - INFO - [Train] step: 33699, loss_mpn: 0.015043, loss_rec: 0.039189, loss_semantic: 0.461344, loss_idmrf: 2.188730, loss_adv_gen: -57.981586
|
519 |
+
2023-02-13 04:43:45,968 - INFO - [Train] step: 33799, loss_adv_disc: -2.260951
|
520 |
+
2023-02-13 04:43:46,177 - INFO - [Train] step: 33799, loss_mpn: 0.012627, loss_rec: 0.029967, loss_semantic: 0.422946, loss_idmrf: 1.669225, loss_adv_gen: -42.562790
|
521 |
+
2023-02-13 04:44:15,423 - INFO - [Train] step: 33899, loss_adv_disc: -2.908595
|
522 |
+
2023-02-13 04:44:15,631 - INFO - [Train] step: 33899, loss_mpn: 0.009963, loss_rec: 0.026809, loss_semantic: 0.331871, loss_idmrf: 4.898873, loss_adv_gen: -28.988209
|
523 |
+
2023-02-13 04:44:45,067 - INFO - [Train] step: 33999, loss_adv_disc: -3.627460
|
524 |
+
2023-02-13 04:44:45,276 - INFO - [Train] step: 33999, loss_mpn: 0.011132, loss_rec: 0.029245, loss_semantic: 0.385511, loss_idmrf: 2.419443, loss_adv_gen: -38.314835
|
525 |
+
2023-02-13 04:44:50,427 - INFO - [Eval] step: 33999, bce: 0.493806, psnr: 24.006083, ssim: 0.933177
|
526 |
+
2023-02-13 04:45:20,720 - INFO - [Train] step: 34099, loss_adv_disc: -2.766464
|
527 |
+
2023-02-13 04:45:20,929 - INFO - [Train] step: 34099, loss_mpn: 0.012229, loss_rec: 0.027476, loss_semantic: 0.361568, loss_idmrf: 1.908972, loss_adv_gen: -34.426533
|
528 |
+
2023-02-13 04:45:50,178 - INFO - [Train] step: 34199, loss_adv_disc: -1.794028
|
529 |
+
2023-02-13 04:45:50,387 - INFO - [Train] step: 34199, loss_mpn: 0.011440, loss_rec: 0.023251, loss_semantic: 0.366893, loss_idmrf: 2.413965, loss_adv_gen: -36.843105
|
530 |
+
2023-02-13 04:46:19,819 - INFO - [Train] step: 34299, loss_adv_disc: -3.664385
|
531 |
+
2023-02-13 04:46:20,028 - INFO - [Train] step: 34299, loss_mpn: 0.005965, loss_rec: 0.025664, loss_semantic: 0.328096, loss_idmrf: 1.862936, loss_adv_gen: -62.822632
|
532 |
+
2023-02-13 04:46:49,273 - INFO - [Train] step: 34399, loss_adv_disc: -4.225115
|
533 |
+
2023-02-13 04:46:49,481 - INFO - [Train] step: 34399, loss_mpn: 0.005665, loss_rec: 0.026986, loss_semantic: 0.313127, loss_idmrf: 1.985018, loss_adv_gen: -42.136913
|
534 |
+
2023-02-13 04:47:18,742 - INFO - [Train] step: 34499, loss_adv_disc: -2.175828
|
535 |
+
2023-02-13 04:47:18,951 - INFO - [Train] step: 34499, loss_mpn: 0.009428, loss_rec: 0.028373, loss_semantic: 0.334881, loss_idmrf: 2.210392, loss_adv_gen: -56.427872
|
536 |
+
2023-02-13 04:47:48,390 - INFO - [Train] step: 34599, loss_adv_disc: -2.655634
|
537 |
+
2023-02-13 04:47:48,599 - INFO - [Train] step: 34599, loss_mpn: 0.007422, loss_rec: 0.026994, loss_semantic: 0.377276, loss_idmrf: 2.321023, loss_adv_gen: -45.277206
|
538 |
+
2023-02-13 04:48:17,863 - INFO - [Train] step: 34699, loss_adv_disc: -2.667136
|
539 |
+
2023-02-13 04:48:18,071 - INFO - [Train] step: 34699, loss_mpn: 0.012077, loss_rec: 0.026902, loss_semantic: 0.367508, loss_idmrf: 1.815201, loss_adv_gen: -30.165154
|
540 |
+
2023-02-13 04:48:47,339 - INFO - [Train] step: 34799, loss_adv_disc: -0.941228
|
541 |
+
2023-02-13 04:48:47,547 - INFO - [Train] step: 34799, loss_mpn: 0.008583, loss_rec: 0.029322, loss_semantic: 0.366778, loss_idmrf: 1.744337, loss_adv_gen: -32.191875
|
542 |
+
2023-02-13 04:49:16,808 - INFO - [Train] step: 34899, loss_adv_disc: -2.496438
|
543 |
+
2023-02-13 04:49:17,016 - INFO - [Train] step: 34899, loss_mpn: 0.008992, loss_rec: 0.033850, loss_semantic: 0.455653, loss_idmrf: 2.395776, loss_adv_gen: -40.789753
|
544 |
+
2023-02-13 04:49:46,483 - INFO - [Train] step: 34999, loss_adv_disc: -3.898407
|
545 |
+
2023-02-13 04:49:46,691 - INFO - [Train] step: 34999, loss_mpn: 0.007672, loss_rec: 0.029516, loss_semantic: 0.391788, loss_idmrf: 3.343580, loss_adv_gen: -49.658260
|
546 |
+
2023-02-13 04:49:51,834 - INFO - [Eval] step: 34999, bce: 0.492687, psnr: 23.920872, ssim: 0.932854
|
547 |
+
2023-02-13 04:50:22,250 - INFO - [Train] step: 35099, loss_adv_disc: -3.315386
|
548 |
+
2023-02-13 04:50:22,458 - INFO - [Train] step: 35099, loss_mpn: 0.007049, loss_rec: 0.026398, loss_semantic: 0.327736, loss_idmrf: 2.569352, loss_adv_gen: -44.722576
|
549 |
+
2023-02-13 04:50:51,704 - INFO - [Train] step: 35199, loss_adv_disc: -1.372233
|
550 |
+
2023-02-13 04:50:51,912 - INFO - [Train] step: 35199, loss_mpn: 0.005760, loss_rec: 0.020659, loss_semantic: 0.290627, loss_idmrf: 1.633903, loss_adv_gen: -40.939323
|
551 |
+
2023-02-13 04:51:21,346 - INFO - [Train] step: 35299, loss_adv_disc: -2.340883
|
552 |
+
2023-02-13 04:51:21,555 - INFO - [Train] step: 35299, loss_mpn: 0.005978, loss_rec: 0.021761, loss_semantic: 0.300960, loss_idmrf: 2.090793, loss_adv_gen: -32.323677
|
553 |
+
2023-02-13 04:51:50,793 - INFO - [Train] step: 35399, loss_adv_disc: -2.734210
|
554 |
+
2023-02-13 04:51:51,001 - INFO - [Train] step: 35399, loss_mpn: 0.009728, loss_rec: 0.029348, loss_semantic: 0.389291, loss_idmrf: 3.652727, loss_adv_gen: -43.234627
|
555 |
+
2023-02-13 04:52:20,240 - INFO - [Train] step: 35499, loss_adv_disc: -1.138059
|
556 |
+
2023-02-13 04:52:20,449 - INFO - [Train] step: 35499, loss_mpn: 0.005424, loss_rec: 0.022387, loss_semantic: 0.280479, loss_idmrf: 4.068128, loss_adv_gen: -42.517578
|
557 |
+
2023-02-13 04:52:49,873 - INFO - [Train] step: 35599, loss_adv_disc: -2.349144
|
558 |
+
2023-02-13 04:52:50,082 - INFO - [Train] step: 35599, loss_mpn: 0.004034, loss_rec: 0.023441, loss_semantic: 0.325334, loss_idmrf: 1.199162, loss_adv_gen: -37.645535
|
559 |
+
2023-02-13 04:53:19,322 - INFO - [Train] step: 35699, loss_adv_disc: -3.336832
|
560 |
+
2023-02-13 04:53:19,531 - INFO - [Train] step: 35699, loss_mpn: 0.009035, loss_rec: 0.032002, loss_semantic: 0.351620, loss_idmrf: 2.143802, loss_adv_gen: -48.266075
|
561 |
+
2023-02-13 04:53:48,771 - INFO - [Train] step: 35799, loss_adv_disc: -5.475993
|
562 |
+
2023-02-13 04:53:48,981 - INFO - [Train] step: 35799, loss_mpn: 0.015121, loss_rec: 0.043468, loss_semantic: 0.448060, loss_idmrf: 4.249763, loss_adv_gen: -59.551453
|
563 |
+
2023-02-13 04:54:18,420 - INFO - [Train] step: 35899, loss_adv_disc: -1.503309
|
564 |
+
2023-02-13 04:54:18,629 - INFO - [Train] step: 35899, loss_mpn: 0.005271, loss_rec: 0.021715, loss_semantic: 0.305199, loss_idmrf: 3.841014, loss_adv_gen: -56.835781
|
565 |
+
2023-02-13 04:54:47,871 - INFO - [Train] step: 35999, loss_adv_disc: -1.925685
|
566 |
+
2023-02-13 04:54:48,079 - INFO - [Train] step: 35999, loss_mpn: 0.007125, loss_rec: 0.024240, loss_semantic: 0.358235, loss_idmrf: 2.990580, loss_adv_gen: -49.677704
|
567 |
+
2023-02-13 04:54:53,246 - INFO - [Eval] step: 35999, bce: 0.451248, psnr: 24.089697, ssim: 0.933519
|
568 |
+
2023-02-13 04:55:23,765 - INFO - [Train] step: 36099, loss_adv_disc: -2.582094
|
569 |
+
2023-02-13 04:55:23,974 - INFO - [Train] step: 36099, loss_mpn: 0.010654, loss_rec: 0.032407, loss_semantic: 0.359470, loss_idmrf: 4.383611, loss_adv_gen: -51.658825
|
570 |
+
2023-02-13 04:55:53,217 - INFO - [Train] step: 36199, loss_adv_disc: -1.860945
|
571 |
+
2023-02-13 04:55:53,426 - INFO - [Train] step: 36199, loss_mpn: 0.005998, loss_rec: 0.023951, loss_semantic: 0.332017, loss_idmrf: 2.729879, loss_adv_gen: -52.860546
|
572 |
+
2023-02-13 04:56:22,869 - INFO - [Train] step: 36299, loss_adv_disc: -3.733282
|
573 |
+
2023-02-13 04:56:23,077 - INFO - [Train] step: 36299, loss_mpn: 0.005556, loss_rec: 0.029372, loss_semantic: 0.412161, loss_idmrf: 4.497084, loss_adv_gen: -80.765701
|
574 |
+
2023-02-13 04:56:52,330 - INFO - [Train] step: 36399, loss_adv_disc: -1.141786
|
575 |
+
2023-02-13 04:56:52,539 - INFO - [Train] step: 36399, loss_mpn: 0.005013, loss_rec: 0.021536, loss_semantic: 0.305248, loss_idmrf: 1.969661, loss_adv_gen: -45.550240
|
576 |
+
2023-02-13 04:57:21,785 - INFO - [Train] step: 36499, loss_adv_disc: -1.790366
|
577 |
+
2023-02-13 04:57:21,994 - INFO - [Train] step: 36499, loss_mpn: 0.005966, loss_rec: 0.025548, loss_semantic: 0.332869, loss_idmrf: 2.200314, loss_adv_gen: -37.936630
|
578 |
+
2023-02-13 04:57:51,440 - INFO - [Train] step: 36599, loss_adv_disc: -3.809116
|
579 |
+
2023-02-13 04:57:51,649 - INFO - [Train] step: 36599, loss_mpn: 0.005486, loss_rec: 0.024083, loss_semantic: 0.316460, loss_idmrf: 2.768273, loss_adv_gen: -49.244308
|
580 |
+
2023-02-13 04:58:20,920 - INFO - [Train] step: 36699, loss_adv_disc: -2.374944
|
581 |
+
2023-02-13 04:58:21,129 - INFO - [Train] step: 36699, loss_mpn: 0.005127, loss_rec: 0.021198, loss_semantic: 0.294804, loss_idmrf: 1.192759, loss_adv_gen: -24.119633
|
582 |
+
2023-02-13 04:58:50,378 - INFO - [Train] step: 36799, loss_adv_disc: -3.080187
|
583 |
+
2023-02-13 04:58:50,587 - INFO - [Train] step: 36799, loss_mpn: 0.009575, loss_rec: 0.025953, loss_semantic: 0.377960, loss_idmrf: 1.418940, loss_adv_gen: -53.199898
|
584 |
+
2023-02-13 04:59:20,041 - INFO - [Train] step: 36899, loss_adv_disc: -1.956524
|
585 |
+
2023-02-13 04:59:20,250 - INFO - [Train] step: 36899, loss_mpn: 0.004222, loss_rec: 0.018766, loss_semantic: 0.279596, loss_idmrf: 3.066632, loss_adv_gen: -43.155220
|
586 |
+
2023-02-13 04:59:49,499 - INFO - [Train] step: 36999, loss_adv_disc: -2.218104
|
587 |
+
2023-02-13 04:59:49,709 - INFO - [Train] step: 36999, loss_mpn: 0.006558, loss_rec: 0.021726, loss_semantic: 0.322024, loss_idmrf: 3.109234, loss_adv_gen: -44.358009
|
588 |
+
2023-02-13 04:59:54,854 - INFO - [Eval] step: 36999, bce: 0.511608, psnr: 24.066105, ssim: 0.933602
|
589 |
+
2023-02-13 05:00:25,154 - INFO - [Train] step: 37099, loss_adv_disc: -3.587424
|
590 |
+
2023-02-13 05:00:25,363 - INFO - [Train] step: 37099, loss_mpn: 0.008115, loss_rec: 0.029436, loss_semantic: 0.385246, loss_idmrf: 1.943526, loss_adv_gen: -34.469177
|
591 |
+
2023-02-13 05:00:54,625 - INFO - [Train] step: 37199, loss_adv_disc: -2.477960
|
592 |
+
2023-02-13 05:00:54,833 - INFO - [Train] step: 37199, loss_mpn: 0.007001, loss_rec: 0.030382, loss_semantic: 0.373161, loss_idmrf: 7.979999, loss_adv_gen: -34.278618
|
593 |
+
2023-02-13 05:01:24,262 - INFO - [Train] step: 37299, loss_adv_disc: -1.760101
|
594 |
+
2023-02-13 05:01:24,471 - INFO - [Train] step: 37299, loss_mpn: 0.006662, loss_rec: 0.020768, loss_semantic: 0.294572, loss_idmrf: 2.288859, loss_adv_gen: -34.755966
|
595 |
+
2023-02-13 05:01:53,723 - INFO - [Train] step: 37399, loss_adv_disc: -4.853773
|
596 |
+
2023-02-13 05:01:53,932 - INFO - [Train] step: 37399, loss_mpn: 0.010881, loss_rec: 0.037399, loss_semantic: 0.408028, loss_idmrf: 5.442630, loss_adv_gen: -53.623989
|
597 |
+
2023-02-13 05:02:23,179 - INFO - [Train] step: 37499, loss_adv_disc: -2.969615
|
598 |
+
2023-02-13 05:02:23,387 - INFO - [Train] step: 37499, loss_mpn: 0.006965, loss_rec: 0.025584, loss_semantic: 0.342903, loss_idmrf: 2.896766, loss_adv_gen: -56.246437
|
599 |
+
2023-02-13 05:02:52,832 - INFO - [Train] step: 37599, loss_adv_disc: -1.794725
|
600 |
+
2023-02-13 05:02:53,041 - INFO - [Train] step: 37599, loss_mpn: 0.008978, loss_rec: 0.024542, loss_semantic: 0.356401, loss_idmrf: 1.502892, loss_adv_gen: -49.551048
|
601 |
+
2023-02-13 05:03:22,289 - INFO - [Train] step: 37699, loss_adv_disc: -2.146141
|
602 |
+
2023-02-13 05:03:22,497 - INFO - [Train] step: 37699, loss_mpn: 0.007144, loss_rec: 0.028536, loss_semantic: 0.323880, loss_idmrf: 2.761802, loss_adv_gen: -54.818451
|
603 |
+
2023-02-13 05:03:51,758 - INFO - [Train] step: 37799, loss_adv_disc: -3.474980
|
604 |
+
2023-02-13 05:03:51,966 - INFO - [Train] step: 37799, loss_mpn: 0.006934, loss_rec: 0.036860, loss_semantic: 0.379891, loss_idmrf: 2.068331, loss_adv_gen: -32.999737
|
605 |
+
2023-02-13 05:04:21,416 - INFO - [Train] step: 37899, loss_adv_disc: -3.225597
|
606 |
+
2023-02-13 05:04:21,625 - INFO - [Train] step: 37899, loss_mpn: 0.010186, loss_rec: 0.033910, loss_semantic: 0.345942, loss_idmrf: 3.703231, loss_adv_gen: -37.339417
|
607 |
+
2023-02-13 05:04:50,873 - INFO - [Train] step: 37999, loss_adv_disc: -1.842927
|
608 |
+
2023-02-13 05:04:51,082 - INFO - [Train] step: 37999, loss_mpn: 0.004289, loss_rec: 0.022910, loss_semantic: 0.307499, loss_idmrf: 0.953175, loss_adv_gen: -35.079369
|
609 |
+
2023-02-13 05:04:56,221 - INFO - [Eval] step: 37999, bce: 0.550441, psnr: 23.929970, ssim: 0.932575
|
610 |
+
2023-02-13 05:05:26,524 - INFO - [Train] step: 38099, loss_adv_disc: -3.334216
|
611 |
+
2023-02-13 05:05:26,732 - INFO - [Train] step: 38099, loss_mpn: 0.010205, loss_rec: 0.032689, loss_semantic: 0.379243, loss_idmrf: 2.807850, loss_adv_gen: -46.448601
|
612 |
+
2023-02-13 05:05:55,991 - INFO - [Train] step: 38199, loss_adv_disc: -4.326716
|
613 |
+
2023-02-13 05:05:56,199 - INFO - [Train] step: 38199, loss_mpn: 0.009495, loss_rec: 0.031506, loss_semantic: 0.367149, loss_idmrf: 7.712746, loss_adv_gen: -44.628571
|
614 |
+
2023-02-13 05:06:25,633 - INFO - [Train] step: 38299, loss_adv_disc: -4.521143
|
615 |
+
2023-02-13 05:06:25,841 - INFO - [Train] step: 38299, loss_mpn: 0.010475, loss_rec: 0.033682, loss_semantic: 0.389523, loss_idmrf: 2.941085, loss_adv_gen: -46.803600
|
616 |
+
2023-02-13 05:06:55,090 - INFO - [Train] step: 38399, loss_adv_disc: -3.862583
|
617 |
+
2023-02-13 05:06:55,299 - INFO - [Train] step: 38399, loss_mpn: 0.008165, loss_rec: 0.033486, loss_semantic: 0.434651, loss_idmrf: 1.601594, loss_adv_gen: -46.221794
|
618 |
+
2023-02-13 05:07:24,538 - INFO - [Train] step: 38499, loss_adv_disc: -2.403943
|
619 |
+
2023-02-13 05:07:24,747 - INFO - [Train] step: 38499, loss_mpn: 0.007161, loss_rec: 0.027608, loss_semantic: 0.396924, loss_idmrf: 1.407646, loss_adv_gen: -39.244766
|
620 |
+
2023-02-13 05:07:54,217 - INFO - [Train] step: 38599, loss_adv_disc: -3.030591
|
621 |
+
2023-02-13 05:07:54,425 - INFO - [Train] step: 38599, loss_mpn: 0.005753, loss_rec: 0.027395, loss_semantic: 0.364165, loss_idmrf: 1.339588, loss_adv_gen: -38.628998
|
622 |
+
2023-02-13 05:08:23,670 - INFO - [Train] step: 38699, loss_adv_disc: -3.780364
|
623 |
+
2023-02-13 05:08:23,879 - INFO - [Train] step: 38699, loss_mpn: 0.009336, loss_rec: 0.031163, loss_semantic: 0.372193, loss_idmrf: 3.190570, loss_adv_gen: -35.400978
|
624 |
+
2023-02-13 05:08:53,137 - INFO - [Train] step: 38799, loss_adv_disc: -3.990635
|
625 |
+
2023-02-13 05:08:53,346 - INFO - [Train] step: 38799, loss_mpn: 0.011644, loss_rec: 0.034559, loss_semantic: 0.392621, loss_idmrf: 2.479134, loss_adv_gen: -49.558567
|
626 |
+
2023-02-13 05:09:22,789 - INFO - [Train] step: 38899, loss_adv_disc: -1.296090
|
627 |
+
2023-02-13 05:09:22,998 - INFO - [Train] step: 38899, loss_mpn: 0.002924, loss_rec: 0.020468, loss_semantic: 0.320249, loss_idmrf: 2.698936, loss_adv_gen: -34.255341
|
628 |
+
2023-02-13 05:09:52,241 - INFO - [Train] step: 38999, loss_adv_disc: -1.709500
|
629 |
+
2023-02-13 05:09:52,450 - INFO - [Train] step: 38999, loss_mpn: 0.007364, loss_rec: 0.024880, loss_semantic: 0.339524, loss_idmrf: 2.538348, loss_adv_gen: -32.774738
|
630 |
+
2023-02-13 05:09:57,599 - INFO - [Eval] step: 38999, bce: 0.543258, psnr: 23.938129, ssim: 0.933121
|
631 |
+
2023-02-13 05:10:27,906 - INFO - [Train] step: 39099, loss_adv_disc: -4.103148
|
632 |
+
2023-02-13 05:10:28,115 - INFO - [Train] step: 39099, loss_mpn: 0.011113, loss_rec: 0.030247, loss_semantic: 0.417251, loss_idmrf: 1.839109, loss_adv_gen: -40.379929
|
633 |
+
2023-02-13 05:10:57,376 - INFO - [Train] step: 39199, loss_adv_disc: -2.376977
|
634 |
+
2023-02-13 05:10:57,585 - INFO - [Train] step: 39199, loss_mpn: 0.006472, loss_rec: 0.027119, loss_semantic: 0.307704, loss_idmrf: 2.029118, loss_adv_gen: -41.603371
|
635 |
+
2023-02-13 05:11:27,026 - INFO - [Train] step: 39299, loss_adv_disc: -2.299066
|
636 |
+
2023-02-13 05:11:27,236 - INFO - [Train] step: 39299, loss_mpn: 0.008503, loss_rec: 0.022436, loss_semantic: 0.329789, loss_idmrf: 1.689665, loss_adv_gen: -55.237625
|
637 |
+
2023-02-13 05:11:56,492 - INFO - [Train] step: 39399, loss_adv_disc: -1.676680
|
638 |
+
2023-02-13 05:11:56,701 - INFO - [Train] step: 39399, loss_mpn: 0.007318, loss_rec: 0.025969, loss_semantic: 0.338832, loss_idmrf: 2.227033, loss_adv_gen: -35.137005
|
639 |
+
2023-02-13 05:12:25,968 - INFO - [Train] step: 39499, loss_adv_disc: -3.850043
|
640 |
+
2023-02-13 05:12:26,177 - INFO - [Train] step: 39499, loss_mpn: 0.007011, loss_rec: 0.025364, loss_semantic: 0.338889, loss_idmrf: 2.077760, loss_adv_gen: -42.100559
|
641 |
+
2023-02-13 05:12:55,617 - INFO - [Train] step: 39599, loss_adv_disc: -2.115081
|
642 |
+
2023-02-13 05:12:55,827 - INFO - [Train] step: 39599, loss_mpn: 0.005659, loss_rec: 0.023901, loss_semantic: 0.306843, loss_idmrf: 1.831819, loss_adv_gen: -40.354424
|
643 |
+
2023-02-13 05:13:25,069 - INFO - [Train] step: 39699, loss_adv_disc: -1.994590
|
644 |
+
2023-02-13 05:13:25,278 - INFO - [Train] step: 39699, loss_mpn: 0.012201, loss_rec: 0.024626, loss_semantic: 0.370962, loss_idmrf: 1.804987, loss_adv_gen: -38.031910
|
645 |
+
2023-02-13 05:13:54,538 - INFO - [Train] step: 39799, loss_adv_disc: -1.271324
|
646 |
+
2023-02-13 05:13:54,747 - INFO - [Train] step: 39799, loss_mpn: 0.007453, loss_rec: 0.022167, loss_semantic: 0.320188, loss_idmrf: 2.214860, loss_adv_gen: -37.319176
|
647 |
+
2023-02-13 05:14:24,190 - INFO - [Train] step: 39899, loss_adv_disc: -1.676649
|
648 |
+
2023-02-13 05:14:24,399 - INFO - [Train] step: 39899, loss_mpn: 0.004491, loss_rec: 0.022268, loss_semantic: 0.293991, loss_idmrf: 2.233423, loss_adv_gen: -52.407726
|
649 |
+
2023-02-13 05:14:53,659 - INFO - [Train] step: 39999, loss_adv_disc: -2.623639
|
650 |
+
2023-02-13 05:14:53,867 - INFO - [Train] step: 39999, loss_mpn: 0.005937, loss_rec: 0.026519, loss_semantic: 0.336726, loss_idmrf: 1.825681, loss_adv_gen: -35.927650
|
651 |
+
2023-02-13 05:14:59,026 - INFO - [Eval] step: 39999, bce: 0.525641, psnr: 23.983196, ssim: 0.932836
|
652 |
+
2023-02-13 05:15:29,452 - INFO - [Train] step: 40099, loss_adv_disc: -2.439545
|
653 |
+
2023-02-13 05:15:29,661 - INFO - [Train] step: 40099, loss_mpn: 0.003933, loss_rec: 0.025905, loss_semantic: 0.298822, loss_idmrf: 1.251885, loss_adv_gen: -25.356033
|
654 |
+
2023-02-13 05:15:58,913 - INFO - [Train] step: 40199, loss_adv_disc: -2.549628
|
655 |
+
2023-02-13 05:15:59,122 - INFO - [Train] step: 40199, loss_mpn: 0.003956, loss_rec: 0.025239, loss_semantic: 0.325746, loss_idmrf: 2.043177, loss_adv_gen: -53.137932
|
656 |
+
2023-02-13 05:16:28,584 - INFO - [Train] step: 40299, loss_adv_disc: -2.381817
|
657 |
+
2023-02-13 05:16:28,792 - INFO - [Train] step: 40299, loss_mpn: 0.005729, loss_rec: 0.019757, loss_semantic: 0.314869, loss_idmrf: 1.337579, loss_adv_gen: -36.087315
|
658 |
+
2023-02-13 05:16:58,059 - INFO - [Train] step: 40399, loss_adv_disc: -1.973938
|
659 |
+
2023-02-13 05:16:58,267 - INFO - [Train] step: 40399, loss_mpn: 0.005751, loss_rec: 0.021610, loss_semantic: 0.290376, loss_idmrf: 1.547544, loss_adv_gen: -47.893143
|
660 |
+
2023-02-13 05:17:27,521 - INFO - [Train] step: 40499, loss_adv_disc: -1.127224
|
661 |
+
2023-02-13 05:17:27,730 - INFO - [Train] step: 40499, loss_mpn: 0.008527, loss_rec: 0.023455, loss_semantic: 0.340019, loss_idmrf: 4.099547, loss_adv_gen: -28.373253
|
662 |
+
2023-02-13 05:17:57,179 - INFO - [Train] step: 40599, loss_adv_disc: -1.630693
|
663 |
+
2023-02-13 05:17:57,388 - INFO - [Train] step: 40599, loss_mpn: 0.004528, loss_rec: 0.022425, loss_semantic: 0.311950, loss_idmrf: 2.045025, loss_adv_gen: -42.663849
|
664 |
+
2023-02-13 05:18:26,648 - INFO - [Train] step: 40699, loss_adv_disc: -4.531513
|
665 |
+
2023-02-13 05:18:26,856 - INFO - [Train] step: 40699, loss_mpn: 0.006523, loss_rec: 0.032161, loss_semantic: 0.387854, loss_idmrf: 2.429147, loss_adv_gen: -54.620087
|
666 |
+
2023-02-13 05:18:56,106 - INFO - [Train] step: 40799, loss_adv_disc: -1.792886
|
667 |
+
2023-02-13 05:18:56,316 - INFO - [Train] step: 40799, loss_mpn: 0.004772, loss_rec: 0.020073, loss_semantic: 0.275349, loss_idmrf: 2.377422, loss_adv_gen: -30.988853
|
668 |
+
2023-02-13 05:19:25,749 - INFO - [Train] step: 40899, loss_adv_disc: -2.539077
|
669 |
+
2023-02-13 05:19:25,958 - INFO - [Train] step: 40899, loss_mpn: 0.006153, loss_rec: 0.024839, loss_semantic: 0.301111, loss_idmrf: 1.659654, loss_adv_gen: -32.603027
|
670 |
+
2023-02-13 05:19:55,216 - INFO - [Train] step: 40999, loss_adv_disc: -2.101744
|
671 |
+
2023-02-13 05:19:55,425 - INFO - [Train] step: 40999, loss_mpn: 0.006380, loss_rec: 0.018790, loss_semantic: 0.273095, loss_idmrf: 0.847124, loss_adv_gen: -29.596489
|
672 |
+
2023-02-13 05:20:00,565 - INFO - [Eval] step: 40999, bce: 0.468785, psnr: 23.891504, ssim: 0.933626
|
673 |
+
2023-02-13 05:20:30,853 - INFO - [Train] step: 41099, loss_adv_disc: -1.681476
|
674 |
+
2023-02-13 05:20:31,061 - INFO - [Train] step: 41099, loss_mpn: 0.007705, loss_rec: 0.024250, loss_semantic: 0.294493, loss_idmrf: 2.501994, loss_adv_gen: -23.292068
|
675 |
+
2023-02-13 05:21:00,315 - INFO - [Train] step: 41199, loss_adv_disc: -2.958653
|
676 |
+
2023-02-13 05:21:00,523 - INFO - [Train] step: 41199, loss_mpn: 0.006746, loss_rec: 0.025472, loss_semantic: 0.317826, loss_idmrf: 1.403564, loss_adv_gen: -49.742279
|
677 |
+
2023-02-13 05:21:29,965 - INFO - [Train] step: 41299, loss_adv_disc: -4.011607
|
678 |
+
2023-02-13 05:21:30,174 - INFO - [Train] step: 41299, loss_mpn: 0.004297, loss_rec: 0.028533, loss_semantic: 0.317103, loss_idmrf: 2.485311, loss_adv_gen: -38.421883
|
679 |
+
2023-02-13 05:21:59,436 - INFO - [Train] step: 41399, loss_adv_disc: -2.888487
|
680 |
+
2023-02-13 05:21:59,645 - INFO - [Train] step: 41399, loss_mpn: 0.003805, loss_rec: 0.019754, loss_semantic: 0.258992, loss_idmrf: 1.701972, loss_adv_gen: -30.434830
|
downstream-shadow-removal/samples/step000999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step001999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step002999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step003999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step004999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step005999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step006999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step007999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step008999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/samples/step009999.png
ADDED
![]() |
Git LFS Details
|
downstream-shadow-removal/tensorboard/events.out.tfevents.1676208201.jason-system.49917.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:491cda9bfb4a46b18a438a1bbacd2e388cb63d31f930b3d473567adb85b850e9
|
3 |
+
size 3340690
|
downstream-shadow-removal/tensorboard/events.out.tfevents.1676256254.jason-system.54160.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:facf7f8557133fa989d8d1fa01dda6946b885368cbb6ca334dde150e7db61914
|
3 |
+
size 10672223
|
downstream-watermark-removal/ckpt/step099999.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2af74105dabcebf32f3b700898727a4070b3d5afbe20a5e592d74c4495f0e132
|
3 |
+
size 160487111
|
downstream-watermark-removal/config-2023-02-12-14-28-38.yaml
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
DATA:
|
2 |
+
DATAROOT: /data/LOGO/
|
3 |
+
IMG_SIZE: 256
|
4 |
+
NAME: LOGO_30K
|
5 |
+
DATALOADER:
|
6 |
+
BATCH_SIZE: 4
|
7 |
+
MICRO_BATCH: 0
|
8 |
+
NUM_WORKERS: 4
|
9 |
+
PIN_MEMORY: true
|
10 |
+
PREFETCH_FACTOR: 2
|
11 |
+
EVALUATE:
|
12 |
+
N_EVAL: 10000
|
13 |
+
MASK:
|
14 |
+
BRUSH_LENGTH_RATIO:
|
15 |
+
- 0.1
|
16 |
+
- 0.25
|
17 |
+
BRUSH_NUM: &id001
|
18 |
+
- 0
|
19 |
+
- 0
|
20 |
+
BRUSH_TURNS:
|
21 |
+
- 4
|
22 |
+
- 18
|
23 |
+
BRUSH_WIDTH_RATIO:
|
24 |
+
- 0.02
|
25 |
+
- 0.1
|
26 |
+
MASK_DIR: null
|
27 |
+
NOISE_DATASETS: []
|
28 |
+
RECT_LENGTH_RATIO:
|
29 |
+
- 0.2
|
30 |
+
- 0.8
|
31 |
+
RECT_NUM: *id001
|
32 |
+
SMOOTH_ITERS: 4
|
33 |
+
SMOOTH_KERNEL_SIZE: 15
|
34 |
+
SMOOTH_SIGMA: 4
|
35 |
+
MODEL:
|
36 |
+
MPN:
|
37 |
+
BASE_N_CHANNELS: 64
|
38 |
+
NECK_N_CHANNELS: 128
|
39 |
+
RIN:
|
40 |
+
BASE_N_CHANNELS: 32
|
41 |
+
NECK_N_CHANNELS: 128
|
42 |
+
WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
|
43 |
+
SAMPLE:
|
44 |
+
N_SAMPLES: 100
|
45 |
+
RANDOM: false
|
46 |
+
SAVE_DIR: ./samples/
|
47 |
+
SEED: 1234
|
48 |
+
TRAIN:
|
49 |
+
COEF_ADV: 0.001
|
50 |
+
COEF_IDMRF: 0.001
|
51 |
+
COEF_MPN: 2.0
|
52 |
+
COEF_REC: 1.4
|
53 |
+
COEF_SEMANTIC: 0.0001
|
54 |
+
EVAL_FREQ: 1000
|
55 |
+
OPTIM_DISC:
|
56 |
+
BETAS: &id002
|
57 |
+
- 0.5
|
58 |
+
- 0.9
|
59 |
+
LR: 1.0e-05
|
60 |
+
OPTIM_MPN:
|
61 |
+
BETAS: *id002
|
62 |
+
LR: 1.0e-05
|
63 |
+
OPTIM_RIN:
|
64 |
+
BETAS: *id002
|
65 |
+
LR: 1.0e-05
|
66 |
+
PRINT_FREQ: 100
|
67 |
+
RESUME: null
|
68 |
+
SAMPLE_FREQ: 1000
|
69 |
+
SAVE_FREQ: 5000
|
70 |
+
TRAIN_STEPS: 10000
|
downstream-watermark-removal/config-2023-02-12-15-41-28.yaml
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
DATA:
|
2 |
+
DATAROOT: /data/LOGO/
|
3 |
+
IMG_SIZE: 256
|
4 |
+
NAME: LOGO_30K
|
5 |
+
DATALOADER:
|
6 |
+
BATCH_SIZE: 4
|
7 |
+
MICRO_BATCH: 0
|
8 |
+
NUM_WORKERS: 4
|
9 |
+
PIN_MEMORY: true
|
10 |
+
PREFETCH_FACTOR: 2
|
11 |
+
EVALUATE:
|
12 |
+
N_EVAL: 10000
|
13 |
+
MASK:
|
14 |
+
BRUSH_LENGTH_RATIO:
|
15 |
+
- 0.1
|
16 |
+
- 0.25
|
17 |
+
BRUSH_NUM: &id001
|
18 |
+
- 0
|
19 |
+
- 0
|
20 |
+
BRUSH_TURNS:
|
21 |
+
- 4
|
22 |
+
- 18
|
23 |
+
BRUSH_WIDTH_RATIO:
|
24 |
+
- 0.02
|
25 |
+
- 0.1
|
26 |
+
MASK_DIR: null
|
27 |
+
NOISE_DATASETS: []
|
28 |
+
RECT_LENGTH_RATIO:
|
29 |
+
- 0.2
|
30 |
+
- 0.8
|
31 |
+
RECT_NUM: *id001
|
32 |
+
SMOOTH_ITERS: 4
|
33 |
+
SMOOTH_KERNEL_SIZE: 15
|
34 |
+
SMOOTH_SIGMA: 4
|
35 |
+
MODEL:
|
36 |
+
MPN:
|
37 |
+
BASE_N_CHANNELS: 64
|
38 |
+
NECK_N_CHANNELS: 128
|
39 |
+
RIN:
|
40 |
+
BASE_N_CHANNELS: 32
|
41 |
+
NECK_N_CHANNELS: 128
|
42 |
+
WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
|
43 |
+
SAMPLE:
|
44 |
+
N_SAMPLES: 100
|
45 |
+
RANDOM: false
|
46 |
+
SAVE_DIR: ./samples/
|
47 |
+
SEED: 1234
|
48 |
+
TRAIN:
|
49 |
+
COEF_ADV: 0.001
|
50 |
+
COEF_IDMRF: 0.001
|
51 |
+
COEF_MPN: 2.0
|
52 |
+
COEF_REC: 1.4
|
53 |
+
COEF_SEMANTIC: 0.0001
|
54 |
+
EVAL_FREQ: 1000
|
55 |
+
OPTIM_DISC:
|
56 |
+
BETAS: &id002
|
57 |
+
- 0.5
|
58 |
+
- 0.9
|
59 |
+
LR: 1.0e-05
|
60 |
+
OPTIM_MPN:
|
61 |
+
BETAS: *id002
|
62 |
+
LR: 1.0e-05
|
63 |
+
OPTIM_RIN:
|
64 |
+
BETAS: *id002
|
65 |
+
LR: 1.0e-05
|
66 |
+
PRINT_FREQ: 100
|
67 |
+
RESUME: latest
|
68 |
+
SAMPLE_FREQ: 1000
|
69 |
+
SAVE_FREQ: 5000
|
70 |
+
TRAIN_STEPS: 100000
|
downstream-watermark-removal/output-2023-02-12-14-28-38.log
ADDED
@@ -0,0 +1,223 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-02-12 14:28:40,985 - INFO - Experiment directory: runs/downstream-watermark-removal
|
2 |
+
2023-02-12 14:28:40,985 - INFO - Device: cuda
|
3 |
+
2023-02-12 14:28:40,985 - INFO - Number of devices: 1
|
4 |
+
2023-02-12 14:28:41,349 - INFO - Size of training set: 28352
|
5 |
+
2023-02-12 14:28:41,349 - INFO - Size of validation set: 4051
|
6 |
+
2023-02-12 14:28:41,349 - INFO - Batch size per device: 4
|
7 |
+
2023-02-12 14:28:41,349 - INFO - Effective batch size: 4
|
8 |
+
2023-02-12 14:28:42,908 - INFO - Successfully load mpn from ./runs/places-joint/ckpt/step149999.pt
|
9 |
+
2023-02-12 14:28:42,913 - INFO - Successfully load rin from ./runs/places-joint/ckpt/step149999.pt
|
10 |
+
2023-02-12 14:28:42,914 - INFO - Successfully load disc from ./runs/places-joint/ckpt/step149999.pt
|
11 |
+
2023-02-12 14:28:42,916 - INFO - Successfully load pdisc from ./runs/places-joint/ckpt/step149999.pt
|
12 |
+
2023-02-12 14:28:42,918 - INFO - Start training...
|
13 |
+
2023-02-12 14:29:14,414 - INFO - [Train] step: 99, loss_adv_disc: 1.299828
|
14 |
+
2023-02-12 14:29:14,621 - INFO - [Train] step: 99, loss_mpn: 0.037874, loss_rec: 0.041837, loss_semantic: 0.520840, loss_idmrf: 1.336823, loss_adv_gen: -128.098953
|
15 |
+
2023-02-12 14:29:43,591 - INFO - [Train] step: 199, loss_adv_disc: -1.303024
|
16 |
+
2023-02-12 14:29:43,800 - INFO - [Train] step: 199, loss_mpn: 0.036802, loss_rec: 0.040308, loss_semantic: 0.500949, loss_idmrf: 1.170398, loss_adv_gen: -188.719955
|
17 |
+
2023-02-12 14:30:12,854 - INFO - [Train] step: 299, loss_adv_disc: 0.880611
|
18 |
+
2023-02-12 14:30:13,062 - INFO - [Train] step: 299, loss_mpn: 0.022952, loss_rec: 0.032488, loss_semantic: 0.472941, loss_idmrf: 0.716021, loss_adv_gen: -164.942017
|
19 |
+
2023-02-12 14:30:42,137 - INFO - [Train] step: 399, loss_adv_disc: 1.241273
|
20 |
+
2023-02-12 14:30:42,346 - INFO - [Train] step: 399, loss_mpn: 0.020866, loss_rec: 0.041395, loss_semantic: 0.461653, loss_idmrf: 1.742797, loss_adv_gen: -94.545654
|
21 |
+
2023-02-12 14:31:11,414 - INFO - [Train] step: 499, loss_adv_disc: -4.018159
|
22 |
+
2023-02-12 14:31:11,624 - INFO - [Train] step: 499, loss_mpn: 0.042017, loss_rec: 0.038558, loss_semantic: 0.466815, loss_idmrf: 1.169562, loss_adv_gen: -143.637222
|
23 |
+
2023-02-12 14:31:40,708 - INFO - [Train] step: 599, loss_adv_disc: 0.942124
|
24 |
+
2023-02-12 14:31:40,916 - INFO - [Train] step: 599, loss_mpn: 0.012776, loss_rec: 0.029491, loss_semantic: 0.450419, loss_idmrf: 0.787979, loss_adv_gen: -97.693977
|
25 |
+
2023-02-12 14:32:10,001 - INFO - [Train] step: 699, loss_adv_disc: 3.814938
|
26 |
+
2023-02-12 14:32:10,211 - INFO - [Train] step: 699, loss_mpn: 0.028852, loss_rec: 0.043598, loss_semantic: 0.497849, loss_idmrf: 1.739692, loss_adv_gen: -155.648834
|
27 |
+
2023-02-12 14:32:39,303 - INFO - [Train] step: 799, loss_adv_disc: -1.610433
|
28 |
+
2023-02-12 14:32:39,512 - INFO - [Train] step: 799, loss_mpn: 0.029309, loss_rec: 0.045206, loss_semantic: 0.510565, loss_idmrf: 3.046881, loss_adv_gen: -96.266266
|
29 |
+
2023-02-12 14:33:08,610 - INFO - [Train] step: 899, loss_adv_disc: 1.118233
|
30 |
+
2023-02-12 14:33:08,819 - INFO - [Train] step: 899, loss_mpn: 0.012930, loss_rec: 0.022580, loss_semantic: 0.349494, loss_idmrf: 0.902996, loss_adv_gen: -88.348045
|
31 |
+
2023-02-12 14:33:37,905 - INFO - [Train] step: 999, loss_adv_disc: 0.263744
|
32 |
+
2023-02-12 14:33:38,114 - INFO - [Train] step: 999, loss_mpn: 0.023001, loss_rec: 0.043346, loss_semantic: 0.504102, loss_idmrf: 2.317935, loss_adv_gen: -102.413116
|
33 |
+
2023-02-12 14:34:15,528 - INFO - [Eval] step: 999, bce: 0.338886, psnr: 26.837254, ssim: 0.944074
|
34 |
+
2023-02-12 14:34:46,223 - INFO - [Train] step: 1099, loss_adv_disc: -2.614538
|
35 |
+
2023-02-12 14:34:46,432 - INFO - [Train] step: 1099, loss_mpn: 0.049012, loss_rec: 0.041178, loss_semantic: 0.505731, loss_idmrf: 1.993219, loss_adv_gen: -61.293377
|
36 |
+
2023-02-12 14:35:15,523 - INFO - [Train] step: 1199, loss_adv_disc: -3.433872
|
37 |
+
2023-02-12 14:35:15,731 - INFO - [Train] step: 1199, loss_mpn: 0.033315, loss_rec: 0.036747, loss_semantic: 0.420346, loss_idmrf: 1.440817, loss_adv_gen: -119.410286
|
38 |
+
2023-02-12 14:35:44,803 - INFO - [Train] step: 1299, loss_adv_disc: -1.207434
|
39 |
+
2023-02-12 14:35:45,012 - INFO - [Train] step: 1299, loss_mpn: 0.015608, loss_rec: 0.038091, loss_semantic: 0.484080, loss_idmrf: 0.916523, loss_adv_gen: -110.312073
|
40 |
+
2023-02-12 14:36:14,077 - INFO - [Train] step: 1399, loss_adv_disc: -5.074930
|
41 |
+
2023-02-12 14:36:14,286 - INFO - [Train] step: 1399, loss_mpn: 0.040863, loss_rec: 0.052102, loss_semantic: 0.532582, loss_idmrf: 1.521286, loss_adv_gen: -111.052315
|
42 |
+
2023-02-12 14:36:43,350 - INFO - [Train] step: 1499, loss_adv_disc: -1.698674
|
43 |
+
2023-02-12 14:36:43,558 - INFO - [Train] step: 1499, loss_mpn: 0.019298, loss_rec: 0.017522, loss_semantic: 0.335085, loss_idmrf: 1.877787, loss_adv_gen: -124.831696
|
44 |
+
2023-02-12 14:37:12,635 - INFO - [Train] step: 1599, loss_adv_disc: 0.078364
|
45 |
+
2023-02-12 14:37:12,844 - INFO - [Train] step: 1599, loss_mpn: 0.032950, loss_rec: 0.033342, loss_semantic: 0.452769, loss_idmrf: 1.889498, loss_adv_gen: -102.988449
|
46 |
+
2023-02-12 14:37:41,908 - INFO - [Train] step: 1699, loss_adv_disc: 3.190553
|
47 |
+
2023-02-12 14:37:42,117 - INFO - [Train] step: 1699, loss_mpn: 0.024712, loss_rec: 0.037171, loss_semantic: 0.477129, loss_idmrf: 1.474838, loss_adv_gen: -107.239792
|
48 |
+
2023-02-12 14:38:11,183 - INFO - [Train] step: 1799, loss_adv_disc: -0.911062
|
49 |
+
2023-02-12 14:38:11,391 - INFO - [Train] step: 1799, loss_mpn: 0.018543, loss_rec: 0.032870, loss_semantic: 0.416096, loss_idmrf: 2.942292, loss_adv_gen: -97.946312
|
50 |
+
2023-02-12 14:38:40,450 - INFO - [Train] step: 1899, loss_adv_disc: -2.992178
|
51 |
+
2023-02-12 14:38:40,658 - INFO - [Train] step: 1899, loss_mpn: 0.044249, loss_rec: 0.053050, loss_semantic: 0.493338, loss_idmrf: 1.900263, loss_adv_gen: -102.013123
|
52 |
+
2023-02-12 14:39:09,716 - INFO - [Train] step: 1999, loss_adv_disc: 1.107126
|
53 |
+
2023-02-12 14:39:09,925 - INFO - [Train] step: 1999, loss_mpn: 0.049591, loss_rec: 0.039920, loss_semantic: 0.413558, loss_idmrf: 1.118119, loss_adv_gen: -138.656616
|
54 |
+
2023-02-12 14:39:47,107 - INFO - [Eval] step: 1999, bce: 0.373344, psnr: 27.043365, ssim: 0.946821
|
55 |
+
2023-02-12 14:40:17,430 - INFO - [Train] step: 2099, loss_adv_disc: -2.116822
|
56 |
+
2023-02-12 14:40:17,638 - INFO - [Train] step: 2099, loss_mpn: 0.022930, loss_rec: 0.023332, loss_semantic: 0.383731, loss_idmrf: 1.122571, loss_adv_gen: -93.001617
|
57 |
+
2023-02-12 14:40:46,705 - INFO - [Train] step: 2199, loss_adv_disc: 1.641629
|
58 |
+
2023-02-12 14:40:46,913 - INFO - [Train] step: 2199, loss_mpn: 0.015344, loss_rec: 0.028494, loss_semantic: 0.344031, loss_idmrf: 1.670086, loss_adv_gen: -85.842712
|
59 |
+
2023-02-12 14:41:15,969 - INFO - [Train] step: 2299, loss_adv_disc: -1.291130
|
60 |
+
2023-02-12 14:41:16,177 - INFO - [Train] step: 2299, loss_mpn: 0.013970, loss_rec: 0.028013, loss_semantic: 0.418684, loss_idmrf: 1.546552, loss_adv_gen: -87.390312
|
61 |
+
2023-02-12 14:41:45,234 - INFO - [Train] step: 2399, loss_adv_disc: 2.109378
|
62 |
+
2023-02-12 14:41:45,443 - INFO - [Train] step: 2399, loss_mpn: 0.038823, loss_rec: 0.049957, loss_semantic: 0.547463, loss_idmrf: 1.906920, loss_adv_gen: -140.577087
|
63 |
+
2023-02-12 14:42:14,492 - INFO - [Train] step: 2499, loss_adv_disc: -2.110101
|
64 |
+
2023-02-12 14:42:14,700 - INFO - [Train] step: 2499, loss_mpn: 0.028055, loss_rec: 0.036691, loss_semantic: 0.471377, loss_idmrf: 1.584231, loss_adv_gen: -66.208374
|
65 |
+
2023-02-12 14:42:43,743 - INFO - [Train] step: 2599, loss_adv_disc: 1.007918
|
66 |
+
2023-02-12 14:42:43,951 - INFO - [Train] step: 2599, loss_mpn: 0.011720, loss_rec: 0.039426, loss_semantic: 0.476948, loss_idmrf: 1.416666, loss_adv_gen: -108.188354
|
67 |
+
2023-02-12 14:43:12,995 - INFO - [Train] step: 2699, loss_adv_disc: 0.309885
|
68 |
+
2023-02-12 14:43:13,203 - INFO - [Train] step: 2699, loss_mpn: 0.043115, loss_rec: 0.048996, loss_semantic: 0.506540, loss_idmrf: 1.823579, loss_adv_gen: -79.640060
|
69 |
+
2023-02-12 14:43:42,259 - INFO - [Train] step: 2799, loss_adv_disc: 1.769973
|
70 |
+
2023-02-12 14:43:42,470 - INFO - [Train] step: 2799, loss_mpn: 0.029426, loss_rec: 0.040913, loss_semantic: 0.455353, loss_idmrf: 1.369126, loss_adv_gen: -82.098961
|
71 |
+
2023-02-12 14:44:11,518 - INFO - [Train] step: 2899, loss_adv_disc: -2.826468
|
72 |
+
2023-02-12 14:44:11,727 - INFO - [Train] step: 2899, loss_mpn: 0.024427, loss_rec: 0.036017, loss_semantic: 0.450512, loss_idmrf: 1.423253, loss_adv_gen: -119.713104
|
73 |
+
2023-02-12 14:44:40,782 - INFO - [Train] step: 2999, loss_adv_disc: -0.511095
|
74 |
+
2023-02-12 14:44:40,992 - INFO - [Train] step: 2999, loss_mpn: 0.019680, loss_rec: 0.028514, loss_semantic: 0.381640, loss_idmrf: 1.594770, loss_adv_gen: -121.859711
|
75 |
+
2023-02-12 14:45:18,178 - INFO - [Eval] step: 2999, bce: 0.325142, psnr: 27.303087, ssim: 0.947815
|
76 |
+
2023-02-12 14:45:48,502 - INFO - [Train] step: 3099, loss_adv_disc: 0.043045
|
77 |
+
2023-02-12 14:45:48,711 - INFO - [Train] step: 3099, loss_mpn: 0.016258, loss_rec: 0.021658, loss_semantic: 0.378096, loss_idmrf: 1.504422, loss_adv_gen: -198.545273
|
78 |
+
2023-02-12 14:46:17,756 - INFO - [Train] step: 3199, loss_adv_disc: -1.038017
|
79 |
+
2023-02-12 14:46:17,965 - INFO - [Train] step: 3199, loss_mpn: 0.012933, loss_rec: 0.026565, loss_semantic: 0.426411, loss_idmrf: 1.404650, loss_adv_gen: -96.168388
|
80 |
+
2023-02-12 14:46:47,018 - INFO - [Train] step: 3299, loss_adv_disc: 0.046620
|
81 |
+
2023-02-12 14:46:47,226 - INFO - [Train] step: 3299, loss_mpn: 0.011906, loss_rec: 0.014265, loss_semantic: 0.269762, loss_idmrf: 0.559204, loss_adv_gen: -99.662216
|
82 |
+
2023-02-12 14:47:16,291 - INFO - [Train] step: 3399, loss_adv_disc: -0.021643
|
83 |
+
2023-02-12 14:47:16,499 - INFO - [Train] step: 3399, loss_mpn: 0.017540, loss_rec: 0.029993, loss_semantic: 0.392757, loss_idmrf: 1.623783, loss_adv_gen: -72.550812
|
84 |
+
2023-02-12 14:47:45,556 - INFO - [Train] step: 3499, loss_adv_disc: 0.026034
|
85 |
+
2023-02-12 14:47:45,764 - INFO - [Train] step: 3499, loss_mpn: 0.024654, loss_rec: 0.025382, loss_semantic: 0.367133, loss_idmrf: 1.421693, loss_adv_gen: -106.288788
|
86 |
+
2023-02-12 14:48:14,815 - INFO - [Train] step: 3599, loss_adv_disc: -0.349758
|
87 |
+
2023-02-12 14:48:15,023 - INFO - [Train] step: 3599, loss_mpn: 0.032561, loss_rec: 0.033780, loss_semantic: 0.448922, loss_idmrf: 1.408511, loss_adv_gen: -133.010773
|
88 |
+
2023-02-12 14:48:44,080 - INFO - [Train] step: 3699, loss_adv_disc: -3.845224
|
89 |
+
2023-02-12 14:48:44,288 - INFO - [Train] step: 3699, loss_mpn: 0.046263, loss_rec: 0.035030, loss_semantic: 0.449073, loss_idmrf: 1.693964, loss_adv_gen: -101.824783
|
90 |
+
2023-02-12 14:49:13,351 - INFO - [Train] step: 3799, loss_adv_disc: 0.691240
|
91 |
+
2023-02-12 14:49:13,559 - INFO - [Train] step: 3799, loss_mpn: 0.033147, loss_rec: 0.032454, loss_semantic: 0.412963, loss_idmrf: 1.078437, loss_adv_gen: -133.369232
|
92 |
+
2023-02-12 14:49:42,612 - INFO - [Train] step: 3899, loss_adv_disc: -1.092527
|
93 |
+
2023-02-12 14:49:42,820 - INFO - [Train] step: 3899, loss_mpn: 0.030574, loss_rec: 0.035086, loss_semantic: 0.450594, loss_idmrf: 1.633320, loss_adv_gen: -157.005127
|
94 |
+
2023-02-12 14:50:11,886 - INFO - [Train] step: 3999, loss_adv_disc: 0.675812
|
95 |
+
2023-02-12 14:50:12,095 - INFO - [Train] step: 3999, loss_mpn: 0.021050, loss_rec: 0.033965, loss_semantic: 0.449186, loss_idmrf: 1.248645, loss_adv_gen: -97.598892
|
96 |
+
2023-02-12 14:50:49,252 - INFO - [Eval] step: 3999, bce: 0.346579, psnr: 27.542393, ssim: 0.949793
|
97 |
+
2023-02-12 14:51:19,568 - INFO - [Train] step: 4099, loss_adv_disc: 0.768379
|
98 |
+
2023-02-12 14:51:19,777 - INFO - [Train] step: 4099, loss_mpn: 0.013409, loss_rec: 0.029262, loss_semantic: 0.393055, loss_idmrf: 1.478358, loss_adv_gen: -109.265968
|
99 |
+
2023-02-12 14:51:48,839 - INFO - [Train] step: 4199, loss_adv_disc: -0.045594
|
100 |
+
2023-02-12 14:51:49,047 - INFO - [Train] step: 4199, loss_mpn: 0.024171, loss_rec: 0.017396, loss_semantic: 0.286983, loss_idmrf: 0.789566, loss_adv_gen: -148.535095
|
101 |
+
2023-02-12 14:52:18,101 - INFO - [Train] step: 4299, loss_adv_disc: -0.430947
|
102 |
+
2023-02-12 14:52:18,310 - INFO - [Train] step: 4299, loss_mpn: 0.014691, loss_rec: 0.021662, loss_semantic: 0.326292, loss_idmrf: 0.946436, loss_adv_gen: -91.638748
|
103 |
+
2023-02-12 14:52:47,360 - INFO - [Train] step: 4399, loss_adv_disc: -0.410857
|
104 |
+
2023-02-12 14:52:47,568 - INFO - [Train] step: 4399, loss_mpn: 0.011707, loss_rec: 0.021423, loss_semantic: 0.334674, loss_idmrf: 0.605388, loss_adv_gen: -137.055222
|
105 |
+
2023-02-12 14:53:16,621 - INFO - [Train] step: 4499, loss_adv_disc: -0.533930
|
106 |
+
2023-02-12 14:53:16,829 - INFO - [Train] step: 4499, loss_mpn: 0.012462, loss_rec: 0.036778, loss_semantic: 0.428064, loss_idmrf: 1.709253, loss_adv_gen: -201.091476
|
107 |
+
2023-02-12 14:53:45,873 - INFO - [Train] step: 4599, loss_adv_disc: -1.762612
|
108 |
+
2023-02-12 14:53:46,082 - INFO - [Train] step: 4599, loss_mpn: 0.026616, loss_rec: 0.052115, loss_semantic: 0.480837, loss_idmrf: 1.606442, loss_adv_gen: -97.646194
|
109 |
+
2023-02-12 14:54:15,132 - INFO - [Train] step: 4699, loss_adv_disc: -0.790706
|
110 |
+
2023-02-12 14:54:15,340 - INFO - [Train] step: 4699, loss_mpn: 0.018837, loss_rec: 0.028844, loss_semantic: 0.392761, loss_idmrf: 1.836121, loss_adv_gen: -98.939705
|
111 |
+
2023-02-12 14:54:44,391 - INFO - [Train] step: 4799, loss_adv_disc: -0.965508
|
112 |
+
2023-02-12 14:54:44,601 - INFO - [Train] step: 4799, loss_mpn: 0.032487, loss_rec: 0.035681, loss_semantic: 0.437911, loss_idmrf: 1.694472, loss_adv_gen: -81.396034
|
113 |
+
2023-02-12 14:55:13,651 - INFO - [Train] step: 4899, loss_adv_disc: -1.041402
|
114 |
+
2023-02-12 14:55:13,860 - INFO - [Train] step: 4899, loss_mpn: 0.028312, loss_rec: 0.028561, loss_semantic: 0.396117, loss_idmrf: 1.250476, loss_adv_gen: -97.100685
|
115 |
+
2023-02-12 14:55:42,913 - INFO - [Train] step: 4999, loss_adv_disc: 0.711361
|
116 |
+
2023-02-12 14:55:43,122 - INFO - [Train] step: 4999, loss_mpn: 0.021475, loss_rec: 0.026028, loss_semantic: 0.379468, loss_idmrf: 1.322870, loss_adv_gen: -115.331512
|
117 |
+
2023-02-12 14:56:20,330 - INFO - [Eval] step: 4999, bce: 0.351477, psnr: 27.780710, ssim: 0.950434
|
118 |
+
2023-02-12 14:56:50,783 - INFO - [Train] step: 5099, loss_adv_disc: -2.985783
|
119 |
+
2023-02-12 14:56:50,991 - INFO - [Train] step: 5099, loss_mpn: 0.019004, loss_rec: 0.041138, loss_semantic: 0.508738, loss_idmrf: 0.819962, loss_adv_gen: -182.782623
|
120 |
+
2023-02-12 14:57:20,046 - INFO - [Train] step: 5199, loss_adv_disc: -1.645151
|
121 |
+
2023-02-12 14:57:20,256 - INFO - [Train] step: 5199, loss_mpn: 0.016225, loss_rec: 0.025021, loss_semantic: 0.370635, loss_idmrf: 0.914956, loss_adv_gen: -103.670837
|
122 |
+
2023-02-12 14:57:49,304 - INFO - [Train] step: 5299, loss_adv_disc: -0.446883
|
123 |
+
2023-02-12 14:57:49,512 - INFO - [Train] step: 5299, loss_mpn: 0.025473, loss_rec: 0.035687, loss_semantic: 0.464093, loss_idmrf: 1.667971, loss_adv_gen: -153.179428
|
124 |
+
2023-02-12 14:58:18,560 - INFO - [Train] step: 5399, loss_adv_disc: 0.041118
|
125 |
+
2023-02-12 14:58:18,767 - INFO - [Train] step: 5399, loss_mpn: 0.022868, loss_rec: 0.027386, loss_semantic: 0.379084, loss_idmrf: 1.039361, loss_adv_gen: -118.277161
|
126 |
+
2023-02-12 14:58:47,819 - INFO - [Train] step: 5499, loss_adv_disc: -0.695383
|
127 |
+
2023-02-12 14:58:48,028 - INFO - [Train] step: 5499, loss_mpn: 0.017166, loss_rec: 0.023030, loss_semantic: 0.343284, loss_idmrf: 1.867031, loss_adv_gen: -132.320648
|
128 |
+
2023-02-12 14:59:17,075 - INFO - [Train] step: 5599, loss_adv_disc: -0.396028
|
129 |
+
2023-02-12 14:59:17,283 - INFO - [Train] step: 5599, loss_mpn: 0.016136, loss_rec: 0.024782, loss_semantic: 0.387950, loss_idmrf: 1.010486, loss_adv_gen: -124.445305
|
130 |
+
2023-02-12 14:59:46,331 - INFO - [Train] step: 5699, loss_adv_disc: -2.364127
|
131 |
+
2023-02-12 14:59:46,539 - INFO - [Train] step: 5699, loss_mpn: 0.022420, loss_rec: 0.027926, loss_semantic: 0.412665, loss_idmrf: 1.334894, loss_adv_gen: -62.740250
|
132 |
+
2023-02-12 15:00:15,585 - INFO - [Train] step: 5799, loss_adv_disc: -0.426686
|
133 |
+
2023-02-12 15:00:15,793 - INFO - [Train] step: 5799, loss_mpn: 0.020284, loss_rec: 0.029912, loss_semantic: 0.416979, loss_idmrf: 2.073402, loss_adv_gen: -149.003220
|
134 |
+
2023-02-12 15:00:44,848 - INFO - [Train] step: 5899, loss_adv_disc: -0.774378
|
135 |
+
2023-02-12 15:00:45,057 - INFO - [Train] step: 5899, loss_mpn: 0.013465, loss_rec: 0.031138, loss_semantic: 0.415459, loss_idmrf: 0.909214, loss_adv_gen: -113.032188
|
136 |
+
2023-02-12 15:01:14,101 - INFO - [Train] step: 5999, loss_adv_disc: -0.389311
|
137 |
+
2023-02-12 15:01:14,312 - INFO - [Train] step: 5999, loss_mpn: 0.039259, loss_rec: 0.035546, loss_semantic: 0.422057, loss_idmrf: 1.271151, loss_adv_gen: -117.240402
|
138 |
+
2023-02-12 15:01:51,501 - INFO - [Eval] step: 5999, bce: 0.340868, psnr: 27.871330, ssim: 0.950925
|
139 |
+
2023-02-12 15:02:21,828 - INFO - [Train] step: 6099, loss_adv_disc: -0.167378
|
140 |
+
2023-02-12 15:02:22,037 - INFO - [Train] step: 6099, loss_mpn: 0.016576, loss_rec: 0.036952, loss_semantic: 0.467681, loss_idmrf: 1.342181, loss_adv_gen: -104.226501
|
141 |
+
2023-02-12 15:02:51,091 - INFO - [Train] step: 6199, loss_adv_disc: -0.683909
|
142 |
+
2023-02-12 15:02:51,300 - INFO - [Train] step: 6199, loss_mpn: 0.018274, loss_rec: 0.032154, loss_semantic: 0.425463, loss_idmrf: 0.816652, loss_adv_gen: -86.657425
|
143 |
+
2023-02-12 15:03:20,351 - INFO - [Train] step: 6299, loss_adv_disc: -1.493726
|
144 |
+
2023-02-12 15:03:20,560 - INFO - [Train] step: 6299, loss_mpn: 0.023145, loss_rec: 0.033411, loss_semantic: 0.432057, loss_idmrf: 0.897825, loss_adv_gen: -139.223816
|
145 |
+
2023-02-12 15:03:49,616 - INFO - [Train] step: 6399, loss_adv_disc: -0.277531
|
146 |
+
2023-02-12 15:03:49,824 - INFO - [Train] step: 6399, loss_mpn: 0.017375, loss_rec: 0.029168, loss_semantic: 0.402245, loss_idmrf: 0.622658, loss_adv_gen: -142.869843
|
147 |
+
2023-02-12 15:04:18,892 - INFO - [Train] step: 6499, loss_adv_disc: -0.235311
|
148 |
+
2023-02-12 15:04:19,101 - INFO - [Train] step: 6499, loss_mpn: 0.012269, loss_rec: 0.022133, loss_semantic: 0.319729, loss_idmrf: 1.260930, loss_adv_gen: -171.630920
|
149 |
+
2023-02-12 15:04:48,160 - INFO - [Train] step: 6599, loss_adv_disc: -0.469994
|
150 |
+
2023-02-12 15:04:48,370 - INFO - [Train] step: 6599, loss_mpn: 0.008784, loss_rec: 0.019133, loss_semantic: 0.306887, loss_idmrf: 0.801923, loss_adv_gen: -64.075722
|
151 |
+
2023-02-12 15:05:17,426 - INFO - [Train] step: 6699, loss_adv_disc: -1.466449
|
152 |
+
2023-02-12 15:05:17,634 - INFO - [Train] step: 6699, loss_mpn: 0.021287, loss_rec: 0.035671, loss_semantic: 0.454560, loss_idmrf: 1.030252, loss_adv_gen: -120.588272
|
153 |
+
2023-02-12 15:05:46,695 - INFO - [Train] step: 6799, loss_adv_disc: 0.515251
|
154 |
+
2023-02-12 15:05:46,904 - INFO - [Train] step: 6799, loss_mpn: 0.013280, loss_rec: 0.027942, loss_semantic: 0.360047, loss_idmrf: 1.654778, loss_adv_gen: -180.273392
|
155 |
+
2023-02-12 15:06:15,963 - INFO - [Train] step: 6899, loss_adv_disc: -0.489713
|
156 |
+
2023-02-12 15:06:16,171 - INFO - [Train] step: 6899, loss_mpn: 0.018902, loss_rec: 0.038339, loss_semantic: 0.442717, loss_idmrf: 2.431321, loss_adv_gen: -115.624718
|
157 |
+
2023-02-12 15:06:45,239 - INFO - [Train] step: 6999, loss_adv_disc: -1.769455
|
158 |
+
2023-02-12 15:06:45,447 - INFO - [Train] step: 6999, loss_mpn: 0.019381, loss_rec: 0.028560, loss_semantic: 0.383751, loss_idmrf: 1.163507, loss_adv_gen: -153.030273
|
159 |
+
2023-02-12 15:07:22,635 - INFO - [Eval] step: 6999, bce: 0.306903, psnr: 28.013441, ssim: 0.951795
|
160 |
+
2023-02-12 15:07:53,174 - INFO - [Train] step: 7099, loss_adv_disc: 1.659884
|
161 |
+
2023-02-12 15:07:53,383 - INFO - [Train] step: 7099, loss_mpn: 0.028287, loss_rec: 0.037931, loss_semantic: 0.468431, loss_idmrf: 1.497301, loss_adv_gen: -132.254868
|
162 |
+
2023-02-12 15:08:22,426 - INFO - [Train] step: 7199, loss_adv_disc: 0.472226
|
163 |
+
2023-02-12 15:08:22,636 - INFO - [Train] step: 7199, loss_mpn: 0.013172, loss_rec: 0.024137, loss_semantic: 0.364234, loss_idmrf: 1.162045, loss_adv_gen: -132.014572
|
164 |
+
2023-02-12 15:08:51,699 - INFO - [Train] step: 7299, loss_adv_disc: -0.581740
|
165 |
+
2023-02-12 15:08:51,908 - INFO - [Train] step: 7299, loss_mpn: 0.012158, loss_rec: 0.031973, loss_semantic: 0.362457, loss_idmrf: 1.268194, loss_adv_gen: -137.127380
|
166 |
+
2023-02-12 15:09:20,969 - INFO - [Train] step: 7399, loss_adv_disc: -0.978511
|
167 |
+
2023-02-12 15:09:21,178 - INFO - [Train] step: 7399, loss_mpn: 0.022330, loss_rec: 0.024040, loss_semantic: 0.309038, loss_idmrf: 0.792379, loss_adv_gen: -106.908493
|
168 |
+
2023-02-12 15:09:50,228 - INFO - [Train] step: 7499, loss_adv_disc: -0.073679
|
169 |
+
2023-02-12 15:09:50,436 - INFO - [Train] step: 7499, loss_mpn: 0.018063, loss_rec: 0.022740, loss_semantic: 0.319768, loss_idmrf: 1.349909, loss_adv_gen: -165.113327
|
170 |
+
2023-02-12 15:10:19,481 - INFO - [Train] step: 7599, loss_adv_disc: 0.049674
|
171 |
+
2023-02-12 15:10:19,689 - INFO - [Train] step: 7599, loss_mpn: 0.010805, loss_rec: 0.025964, loss_semantic: 0.350963, loss_idmrf: 0.961297, loss_adv_gen: -113.559540
|
172 |
+
2023-02-12 15:10:48,747 - INFO - [Train] step: 7699, loss_adv_disc: -0.608275
|
173 |
+
2023-02-12 15:10:48,955 - INFO - [Train] step: 7699, loss_mpn: 0.016553, loss_rec: 0.027370, loss_semantic: 0.380454, loss_idmrf: 1.702928, loss_adv_gen: -108.241295
|
174 |
+
2023-02-12 15:11:18,013 - INFO - [Train] step: 7799, loss_adv_disc: 1.413847
|
175 |
+
2023-02-12 15:11:18,221 - INFO - [Train] step: 7799, loss_mpn: 0.021216, loss_rec: 0.033353, loss_semantic: 0.390046, loss_idmrf: 1.304851, loss_adv_gen: -118.235794
|
176 |
+
2023-02-12 15:11:47,280 - INFO - [Train] step: 7899, loss_adv_disc: -0.615377
|
177 |
+
2023-02-12 15:11:47,489 - INFO - [Train] step: 7899, loss_mpn: 0.016149, loss_rec: 0.024272, loss_semantic: 0.358250, loss_idmrf: 0.852178, loss_adv_gen: -133.236221
|
178 |
+
2023-02-12 15:12:16,555 - INFO - [Train] step: 7999, loss_adv_disc: -0.062408
|
179 |
+
2023-02-12 15:12:16,764 - INFO - [Train] step: 7999, loss_mpn: 0.013659, loss_rec: 0.026106, loss_semantic: 0.386249, loss_idmrf: 1.021867, loss_adv_gen: -147.716049
|
180 |
+
2023-02-12 15:12:53,951 - INFO - [Eval] step: 7999, bce: 0.288697, psnr: 28.163181, ssim: 0.952470
|
181 |
+
2023-02-12 15:13:24,273 - INFO - [Train] step: 8099, loss_adv_disc: -1.120922
|
182 |
+
2023-02-12 15:13:24,481 - INFO - [Train] step: 8099, loss_mpn: 0.004955, loss_rec: 0.020915, loss_semantic: 0.322940, loss_idmrf: 1.106715, loss_adv_gen: -151.821106
|
183 |
+
2023-02-12 15:13:53,543 - INFO - [Train] step: 8199, loss_adv_disc: 0.029411
|
184 |
+
2023-02-12 15:13:53,753 - INFO - [Train] step: 8199, loss_mpn: 0.026964, loss_rec: 0.028958, loss_semantic: 0.445343, loss_idmrf: 1.209411, loss_adv_gen: -102.843552
|
185 |
+
2023-02-12 15:14:22,805 - INFO - [Train] step: 8299, loss_adv_disc: -4.315375
|
186 |
+
2023-02-12 15:14:23,012 - INFO - [Train] step: 8299, loss_mpn: 0.019549, loss_rec: 0.034941, loss_semantic: 0.462220, loss_idmrf: 0.890771, loss_adv_gen: -154.584610
|
187 |
+
2023-02-12 15:14:52,068 - INFO - [Train] step: 8399, loss_adv_disc: -1.995814
|
188 |
+
2023-02-12 15:14:52,276 - INFO - [Train] step: 8399, loss_mpn: 0.012006, loss_rec: 0.034098, loss_semantic: 0.396413, loss_idmrf: 0.811635, loss_adv_gen: -96.562088
|
189 |
+
2023-02-12 15:15:21,323 - INFO - [Train] step: 8499, loss_adv_disc: 0.820693
|
190 |
+
2023-02-12 15:15:21,532 - INFO - [Train] step: 8499, loss_mpn: 0.013124, loss_rec: 0.029301, loss_semantic: 0.393588, loss_idmrf: 1.013259, loss_adv_gen: -89.326385
|
191 |
+
2023-02-12 15:15:50,592 - INFO - [Train] step: 8599, loss_adv_disc: 0.393147
|
192 |
+
2023-02-12 15:15:50,800 - INFO - [Train] step: 8599, loss_mpn: 0.011680, loss_rec: 0.017792, loss_semantic: 0.300831, loss_idmrf: 0.998240, loss_adv_gen: -184.119385
|
193 |
+
2023-02-12 15:16:19,842 - INFO - [Train] step: 8699, loss_adv_disc: -0.531485
|
194 |
+
2023-02-12 15:16:20,051 - INFO - [Train] step: 8699, loss_mpn: 0.015221, loss_rec: 0.025195, loss_semantic: 0.341359, loss_idmrf: 1.062933, loss_adv_gen: -110.207268
|
195 |
+
2023-02-12 15:16:49,123 - INFO - [Train] step: 8799, loss_adv_disc: -0.491042
|
196 |
+
2023-02-12 15:16:49,332 - INFO - [Train] step: 8799, loss_mpn: 0.010106, loss_rec: 0.024656, loss_semantic: 0.360912, loss_idmrf: 0.767988, loss_adv_gen: -151.354294
|
197 |
+
2023-02-12 15:17:18,385 - INFO - [Train] step: 8899, loss_adv_disc: -0.419860
|
198 |
+
2023-02-12 15:17:18,594 - INFO - [Train] step: 8899, loss_mpn: 0.013824, loss_rec: 0.030028, loss_semantic: 0.417085, loss_idmrf: 1.085957, loss_adv_gen: -207.431549
|
199 |
+
2023-02-12 15:17:47,642 - INFO - [Train] step: 8999, loss_adv_disc: -0.628583
|
200 |
+
2023-02-12 15:17:47,850 - INFO - [Train] step: 8999, loss_mpn: 0.024702, loss_rec: 0.037819, loss_semantic: 0.416768, loss_idmrf: 2.331512, loss_adv_gen: -108.836723
|
201 |
+
2023-02-12 15:18:25,047 - INFO - [Eval] step: 8999, bce: 0.302010, psnr: 28.254166, ssim: 0.952631
|
202 |
+
2023-02-12 15:18:55,405 - INFO - [Train] step: 9099, loss_adv_disc: 0.632391
|
203 |
+
2023-02-12 15:18:55,616 - INFO - [Train] step: 9099, loss_mpn: 0.014931, loss_rec: 0.024709, loss_semantic: 0.325625, loss_idmrf: 1.853235, loss_adv_gen: -115.922569
|
204 |
+
2023-02-12 15:19:24,673 - INFO - [Train] step: 9199, loss_adv_disc: -1.815902
|
205 |
+
2023-02-12 15:19:24,881 - INFO - [Train] step: 9199, loss_mpn: 0.016712, loss_rec: 0.022784, loss_semantic: 0.345748, loss_idmrf: 1.113174, loss_adv_gen: -51.955341
|
206 |
+
2023-02-12 15:19:53,938 - INFO - [Train] step: 9299, loss_adv_disc: -0.156092
|
207 |
+
2023-02-12 15:19:54,147 - INFO - [Train] step: 9299, loss_mpn: 0.011803, loss_rec: 0.023657, loss_semantic: 0.339971, loss_idmrf: 1.318946, loss_adv_gen: -127.398438
|
208 |
+
2023-02-12 15:20:23,205 - INFO - [Train] step: 9399, loss_adv_disc: -0.345600
|
209 |
+
2023-02-12 15:20:23,415 - INFO - [Train] step: 9399, loss_mpn: 0.036090, loss_rec: 0.027977, loss_semantic: 0.409671, loss_idmrf: 1.271996, loss_adv_gen: -96.280853
|
210 |
+
2023-02-12 15:20:52,463 - INFO - [Train] step: 9499, loss_adv_disc: -0.154040
|
211 |
+
2023-02-12 15:20:52,671 - INFO - [Train] step: 9499, loss_mpn: 0.008822, loss_rec: 0.023087, loss_semantic: 0.333166, loss_idmrf: 1.134625, loss_adv_gen: -172.469971
|
212 |
+
2023-02-12 15:21:21,728 - INFO - [Train] step: 9599, loss_adv_disc: -2.609488
|
213 |
+
2023-02-12 15:21:21,937 - INFO - [Train] step: 9599, loss_mpn: 0.018622, loss_rec: 0.031683, loss_semantic: 0.426192, loss_idmrf: 0.870782, loss_adv_gen: -63.231483
|
214 |
+
2023-02-12 15:21:50,997 - INFO - [Train] step: 9699, loss_adv_disc: -1.726371
|
215 |
+
2023-02-12 15:21:51,205 - INFO - [Train] step: 9699, loss_mpn: 0.016741, loss_rec: 0.030167, loss_semantic: 0.398618, loss_idmrf: 1.257529, loss_adv_gen: -196.563248
|
216 |
+
2023-02-12 15:22:20,258 - INFO - [Train] step: 9799, loss_adv_disc: -2.347822
|
217 |
+
2023-02-12 15:22:20,466 - INFO - [Train] step: 9799, loss_mpn: 0.021154, loss_rec: 0.028975, loss_semantic: 0.403203, loss_idmrf: 0.936794, loss_adv_gen: -92.555832
|
218 |
+
2023-02-12 15:22:49,519 - INFO - [Train] step: 9899, loss_adv_disc: -0.370060
|
219 |
+
2023-02-12 15:22:49,728 - INFO - [Train] step: 9899, loss_mpn: 0.013475, loss_rec: 0.026137, loss_semantic: 0.385814, loss_idmrf: 0.893062, loss_adv_gen: -135.118530
|
220 |
+
2023-02-12 15:23:18,785 - INFO - [Train] step: 9999, loss_adv_disc: -1.345161
|
221 |
+
2023-02-12 15:23:18,993 - INFO - [Train] step: 9999, loss_mpn: 0.022469, loss_rec: 0.028983, loss_semantic: 0.421686, loss_idmrf: 1.735447, loss_adv_gen: -143.855499
|
222 |
+
2023-02-12 15:23:56,181 - INFO - [Eval] step: 9999, bce: 0.237329, psnr: 28.375944, ssim: 0.953470
|
223 |
+
2023-02-12 15:23:57,541 - INFO - End of training
|
downstream-watermark-removal/output-2023-02-12-15-41-28.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
downstream-watermark-removal/samples/step009999.png
ADDED
![]() |
Git LFS Details
|
downstream-watermark-removal/samples/step099999.png
ADDED
![]() |
Git LFS Details
|
downstream-watermark-removal/tensorboard/events.out.tfevents.1676212121.jason-system.50864.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:72f66be949f0e94261734eec3cbad480c18474ca8baa123a2a45c385766c716d
|
3 |
+
size 3340690
|
downstream-watermark-removal/tensorboard/events.out.tfevents.1676216490.jason-system.51652.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e75c5e9ee010d9336b6c412ca566ae68faaedcece11d58e810a0f37f1a197be9
|
3 |
+
size 30574858
|
joint_ffhq_brush_realnoise/ckpt/step149999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
|
3 |
+
size 425
|
joint_ffhq_brush_realnoise/ckpt/step149999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d809ab838d9255922d407abef2cba547971c2a927491a17be7c50c2a330c4197
|
3 |
+
size 53562757
|
joint_ffhq_brush_realnoise/ckpt/step149999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:48843df6a56026b7b32b5df254a7920bfd82ec587d071398a38d2e620fd29b5d
|
3 |
+
size 106908613
|
joint_ffhq_brush_realnoise/config-2023-03-20-11-52-44.yaml
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
data:
|
2 |
+
dataroot: /amax/xyf/data/FFHQ/
|
3 |
+
img_size: 256
|
4 |
+
name: FFHQ
|
5 |
+
dataloader:
|
6 |
+
micro_batch: 0
|
7 |
+
num_workers: 4
|
8 |
+
pin_memory: true
|
9 |
+
prefetch_factor: 2
|
10 |
+
mask:
|
11 |
+
mask_type: brush
|
12 |
+
noise_type: real
|
13 |
+
real_dataset:
|
14 |
+
- dataroot: /amax/xyf/data/CelebA-HQ/
|
15 |
+
img_size: 256
|
16 |
+
name: CelebA-HQ
|
17 |
+
- dataroot: /amax/xyf/data/ImageNet/
|
18 |
+
img_size: 256
|
19 |
+
name: ImageNet
|
20 |
+
smooth_type: iterative_gaussian
|
21 |
+
model:
|
22 |
+
mpn:
|
23 |
+
base_n_channels: 64
|
24 |
+
neck_n_channels: 128
|
25 |
+
rin:
|
26 |
+
base_n_channels: 32
|
27 |
+
neck_n_channels: 128
|
28 |
+
seed: 1234
|
29 |
+
train:
|
30 |
+
batch_size: 4
|
31 |
+
coef_adv: 0.001
|
32 |
+
coef_idmrf: 0.001
|
33 |
+
coef_mpn: 2.0
|
34 |
+
coef_rec: 1.4
|
35 |
+
coef_semantic: 0.0001
|
36 |
+
eval_freq: 1000
|
37 |
+
n_steps: 150000
|
38 |
+
optim_disc:
|
39 |
+
betas:
|
40 |
+
- 0.5
|
41 |
+
- 0.9
|
42 |
+
lr: 1.0e-05
|
43 |
+
optim_mpn:
|
44 |
+
betas:
|
45 |
+
- 0.5
|
46 |
+
- 0.9
|
47 |
+
lr: 1.0e-05
|
48 |
+
optim_rin:
|
49 |
+
betas:
|
50 |
+
- 0.5
|
51 |
+
- 0.9
|
52 |
+
lr: 1.0e-05
|
53 |
+
pretrained: ./runs/separate_ffhq_brush_realnoise/ckpt/step079999/model.pt
|
54 |
+
print_freq: 100
|
55 |
+
resume: null
|
56 |
+
sample_freq: 1000
|
57 |
+
save_freq: 5000
|
joint_ffhq_brush_realnoise/output-2023-03-20-11-52-44.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
joint_ffhq_brush_realnoise/samples/step149999.png
ADDED
![]() |
Git LFS Details
|
joint_ffhq_brush_realnoise/tensorboard/events.out.tfevents.1679284370.admin.cluster.local.35162.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6ba10d8f4626a47316c7d9f6174a6335bb323bbb5add80852704d3b7589d7b6
|
3 |
+
size 74423266
|
joint_ffhq_maskdir_realnoise/ckpt/step149999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
|
3 |
+
size 425
|
joint_ffhq_maskdir_realnoise/ckpt/step149999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:70bbc5fd9861e9f8898e85a15932d529b39c3e72d9dd85671754d09250ff1063
|
3 |
+
size 53562757
|
joint_ffhq_maskdir_realnoise/ckpt/step149999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f4bc75ed20245fd2b74310ee4f6d2f5a66e52a25cd348184ace9d1a8215b8155
|
3 |
+
size 106908613
|
joint_ffhq_maskdir_realnoise/config-2023-03-20-11-55-31.yaml
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
data:
|
2 |
+
dataroot: /amax/xyf/data/FFHQ/
|
3 |
+
img_size: 256
|
4 |
+
name: FFHQ
|
5 |
+
dataloader:
|
6 |
+
micro_batch: 0
|
7 |
+
num_workers: 4
|
8 |
+
pin_memory: true
|
9 |
+
prefetch_factor: 2
|
10 |
+
mask:
|
11 |
+
dir_invert_color: true
|
12 |
+
dir_path: /amax/xyf/data/NVIDIAIrregularMaskDataset/train/
|
13 |
+
mask_type: dir
|
14 |
+
noise_type: real
|
15 |
+
real_dataset:
|
16 |
+
- dataroot: /amax/xyf/data/CelebA-HQ/
|
17 |
+
img_size: 256
|
18 |
+
name: CelebA-HQ
|
19 |
+
- dataroot: /amax/xyf/data/ImageNet/
|
20 |
+
img_size: 256
|
21 |
+
name: ImageNet
|
22 |
+
smooth_kernel_size: 13
|
23 |
+
smooth_sigma: 3
|
24 |
+
smooth_type: gaussian
|
25 |
+
model:
|
26 |
+
mpn:
|
27 |
+
base_n_channels: 64
|
28 |
+
neck_n_channels: 128
|
29 |
+
rin:
|
30 |
+
base_n_channels: 32
|
31 |
+
neck_n_channels: 128
|
32 |
+
seed: 1234
|
33 |
+
train:
|
34 |
+
batch_size: 4
|
35 |
+
coef_adv: 0.001
|
36 |
+
coef_idmrf: 0.001
|
37 |
+
coef_mpn: 2.0
|
38 |
+
coef_rec: 1.4
|
39 |
+
coef_semantic: 0.0001
|
40 |
+
eval_freq: 1000
|
41 |
+
n_steps: 150000
|
42 |
+
optim_disc:
|
43 |
+
betas:
|
44 |
+
- 0.5
|
45 |
+
- 0.9
|
46 |
+
lr: 1.0e-05
|
47 |
+
optim_mpn:
|
48 |
+
betas:
|
49 |
+
- 0.5
|
50 |
+
- 0.9
|
51 |
+
lr: 1.0e-05
|
52 |
+
optim_rin:
|
53 |
+
betas:
|
54 |
+
- 0.5
|
55 |
+
- 0.9
|
56 |
+
lr: 1.0e-05
|
57 |
+
pretrained: ./runs/separate_ffhq_maskdir_realnoise/ckpt/step079999/model.pt
|
58 |
+
print_freq: 100
|
59 |
+
resume: null
|
60 |
+
sample_freq: 1000
|
61 |
+
save_freq: 5000
|
joint_ffhq_maskdir_realnoise/output-2023-03-20-11-55-31.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
joint_ffhq_maskdir_realnoise/samples/step149999.png
ADDED
![]() |
Git LFS Details
|
joint_ffhq_maskdir_realnoise/tensorboard/events.out.tfevents.1679284533.admin.cluster.local.727.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ada1bcfab3f9ee5ff8c747daa086e98cdc3d3a368d73e703a3a0ff7a1c3c6b50
|
3 |
+
size 74423266
|
joint_imagenet_brush_realnoise/ckpt/step149999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
|
3 |
+
size 425
|
joint_imagenet_brush_realnoise/ckpt/step149999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:248ca4c17d6ef0894d98d0bbf7556c025362489ba27800b2b175e45707e4b604
|
3 |
+
size 53562757
|
joint_imagenet_brush_realnoise/ckpt/step149999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09d31cf2963d5ace5747cc410bbfac3c66b5263fd75ee3850f7d332f2ba1343e
|
3 |
+
size 106908613
|
joint_imagenet_brush_realnoise/config-2023-03-22-15-20-53.yaml
ADDED
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
data:
|
2 |
+
dataroot: /amax/xyf/data/ImageNet/
|
3 |
+
img_size: 256
|
4 |
+
name: ImageNet
|
5 |
+
dataloader:
|
6 |
+
micro_batch: 0
|
7 |
+
num_workers: 4
|
8 |
+
pin_memory: true
|
9 |
+
prefetch_factor: 2
|
10 |
+
mask:
|
11 |
+
mask_type: brush
|
12 |
+
noise_type: real
|
13 |
+
real_dataset:
|
14 |
+
- dataroot: /amax/xyf/data/Places365/
|
15 |
+
img_size: 256
|
16 |
+
name: Places365
|
17 |
+
smooth_type: iterative_gaussian
|
18 |
+
model:
|
19 |
+
mpn:
|
20 |
+
base_n_channels: 64
|
21 |
+
neck_n_channels: 128
|
22 |
+
rin:
|
23 |
+
base_n_channels: 32
|
24 |
+
neck_n_channels: 128
|
25 |
+
seed: 1234
|
26 |
+
train:
|
27 |
+
batch_size: 4
|
28 |
+
coef_adv: 0.001
|
29 |
+
coef_idmrf: 0.001
|
30 |
+
coef_mpn: 2.0
|
31 |
+
coef_rec: 1.4
|
32 |
+
coef_semantic: 0.0001
|
33 |
+
eval_freq: 1000
|
34 |
+
n_steps: 150000
|
35 |
+
optim_disc:
|
36 |
+
betas:
|
37 |
+
- 0.5
|
38 |
+
- 0.9
|
39 |
+
lr: 1.0e-05
|
40 |
+
optim_mpn:
|
41 |
+
betas:
|
42 |
+
- 0.5
|
43 |
+
- 0.9
|
44 |
+
lr: 1.0e-05
|
45 |
+
optim_rin:
|
46 |
+
betas:
|
47 |
+
- 0.5
|
48 |
+
- 0.9
|
49 |
+
lr: 1.0e-05
|
50 |
+
pretrained: ./runs/separate_imagenet_brush_realnoise/ckpt/step149999/model.pt
|
51 |
+
print_freq: 100
|
52 |
+
resume: null
|
53 |
+
sample_freq: 1000
|
54 |
+
save_freq: 5000
|
joint_imagenet_brush_realnoise/output-2023-03-22-15-20-53.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
joint_imagenet_brush_realnoise/samples/step149999.png
ADDED
![]() |
Git LFS Details
|
joint_imagenet_brush_realnoise/tensorboard/events.out.tfevents.1679469675.admin.cluster.local.8096.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfbcea7719488dd19a69e053cbd84bd0e8bcc89471b67219ebf9fd2541691c98
|
3 |
+
size 74430750
|
joint_imagenet_maskdir_realnoise/ckpt/step149999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
|
3 |
+
size 425
|
joint_imagenet_maskdir_realnoise/ckpt/step149999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02f470d4182208054de483d2729e56f03904a2e64526f9fb61362dc7a69add86
|
3 |
+
size 53562757
|