xyfJASON commited on
Commit
03e1438
·
verified ·
1 Parent(s): 423b14d

Upload checkpoints and training logs

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +24 -0
  2. downstream-shadow-removal/ckpt/step039999.pt +3 -0
  3. downstream-shadow-removal/config-2023-02-12-13-23-18.yaml +70 -0
  4. downstream-shadow-removal/config-2023-02-13-02-44-14.yaml +70 -0
  5. downstream-shadow-removal/output-2023-02-12-13-23-18.log +224 -0
  6. downstream-shadow-removal/output-2023-02-13-02-44-14.log +680 -0
  7. downstream-shadow-removal/samples/step000999.png +3 -0
  8. downstream-shadow-removal/samples/step001999.png +3 -0
  9. downstream-shadow-removal/samples/step002999.png +3 -0
  10. downstream-shadow-removal/samples/step003999.png +3 -0
  11. downstream-shadow-removal/samples/step004999.png +3 -0
  12. downstream-shadow-removal/samples/step005999.png +3 -0
  13. downstream-shadow-removal/samples/step006999.png +3 -0
  14. downstream-shadow-removal/samples/step007999.png +3 -0
  15. downstream-shadow-removal/samples/step008999.png +3 -0
  16. downstream-shadow-removal/samples/step009999.png +3 -0
  17. downstream-shadow-removal/tensorboard/events.out.tfevents.1676208201.jason-system.49917.0 +3 -0
  18. downstream-shadow-removal/tensorboard/events.out.tfevents.1676256254.jason-system.54160.0 +3 -0
  19. downstream-watermark-removal/ckpt/step099999.pt +3 -0
  20. downstream-watermark-removal/config-2023-02-12-14-28-38.yaml +70 -0
  21. downstream-watermark-removal/config-2023-02-12-15-41-28.yaml +70 -0
  22. downstream-watermark-removal/output-2023-02-12-14-28-38.log +223 -0
  23. downstream-watermark-removal/output-2023-02-12-15-41-28.log +0 -0
  24. downstream-watermark-removal/samples/step009999.png +3 -0
  25. downstream-watermark-removal/samples/step099999.png +3 -0
  26. downstream-watermark-removal/tensorboard/events.out.tfevents.1676212121.jason-system.50864.0 +3 -0
  27. downstream-watermark-removal/tensorboard/events.out.tfevents.1676216490.jason-system.51652.0 +3 -0
  28. joint_ffhq_brush_realnoise/ckpt/step149999/meta.pt +3 -0
  29. joint_ffhq_brush_realnoise/ckpt/step149999/model.pt +3 -0
  30. joint_ffhq_brush_realnoise/ckpt/step149999/optimizer.pt +3 -0
  31. joint_ffhq_brush_realnoise/config-2023-03-20-11-52-44.yaml +57 -0
  32. joint_ffhq_brush_realnoise/output-2023-03-20-11-52-44.log +0 -0
  33. joint_ffhq_brush_realnoise/samples/step149999.png +3 -0
  34. joint_ffhq_brush_realnoise/tensorboard/events.out.tfevents.1679284370.admin.cluster.local.35162.0 +3 -0
  35. joint_ffhq_maskdir_realnoise/ckpt/step149999/meta.pt +3 -0
  36. joint_ffhq_maskdir_realnoise/ckpt/step149999/model.pt +3 -0
  37. joint_ffhq_maskdir_realnoise/ckpt/step149999/optimizer.pt +3 -0
  38. joint_ffhq_maskdir_realnoise/config-2023-03-20-11-55-31.yaml +61 -0
  39. joint_ffhq_maskdir_realnoise/output-2023-03-20-11-55-31.log +0 -0
  40. joint_ffhq_maskdir_realnoise/samples/step149999.png +3 -0
  41. joint_ffhq_maskdir_realnoise/tensorboard/events.out.tfevents.1679284533.admin.cluster.local.727.0 +3 -0
  42. joint_imagenet_brush_realnoise/ckpt/step149999/meta.pt +3 -0
  43. joint_imagenet_brush_realnoise/ckpt/step149999/model.pt +3 -0
  44. joint_imagenet_brush_realnoise/ckpt/step149999/optimizer.pt +3 -0
  45. joint_imagenet_brush_realnoise/config-2023-03-22-15-20-53.yaml +54 -0
  46. joint_imagenet_brush_realnoise/output-2023-03-22-15-20-53.log +0 -0
  47. joint_imagenet_brush_realnoise/samples/step149999.png +3 -0
  48. joint_imagenet_brush_realnoise/tensorboard/events.out.tfevents.1679469675.admin.cluster.local.8096.0 +3 -0
  49. joint_imagenet_maskdir_realnoise/ckpt/step149999/meta.pt +3 -0
  50. joint_imagenet_maskdir_realnoise/ckpt/step149999/model.pt +3 -0
.gitattributes CHANGED
@@ -33,3 +33,27 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ downstream-shadow-removal/samples/step000999.png filter=lfs diff=lfs merge=lfs -text
37
+ downstream-shadow-removal/samples/step001999.png filter=lfs diff=lfs merge=lfs -text
38
+ downstream-shadow-removal/samples/step002999.png filter=lfs diff=lfs merge=lfs -text
39
+ downstream-shadow-removal/samples/step003999.png filter=lfs diff=lfs merge=lfs -text
40
+ downstream-shadow-removal/samples/step004999.png filter=lfs diff=lfs merge=lfs -text
41
+ downstream-shadow-removal/samples/step005999.png filter=lfs diff=lfs merge=lfs -text
42
+ downstream-shadow-removal/samples/step006999.png filter=lfs diff=lfs merge=lfs -text
43
+ downstream-shadow-removal/samples/step007999.png filter=lfs diff=lfs merge=lfs -text
44
+ downstream-shadow-removal/samples/step008999.png filter=lfs diff=lfs merge=lfs -text
45
+ downstream-shadow-removal/samples/step009999.png filter=lfs diff=lfs merge=lfs -text
46
+ downstream-watermark-removal/samples/step009999.png filter=lfs diff=lfs merge=lfs -text
47
+ downstream-watermark-removal/samples/step099999.png filter=lfs diff=lfs merge=lfs -text
48
+ joint_ffhq_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
49
+ joint_ffhq_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
50
+ joint_imagenet_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
51
+ joint_imagenet_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
52
+ joint_places_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
53
+ joint_places_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
54
+ separate_ffhq_brush_realnoise/samples/step079999.png filter=lfs diff=lfs merge=lfs -text
55
+ separate_ffhq_maskdir_realnoise/samples/step079999.png filter=lfs diff=lfs merge=lfs -text
56
+ separate_imagenet_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
57
+ separate_imagenet_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
58
+ separate_places_brush_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
59
+ separate_places_maskdir_realnoise/samples/step149999.png filter=lfs diff=lfs merge=lfs -text
downstream-shadow-removal/ckpt/step039999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:550ed527f61244a228d83cbf3c8d3afbbb59238b61eb01aef029a3d73c4f10e6
3
+ size 160487047
downstream-shadow-removal/config-2023-02-12-13-23-18.yaml ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DATA:
2
+ DATAROOT: /data/ISTD/
3
+ IMG_SIZE: 256
4
+ NAME: ISTD
5
+ DATALOADER:
6
+ BATCH_SIZE: 4
7
+ MICRO_BATCH: 0
8
+ NUM_WORKERS: 4
9
+ PIN_MEMORY: true
10
+ PREFETCH_FACTOR: 2
11
+ EVALUATE:
12
+ N_EVAL: 10000
13
+ MASK:
14
+ BRUSH_LENGTH_RATIO:
15
+ - 0.1
16
+ - 0.25
17
+ BRUSH_NUM: &id001
18
+ - 0
19
+ - 0
20
+ BRUSH_TURNS:
21
+ - 4
22
+ - 18
23
+ BRUSH_WIDTH_RATIO:
24
+ - 0.02
25
+ - 0.1
26
+ MASK_DIR: null
27
+ NOISE_DATASETS: []
28
+ RECT_LENGTH_RATIO:
29
+ - 0.2
30
+ - 0.8
31
+ RECT_NUM: *id001
32
+ SMOOTH_ITERS: 4
33
+ SMOOTH_KERNEL_SIZE: 15
34
+ SMOOTH_SIGMA: 4
35
+ MODEL:
36
+ MPN:
37
+ BASE_N_CHANNELS: 64
38
+ NECK_N_CHANNELS: 128
39
+ RIN:
40
+ BASE_N_CHANNELS: 32
41
+ NECK_N_CHANNELS: 128
42
+ WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
43
+ SAMPLE:
44
+ N_SAMPLES: 100
45
+ RANDOM: false
46
+ SAVE_DIR: ./samples/
47
+ SEED: 1234
48
+ TRAIN:
49
+ COEF_ADV: 0.001
50
+ COEF_IDMRF: 0.001
51
+ COEF_MPN: 2.0
52
+ COEF_REC: 1.4
53
+ COEF_SEMANTIC: 0.0001
54
+ EVAL_FREQ: 1000
55
+ OPTIM_DISC:
56
+ BETAS: &id002
57
+ - 0.5
58
+ - 0.9
59
+ LR: 1.0e-05
60
+ OPTIM_MPN:
61
+ BETAS: *id002
62
+ LR: 1.0e-05
63
+ OPTIM_RIN:
64
+ BETAS: *id002
65
+ LR: 1.0e-05
66
+ PRINT_FREQ: 100
67
+ RESUME: null
68
+ SAMPLE_FREQ: 1000
69
+ SAVE_FREQ: 5000
70
+ TRAIN_STEPS: 10000
downstream-shadow-removal/config-2023-02-13-02-44-14.yaml ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DATA:
2
+ DATAROOT: /data/ISTD/
3
+ IMG_SIZE: 256
4
+ NAME: ISTD
5
+ DATALOADER:
6
+ BATCH_SIZE: 4
7
+ MICRO_BATCH: 0
8
+ NUM_WORKERS: 4
9
+ PIN_MEMORY: true
10
+ PREFETCH_FACTOR: 2
11
+ EVALUATE:
12
+ N_EVAL: 10000
13
+ MASK:
14
+ BRUSH_LENGTH_RATIO:
15
+ - 0.1
16
+ - 0.25
17
+ BRUSH_NUM: &id001
18
+ - 0
19
+ - 0
20
+ BRUSH_TURNS:
21
+ - 4
22
+ - 18
23
+ BRUSH_WIDTH_RATIO:
24
+ - 0.02
25
+ - 0.1
26
+ MASK_DIR: null
27
+ NOISE_DATASETS: []
28
+ RECT_LENGTH_RATIO:
29
+ - 0.2
30
+ - 0.8
31
+ RECT_NUM: *id001
32
+ SMOOTH_ITERS: 4
33
+ SMOOTH_KERNEL_SIZE: 15
34
+ SMOOTH_SIGMA: 4
35
+ MODEL:
36
+ MPN:
37
+ BASE_N_CHANNELS: 64
38
+ NECK_N_CHANNELS: 128
39
+ RIN:
40
+ BASE_N_CHANNELS: 32
41
+ NECK_N_CHANNELS: 128
42
+ WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
43
+ SAMPLE:
44
+ N_SAMPLES: 100
45
+ RANDOM: false
46
+ SAVE_DIR: ./samples/
47
+ SEED: 1234
48
+ TRAIN:
49
+ COEF_ADV: 0.001
50
+ COEF_IDMRF: 0.001
51
+ COEF_MPN: 2.0
52
+ COEF_REC: 1.4
53
+ COEF_SEMANTIC: 0.0001
54
+ EVAL_FREQ: 1000
55
+ OPTIM_DISC:
56
+ BETAS: &id002
57
+ - 0.5
58
+ - 0.9
59
+ LR: 1.0e-05
60
+ OPTIM_MPN:
61
+ BETAS: *id002
62
+ LR: 1.0e-05
63
+ OPTIM_RIN:
64
+ BETAS: *id002
65
+ LR: 1.0e-05
66
+ PRINT_FREQ: 100
67
+ RESUME: latest
68
+ SAMPLE_FREQ: 1000
69
+ SAVE_FREQ: 5000
70
+ TRAIN_STEPS: 100000
downstream-shadow-removal/output-2023-02-12-13-23-18.log ADDED
@@ -0,0 +1,224 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-02-12 13:23:21,129 - INFO - Experiment directory: runs/downstream-shadow-removal
2
+ 2023-02-12 13:23:21,129 - INFO - Device: cuda
3
+ 2023-02-12 13:23:21,129 - INFO - Number of devices: 1
4
+ 2023-02-12 13:23:21,135 - WARNING - Replace split `valid` with split `test`
5
+ 2023-02-12 13:23:21,137 - INFO - Size of training set: 1330
6
+ 2023-02-12 13:23:21,137 - INFO - Size of validation set: 540
7
+ 2023-02-12 13:23:21,137 - INFO - Batch size per device: 4
8
+ 2023-02-12 13:23:21,137 - INFO - Effective batch size: 4
9
+ 2023-02-12 13:23:22,639 - INFO - Successfully load mpn from ./runs/places-joint/ckpt/step149999.pt
10
+ 2023-02-12 13:23:22,643 - INFO - Successfully load rin from ./runs/places-joint/ckpt/step149999.pt
11
+ 2023-02-12 13:23:22,645 - INFO - Successfully load disc from ./runs/places-joint/ckpt/step149999.pt
12
+ 2023-02-12 13:23:22,647 - INFO - Successfully load pdisc from ./runs/places-joint/ckpt/step149999.pt
13
+ 2023-02-12 13:23:22,650 - INFO - Start training...
14
+ 2023-02-12 13:23:54,134 - INFO - [Train] step: 99, loss_adv_disc: -1.281875
15
+ 2023-02-12 13:23:54,342 - INFO - [Train] step: 99, loss_mpn: 0.162302, loss_rec: 0.088649, loss_semantic: 0.554951, loss_idmrf: 7.990071, loss_adv_gen: -119.046860
16
+ 2023-02-12 13:24:23,357 - INFO - [Train] step: 199, loss_adv_disc: -4.481179
17
+ 2023-02-12 13:24:23,565 - INFO - [Train] step: 199, loss_mpn: 0.074521, loss_rec: 0.074638, loss_semantic: 0.563851, loss_idmrf: 7.788015, loss_adv_gen: -152.929199
18
+ 2023-02-12 13:24:52,586 - INFO - [Train] step: 299, loss_adv_disc: 10.212236
19
+ 2023-02-12 13:24:52,793 - INFO - [Train] step: 299, loss_mpn: 0.058758, loss_rec: 0.090663, loss_semantic: 0.657425, loss_idmrf: 8.309335, loss_adv_gen: -154.783676
20
+ 2023-02-12 13:25:21,980 - INFO - [Train] step: 399, loss_adv_disc: -1.535188
21
+ 2023-02-12 13:25:22,187 - INFO - [Train] step: 399, loss_mpn: 0.035261, loss_rec: 0.079162, loss_semantic: 0.496730, loss_idmrf: 4.807649, loss_adv_gen: -116.015060
22
+ 2023-02-12 13:25:51,186 - INFO - [Train] step: 499, loss_adv_disc: -12.597924
23
+ 2023-02-12 13:25:51,393 - INFO - [Train] step: 499, loss_mpn: 0.090085, loss_rec: 0.132387, loss_semantic: 0.493011, loss_idmrf: 6.727902, loss_adv_gen: -152.829514
24
+ 2023-02-12 13:26:20,392 - INFO - [Train] step: 599, loss_adv_disc: -1.783383
25
+ 2023-02-12 13:26:20,599 - INFO - [Train] step: 599, loss_mpn: 0.072984, loss_rec: 0.063185, loss_semantic: 0.514825, loss_idmrf: 5.854929, loss_adv_gen: -98.764038
26
+ 2023-02-12 13:26:49,784 - INFO - [Train] step: 699, loss_adv_disc: -0.208335
27
+ 2023-02-12 13:26:49,991 - INFO - [Train] step: 699, loss_mpn: 0.039042, loss_rec: 0.099020, loss_semantic: 0.674136, loss_idmrf: 6.369217, loss_adv_gen: -154.383133
28
+ 2023-02-12 13:27:18,971 - INFO - [Train] step: 799, loss_adv_disc: 0.360612
29
+ 2023-02-12 13:27:19,179 - INFO - [Train] step: 799, loss_mpn: 0.025015, loss_rec: 0.058072, loss_semantic: 0.417860, loss_idmrf: 3.865017, loss_adv_gen: -109.052032
30
+ 2023-02-12 13:27:48,162 - INFO - [Train] step: 899, loss_adv_disc: -1.592671
31
+ 2023-02-12 13:27:48,370 - INFO - [Train] step: 899, loss_mpn: 0.020442, loss_rec: 0.082523, loss_semantic: 0.448917, loss_idmrf: 5.112060, loss_adv_gen: -113.044441
32
+ 2023-02-12 13:28:17,555 - INFO - [Train] step: 999, loss_adv_disc: -3.951293
33
+ 2023-02-12 13:28:17,763 - INFO - [Train] step: 999, loss_mpn: 0.089411, loss_rec: 0.061914, loss_semantic: 0.476440, loss_idmrf: 5.286564, loss_adv_gen: -118.216484
34
+ 2023-02-12 13:28:22,927 - INFO - [Eval] step: 999, bce: 0.387083, psnr: 22.987576, ssim: 0.910526
35
+ 2023-02-12 13:28:53,546 - INFO - [Train] step: 1099, loss_adv_disc: -1.074106
36
+ 2023-02-12 13:28:53,753 - INFO - [Train] step: 1099, loss_mpn: 0.014869, loss_rec: 0.055306, loss_semantic: 0.405501, loss_idmrf: 3.033794, loss_adv_gen: -105.430122
37
+ 2023-02-12 13:29:22,756 - INFO - [Train] step: 1199, loss_adv_disc: -9.270275
38
+ 2023-02-12 13:29:22,963 - INFO - [Train] step: 1199, loss_mpn: 0.025306, loss_rec: 0.071859, loss_semantic: 0.442767, loss_idmrf: 8.446411, loss_adv_gen: -92.132133
39
+ 2023-02-12 13:29:51,972 - INFO - [Train] step: 1299, loss_adv_disc: -14.562277
40
+ 2023-02-12 13:29:52,179 - INFO - [Train] step: 1299, loss_mpn: 0.049613, loss_rec: 0.083995, loss_semantic: 0.550962, loss_idmrf: 12.367452, loss_adv_gen: -142.591858
41
+ 2023-02-12 13:30:21,389 - INFO - [Train] step: 1399, loss_adv_disc: 1.822066
42
+ 2023-02-12 13:30:21,596 - INFO - [Train] step: 1399, loss_mpn: 0.029735, loss_rec: 0.060577, loss_semantic: 0.492058, loss_idmrf: 9.837570, loss_adv_gen: -199.041153
43
+ 2023-02-12 13:30:50,607 - INFO - [Train] step: 1499, loss_adv_disc: -6.002779
44
+ 2023-02-12 13:30:50,814 - INFO - [Train] step: 1499, loss_mpn: 0.030221, loss_rec: 0.072016, loss_semantic: 0.597045, loss_idmrf: 4.388852, loss_adv_gen: -157.418549
45
+ 2023-02-12 13:31:19,828 - INFO - [Train] step: 1599, loss_adv_disc: -7.131116
46
+ 2023-02-12 13:31:20,035 - INFO - [Train] step: 1599, loss_mpn: 0.026604, loss_rec: 0.073666, loss_semantic: 0.473443, loss_idmrf: 6.644470, loss_adv_gen: -159.523529
47
+ 2023-02-12 13:31:49,247 - INFO - [Train] step: 1699, loss_adv_disc: 6.895418
48
+ 2023-02-12 13:31:49,455 - INFO - [Train] step: 1699, loss_mpn: 0.018450, loss_rec: 0.074060, loss_semantic: 0.419208, loss_idmrf: 5.020046, loss_adv_gen: -115.342003
49
+ 2023-02-12 13:32:18,483 - INFO - [Train] step: 1799, loss_adv_disc: -5.128633
50
+ 2023-02-12 13:32:18,690 - INFO - [Train] step: 1799, loss_mpn: 0.025845, loss_rec: 0.077207, loss_semantic: 0.518926, loss_idmrf: 7.905535, loss_adv_gen: -180.514923
51
+ 2023-02-12 13:32:47,721 - INFO - [Train] step: 1899, loss_adv_disc: -7.722089
52
+ 2023-02-12 13:32:47,928 - INFO - [Train] step: 1899, loss_mpn: 0.019188, loss_rec: 0.056849, loss_semantic: 0.458317, loss_idmrf: 2.534275, loss_adv_gen: -159.524200
53
+ 2023-02-12 13:33:17,162 - INFO - [Train] step: 1999, loss_adv_disc: -13.602573
54
+ 2023-02-12 13:33:17,369 - INFO - [Train] step: 1999, loss_mpn: 0.048309, loss_rec: 0.095544, loss_semantic: 0.487569, loss_idmrf: 6.279430, loss_adv_gen: -165.249130
55
+ 2023-02-12 13:33:22,523 - INFO - [Eval] step: 1999, bce: 0.361587, psnr: 23.048809, ssim: 0.913292
56
+ 2023-02-12 13:33:52,822 - INFO - [Train] step: 2099, loss_adv_disc: 4.722687
57
+ 2023-02-12 13:33:53,029 - INFO - [Train] step: 2099, loss_mpn: 0.020883, loss_rec: 0.064225, loss_semantic: 0.439803, loss_idmrf: 5.173006, loss_adv_gen: -127.306946
58
+ 2023-02-12 13:34:22,033 - INFO - [Train] step: 2199, loss_adv_disc: -7.286263
59
+ 2023-02-12 13:34:22,241 - INFO - [Train] step: 2199, loss_mpn: 0.025944, loss_rec: 0.060299, loss_semantic: 0.436272, loss_idmrf: 3.550529, loss_adv_gen: -123.120049
60
+ 2023-02-12 13:34:51,247 - INFO - [Train] step: 2299, loss_adv_disc: -0.899125
61
+ 2023-02-12 13:34:51,454 - INFO - [Train] step: 2299, loss_mpn: 0.015718, loss_rec: 0.044088, loss_semantic: 0.435449, loss_idmrf: 6.796015, loss_adv_gen: -150.392487
62
+ 2023-02-12 13:35:20,660 - INFO - [Train] step: 2399, loss_adv_disc: 1.281568
63
+ 2023-02-12 13:35:20,868 - INFO - [Train] step: 2399, loss_mpn: 0.036071, loss_rec: 0.046558, loss_semantic: 0.442822, loss_idmrf: 7.079724, loss_adv_gen: -145.621643
64
+ 2023-02-12 13:35:49,875 - INFO - [Train] step: 2499, loss_adv_disc: -0.852216
65
+ 2023-02-12 13:35:50,082 - INFO - [Train] step: 2499, loss_mpn: 0.025623, loss_rec: 0.052791, loss_semantic: 0.480111, loss_idmrf: 6.043885, loss_adv_gen: -119.099579
66
+ 2023-02-12 13:36:19,086 - INFO - [Train] step: 2599, loss_adv_disc: 0.180318
67
+ 2023-02-12 13:36:19,293 - INFO - [Train] step: 2599, loss_mpn: 0.016104, loss_rec: 0.048713, loss_semantic: 0.442174, loss_idmrf: 9.205061, loss_adv_gen: -186.506836
68
+ 2023-02-12 13:36:48,498 - INFO - [Train] step: 2699, loss_adv_disc: 1.399606
69
+ 2023-02-12 13:36:48,705 - INFO - [Train] step: 2699, loss_mpn: 0.009911, loss_rec: 0.038589, loss_semantic: 0.336942, loss_idmrf: 2.168548, loss_adv_gen: -112.174011
70
+ 2023-02-12 13:37:17,709 - INFO - [Train] step: 2799, loss_adv_disc: -2.826076
71
+ 2023-02-12 13:37:17,916 - INFO - [Train] step: 2799, loss_mpn: 0.019236, loss_rec: 0.056004, loss_semantic: 0.425428, loss_idmrf: 4.864495, loss_adv_gen: -128.027161
72
+ 2023-02-12 13:37:46,925 - INFO - [Train] step: 2899, loss_adv_disc: -8.600943
73
+ 2023-02-12 13:37:47,133 - INFO - [Train] step: 2899, loss_mpn: 0.035397, loss_rec: 0.070796, loss_semantic: 0.379691, loss_idmrf: 3.269684, loss_adv_gen: -134.056183
74
+ 2023-02-12 13:38:16,347 - INFO - [Train] step: 2999, loss_adv_disc: -4.140905
75
+ 2023-02-12 13:38:16,555 - INFO - [Train] step: 2999, loss_mpn: 0.015209, loss_rec: 0.055310, loss_semantic: 0.407628, loss_idmrf: 9.434046, loss_adv_gen: -126.835373
76
+ 2023-02-12 13:38:21,704 - INFO - [Eval] step: 2999, bce: 0.289332, psnr: 23.335571, ssim: 0.917456
77
+ 2023-02-12 13:38:51,972 - INFO - [Train] step: 3099, loss_adv_disc: 0.191693
78
+ 2023-02-12 13:38:52,179 - INFO - [Train] step: 3099, loss_mpn: 0.010881, loss_rec: 0.039351, loss_semantic: 0.360058, loss_idmrf: 5.650393, loss_adv_gen: -82.992958
79
+ 2023-02-12 13:39:21,195 - INFO - [Train] step: 3199, loss_adv_disc: 0.658779
80
+ 2023-02-12 13:39:21,402 - INFO - [Train] step: 3199, loss_mpn: 0.014635, loss_rec: 0.053971, loss_semantic: 0.422246, loss_idmrf: 3.753058, loss_adv_gen: -134.257843
81
+ 2023-02-12 13:39:50,413 - INFO - [Train] step: 3299, loss_adv_disc: 3.345398
82
+ 2023-02-12 13:39:50,621 - INFO - [Train] step: 3299, loss_mpn: 0.011842, loss_rec: 0.055801, loss_semantic: 0.412139, loss_idmrf: 3.589517, loss_adv_gen: -142.031662
83
+ 2023-02-12 13:40:19,828 - INFO - [Train] step: 3399, loss_adv_disc: 0.072930
84
+ 2023-02-12 13:40:20,035 - INFO - [Train] step: 3399, loss_mpn: 0.011662, loss_rec: 0.046599, loss_semantic: 0.414424, loss_idmrf: 6.025554, loss_adv_gen: -136.393219
85
+ 2023-02-12 13:40:49,051 - INFO - [Train] step: 3499, loss_adv_disc: -2.746962
86
+ 2023-02-12 13:40:49,258 - INFO - [Train] step: 3499, loss_mpn: 0.011284, loss_rec: 0.072484, loss_semantic: 0.463066, loss_idmrf: 4.362463, loss_adv_gen: -117.525566
87
+ 2023-02-12 13:41:18,282 - INFO - [Train] step: 3599, loss_adv_disc: -5.855626
88
+ 2023-02-12 13:41:18,489 - INFO - [Train] step: 3599, loss_mpn: 0.036952, loss_rec: 0.063538, loss_semantic: 0.474658, loss_idmrf: 6.167106, loss_adv_gen: -145.618546
89
+ 2023-02-12 13:41:47,702 - INFO - [Train] step: 3699, loss_adv_disc: -1.438740
90
+ 2023-02-12 13:41:47,910 - INFO - [Train] step: 3699, loss_mpn: 0.021134, loss_rec: 0.054315, loss_semantic: 0.490676, loss_idmrf: 3.553780, loss_adv_gen: -86.800690
91
+ 2023-02-12 13:42:16,919 - INFO - [Train] step: 3799, loss_adv_disc: -6.214058
92
+ 2023-02-12 13:42:17,126 - INFO - [Train] step: 3799, loss_mpn: 0.073864, loss_rec: 0.111120, loss_semantic: 0.516712, loss_idmrf: 7.677469, loss_adv_gen: -118.615211
93
+ 2023-02-12 13:42:46,136 - INFO - [Train] step: 3899, loss_adv_disc: 0.043025
94
+ 2023-02-12 13:42:46,343 - INFO - [Train] step: 3899, loss_mpn: 0.018508, loss_rec: 0.053039, loss_semantic: 0.664957, loss_idmrf: 3.672100, loss_adv_gen: -142.985031
95
+ 2023-02-12 13:43:15,559 - INFO - [Train] step: 3999, loss_adv_disc: 0.784531
96
+ 2023-02-12 13:43:15,767 - INFO - [Train] step: 3999, loss_mpn: 0.012252, loss_rec: 0.043319, loss_semantic: 0.343232, loss_idmrf: 3.316679, loss_adv_gen: -110.461723
97
+ 2023-02-12 13:43:20,880 - INFO - [Eval] step: 3999, bce: 0.354661, psnr: 22.951654, ssim: 0.917451
98
+ 2023-02-12 13:43:50,951 - INFO - [Train] step: 4099, loss_adv_disc: -0.960966
99
+ 2023-02-12 13:43:51,157 - INFO - [Train] step: 4099, loss_mpn: 0.031528, loss_rec: 0.049657, loss_semantic: 0.404885, loss_idmrf: 4.565471, loss_adv_gen: -120.507851
100
+ 2023-02-12 13:44:20,173 - INFO - [Train] step: 4199, loss_adv_disc: 2.489911
101
+ 2023-02-12 13:44:20,381 - INFO - [Train] step: 4199, loss_mpn: 0.011535, loss_rec: 0.054218, loss_semantic: 0.445787, loss_idmrf: 3.776466, loss_adv_gen: -103.148987
102
+ 2023-02-12 13:44:49,396 - INFO - [Train] step: 4299, loss_adv_disc: -9.166822
103
+ 2023-02-12 13:44:49,603 - INFO - [Train] step: 4299, loss_mpn: 0.024727, loss_rec: 0.083794, loss_semantic: 0.552959, loss_idmrf: 8.386415, loss_adv_gen: -101.256668
104
+ 2023-02-12 13:45:18,813 - INFO - [Train] step: 4399, loss_adv_disc: -9.202044
105
+ 2023-02-12 13:45:19,022 - INFO - [Train] step: 4399, loss_mpn: 0.011078, loss_rec: 0.059956, loss_semantic: 0.367525, loss_idmrf: 4.487533, loss_adv_gen: -93.162827
106
+ 2023-02-12 13:45:48,031 - INFO - [Train] step: 4499, loss_adv_disc: -2.651873
107
+ 2023-02-12 13:45:48,239 - INFO - [Train] step: 4499, loss_mpn: 0.017707, loss_rec: 0.047575, loss_semantic: 0.423964, loss_idmrf: 7.904148, loss_adv_gen: -92.877556
108
+ 2023-02-12 13:46:17,249 - INFO - [Train] step: 4599, loss_adv_disc: -1.531205
109
+ 2023-02-12 13:46:17,456 - INFO - [Train] step: 4599, loss_mpn: 0.020947, loss_rec: 0.068436, loss_semantic: 0.545828, loss_idmrf: 4.993389, loss_adv_gen: -90.154297
110
+ 2023-02-12 13:46:46,710 - INFO - [Train] step: 4699, loss_adv_disc: 4.040330
111
+ 2023-02-12 13:46:46,918 - INFO - [Train] step: 4699, loss_mpn: 0.014810, loss_rec: 0.070468, loss_semantic: 0.525938, loss_idmrf: 5.143922, loss_adv_gen: -120.824249
112
+ 2023-02-12 13:47:15,956 - INFO - [Train] step: 4799, loss_adv_disc: -1.453163
113
+ 2023-02-12 13:47:16,164 - INFO - [Train] step: 4799, loss_mpn: 0.015021, loss_rec: 0.053999, loss_semantic: 0.496558, loss_idmrf: 3.925681, loss_adv_gen: -118.246658
114
+ 2023-02-12 13:47:45,201 - INFO - [Train] step: 4899, loss_adv_disc: -0.842131
115
+ 2023-02-12 13:47:45,408 - INFO - [Train] step: 4899, loss_mpn: 0.013811, loss_rec: 0.038493, loss_semantic: 0.386875, loss_idmrf: 6.477338, loss_adv_gen: -102.733505
116
+ 2023-02-12 13:48:14,643 - INFO - [Train] step: 4999, loss_adv_disc: 1.860435
117
+ 2023-02-12 13:48:14,851 - INFO - [Train] step: 4999, loss_mpn: 0.026306, loss_rec: 0.045528, loss_semantic: 0.423802, loss_idmrf: 4.491714, loss_adv_gen: -81.664734
118
+ 2023-02-12 13:48:20,010 - INFO - [Eval] step: 4999, bce: 0.390663, psnr: 23.292404, ssim: 0.920074
119
+ 2023-02-12 13:48:50,229 - INFO - [Train] step: 5099, loss_adv_disc: 4.181933
120
+ 2023-02-12 13:48:50,436 - INFO - [Train] step: 5099, loss_mpn: 0.018531, loss_rec: 0.056876, loss_semantic: 0.372196, loss_idmrf: 5.690077, loss_adv_gen: -91.499268
121
+ 2023-02-12 13:49:19,440 - INFO - [Train] step: 5199, loss_adv_disc: 3.452145
122
+ 2023-02-12 13:49:19,648 - INFO - [Train] step: 5199, loss_mpn: 0.012540, loss_rec: 0.050392, loss_semantic: 0.367254, loss_idmrf: 5.612234, loss_adv_gen: -104.571259
123
+ 2023-02-12 13:49:48,664 - INFO - [Train] step: 5299, loss_adv_disc: 0.072335
124
+ 2023-02-12 13:49:48,871 - INFO - [Train] step: 5299, loss_mpn: 0.009281, loss_rec: 0.034018, loss_semantic: 0.373055, loss_idmrf: 9.307406, loss_adv_gen: -86.459152
125
+ 2023-02-12 13:50:18,082 - INFO - [Train] step: 5399, loss_adv_disc: -1.546246
126
+ 2023-02-12 13:50:18,289 - INFO - [Train] step: 5399, loss_mpn: 0.015896, loss_rec: 0.049707, loss_semantic: 0.560504, loss_idmrf: 4.826586, loss_adv_gen: -86.684578
127
+ 2023-02-12 13:50:47,301 - INFO - [Train] step: 5499, loss_adv_disc: -3.722799
128
+ 2023-02-12 13:50:47,508 - INFO - [Train] step: 5499, loss_mpn: 0.012618, loss_rec: 0.040985, loss_semantic: 0.321694, loss_idmrf: 3.553346, loss_adv_gen: -95.202797
129
+ 2023-02-12 13:51:16,520 - INFO - [Train] step: 5599, loss_adv_disc: -1.551634
130
+ 2023-02-12 13:51:16,727 - INFO - [Train] step: 5599, loss_mpn: 0.010603, loss_rec: 0.041937, loss_semantic: 0.368314, loss_idmrf: 6.840730, loss_adv_gen: -122.285248
131
+ 2023-02-12 13:51:45,929 - INFO - [Train] step: 5699, loss_adv_disc: 1.480345
132
+ 2023-02-12 13:51:46,137 - INFO - [Train] step: 5699, loss_mpn: 0.022207, loss_rec: 0.033643, loss_semantic: 0.418103, loss_idmrf: 3.735212, loss_adv_gen: -70.136971
133
+ 2023-02-12 13:52:15,152 - INFO - [Train] step: 5799, loss_adv_disc: -5.609029
134
+ 2023-02-12 13:52:15,358 - INFO - [Train] step: 5799, loss_mpn: 0.018745, loss_rec: 0.051710, loss_semantic: 0.431293, loss_idmrf: 6.607909, loss_adv_gen: -109.463600
135
+ 2023-02-12 13:52:44,366 - INFO - [Train] step: 5899, loss_adv_disc: 1.087741
136
+ 2023-02-12 13:52:44,574 - INFO - [Train] step: 5899, loss_mpn: 0.019051, loss_rec: 0.037814, loss_semantic: 0.404567, loss_idmrf: 2.744005, loss_adv_gen: -85.043411
137
+ 2023-02-12 13:53:13,796 - INFO - [Train] step: 5999, loss_adv_disc: 1.313997
138
+ 2023-02-12 13:53:14,004 - INFO - [Train] step: 5999, loss_mpn: 0.007717, loss_rec: 0.039593, loss_semantic: 0.355299, loss_idmrf: 3.153148, loss_adv_gen: -92.488480
139
+ 2023-02-12 13:53:19,152 - INFO - [Eval] step: 5999, bce: 0.360973, psnr: 23.440897, ssim: 0.921864
140
+ 2023-02-12 13:53:49,448 - INFO - [Train] step: 6099, loss_adv_disc: -6.527198
141
+ 2023-02-12 13:53:49,655 - INFO - [Train] step: 6099, loss_mpn: 0.012591, loss_rec: 0.060579, loss_semantic: 0.470930, loss_idmrf: 5.310204, loss_adv_gen: -102.355423
142
+ 2023-02-12 13:54:18,664 - INFO - [Train] step: 6199, loss_adv_disc: -1.580616
143
+ 2023-02-12 13:54:18,872 - INFO - [Train] step: 6199, loss_mpn: 0.014308, loss_rec: 0.048241, loss_semantic: 0.396192, loss_idmrf: 6.307591, loss_adv_gen: -101.506081
144
+ 2023-02-12 13:54:47,889 - INFO - [Train] step: 6299, loss_adv_disc: -0.838180
145
+ 2023-02-12 13:54:48,096 - INFO - [Train] step: 6299, loss_mpn: 0.014393, loss_rec: 0.046224, loss_semantic: 0.443932, loss_idmrf: 4.024370, loss_adv_gen: -59.003387
146
+ 2023-02-12 13:55:17,334 - INFO - [Train] step: 6399, loss_adv_disc: -1.536690
147
+ 2023-02-12 13:55:17,541 - INFO - [Train] step: 6399, loss_mpn: 0.017005, loss_rec: 0.041010, loss_semantic: 0.443276, loss_idmrf: 5.697750, loss_adv_gen: -82.131157
148
+ 2023-02-12 13:55:46,587 - INFO - [Train] step: 6499, loss_adv_disc: -2.835433
149
+ 2023-02-12 13:55:46,795 - INFO - [Train] step: 6499, loss_mpn: 0.008315, loss_rec: 0.038857, loss_semantic: 0.353050, loss_idmrf: 3.457996, loss_adv_gen: -78.603981
150
+ 2023-02-12 13:56:15,842 - INFO - [Train] step: 6599, loss_adv_disc: -2.492158
151
+ 2023-02-12 13:56:16,049 - INFO - [Train] step: 6599, loss_mpn: 0.008221, loss_rec: 0.046404, loss_semantic: 0.431954, loss_idmrf: 4.713380, loss_adv_gen: -86.217865
152
+ 2023-02-12 13:56:45,275 - INFO - [Train] step: 6699, loss_adv_disc: -0.614049
153
+ 2023-02-12 13:56:45,482 - INFO - [Train] step: 6699, loss_mpn: 0.010530, loss_rec: 0.046490, loss_semantic: 0.398438, loss_idmrf: 2.225234, loss_adv_gen: -66.221626
154
+ 2023-02-12 13:57:14,500 - INFO - [Train] step: 6799, loss_adv_disc: -2.154045
155
+ 2023-02-12 13:57:14,707 - INFO - [Train] step: 6799, loss_mpn: 0.014369, loss_rec: 0.050858, loss_semantic: 0.448220, loss_idmrf: 3.556105, loss_adv_gen: -105.186676
156
+ 2023-02-12 13:57:43,719 - INFO - [Train] step: 6899, loss_adv_disc: -1.190847
157
+ 2023-02-12 13:57:43,927 - INFO - [Train] step: 6899, loss_mpn: 0.028931, loss_rec: 0.051065, loss_semantic: 0.450494, loss_idmrf: 6.116441, loss_adv_gen: -104.228516
158
+ 2023-02-12 13:58:13,150 - INFO - [Train] step: 6999, loss_adv_disc: -4.266356
159
+ 2023-02-12 13:58:13,357 - INFO - [Train] step: 6999, loss_mpn: 0.013678, loss_rec: 0.041581, loss_semantic: 0.424674, loss_idmrf: 4.031384, loss_adv_gen: -84.939545
160
+ 2023-02-12 13:58:18,504 - INFO - [Eval] step: 6999, bce: 0.392871, psnr: 23.324829, ssim: 0.921624
161
+ 2023-02-12 13:58:48,580 - INFO - [Train] step: 7099, loss_adv_disc: -4.838001
162
+ 2023-02-12 13:58:48,788 - INFO - [Train] step: 7099, loss_mpn: 0.006589, loss_rec: 0.039819, loss_semantic: 0.366608, loss_idmrf: 3.555822, loss_adv_gen: -76.142303
163
+ 2023-02-12 13:59:17,801 - INFO - [Train] step: 7199, loss_adv_disc: -1.795486
164
+ 2023-02-12 13:59:18,008 - INFO - [Train] step: 7199, loss_mpn: 0.007781, loss_rec: 0.037948, loss_semantic: 0.380204, loss_idmrf: 3.421008, loss_adv_gen: -98.430969
165
+ 2023-02-12 13:59:47,024 - INFO - [Train] step: 7299, loss_adv_disc: -5.008701
166
+ 2023-02-12 13:59:47,231 - INFO - [Train] step: 7299, loss_mpn: 0.006212, loss_rec: 0.050326, loss_semantic: 0.343246, loss_idmrf: 3.782480, loss_adv_gen: -89.840332
167
+ 2023-02-12 14:00:16,439 - INFO - [Train] step: 7399, loss_adv_disc: 0.952745
168
+ 2023-02-12 14:00:16,647 - INFO - [Train] step: 7399, loss_mpn: 0.011856, loss_rec: 0.037551, loss_semantic: 0.361128, loss_idmrf: 4.399923, loss_adv_gen: -75.590004
169
+ 2023-02-12 14:00:45,662 - INFO - [Train] step: 7499, loss_adv_disc: -0.309092
170
+ 2023-02-12 14:00:45,869 - INFO - [Train] step: 7499, loss_mpn: 0.032277, loss_rec: 0.040263, loss_semantic: 0.608975, loss_idmrf: 3.067678, loss_adv_gen: -71.062065
171
+ 2023-02-12 14:01:14,880 - INFO - [Train] step: 7599, loss_adv_disc: -3.749333
172
+ 2023-02-12 14:01:15,087 - INFO - [Train] step: 7599, loss_mpn: 0.025242, loss_rec: 0.047163, loss_semantic: 0.448062, loss_idmrf: 3.615059, loss_adv_gen: -102.538254
173
+ 2023-02-12 14:01:44,291 - INFO - [Train] step: 7699, loss_adv_disc: -1.246231
174
+ 2023-02-12 14:01:44,499 - INFO - [Train] step: 7699, loss_mpn: 0.011135, loss_rec: 0.053402, loss_semantic: 0.458806, loss_idmrf: 2.834428, loss_adv_gen: -89.317276
175
+ 2023-02-12 14:02:13,491 - INFO - [Train] step: 7799, loss_adv_disc: -0.799770
176
+ 2023-02-12 14:02:13,698 - INFO - [Train] step: 7799, loss_mpn: 0.024141, loss_rec: 0.038572, loss_semantic: 0.450097, loss_idmrf: 4.533717, loss_adv_gen: -91.607513
177
+ 2023-02-12 14:02:42,691 - INFO - [Train] step: 7899, loss_adv_disc: -1.562794
178
+ 2023-02-12 14:02:42,898 - INFO - [Train] step: 7899, loss_mpn: 0.014830, loss_rec: 0.039014, loss_semantic: 0.443213, loss_idmrf: 7.125534, loss_adv_gen: -99.803688
179
+ 2023-02-12 14:03:12,101 - INFO - [Train] step: 7999, loss_adv_disc: -1.292525
180
+ 2023-02-12 14:03:12,308 - INFO - [Train] step: 7999, loss_mpn: 0.009162, loss_rec: 0.035810, loss_semantic: 0.339180, loss_idmrf: 4.937554, loss_adv_gen: -66.376938
181
+ 2023-02-12 14:03:17,463 - INFO - [Eval] step: 7999, bce: 0.410713, psnr: 23.380175, ssim: 0.922875
182
+ 2023-02-12 14:03:47,540 - INFO - [Train] step: 8099, loss_adv_disc: -1.598471
183
+ 2023-02-12 14:03:47,747 - INFO - [Train] step: 8099, loss_mpn: 0.008017, loss_rec: 0.038835, loss_semantic: 0.342724, loss_idmrf: 3.858693, loss_adv_gen: -72.846497
184
+ 2023-02-12 14:04:16,766 - INFO - [Train] step: 8199, loss_adv_disc: -1.526141
185
+ 2023-02-12 14:04:16,974 - INFO - [Train] step: 8199, loss_mpn: 0.007769, loss_rec: 0.048700, loss_semantic: 0.370412, loss_idmrf: 4.246994, loss_adv_gen: -87.642441
186
+ 2023-02-12 14:04:45,993 - INFO - [Train] step: 8299, loss_adv_disc: -4.967025
187
+ 2023-02-12 14:04:46,200 - INFO - [Train] step: 8299, loss_mpn: 0.017493, loss_rec: 0.051434, loss_semantic: 0.451939, loss_idmrf: 4.851607, loss_adv_gen: -92.471909
188
+ 2023-02-12 14:05:15,432 - INFO - [Train] step: 8399, loss_adv_disc: -0.423972
189
+ 2023-02-12 14:05:15,639 - INFO - [Train] step: 8399, loss_mpn: 0.009851, loss_rec: 0.035539, loss_semantic: 0.387267, loss_idmrf: 4.063365, loss_adv_gen: -88.120407
190
+ 2023-02-12 14:05:44,657 - INFO - [Train] step: 8499, loss_adv_disc: 0.920736
191
+ 2023-02-12 14:05:44,864 - INFO - [Train] step: 8499, loss_mpn: 0.008419, loss_rec: 0.030417, loss_semantic: 0.338675, loss_idmrf: 3.181174, loss_adv_gen: -59.173306
192
+ 2023-02-12 14:06:13,877 - INFO - [Train] step: 8599, loss_adv_disc: -0.524680
193
+ 2023-02-12 14:06:14,084 - INFO - [Train] step: 8599, loss_mpn: 0.009725, loss_rec: 0.048215, loss_semantic: 0.435598, loss_idmrf: 2.502195, loss_adv_gen: -71.777512
194
+ 2023-02-12 14:06:43,310 - INFO - [Train] step: 8699, loss_adv_disc: -1.867347
195
+ 2023-02-12 14:06:43,517 - INFO - [Train] step: 8699, loss_mpn: 0.005162, loss_rec: 0.028492, loss_semantic: 0.312991, loss_idmrf: 4.448667, loss_adv_gen: -58.639214
196
+ 2023-02-12 14:07:12,536 - INFO - [Train] step: 8799, loss_adv_disc: 0.215890
197
+ 2023-02-12 14:07:12,743 - INFO - [Train] step: 8799, loss_mpn: 0.007085, loss_rec: 0.027526, loss_semantic: 0.321581, loss_idmrf: 2.571146, loss_adv_gen: -79.483131
198
+ 2023-02-12 14:07:41,757 - INFO - [Train] step: 8899, loss_adv_disc: -5.541975
199
+ 2023-02-12 14:07:41,964 - INFO - [Train] step: 8899, loss_mpn: 0.014692, loss_rec: 0.058386, loss_semantic: 0.391197, loss_idmrf: 6.332783, loss_adv_gen: -77.094063
200
+ 2023-02-12 14:08:11,177 - INFO - [Train] step: 8999, loss_adv_disc: -4.298098
201
+ 2023-02-12 14:08:11,385 - INFO - [Train] step: 8999, loss_mpn: 0.009263, loss_rec: 0.041659, loss_semantic: 0.368018, loss_idmrf: 3.096043, loss_adv_gen: -94.624344
202
+ 2023-02-12 14:08:16,542 - INFO - [Eval] step: 8999, bce: 0.350487, psnr: 23.695768, ssim: 0.925011
203
+ 2023-02-12 14:08:46,831 - INFO - [Train] step: 9099, loss_adv_disc: -0.096787
204
+ 2023-02-12 14:08:47,038 - INFO - [Train] step: 9099, loss_mpn: 0.007456, loss_rec: 0.050012, loss_semantic: 0.391861, loss_idmrf: 4.834039, loss_adv_gen: -34.803146
205
+ 2023-02-12 14:09:16,052 - INFO - [Train] step: 9199, loss_adv_disc: -2.421654
206
+ 2023-02-12 14:09:16,260 - INFO - [Train] step: 9199, loss_mpn: 0.010739, loss_rec: 0.043619, loss_semantic: 0.375820, loss_idmrf: 2.879037, loss_adv_gen: -95.356018
207
+ 2023-02-12 14:09:45,468 - INFO - [Train] step: 9299, loss_adv_disc: -0.639228
208
+ 2023-02-12 14:09:45,675 - INFO - [Train] step: 9299, loss_mpn: 0.009492, loss_rec: 0.041271, loss_semantic: 0.385236, loss_idmrf: 3.942233, loss_adv_gen: -81.359604
209
+ 2023-02-12 14:10:14,686 - INFO - [Train] step: 9399, loss_adv_disc: -0.824157
210
+ 2023-02-12 14:10:14,893 - INFO - [Train] step: 9399, loss_mpn: 0.007119, loss_rec: 0.035177, loss_semantic: 0.354077, loss_idmrf: 4.972061, loss_adv_gen: -74.898308
211
+ 2023-02-12 14:10:43,902 - INFO - [Train] step: 9499, loss_adv_disc: -2.232660
212
+ 2023-02-12 14:10:44,110 - INFO - [Train] step: 9499, loss_mpn: 0.014257, loss_rec: 0.035584, loss_semantic: 0.429310, loss_idmrf: 5.182566, loss_adv_gen: -64.690308
213
+ 2023-02-12 14:11:13,120 - INFO - [Train] step: 9599, loss_adv_disc: -1.639446
214
+ 2023-02-12 14:11:13,327 - INFO - [Train] step: 9599, loss_mpn: 0.019186, loss_rec: 0.041353, loss_semantic: 0.504692, loss_idmrf: 3.800310, loss_adv_gen: -68.249176
215
+ 2023-02-12 14:11:42,542 - INFO - [Train] step: 9699, loss_adv_disc: -3.789857
216
+ 2023-02-12 14:11:42,750 - INFO - [Train] step: 9699, loss_mpn: 0.021227, loss_rec: 0.046244, loss_semantic: 0.464764, loss_idmrf: 4.286100, loss_adv_gen: -80.508911
217
+ 2023-02-12 14:12:11,774 - INFO - [Train] step: 9799, loss_adv_disc: -1.276845
218
+ 2023-02-12 14:12:11,981 - INFO - [Train] step: 9799, loss_mpn: 0.008188, loss_rec: 0.036653, loss_semantic: 0.358453, loss_idmrf: 2.212314, loss_adv_gen: -55.353634
219
+ 2023-02-12 14:12:41,001 - INFO - [Train] step: 9899, loss_adv_disc: -1.013932
220
+ 2023-02-12 14:12:41,209 - INFO - [Train] step: 9899, loss_mpn: 0.008457, loss_rec: 0.037062, loss_semantic: 0.404350, loss_idmrf: 2.998356, loss_adv_gen: -77.622223
221
+ 2023-02-12 14:13:10,420 - INFO - [Train] step: 9999, loss_adv_disc: -2.420930
222
+ 2023-02-12 14:13:10,628 - INFO - [Train] step: 9999, loss_mpn: 0.009475, loss_rec: 0.047726, loss_semantic: 0.396993, loss_idmrf: 5.336214, loss_adv_gen: -73.804047
223
+ 2023-02-12 14:13:15,773 - INFO - [Eval] step: 9999, bce: 0.441264, psnr: 23.645063, ssim: 0.924997
224
+ 2023-02-12 14:13:16,954 - INFO - End of training
downstream-shadow-removal/output-2023-02-13-02-44-14.log ADDED
@@ -0,0 +1,680 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-02-13 02:44:14,898 - INFO - Experiment directory: runs/downstream-shadow-removal
2
+ 2023-02-13 02:44:14,898 - INFO - Device: cuda
3
+ 2023-02-13 02:44:14,898 - INFO - Number of devices: 1
4
+ 2023-02-13 02:44:14,904 - WARNING - Replace split `valid` with split `test`
5
+ 2023-02-13 02:44:14,907 - INFO - Size of training set: 1330
6
+ 2023-02-13 02:44:14,907 - INFO - Size of validation set: 540
7
+ 2023-02-13 02:44:14,907 - INFO - Batch size per device: 4
8
+ 2023-02-13 02:44:14,907 - INFO - Effective batch size: 4
9
+ 2023-02-13 02:44:16,374 - INFO - Successfully load mpn from ./runs/places-joint/ckpt/step149999.pt
10
+ 2023-02-13 02:44:16,378 - INFO - Successfully load rin from ./runs/places-joint/ckpt/step149999.pt
11
+ 2023-02-13 02:44:16,384 - INFO - Successfully load disc from ./runs/places-joint/ckpt/step149999.pt
12
+ 2023-02-13 02:44:16,385 - INFO - Successfully load pdisc from ./runs/places-joint/ckpt/step149999.pt
13
+ 2023-02-13 02:44:16,388 - INFO - Resume from runs/downstream-shadow-removal/ckpt/step009999.pt
14
+ 2023-02-13 02:44:16,472 - INFO - Successfully load mpn from runs/downstream-shadow-removal/ckpt/step009999.pt
15
+ 2023-02-13 02:44:16,476 - INFO - Successfully load rin from runs/downstream-shadow-removal/ckpt/step009999.pt
16
+ 2023-02-13 02:44:16,478 - INFO - Successfully load disc from runs/downstream-shadow-removal/ckpt/step009999.pt
17
+ 2023-02-13 02:44:16,480 - INFO - Successfully load pdisc from runs/downstream-shadow-removal/ckpt/step009999.pt
18
+ 2023-02-13 02:44:16,515 - INFO - Successfully load optimizers from runs/downstream-shadow-removal/ckpt/step009999.pt
19
+ 2023-02-13 02:44:16,515 - INFO - Restart training at step 10000
20
+ 2023-02-13 02:44:16,515 - INFO - Best psnr so far: 23.695768356323242
21
+ 2023-02-13 02:44:16,518 - INFO - Start training...
22
+ 2023-02-13 02:44:48,197 - INFO - [Train] step: 10099, loss_adv_disc: 0.078588
23
+ 2023-02-13 02:44:48,404 - INFO - [Train] step: 10099, loss_mpn: 0.011509, loss_rec: 0.039201, loss_semantic: 0.377556, loss_idmrf: 3.615375, loss_adv_gen: -48.026169
24
+ 2023-02-13 02:45:17,488 - INFO - [Train] step: 10199, loss_adv_disc: -1.947309
25
+ 2023-02-13 02:45:17,697 - INFO - [Train] step: 10199, loss_mpn: 0.012715, loss_rec: 0.030864, loss_semantic: 0.360930, loss_idmrf: 2.600466, loss_adv_gen: -85.842705
26
+ 2023-02-13 02:45:46,859 - INFO - [Train] step: 10299, loss_adv_disc: -0.551243
27
+ 2023-02-13 02:45:47,067 - INFO - [Train] step: 10299, loss_mpn: 0.014206, loss_rec: 0.043966, loss_semantic: 0.523468, loss_idmrf: 5.591407, loss_adv_gen: -92.931442
28
+ 2023-02-13 02:46:16,444 - INFO - [Train] step: 10399, loss_adv_disc: -4.300162
29
+ 2023-02-13 02:46:16,652 - INFO - [Train] step: 10399, loss_mpn: 0.008325, loss_rec: 0.037018, loss_semantic: 0.390521, loss_idmrf: 2.957632, loss_adv_gen: -63.995998
30
+ 2023-02-13 02:46:45,872 - INFO - [Train] step: 10499, loss_adv_disc: -0.984805
31
+ 2023-02-13 02:46:46,081 - INFO - [Train] step: 10499, loss_mpn: 0.025511, loss_rec: 0.042500, loss_semantic: 0.398691, loss_idmrf: 4.002370, loss_adv_gen: -72.055634
32
+ 2023-02-13 02:47:15,306 - INFO - [Train] step: 10599, loss_adv_disc: -0.168864
33
+ 2023-02-13 02:47:15,514 - INFO - [Train] step: 10599, loss_mpn: 0.010539, loss_rec: 0.032968, loss_semantic: 0.400717, loss_idmrf: 3.446010, loss_adv_gen: -75.776711
34
+ 2023-02-13 02:47:44,915 - INFO - [Train] step: 10699, loss_adv_disc: -0.191629
35
+ 2023-02-13 02:47:45,123 - INFO - [Train] step: 10699, loss_mpn: 0.010557, loss_rec: 0.038703, loss_semantic: 0.471336, loss_idmrf: 2.717856, loss_adv_gen: -81.432404
36
+ 2023-02-13 02:48:14,342 - INFO - [Train] step: 10799, loss_adv_disc: 1.988065
37
+ 2023-02-13 02:48:14,551 - INFO - [Train] step: 10799, loss_mpn: 0.008534, loss_rec: 0.033567, loss_semantic: 0.337367, loss_idmrf: 2.358154, loss_adv_gen: -62.045174
38
+ 2023-02-13 02:48:43,779 - INFO - [Train] step: 10899, loss_adv_disc: -0.890945
39
+ 2023-02-13 02:48:43,988 - INFO - [Train] step: 10899, loss_mpn: 0.009466, loss_rec: 0.031909, loss_semantic: 0.363756, loss_idmrf: 2.512322, loss_adv_gen: -55.988701
40
+ 2023-02-13 02:49:13,412 - INFO - [Train] step: 10999, loss_adv_disc: -0.436340
41
+ 2023-02-13 02:49:13,620 - INFO - [Train] step: 10999, loss_mpn: 0.013038, loss_rec: 0.034377, loss_semantic: 0.406465, loss_idmrf: 3.428352, loss_adv_gen: -76.155876
42
+ 2023-02-13 02:49:18,810 - INFO - [Eval] step: 10999, bce: 0.356415, psnr: 23.687820, ssim: 0.926956
43
+ 2023-02-13 02:49:49,575 - INFO - [Train] step: 11099, loss_adv_disc: -1.143975
44
+ 2023-02-13 02:49:49,784 - INFO - [Train] step: 11099, loss_mpn: 0.006747, loss_rec: 0.029099, loss_semantic: 0.350319, loss_idmrf: 2.055681, loss_adv_gen: -41.675575
45
+ 2023-02-13 02:50:19,039 - INFO - [Train] step: 11199, loss_adv_disc: -2.175141
46
+ 2023-02-13 02:50:19,248 - INFO - [Train] step: 11199, loss_mpn: 0.014661, loss_rec: 0.030048, loss_semantic: 0.366945, loss_idmrf: 5.886166, loss_adv_gen: -57.670628
47
+ 2023-02-13 02:50:48,495 - INFO - [Train] step: 11299, loss_adv_disc: -2.454169
48
+ 2023-02-13 02:50:48,704 - INFO - [Train] step: 11299, loss_mpn: 0.011991, loss_rec: 0.038149, loss_semantic: 0.407933, loss_idmrf: 8.903237, loss_adv_gen: -93.871811
49
+ 2023-02-13 02:51:18,146 - INFO - [Train] step: 11399, loss_adv_disc: -2.127088
50
+ 2023-02-13 02:51:18,355 - INFO - [Train] step: 11399, loss_mpn: 0.012583, loss_rec: 0.028059, loss_semantic: 0.370067, loss_idmrf: 5.352232, loss_adv_gen: -75.710938
51
+ 2023-02-13 02:51:47,616 - INFO - [Train] step: 11499, loss_adv_disc: -3.866915
52
+ 2023-02-13 02:51:47,825 - INFO - [Train] step: 11499, loss_mpn: 0.010172, loss_rec: 0.039636, loss_semantic: 0.481377, loss_idmrf: 2.577618, loss_adv_gen: -50.608372
53
+ 2023-02-13 02:52:17,063 - INFO - [Train] step: 11599, loss_adv_disc: -0.510452
54
+ 2023-02-13 02:52:17,272 - INFO - [Train] step: 11599, loss_mpn: 0.012594, loss_rec: 0.036184, loss_semantic: 0.380372, loss_idmrf: 4.289717, loss_adv_gen: -74.201294
55
+ 2023-02-13 02:52:46,702 - INFO - [Train] step: 11699, loss_adv_disc: -3.221022
56
+ 2023-02-13 02:52:46,911 - INFO - [Train] step: 11699, loss_mpn: 0.007079, loss_rec: 0.033923, loss_semantic: 0.347901, loss_idmrf: 3.232406, loss_adv_gen: -49.549866
57
+ 2023-02-13 02:53:16,156 - INFO - [Train] step: 11799, loss_adv_disc: -2.184320
58
+ 2023-02-13 02:53:16,365 - INFO - [Train] step: 11799, loss_mpn: 0.009117, loss_rec: 0.041747, loss_semantic: 0.414465, loss_idmrf: 4.049892, loss_adv_gen: -93.858597
59
+ 2023-02-13 02:53:45,611 - INFO - [Train] step: 11899, loss_adv_disc: -0.988897
60
+ 2023-02-13 02:53:45,819 - INFO - [Train] step: 11899, loss_mpn: 0.005853, loss_rec: 0.026617, loss_semantic: 0.351113, loss_idmrf: 1.215564, loss_adv_gen: -80.830864
61
+ 2023-02-13 02:54:15,241 - INFO - [Train] step: 11999, loss_adv_disc: 0.980708
62
+ 2023-02-13 02:54:15,449 - INFO - [Train] step: 11999, loss_mpn: 0.021106, loss_rec: 0.040558, loss_semantic: 0.430699, loss_idmrf: 4.337394, loss_adv_gen: -98.472916
63
+ 2023-02-13 02:54:20,591 - INFO - [Eval] step: 11999, bce: 0.448820, psnr: 23.549391, ssim: 0.926096
64
+ 2023-02-13 02:54:50,910 - INFO - [Train] step: 12099, loss_adv_disc: -0.077401
65
+ 2023-02-13 02:54:51,118 - INFO - [Train] step: 12099, loss_mpn: 0.010991, loss_rec: 0.039699, loss_semantic: 0.377598, loss_idmrf: 3.547094, loss_adv_gen: -59.509033
66
+ 2023-02-13 02:55:20,352 - INFO - [Train] step: 12199, loss_adv_disc: -1.588708
67
+ 2023-02-13 02:55:20,561 - INFO - [Train] step: 12199, loss_mpn: 0.012134, loss_rec: 0.032258, loss_semantic: 0.373746, loss_idmrf: 2.459065, loss_adv_gen: -80.642197
68
+ 2023-02-13 02:55:49,796 - INFO - [Train] step: 12299, loss_adv_disc: -0.440839
69
+ 2023-02-13 02:55:50,005 - INFO - [Train] step: 12299, loss_mpn: 0.007193, loss_rec: 0.028371, loss_semantic: 0.365545, loss_idmrf: 4.296700, loss_adv_gen: -86.054810
70
+ 2023-02-13 02:56:19,442 - INFO - [Train] step: 12399, loss_adv_disc: -1.179346
71
+ 2023-02-13 02:56:19,651 - INFO - [Train] step: 12399, loss_mpn: 0.014047, loss_rec: 0.025998, loss_semantic: 0.354992, loss_idmrf: 3.493587, loss_adv_gen: -57.490608
72
+ 2023-02-13 02:56:48,891 - INFO - [Train] step: 12499, loss_adv_disc: -1.112655
73
+ 2023-02-13 02:56:49,099 - INFO - [Train] step: 12499, loss_mpn: 0.013486, loss_rec: 0.032641, loss_semantic: 0.408824, loss_idmrf: 4.551135, loss_adv_gen: -69.951340
74
+ 2023-02-13 02:57:18,345 - INFO - [Train] step: 12599, loss_adv_disc: 0.841728
75
+ 2023-02-13 02:57:18,554 - INFO - [Train] step: 12599, loss_mpn: 0.008797, loss_rec: 0.028766, loss_semantic: 0.358703, loss_idmrf: 6.030992, loss_adv_gen: -90.953262
76
+ 2023-02-13 02:57:47,993 - INFO - [Train] step: 12699, loss_adv_disc: 0.809182
77
+ 2023-02-13 02:57:48,201 - INFO - [Train] step: 12699, loss_mpn: 0.004429, loss_rec: 0.026492, loss_semantic: 0.300694, loss_idmrf: 1.625929, loss_adv_gen: -77.897903
78
+ 2023-02-13 02:58:17,440 - INFO - [Train] step: 12799, loss_adv_disc: -0.801807
79
+ 2023-02-13 02:58:17,649 - INFO - [Train] step: 12799, loss_mpn: 0.014015, loss_rec: 0.034848, loss_semantic: 0.370866, loss_idmrf: 2.979873, loss_adv_gen: -68.310692
80
+ 2023-02-13 02:58:46,886 - INFO - [Train] step: 12899, loss_adv_disc: -5.047458
81
+ 2023-02-13 02:58:47,095 - INFO - [Train] step: 12899, loss_mpn: 0.010504, loss_rec: 0.033995, loss_semantic: 0.352904, loss_idmrf: 2.596021, loss_adv_gen: -68.085014
82
+ 2023-02-13 02:59:16,521 - INFO - [Train] step: 12999, loss_adv_disc: -6.217038
83
+ 2023-02-13 02:59:16,729 - INFO - [Train] step: 12999, loss_mpn: 0.009560, loss_rec: 0.042248, loss_semantic: 0.351271, loss_idmrf: 7.592016, loss_adv_gen: -72.404373
84
+ 2023-02-13 02:59:21,915 - INFO - [Eval] step: 12999, bce: 0.403072, psnr: 23.723335, ssim: 0.927467
85
+ 2023-02-13 02:59:52,416 - INFO - [Train] step: 13099, loss_adv_disc: -0.344683
86
+ 2023-02-13 02:59:52,625 - INFO - [Train] step: 13099, loss_mpn: 0.008536, loss_rec: 0.027696, loss_semantic: 0.320824, loss_idmrf: 4.384482, loss_adv_gen: -36.807159
87
+ 2023-02-13 03:00:21,858 - INFO - [Train] step: 13199, loss_adv_disc: -0.633883
88
+ 2023-02-13 03:00:22,067 - INFO - [Train] step: 13199, loss_mpn: 0.011513, loss_rec: 0.032611, loss_semantic: 0.386401, loss_idmrf: 3.015574, loss_adv_gen: -84.160339
89
+ 2023-02-13 03:00:51,326 - INFO - [Train] step: 13299, loss_adv_disc: -1.454661
90
+ 2023-02-13 03:00:51,535 - INFO - [Train] step: 13299, loss_mpn: 0.010252, loss_rec: 0.031422, loss_semantic: 0.363536, loss_idmrf: 2.590449, loss_adv_gen: -77.080185
91
+ 2023-02-13 03:01:20,963 - INFO - [Train] step: 13399, loss_adv_disc: -1.952087
92
+ 2023-02-13 03:01:21,172 - INFO - [Train] step: 13399, loss_mpn: 0.006985, loss_rec: 0.028874, loss_semantic: 0.363051, loss_idmrf: 4.615744, loss_adv_gen: -86.813782
93
+ 2023-02-13 03:01:50,420 - INFO - [Train] step: 13499, loss_adv_disc: -2.931163
94
+ 2023-02-13 03:01:50,629 - INFO - [Train] step: 13499, loss_mpn: 0.007509, loss_rec: 0.041209, loss_semantic: 0.391465, loss_idmrf: 2.554505, loss_adv_gen: -65.555023
95
+ 2023-02-13 03:02:19,886 - INFO - [Train] step: 13599, loss_adv_disc: -1.157901
96
+ 2023-02-13 03:02:20,095 - INFO - [Train] step: 13599, loss_mpn: 0.024059, loss_rec: 0.033903, loss_semantic: 0.395873, loss_idmrf: 3.517569, loss_adv_gen: -80.516960
97
+ 2023-02-13 03:02:49,519 - INFO - [Train] step: 13699, loss_adv_disc: -0.466357
98
+ 2023-02-13 03:02:49,728 - INFO - [Train] step: 13699, loss_mpn: 0.009441, loss_rec: 0.038954, loss_semantic: 0.422468, loss_idmrf: 2.220746, loss_adv_gen: -57.919315
99
+ 2023-02-13 03:03:18,965 - INFO - [Train] step: 13799, loss_adv_disc: -1.499573
100
+ 2023-02-13 03:03:19,174 - INFO - [Train] step: 13799, loss_mpn: 0.021366, loss_rec: 0.046169, loss_semantic: 0.447270, loss_idmrf: 5.458885, loss_adv_gen: -85.748489
101
+ 2023-02-13 03:03:48,422 - INFO - [Train] step: 13899, loss_adv_disc: -1.536951
102
+ 2023-02-13 03:03:48,630 - INFO - [Train] step: 13899, loss_mpn: 0.009894, loss_rec: 0.037270, loss_semantic: 0.543508, loss_idmrf: 1.969022, loss_adv_gen: -76.695808
103
+ 2023-02-13 03:04:18,050 - INFO - [Train] step: 13999, loss_adv_disc: -4.334954
104
+ 2023-02-13 03:04:18,258 - INFO - [Train] step: 13999, loss_mpn: 0.008074, loss_rec: 0.032596, loss_semantic: 0.314528, loss_idmrf: 2.552278, loss_adv_gen: -43.856140
105
+ 2023-02-13 03:04:23,391 - INFO - [Eval] step: 13999, bce: 0.459337, psnr: 23.695547, ssim: 0.927995
106
+ 2023-02-13 03:04:53,694 - INFO - [Train] step: 14099, loss_adv_disc: -1.992434
107
+ 2023-02-13 03:04:53,903 - INFO - [Train] step: 14099, loss_mpn: 0.022011, loss_rec: 0.037382, loss_semantic: 0.366072, loss_idmrf: 3.695474, loss_adv_gen: -59.897938
108
+ 2023-02-13 03:05:23,137 - INFO - [Train] step: 14199, loss_adv_disc: -1.008330
109
+ 2023-02-13 03:05:23,346 - INFO - [Train] step: 14199, loss_mpn: 0.006269, loss_rec: 0.036688, loss_semantic: 0.391129, loss_idmrf: 2.103096, loss_adv_gen: -51.805092
110
+ 2023-02-13 03:05:52,601 - INFO - [Train] step: 14299, loss_adv_disc: -4.214204
111
+ 2023-02-13 03:05:52,809 - INFO - [Train] step: 14299, loss_mpn: 0.014387, loss_rec: 0.059094, loss_semantic: 0.497419, loss_idmrf: 6.891026, loss_adv_gen: -68.543396
112
+ 2023-02-13 03:06:22,244 - INFO - [Train] step: 14399, loss_adv_disc: -2.768673
113
+ 2023-02-13 03:06:22,453 - INFO - [Train] step: 14399, loss_mpn: 0.006042, loss_rec: 0.028931, loss_semantic: 0.317719, loss_idmrf: 2.839618, loss_adv_gen: -49.560658
114
+ 2023-02-13 03:06:51,708 - INFO - [Train] step: 14499, loss_adv_disc: -1.587522
115
+ 2023-02-13 03:06:51,916 - INFO - [Train] step: 14499, loss_mpn: 0.013344, loss_rec: 0.031495, loss_semantic: 0.366431, loss_idmrf: 5.531614, loss_adv_gen: -43.433792
116
+ 2023-02-13 03:07:21,159 - INFO - [Train] step: 14599, loss_adv_disc: -3.142626
117
+ 2023-02-13 03:07:21,368 - INFO - [Train] step: 14599, loss_mpn: 0.015624, loss_rec: 0.041417, loss_semantic: 0.470021, loss_idmrf: 3.430955, loss_adv_gen: -44.935692
118
+ 2023-02-13 03:07:50,822 - INFO - [Train] step: 14699, loss_adv_disc: -2.612098
119
+ 2023-02-13 03:07:51,031 - INFO - [Train] step: 14699, loss_mpn: 0.011433, loss_rec: 0.045211, loss_semantic: 0.458431, loss_idmrf: 3.423684, loss_adv_gen: -78.737206
120
+ 2023-02-13 03:08:20,280 - INFO - [Train] step: 14799, loss_adv_disc: -1.572463
121
+ 2023-02-13 03:08:20,489 - INFO - [Train] step: 14799, loss_mpn: 0.010487, loss_rec: 0.042552, loss_semantic: 0.449359, loss_idmrf: 2.908170, loss_adv_gen: -67.777924
122
+ 2023-02-13 03:08:49,729 - INFO - [Train] step: 14899, loss_adv_disc: -0.717800
123
+ 2023-02-13 03:08:49,938 - INFO - [Train] step: 14899, loss_mpn: 0.010333, loss_rec: 0.027488, loss_semantic: 0.339489, loss_idmrf: 5.085710, loss_adv_gen: -69.917725
124
+ 2023-02-13 03:09:19,376 - INFO - [Train] step: 14999, loss_adv_disc: -1.933161
125
+ 2023-02-13 03:09:19,585 - INFO - [Train] step: 14999, loss_mpn: 0.018537, loss_rec: 0.033588, loss_semantic: 0.374348, loss_idmrf: 3.241080, loss_adv_gen: -35.936356
126
+ 2023-02-13 03:09:24,728 - INFO - [Eval] step: 14999, bce: 0.398218, psnr: 23.808777, ssim: 0.929202
127
+ 2023-02-13 03:09:55,338 - INFO - [Train] step: 15099, loss_adv_disc: -1.123843
128
+ 2023-02-13 03:09:55,547 - INFO - [Train] step: 15099, loss_mpn: 0.010467, loss_rec: 0.036722, loss_semantic: 0.335451, loss_idmrf: 3.921709, loss_adv_gen: -65.612755
129
+ 2023-02-13 03:10:24,781 - INFO - [Train] step: 15199, loss_adv_disc: -0.055525
130
+ 2023-02-13 03:10:24,989 - INFO - [Train] step: 15199, loss_mpn: 0.008737, loss_rec: 0.030392, loss_semantic: 0.323074, loss_idmrf: 4.016256, loss_adv_gen: -63.581787
131
+ 2023-02-13 03:10:54,249 - INFO - [Train] step: 15299, loss_adv_disc: -0.904782
132
+ 2023-02-13 03:10:54,457 - INFO - [Train] step: 15299, loss_mpn: 0.007227, loss_rec: 0.025108, loss_semantic: 0.326624, loss_idmrf: 7.887135, loss_adv_gen: -52.554401
133
+ 2023-02-13 03:11:23,871 - INFO - [Train] step: 15399, loss_adv_disc: -2.399804
134
+ 2023-02-13 03:11:24,080 - INFO - [Train] step: 15399, loss_mpn: 0.011453, loss_rec: 0.036506, loss_semantic: 0.499180, loss_idmrf: 3.387949, loss_adv_gen: -48.605957
135
+ 2023-02-13 03:11:53,310 - INFO - [Train] step: 15499, loss_adv_disc: -1.687280
136
+ 2023-02-13 03:11:53,519 - INFO - [Train] step: 15499, loss_mpn: 0.006589, loss_rec: 0.029240, loss_semantic: 0.297095, loss_idmrf: 2.843406, loss_adv_gen: -62.710510
137
+ 2023-02-13 03:12:22,762 - INFO - [Train] step: 15599, loss_adv_disc: -0.255836
138
+ 2023-02-13 03:12:22,970 - INFO - [Train] step: 15599, loss_mpn: 0.005989, loss_rec: 0.031084, loss_semantic: 0.322900, loss_idmrf: 5.402665, loss_adv_gen: -63.991249
139
+ 2023-02-13 03:12:52,388 - INFO - [Train] step: 15699, loss_adv_disc: -0.566187
140
+ 2023-02-13 03:12:52,597 - INFO - [Train] step: 15699, loss_mpn: 0.013880, loss_rec: 0.024043, loss_semantic: 0.370826, loss_idmrf: 2.768170, loss_adv_gen: -44.828705
141
+ 2023-02-13 03:13:21,837 - INFO - [Train] step: 15799, loss_adv_disc: -3.458609
142
+ 2023-02-13 03:13:22,046 - INFO - [Train] step: 15799, loss_mpn: 0.011703, loss_rec: 0.035102, loss_semantic: 0.382987, loss_idmrf: 4.488460, loss_adv_gen: -65.005875
143
+ 2023-02-13 03:13:51,286 - INFO - [Train] step: 15899, loss_adv_disc: -0.599114
144
+ 2023-02-13 03:13:51,495 - INFO - [Train] step: 15899, loss_mpn: 0.012524, loss_rec: 0.025607, loss_semantic: 0.360270, loss_idmrf: 1.651819, loss_adv_gen: -54.153267
145
+ 2023-02-13 03:14:20,919 - INFO - [Train] step: 15999, loss_adv_disc: -0.632309
146
+ 2023-02-13 03:14:21,128 - INFO - [Train] step: 15999, loss_mpn: 0.005059, loss_rec: 0.024989, loss_semantic: 0.313581, loss_idmrf: 2.090605, loss_adv_gen: -50.700420
147
+ 2023-02-13 03:14:26,263 - INFO - [Eval] step: 15999, bce: 0.407530, psnr: 23.910851, ssim: 0.929588
148
+ 2023-02-13 03:14:56,787 - INFO - [Train] step: 16099, loss_adv_disc: -2.858291
149
+ 2023-02-13 03:14:56,996 - INFO - [Train] step: 16099, loss_mpn: 0.009401, loss_rec: 0.040451, loss_semantic: 0.429576, loss_idmrf: 3.903901, loss_adv_gen: -67.689865
150
+ 2023-02-13 03:15:26,239 - INFO - [Train] step: 16199, loss_adv_disc: -4.446222
151
+ 2023-02-13 03:15:26,448 - INFO - [Train] step: 16199, loss_mpn: 0.010294, loss_rec: 0.029341, loss_semantic: 0.350636, loss_idmrf: 4.611250, loss_adv_gen: -58.366482
152
+ 2023-02-13 03:15:55,697 - INFO - [Train] step: 16299, loss_adv_disc: -1.042235
153
+ 2023-02-13 03:15:55,905 - INFO - [Train] step: 16299, loss_mpn: 0.008295, loss_rec: 0.034843, loss_semantic: 0.395846, loss_idmrf: 2.474969, loss_adv_gen: -40.464108
154
+ 2023-02-13 03:16:25,337 - INFO - [Train] step: 16399, loss_adv_disc: -0.666275
155
+ 2023-02-13 03:16:25,546 - INFO - [Train] step: 16399, loss_mpn: 0.008661, loss_rec: 0.032371, loss_semantic: 0.398836, loss_idmrf: 4.959391, loss_adv_gen: -58.231514
156
+ 2023-02-13 03:16:54,787 - INFO - [Train] step: 16499, loss_adv_disc: -3.111873
157
+ 2023-02-13 03:16:54,996 - INFO - [Train] step: 16499, loss_mpn: 0.006355, loss_rec: 0.030116, loss_semantic: 0.322977, loss_idmrf: 2.656320, loss_adv_gen: -59.407661
158
+ 2023-02-13 03:17:24,246 - INFO - [Train] step: 16599, loss_adv_disc: -0.969188
159
+ 2023-02-13 03:17:24,454 - INFO - [Train] step: 16599, loss_mpn: 0.006812, loss_rec: 0.030045, loss_semantic: 0.388530, loss_idmrf: 3.429312, loss_adv_gen: -55.191212
160
+ 2023-02-13 03:17:53,885 - INFO - [Train] step: 16699, loss_adv_disc: -1.384235
161
+ 2023-02-13 03:17:54,094 - INFO - [Train] step: 16699, loss_mpn: 0.007724, loss_rec: 0.036217, loss_semantic: 0.368885, loss_idmrf: 1.676530, loss_adv_gen: -57.781860
162
+ 2023-02-13 03:18:23,332 - INFO - [Train] step: 16799, loss_adv_disc: -1.833328
163
+ 2023-02-13 03:18:23,541 - INFO - [Train] step: 16799, loss_mpn: 0.011292, loss_rec: 0.038726, loss_semantic: 0.407536, loss_idmrf: 2.409158, loss_adv_gen: -69.407043
164
+ 2023-02-13 03:18:52,791 - INFO - [Train] step: 16899, loss_adv_disc: -2.642775
165
+ 2023-02-13 03:18:53,000 - INFO - [Train] step: 16899, loss_mpn: 0.017453, loss_rec: 0.039096, loss_semantic: 0.409500, loss_idmrf: 4.307015, loss_adv_gen: -69.141121
166
+ 2023-02-13 03:19:22,424 - INFO - [Train] step: 16999, loss_adv_disc: -2.500937
167
+ 2023-02-13 03:19:22,633 - INFO - [Train] step: 16999, loss_mpn: 0.011810, loss_rec: 0.031087, loss_semantic: 0.387880, loss_idmrf: 3.024927, loss_adv_gen: -49.219688
168
+ 2023-02-13 03:19:27,776 - INFO - [Eval] step: 16999, bce: 0.381526, psnr: 23.834940, ssim: 0.929512
169
+ 2023-02-13 03:19:58,077 - INFO - [Train] step: 17099, loss_adv_disc: -4.640062
170
+ 2023-02-13 03:19:58,286 - INFO - [Train] step: 17099, loss_mpn: 0.005222, loss_rec: 0.031922, loss_semantic: 0.341579, loss_idmrf: 2.995645, loss_adv_gen: -48.879826
171
+ 2023-02-13 03:20:27,517 - INFO - [Train] step: 17199, loss_adv_disc: -1.268022
172
+ 2023-02-13 03:20:27,726 - INFO - [Train] step: 17199, loss_mpn: 0.006362, loss_rec: 0.028963, loss_semantic: 0.348143, loss_idmrf: 2.592186, loss_adv_gen: -58.502922
173
+ 2023-02-13 03:20:56,971 - INFO - [Train] step: 17299, loss_adv_disc: -3.076265
174
+ 2023-02-13 03:20:57,179 - INFO - [Train] step: 17299, loss_mpn: 0.005794, loss_rec: 0.032290, loss_semantic: 0.320258, loss_idmrf: 3.073877, loss_adv_gen: -75.206711
175
+ 2023-02-13 03:21:26,600 - INFO - [Train] step: 17399, loss_adv_disc: -0.722616
176
+ 2023-02-13 03:21:26,809 - INFO - [Train] step: 17399, loss_mpn: 0.008862, loss_rec: 0.027422, loss_semantic: 0.329777, loss_idmrf: 3.391365, loss_adv_gen: -45.755348
177
+ 2023-02-13 03:21:56,062 - INFO - [Train] step: 17499, loss_adv_disc: -1.951949
178
+ 2023-02-13 03:21:56,271 - INFO - [Train] step: 17499, loss_mpn: 0.022730, loss_rec: 0.036895, loss_semantic: 0.536559, loss_idmrf: 2.266129, loss_adv_gen: -33.774021
179
+ 2023-02-13 03:22:25,513 - INFO - [Train] step: 17599, loss_adv_disc: -2.635029
180
+ 2023-02-13 03:22:25,721 - INFO - [Train] step: 17599, loss_mpn: 0.020863, loss_rec: 0.034489, loss_semantic: 0.406302, loss_idmrf: 2.749728, loss_adv_gen: -68.144875
181
+ 2023-02-13 03:22:55,157 - INFO - [Train] step: 17699, loss_adv_disc: -1.602936
182
+ 2023-02-13 03:22:55,366 - INFO - [Train] step: 17699, loss_mpn: 0.009050, loss_rec: 0.035981, loss_semantic: 0.420545, loss_idmrf: 2.019552, loss_adv_gen: -59.465828
183
+ 2023-02-13 03:23:24,609 - INFO - [Train] step: 17799, loss_adv_disc: -1.937828
184
+ 2023-02-13 03:23:24,817 - INFO - [Train] step: 17799, loss_mpn: 0.016830, loss_rec: 0.032039, loss_semantic: 0.409238, loss_idmrf: 3.658442, loss_adv_gen: -62.794762
185
+ 2023-02-13 03:23:54,047 - INFO - [Train] step: 17899, loss_adv_disc: -1.917367
186
+ 2023-02-13 03:23:54,256 - INFO - [Train] step: 17899, loss_mpn: 0.011632, loss_rec: 0.032111, loss_semantic: 0.397115, loss_idmrf: 5.596826, loss_adv_gen: -54.250221
187
+ 2023-02-13 03:24:23,692 - INFO - [Train] step: 17999, loss_adv_disc: -1.522330
188
+ 2023-02-13 03:24:23,900 - INFO - [Train] step: 17999, loss_mpn: 0.008025, loss_rec: 0.027907, loss_semantic: 0.317568, loss_idmrf: 4.100923, loss_adv_gen: -41.896065
189
+ 2023-02-13 03:24:29,031 - INFO - [Eval] step: 17999, bce: 0.442065, psnr: 23.760124, ssim: 0.929170
190
+ 2023-02-13 03:24:59,321 - INFO - [Train] step: 18099, loss_adv_disc: -0.926828
191
+ 2023-02-13 03:24:59,530 - INFO - [Train] step: 18099, loss_mpn: 0.005604, loss_rec: 0.026934, loss_semantic: 0.317587, loss_idmrf: 3.018780, loss_adv_gen: -43.219757
192
+ 2023-02-13 03:25:28,761 - INFO - [Train] step: 18199, loss_adv_disc: -1.252304
193
+ 2023-02-13 03:25:28,970 - INFO - [Train] step: 18199, loss_mpn: 0.006583, loss_rec: 0.036220, loss_semantic: 0.343108, loss_idmrf: 3.407384, loss_adv_gen: -59.127075
194
+ 2023-02-13 03:25:58,215 - INFO - [Train] step: 18299, loss_adv_disc: -4.361234
195
+ 2023-02-13 03:25:58,423 - INFO - [Train] step: 18299, loss_mpn: 0.015157, loss_rec: 0.037108, loss_semantic: 0.418717, loss_idmrf: 3.749311, loss_adv_gen: -60.900532
196
+ 2023-02-13 03:26:27,850 - INFO - [Train] step: 18399, loss_adv_disc: -2.134433
197
+ 2023-02-13 03:26:28,059 - INFO - [Train] step: 18399, loss_mpn: 0.008116, loss_rec: 0.027292, loss_semantic: 0.342987, loss_idmrf: 2.560256, loss_adv_gen: -48.888443
198
+ 2023-02-13 03:26:57,292 - INFO - [Train] step: 18499, loss_adv_disc: 0.527846
199
+ 2023-02-13 03:26:57,501 - INFO - [Train] step: 18499, loss_mpn: 0.006677, loss_rec: 0.026621, loss_semantic: 0.318599, loss_idmrf: 2.754797, loss_adv_gen: -46.165695
200
+ 2023-02-13 03:27:26,725 - INFO - [Train] step: 18599, loss_adv_disc: -1.308938
201
+ 2023-02-13 03:27:26,934 - INFO - [Train] step: 18599, loss_mpn: 0.007462, loss_rec: 0.039733, loss_semantic: 0.407243, loss_idmrf: 1.858337, loss_adv_gen: -37.666771
202
+ 2023-02-13 03:27:56,360 - INFO - [Train] step: 18699, loss_adv_disc: -2.993707
203
+ 2023-02-13 03:27:56,569 - INFO - [Train] step: 18699, loss_mpn: 0.004185, loss_rec: 0.024976, loss_semantic: 0.292228, loss_idmrf: 3.635601, loss_adv_gen: -41.624527
204
+ 2023-02-13 03:28:25,807 - INFO - [Train] step: 18799, loss_adv_disc: -0.925315
205
+ 2023-02-13 03:28:26,016 - INFO - [Train] step: 18799, loss_mpn: 0.005924, loss_rec: 0.023125, loss_semantic: 0.296790, loss_idmrf: 2.006775, loss_adv_gen: -57.868706
206
+ 2023-02-13 03:28:55,264 - INFO - [Train] step: 18899, loss_adv_disc: -4.718306
207
+ 2023-02-13 03:28:55,472 - INFO - [Train] step: 18899, loss_mpn: 0.010640, loss_rec: 0.041306, loss_semantic: 0.368742, loss_idmrf: 5.275894, loss_adv_gen: -66.048042
208
+ 2023-02-13 03:29:24,901 - INFO - [Train] step: 18999, loss_adv_disc: -3.436166
209
+ 2023-02-13 03:29:25,109 - INFO - [Train] step: 18999, loss_mpn: 0.008324, loss_rec: 0.031876, loss_semantic: 0.347499, loss_idmrf: 2.720975, loss_adv_gen: -78.016502
210
+ 2023-02-13 03:29:30,252 - INFO - [Eval] step: 18999, bce: 0.449807, psnr: 23.802105, ssim: 0.928289
211
+ 2023-02-13 03:30:00,570 - INFO - [Train] step: 19099, loss_adv_disc: -1.636162
212
+ 2023-02-13 03:30:00,779 - INFO - [Train] step: 19099, loss_mpn: 0.006203, loss_rec: 0.037541, loss_semantic: 0.359517, loss_idmrf: 3.860636, loss_adv_gen: -16.672791
213
+ 2023-02-13 03:30:30,016 - INFO - [Train] step: 19199, loss_adv_disc: -3.129387
214
+ 2023-02-13 03:30:30,225 - INFO - [Train] step: 19199, loss_mpn: 0.007494, loss_rec: 0.033616, loss_semantic: 0.351235, loss_idmrf: 2.313427, loss_adv_gen: -65.895172
215
+ 2023-02-13 03:30:59,645 - INFO - [Train] step: 19299, loss_adv_disc: -1.372097
216
+ 2023-02-13 03:30:59,853 - INFO - [Train] step: 19299, loss_mpn: 0.007141, loss_rec: 0.030316, loss_semantic: 0.351007, loss_idmrf: 2.720200, loss_adv_gen: -61.847340
217
+ 2023-02-13 03:31:29,092 - INFO - [Train] step: 19399, loss_adv_disc: -1.721190
218
+ 2023-02-13 03:31:29,301 - INFO - [Train] step: 19399, loss_mpn: 0.006103, loss_rec: 0.029222, loss_semantic: 0.334996, loss_idmrf: 4.253203, loss_adv_gen: -60.822281
219
+ 2023-02-13 03:31:58,551 - INFO - [Train] step: 19499, loss_adv_disc: -1.837099
220
+ 2023-02-13 03:31:58,760 - INFO - [Train] step: 19499, loss_mpn: 0.010932, loss_rec: 0.026487, loss_semantic: 0.384942, loss_idmrf: 3.565819, loss_adv_gen: -45.608253
221
+ 2023-02-13 03:32:28,006 - INFO - [Train] step: 19599, loss_adv_disc: -2.718598
222
+ 2023-02-13 03:32:28,214 - INFO - [Train] step: 19599, loss_mpn: 0.015166, loss_rec: 0.035973, loss_semantic: 0.467589, loss_idmrf: 2.883530, loss_adv_gen: -50.148312
223
+ 2023-02-13 03:32:57,640 - INFO - [Train] step: 19699, loss_adv_disc: -2.816616
224
+ 2023-02-13 03:32:57,848 - INFO - [Train] step: 19699, loss_mpn: 0.016170, loss_rec: 0.035296, loss_semantic: 0.428836, loss_idmrf: 3.616713, loss_adv_gen: -58.801323
225
+ 2023-02-13 03:33:27,079 - INFO - [Train] step: 19799, loss_adv_disc: -0.808167
226
+ 2023-02-13 03:33:27,288 - INFO - [Train] step: 19799, loss_mpn: 0.006861, loss_rec: 0.030523, loss_semantic: 0.333927, loss_idmrf: 1.760766, loss_adv_gen: -48.423855
227
+ 2023-02-13 03:33:56,540 - INFO - [Train] step: 19899, loss_adv_disc: -1.257056
228
+ 2023-02-13 03:33:56,748 - INFO - [Train] step: 19899, loss_mpn: 0.007679, loss_rec: 0.029557, loss_semantic: 0.375780, loss_idmrf: 2.455881, loss_adv_gen: -60.748562
229
+ 2023-02-13 03:34:26,168 - INFO - [Train] step: 19999, loss_adv_disc: -2.131047
230
+ 2023-02-13 03:34:26,376 - INFO - [Train] step: 19999, loss_mpn: 0.007155, loss_rec: 0.033969, loss_semantic: 0.365665, loss_idmrf: 4.215178, loss_adv_gen: -42.286987
231
+ 2023-02-13 03:34:31,535 - INFO - [Eval] step: 19999, bce: 0.532807, psnr: 23.844255, ssim: 0.928977
232
+ 2023-02-13 03:35:01,994 - INFO - [Train] step: 20099, loss_adv_disc: -0.821020
233
+ 2023-02-13 03:35:02,203 - INFO - [Train] step: 20099, loss_mpn: 0.002918, loss_rec: 0.019707, loss_semantic: 0.295061, loss_idmrf: 1.051597, loss_adv_gen: -54.710880
234
+ 2023-02-13 03:35:31,436 - INFO - [Train] step: 20199, loss_adv_disc: -1.382209
235
+ 2023-02-13 03:35:31,645 - INFO - [Train] step: 20199, loss_mpn: 0.008895, loss_rec: 0.024759, loss_semantic: 0.342038, loss_idmrf: 4.306022, loss_adv_gen: -59.216305
236
+ 2023-02-13 03:36:01,086 - INFO - [Train] step: 20299, loss_adv_disc: -1.843528
237
+ 2023-02-13 03:36:01,295 - INFO - [Train] step: 20299, loss_mpn: 0.021462, loss_rec: 0.036182, loss_semantic: 0.377382, loss_idmrf: 4.597513, loss_adv_gen: -64.265610
238
+ 2023-02-13 03:36:30,533 - INFO - [Train] step: 20399, loss_adv_disc: -1.680128
239
+ 2023-02-13 03:36:30,741 - INFO - [Train] step: 20399, loss_mpn: 0.006874, loss_rec: 0.026922, loss_semantic: 0.314610, loss_idmrf: 2.805595, loss_adv_gen: -47.953209
240
+ 2023-02-13 03:36:59,988 - INFO - [Train] step: 20499, loss_adv_disc: -0.908575
241
+ 2023-02-13 03:37:00,197 - INFO - [Train] step: 20499, loss_mpn: 0.012316, loss_rec: 0.027403, loss_semantic: 0.369375, loss_idmrf: 2.440661, loss_adv_gen: -50.821514
242
+ 2023-02-13 03:37:29,441 - INFO - [Train] step: 20599, loss_adv_disc: -0.791389
243
+ 2023-02-13 03:37:29,650 - INFO - [Train] step: 20599, loss_mpn: 0.005344, loss_rec: 0.021288, loss_semantic: 0.310707, loss_idmrf: 5.183637, loss_adv_gen: -62.245640
244
+ 2023-02-13 03:37:59,069 - INFO - [Train] step: 20699, loss_adv_disc: -1.344117
245
+ 2023-02-13 03:37:59,278 - INFO - [Train] step: 20699, loss_mpn: 0.007965, loss_rec: 0.025558, loss_semantic: 0.339915, loss_idmrf: 1.723730, loss_adv_gen: -37.892769
246
+ 2023-02-13 03:38:28,512 - INFO - [Train] step: 20799, loss_adv_disc: -4.988375
247
+ 2023-02-13 03:38:28,720 - INFO - [Train] step: 20799, loss_mpn: 0.012833, loss_rec: 0.037847, loss_semantic: 0.357962, loss_idmrf: 3.132791, loss_adv_gen: -62.692024
248
+ 2023-02-13 03:38:57,955 - INFO - [Train] step: 20899, loss_adv_disc: -1.214686
249
+ 2023-02-13 03:38:58,164 - INFO - [Train] step: 20899, loss_mpn: 0.010181, loss_rec: 0.043168, loss_semantic: 0.396244, loss_idmrf: 3.857690, loss_adv_gen: -67.181183
250
+ 2023-02-13 03:39:27,591 - INFO - [Train] step: 20999, loss_adv_disc: -1.825191
251
+ 2023-02-13 03:39:27,800 - INFO - [Train] step: 20999, loss_mpn: 0.010744, loss_rec: 0.033796, loss_semantic: 0.396960, loss_idmrf: 3.169326, loss_adv_gen: -53.664833
252
+ 2023-02-13 03:39:32,938 - INFO - [Eval] step: 20999, bce: 0.412189, psnr: 23.995857, ssim: 0.931168
253
+ 2023-02-13 03:40:03,449 - INFO - [Train] step: 21099, loss_adv_disc: -2.373983
254
+ 2023-02-13 03:40:03,657 - INFO - [Train] step: 21099, loss_mpn: 0.016065, loss_rec: 0.031946, loss_semantic: 0.375466, loss_idmrf: 3.098600, loss_adv_gen: -42.354683
255
+ 2023-02-13 03:40:32,892 - INFO - [Train] step: 21199, loss_adv_disc: -4.281846
256
+ 2023-02-13 03:40:33,101 - INFO - [Train] step: 21199, loss_mpn: 0.008469, loss_rec: 0.030433, loss_semantic: 0.343839, loss_idmrf: 2.670589, loss_adv_gen: -42.038010
257
+ 2023-02-13 03:41:02,534 - INFO - [Train] step: 21299, loss_adv_disc: -2.047416
258
+ 2023-02-13 03:41:02,743 - INFO - [Train] step: 21299, loss_mpn: 0.007427, loss_rec: 0.032674, loss_semantic: 0.365108, loss_idmrf: 2.909791, loss_adv_gen: -58.266251
259
+ 2023-02-13 03:41:31,982 - INFO - [Train] step: 21399, loss_adv_disc: -0.787121
260
+ 2023-02-13 03:41:32,191 - INFO - [Train] step: 21399, loss_mpn: 0.007601, loss_rec: 0.032305, loss_semantic: 0.341065, loss_idmrf: 1.518853, loss_adv_gen: -49.361404
261
+ 2023-02-13 03:42:01,440 - INFO - [Train] step: 21499, loss_adv_disc: -2.960994
262
+ 2023-02-13 03:42:01,649 - INFO - [Train] step: 21499, loss_mpn: 0.010133, loss_rec: 0.033218, loss_semantic: 0.454507, loss_idmrf: 3.152321, loss_adv_gen: -43.761852
263
+ 2023-02-13 03:42:30,902 - INFO - [Train] step: 21599, loss_adv_disc: -0.531810
264
+ 2023-02-13 03:42:31,111 - INFO - [Train] step: 21599, loss_mpn: 0.006240, loss_rec: 0.026150, loss_semantic: 0.333196, loss_idmrf: 3.342569, loss_adv_gen: -49.895046
265
+ 2023-02-13 03:43:00,552 - INFO - [Train] step: 21699, loss_adv_disc: -0.934840
266
+ 2023-02-13 03:43:00,760 - INFO - [Train] step: 21699, loss_mpn: 0.006143, loss_rec: 0.025827, loss_semantic: 0.347423, loss_idmrf: 5.185640, loss_adv_gen: -61.407597
267
+ 2023-02-13 03:43:30,007 - INFO - [Train] step: 21799, loss_adv_disc: -1.708362
268
+ 2023-02-13 03:43:30,215 - INFO - [Train] step: 21799, loss_mpn: 0.007157, loss_rec: 0.033153, loss_semantic: 0.366003, loss_idmrf: 2.537788, loss_adv_gen: -56.243927
269
+ 2023-02-13 03:43:59,453 - INFO - [Train] step: 21899, loss_adv_disc: -2.662016
270
+ 2023-02-13 03:43:59,662 - INFO - [Train] step: 21899, loss_mpn: 0.011560, loss_rec: 0.034256, loss_semantic: 0.414814, loss_idmrf: 2.540663, loss_adv_gen: -47.157578
271
+ 2023-02-13 03:44:29,088 - INFO - [Train] step: 21999, loss_adv_disc: -1.665895
272
+ 2023-02-13 03:44:29,297 - INFO - [Train] step: 21999, loss_mpn: 0.012481, loss_rec: 0.041681, loss_semantic: 0.410288, loss_idmrf: 1.742691, loss_adv_gen: -62.708092
273
+ 2023-02-13 03:44:34,479 - INFO - [Eval] step: 21999, bce: 0.464714, psnr: 23.809441, ssim: 0.930178
274
+ 2023-02-13 03:45:04,803 - INFO - [Train] step: 22099, loss_adv_disc: -2.987253
275
+ 2023-02-13 03:45:05,012 - INFO - [Train] step: 22099, loss_mpn: 0.007824, loss_rec: 0.031368, loss_semantic: 0.363590, loss_idmrf: 2.717836, loss_adv_gen: -58.619572
276
+ 2023-02-13 03:45:34,241 - INFO - [Train] step: 22199, loss_adv_disc: -0.920993
277
+ 2023-02-13 03:45:34,449 - INFO - [Train] step: 22199, loss_mpn: 0.009114, loss_rec: 0.032073, loss_semantic: 0.363013, loss_idmrf: 3.228037, loss_adv_gen: -51.014500
278
+ 2023-02-13 03:46:03,893 - INFO - [Train] step: 22299, loss_adv_disc: -1.332216
279
+ 2023-02-13 03:46:04,102 - INFO - [Train] step: 22299, loss_mpn: 0.005149, loss_rec: 0.023768, loss_semantic: 0.309608, loss_idmrf: 3.535915, loss_adv_gen: -58.052727
280
+ 2023-02-13 03:46:33,337 - INFO - [Train] step: 22399, loss_adv_disc: -2.954251
281
+ 2023-02-13 03:46:33,546 - INFO - [Train] step: 22399, loss_mpn: 0.008305, loss_rec: 0.042678, loss_semantic: 0.423436, loss_idmrf: 3.085995, loss_adv_gen: -57.386398
282
+ 2023-02-13 03:47:02,796 - INFO - [Train] step: 22499, loss_adv_disc: -2.537665
283
+ 2023-02-13 03:47:03,005 - INFO - [Train] step: 22499, loss_mpn: 0.010203, loss_rec: 0.033067, loss_semantic: 0.427873, loss_idmrf: 3.364084, loss_adv_gen: -31.977356
284
+ 2023-02-13 03:47:32,266 - INFO - [Train] step: 22599, loss_adv_disc: -2.119593
285
+ 2023-02-13 03:47:32,475 - INFO - [Train] step: 22599, loss_mpn: 0.009034, loss_rec: 0.026896, loss_semantic: 0.351944, loss_idmrf: 2.590875, loss_adv_gen: -39.202179
286
+ 2023-02-13 03:48:01,904 - INFO - [Train] step: 22699, loss_adv_disc: -2.218577
287
+ 2023-02-13 03:48:02,113 - INFO - [Train] step: 22699, loss_mpn: 0.005697, loss_rec: 0.031317, loss_semantic: 0.377514, loss_idmrf: 1.533009, loss_adv_gen: -43.550716
288
+ 2023-02-13 03:48:31,363 - INFO - [Train] step: 22799, loss_adv_disc: -3.975915
289
+ 2023-02-13 03:48:31,571 - INFO - [Train] step: 22799, loss_mpn: 0.004963, loss_rec: 0.030778, loss_semantic: 0.357679, loss_idmrf: 1.428168, loss_adv_gen: -64.493042
290
+ 2023-02-13 03:49:00,815 - INFO - [Train] step: 22899, loss_adv_disc: -2.386568
291
+ 2023-02-13 03:49:01,024 - INFO - [Train] step: 22899, loss_mpn: 0.008505, loss_rec: 0.033951, loss_semantic: 0.400746, loss_idmrf: 2.240148, loss_adv_gen: -60.405560
292
+ 2023-02-13 03:49:30,445 - INFO - [Train] step: 22999, loss_adv_disc: -1.425534
293
+ 2023-02-13 03:49:30,654 - INFO - [Train] step: 22999, loss_mpn: 0.004612, loss_rec: 0.022062, loss_semantic: 0.292219, loss_idmrf: 2.723917, loss_adv_gen: -54.489628
294
+ 2023-02-13 03:49:35,816 - INFO - [Eval] step: 22999, bce: 0.496233, psnr: 23.705549, ssim: 0.930028
295
+ 2023-02-13 03:50:06,130 - INFO - [Train] step: 23099, loss_adv_disc: -3.629030
296
+ 2023-02-13 03:50:06,339 - INFO - [Train] step: 23099, loss_mpn: 0.017113, loss_rec: 0.039963, loss_semantic: 0.420710, loss_idmrf: 3.228497, loss_adv_gen: -66.667511
297
+ 2023-02-13 03:50:35,579 - INFO - [Train] step: 23199, loss_adv_disc: -2.341859
298
+ 2023-02-13 03:50:35,787 - INFO - [Train] step: 23199, loss_mpn: 0.007523, loss_rec: 0.030686, loss_semantic: 0.342419, loss_idmrf: 2.144747, loss_adv_gen: -73.500671
299
+ 2023-02-13 03:51:05,245 - INFO - [Train] step: 23299, loss_adv_disc: -1.051663
300
+ 2023-02-13 03:51:05,454 - INFO - [Train] step: 23299, loss_mpn: 0.005567, loss_rec: 0.023323, loss_semantic: 0.330779, loss_idmrf: 2.536284, loss_adv_gen: -54.168274
301
+ 2023-02-13 03:51:34,704 - INFO - [Train] step: 23399, loss_adv_disc: -2.064084
302
+ 2023-02-13 03:51:34,914 - INFO - [Train] step: 23399, loss_mpn: 0.005655, loss_rec: 0.025069, loss_semantic: 0.338070, loss_idmrf: 2.127142, loss_adv_gen: -64.372055
303
+ 2023-02-13 03:52:04,164 - INFO - [Train] step: 23499, loss_adv_disc: -2.628796
304
+ 2023-02-13 03:52:04,373 - INFO - [Train] step: 23499, loss_mpn: 0.007882, loss_rec: 0.035950, loss_semantic: 0.411128, loss_idmrf: 1.505162, loss_adv_gen: -54.485676
305
+ 2023-02-13 03:52:33,627 - INFO - [Train] step: 23599, loss_adv_disc: -2.026811
306
+ 2023-02-13 03:52:33,836 - INFO - [Train] step: 23599, loss_mpn: 0.008700, loss_rec: 0.021994, loss_semantic: 0.314727, loss_idmrf: 2.674738, loss_adv_gen: -50.828640
307
+ 2023-02-13 03:53:03,268 - INFO - [Train] step: 23699, loss_adv_disc: -2.408108
308
+ 2023-02-13 03:53:03,477 - INFO - [Train] step: 23699, loss_mpn: 0.007119, loss_rec: 0.029397, loss_semantic: 0.370820, loss_idmrf: 2.113973, loss_adv_gen: -64.023994
309
+ 2023-02-13 03:53:32,736 - INFO - [Train] step: 23799, loss_adv_disc: -2.448371
310
+ 2023-02-13 03:53:32,945 - INFO - [Train] step: 23799, loss_mpn: 0.009236, loss_rec: 0.038028, loss_semantic: 0.389971, loss_idmrf: 2.496542, loss_adv_gen: -68.355583
311
+ 2023-02-13 03:54:02,191 - INFO - [Train] step: 23899, loss_adv_disc: -2.646384
312
+ 2023-02-13 03:54:02,400 - INFO - [Train] step: 23899, loss_mpn: 0.005686, loss_rec: 0.026619, loss_semantic: 0.305794, loss_idmrf: 2.764585, loss_adv_gen: -58.164246
313
+ 2023-02-13 03:54:31,844 - INFO - [Train] step: 23999, loss_adv_disc: -2.244050
314
+ 2023-02-13 03:54:32,053 - INFO - [Train] step: 23999, loss_mpn: 0.019072, loss_rec: 0.036695, loss_semantic: 0.419800, loss_idmrf: 4.794349, loss_adv_gen: -56.691669
315
+ 2023-02-13 03:54:37,188 - INFO - [Eval] step: 23999, bce: 0.423072, psnr: 23.929384, ssim: 0.931687
316
+ 2023-02-13 03:55:07,474 - INFO - [Train] step: 24099, loss_adv_disc: -1.824091
317
+ 2023-02-13 03:55:07,682 - INFO - [Train] step: 24099, loss_mpn: 0.007323, loss_rec: 0.025409, loss_semantic: 0.367337, loss_idmrf: 1.783256, loss_adv_gen: -41.997749
318
+ 2023-02-13 03:55:36,922 - INFO - [Train] step: 24199, loss_adv_disc: -1.705926
319
+ 2023-02-13 03:55:37,131 - INFO - [Train] step: 24199, loss_mpn: 0.006608, loss_rec: 0.024922, loss_semantic: 0.325300, loss_idmrf: 1.772946, loss_adv_gen: -55.280464
320
+ 2023-02-13 03:56:06,571 - INFO - [Train] step: 24299, loss_adv_disc: -1.387813
321
+ 2023-02-13 03:56:06,780 - INFO - [Train] step: 24299, loss_mpn: 0.008532, loss_rec: 0.024819, loss_semantic: 0.333036, loss_idmrf: 3.660525, loss_adv_gen: -42.061749
322
+ 2023-02-13 03:56:36,009 - INFO - [Train] step: 24399, loss_adv_disc: -3.398212
323
+ 2023-02-13 03:56:36,218 - INFO - [Train] step: 24399, loss_mpn: 0.004310, loss_rec: 0.025555, loss_semantic: 0.320220, loss_idmrf: 0.711168, loss_adv_gen: -52.639893
324
+ 2023-02-13 03:57:05,452 - INFO - [Train] step: 24499, loss_adv_disc: -1.182582
325
+ 2023-02-13 03:57:05,661 - INFO - [Train] step: 24499, loss_mpn: 0.004868, loss_rec: 0.026463, loss_semantic: 0.320026, loss_idmrf: 2.703800, loss_adv_gen: -44.305458
326
+ 2023-02-13 03:57:34,890 - INFO - [Train] step: 24599, loss_adv_disc: -2.179983
327
+ 2023-02-13 03:57:35,098 - INFO - [Train] step: 24599, loss_mpn: 0.008763, loss_rec: 0.021728, loss_semantic: 0.337773, loss_idmrf: 2.351331, loss_adv_gen: -39.223248
328
+ 2023-02-13 03:58:04,537 - INFO - [Train] step: 24699, loss_adv_disc: -1.959560
329
+ 2023-02-13 03:58:04,745 - INFO - [Train] step: 24699, loss_mpn: 0.007004, loss_rec: 0.029935, loss_semantic: 0.346946, loss_idmrf: 2.790915, loss_adv_gen: -56.569611
330
+ 2023-02-13 03:58:33,993 - INFO - [Train] step: 24799, loss_adv_disc: -1.558927
331
+ 2023-02-13 03:58:34,202 - INFO - [Train] step: 24799, loss_mpn: 0.013352, loss_rec: 0.030124, loss_semantic: 0.413617, loss_idmrf: 2.552111, loss_adv_gen: -56.299240
332
+ 2023-02-13 03:59:03,447 - INFO - [Train] step: 24899, loss_adv_disc: -2.618589
333
+ 2023-02-13 03:59:03,656 - INFO - [Train] step: 24899, loss_mpn: 0.012856, loss_rec: 0.033626, loss_semantic: 0.396584, loss_idmrf: 5.144225, loss_adv_gen: -46.024830
334
+ 2023-02-13 03:59:33,088 - INFO - [Train] step: 24999, loss_adv_disc: -2.875488
335
+ 2023-02-13 03:59:33,296 - INFO - [Train] step: 24999, loss_mpn: 0.006294, loss_rec: 0.032758, loss_semantic: 0.382921, loss_idmrf: 4.101409, loss_adv_gen: -38.091202
336
+ 2023-02-13 03:59:38,456 - INFO - [Eval] step: 24999, bce: 0.494293, psnr: 23.910641, ssim: 0.931275
337
+ 2023-02-13 04:00:08,901 - INFO - [Train] step: 25099, loss_adv_disc: -3.002494
338
+ 2023-02-13 04:00:09,110 - INFO - [Train] step: 25099, loss_mpn: 0.012923, loss_rec: 0.042998, loss_semantic: 0.389094, loss_idmrf: 3.376410, loss_adv_gen: -64.067093
339
+ 2023-02-13 04:00:38,346 - INFO - [Train] step: 25199, loss_adv_disc: -5.016815
340
+ 2023-02-13 04:00:38,555 - INFO - [Train] step: 25199, loss_mpn: 0.014368, loss_rec: 0.053279, loss_semantic: 0.451728, loss_idmrf: 3.427337, loss_adv_gen: -61.353497
341
+ 2023-02-13 04:01:07,993 - INFO - [Train] step: 25299, loss_adv_disc: -2.128964
342
+ 2023-02-13 04:01:08,202 - INFO - [Train] step: 25299, loss_mpn: 0.007653, loss_rec: 0.031010, loss_semantic: 0.372408, loss_idmrf: 3.696013, loss_adv_gen: -37.674156
343
+ 2023-02-13 04:01:37,446 - INFO - [Train] step: 25399, loss_adv_disc: -1.170438
344
+ 2023-02-13 04:01:37,655 - INFO - [Train] step: 25399, loss_mpn: 0.007121, loss_rec: 0.023605, loss_semantic: 0.337867, loss_idmrf: 2.371561, loss_adv_gen: -44.922745
345
+ 2023-02-13 04:02:06,886 - INFO - [Train] step: 25499, loss_adv_disc: -2.289455
346
+ 2023-02-13 04:02:07,095 - INFO - [Train] step: 25499, loss_mpn: 0.009439, loss_rec: 0.030207, loss_semantic: 0.341227, loss_idmrf: 3.571287, loss_adv_gen: -52.630623
347
+ 2023-02-13 04:02:36,331 - INFO - [Train] step: 25599, loss_adv_disc: -2.275971
348
+ 2023-02-13 04:02:36,540 - INFO - [Train] step: 25599, loss_mpn: 0.010324, loss_rec: 0.033716, loss_semantic: 0.369033, loss_idmrf: 2.983631, loss_adv_gen: -56.771912
349
+ 2023-02-13 04:03:05,956 - INFO - [Train] step: 25699, loss_adv_disc: -3.929438
350
+ 2023-02-13 04:03:06,165 - INFO - [Train] step: 25699, loss_mpn: 0.005107, loss_rec: 0.023621, loss_semantic: 0.317293, loss_idmrf: 1.977758, loss_adv_gen: -38.194695
351
+ 2023-02-13 04:03:35,407 - INFO - [Train] step: 25799, loss_adv_disc: -2.947953
352
+ 2023-02-13 04:03:35,616 - INFO - [Train] step: 25799, loss_mpn: 0.007025, loss_rec: 0.026580, loss_semantic: 0.322663, loss_idmrf: 1.570272, loss_adv_gen: -59.964291
353
+ 2023-02-13 04:04:04,848 - INFO - [Train] step: 25899, loss_adv_disc: -1.598645
354
+ 2023-02-13 04:04:05,057 - INFO - [Train] step: 25899, loss_mpn: 0.005981, loss_rec: 0.039875, loss_semantic: 0.387087, loss_idmrf: 2.259921, loss_adv_gen: -31.552460
355
+ 2023-02-13 04:04:34,480 - INFO - [Train] step: 25999, loss_adv_disc: -1.000123
356
+ 2023-02-13 04:04:34,688 - INFO - [Train] step: 25999, loss_mpn: 0.011854, loss_rec: 0.030783, loss_semantic: 0.366924, loss_idmrf: 2.419249, loss_adv_gen: -42.398209
357
+ 2023-02-13 04:04:39,819 - INFO - [Eval] step: 25999, bce: 0.379053, psnr: 23.972345, ssim: 0.932233
358
+ 2023-02-13 04:05:10,126 - INFO - [Train] step: 26099, loss_adv_disc: -1.131559
359
+ 2023-02-13 04:05:10,335 - INFO - [Train] step: 26099, loss_mpn: 0.007134, loss_rec: 0.021641, loss_semantic: 0.318042, loss_idmrf: 3.112147, loss_adv_gen: -53.450821
360
+ 2023-02-13 04:05:39,581 - INFO - [Train] step: 26199, loss_adv_disc: -1.814842
361
+ 2023-02-13 04:05:39,789 - INFO - [Train] step: 26199, loss_mpn: 0.006957, loss_rec: 0.031351, loss_semantic: 0.359431, loss_idmrf: 3.939166, loss_adv_gen: -49.258316
362
+ 2023-02-13 04:06:09,217 - INFO - [Train] step: 26299, loss_adv_disc: -1.674930
363
+ 2023-02-13 04:06:09,426 - INFO - [Train] step: 26299, loss_mpn: 0.005781, loss_rec: 0.029166, loss_semantic: 0.291324, loss_idmrf: 3.769212, loss_adv_gen: -36.555950
364
+ 2023-02-13 04:06:38,659 - INFO - [Train] step: 26399, loss_adv_disc: -1.669163
365
+ 2023-02-13 04:06:38,867 - INFO - [Train] step: 26399, loss_mpn: 0.009312, loss_rec: 0.021297, loss_semantic: 0.302380, loss_idmrf: 3.968942, loss_adv_gen: -35.490517
366
+ 2023-02-13 04:07:08,106 - INFO - [Train] step: 26499, loss_adv_disc: -3.611588
367
+ 2023-02-13 04:07:08,314 - INFO - [Train] step: 26499, loss_mpn: 0.012238, loss_rec: 0.030930, loss_semantic: 0.397214, loss_idmrf: 4.898618, loss_adv_gen: -69.384735
368
+ 2023-02-13 04:07:37,547 - INFO - [Train] step: 26599, loss_adv_disc: -4.188798
369
+ 2023-02-13 04:07:37,755 - INFO - [Train] step: 26599, loss_mpn: 0.014178, loss_rec: 0.041108, loss_semantic: 0.457722, loss_idmrf: 3.513246, loss_adv_gen: -88.662796
370
+ 2023-02-13 04:08:07,192 - INFO - [Train] step: 26699, loss_adv_disc: -2.032719
371
+ 2023-02-13 04:08:07,400 - INFO - [Train] step: 26699, loss_mpn: 0.005614, loss_rec: 0.028942, loss_semantic: 0.317826, loss_idmrf: 2.767789, loss_adv_gen: -53.963402
372
+ 2023-02-13 04:08:36,641 - INFO - [Train] step: 26799, loss_adv_disc: -1.779737
373
+ 2023-02-13 04:08:36,850 - INFO - [Train] step: 26799, loss_mpn: 0.005052, loss_rec: 0.023449, loss_semantic: 0.296901, loss_idmrf: 1.286634, loss_adv_gen: -62.788879
374
+ 2023-02-13 04:09:06,077 - INFO - [Train] step: 26899, loss_adv_disc: -3.180129
375
+ 2023-02-13 04:09:06,286 - INFO - [Train] step: 26899, loss_mpn: 0.010187, loss_rec: 0.027996, loss_semantic: 0.343080, loss_idmrf: 2.296761, loss_adv_gen: -52.778481
376
+ 2023-02-13 04:09:35,690 - INFO - [Train] step: 26999, loss_adv_disc: -0.776358
377
+ 2023-02-13 04:09:35,903 - INFO - [Train] step: 26999, loss_mpn: 0.003122, loss_rec: 0.022430, loss_semantic: 0.296052, loss_idmrf: 1.410728, loss_adv_gen: -56.631275
378
+ 2023-02-13 04:09:41,048 - INFO - [Eval] step: 26999, bce: 0.467820, psnr: 23.892040, ssim: 0.931154
379
+ 2023-02-13 04:10:11,354 - INFO - [Train] step: 27099, loss_adv_disc: -3.549163
380
+ 2023-02-13 04:10:11,563 - INFO - [Train] step: 27099, loss_mpn: 0.012723, loss_rec: 0.045948, loss_semantic: 0.476646, loss_idmrf: 3.937117, loss_adv_gen: -45.297981
381
+ 2023-02-13 04:10:40,799 - INFO - [Train] step: 27199, loss_adv_disc: -1.461819
382
+ 2023-02-13 04:10:41,008 - INFO - [Train] step: 27199, loss_mpn: 0.006455, loss_rec: 0.025119, loss_semantic: 0.319218, loss_idmrf: 1.629803, loss_adv_gen: -49.119114
383
+ 2023-02-13 04:11:10,427 - INFO - [Train] step: 27299, loss_adv_disc: -1.982147
384
+ 2023-02-13 04:11:10,636 - INFO - [Train] step: 27299, loss_mpn: 0.007175, loss_rec: 0.028984, loss_semantic: 0.362375, loss_idmrf: 1.994591, loss_adv_gen: -35.928234
385
+ 2023-02-13 04:11:39,875 - INFO - [Train] step: 27399, loss_adv_disc: -1.590581
386
+ 2023-02-13 04:11:40,084 - INFO - [Train] step: 27399, loss_mpn: 0.008258, loss_rec: 0.030676, loss_semantic: 0.359340, loss_idmrf: 5.009028, loss_adv_gen: -66.535286
387
+ 2023-02-13 04:12:09,315 - INFO - [Train] step: 27499, loss_adv_disc: -0.402845
388
+ 2023-02-13 04:12:09,524 - INFO - [Train] step: 27499, loss_mpn: 0.005017, loss_rec: 0.018790, loss_semantic: 0.267146, loss_idmrf: 3.032572, loss_adv_gen: -31.063499
389
+ 2023-02-13 04:12:38,945 - INFO - [Train] step: 27599, loss_adv_disc: -2.334307
390
+ 2023-02-13 04:12:39,153 - INFO - [Train] step: 27599, loss_mpn: 0.009680, loss_rec: 0.031130, loss_semantic: 0.403578, loss_idmrf: 2.780985, loss_adv_gen: -43.041565
391
+ 2023-02-13 04:13:08,383 - INFO - [Train] step: 27699, loss_adv_disc: -3.015743
392
+ 2023-02-13 04:13:08,591 - INFO - [Train] step: 27699, loss_mpn: 0.018975, loss_rec: 0.047783, loss_semantic: 0.425497, loss_idmrf: 3.275642, loss_adv_gen: -46.561310
393
+ 2023-02-13 04:13:37,834 - INFO - [Train] step: 27799, loss_adv_disc: -2.397944
394
+ 2023-02-13 04:13:38,042 - INFO - [Train] step: 27799, loss_mpn: 0.008875, loss_rec: 0.022515, loss_semantic: 0.318298, loss_idmrf: 3.172273, loss_adv_gen: -50.527237
395
+ 2023-02-13 04:14:07,284 - INFO - [Train] step: 27899, loss_adv_disc: -3.253990
396
+ 2023-02-13 04:14:07,492 - INFO - [Train] step: 27899, loss_mpn: 0.008475, loss_rec: 0.030102, loss_semantic: 0.385625, loss_idmrf: 3.359437, loss_adv_gen: -54.444069
397
+ 2023-02-13 04:14:36,918 - INFO - [Train] step: 27999, loss_adv_disc: -1.056982
398
+ 2023-02-13 04:14:37,126 - INFO - [Train] step: 27999, loss_mpn: 0.007983, loss_rec: 0.024990, loss_semantic: 0.312865, loss_idmrf: 1.289099, loss_adv_gen: -40.702690
399
+ 2023-02-13 04:14:42,289 - INFO - [Eval] step: 27999, bce: 0.407889, psnr: 24.050518, ssim: 0.932894
400
+ 2023-02-13 04:15:12,757 - INFO - [Train] step: 28099, loss_adv_disc: -3.248437
401
+ 2023-02-13 04:15:12,965 - INFO - [Train] step: 28099, loss_mpn: 0.006869, loss_rec: 0.028881, loss_semantic: 0.362487, loss_idmrf: 1.182486, loss_adv_gen: -54.308258
402
+ 2023-02-13 04:15:42,196 - INFO - [Train] step: 28199, loss_adv_disc: -2.695652
403
+ 2023-02-13 04:15:42,405 - INFO - [Train] step: 28199, loss_mpn: 0.006013, loss_rec: 0.027336, loss_semantic: 0.359971, loss_idmrf: 1.179807, loss_adv_gen: -31.996471
404
+ 2023-02-13 04:16:11,831 - INFO - [Train] step: 28299, loss_adv_disc: -2.555561
405
+ 2023-02-13 04:16:12,040 - INFO - [Train] step: 28299, loss_mpn: 0.006889, loss_rec: 0.026413, loss_semantic: 0.323360, loss_idmrf: 2.315033, loss_adv_gen: -45.354450
406
+ 2023-02-13 04:16:41,278 - INFO - [Train] step: 28399, loss_adv_disc: -0.930417
407
+ 2023-02-13 04:16:41,486 - INFO - [Train] step: 28399, loss_mpn: 0.003901, loss_rec: 0.020817, loss_semantic: 0.298452, loss_idmrf: 2.879560, loss_adv_gen: -52.044228
408
+ 2023-02-13 04:17:10,724 - INFO - [Train] step: 28499, loss_adv_disc: -1.263307
409
+ 2023-02-13 04:17:10,933 - INFO - [Train] step: 28499, loss_mpn: 0.008175, loss_rec: 0.027284, loss_semantic: 0.342604, loss_idmrf: 1.002726, loss_adv_gen: -47.339424
410
+ 2023-02-13 04:17:40,360 - INFO - [Train] step: 28599, loss_adv_disc: -3.608934
411
+ 2023-02-13 04:17:40,568 - INFO - [Train] step: 28599, loss_mpn: 0.009251, loss_rec: 0.031929, loss_semantic: 0.374587, loss_idmrf: 1.810358, loss_adv_gen: -42.233673
412
+ 2023-02-13 04:18:09,808 - INFO - [Train] step: 28699, loss_adv_disc: -1.356530
413
+ 2023-02-13 04:18:10,017 - INFO - [Train] step: 28699, loss_mpn: 0.006944, loss_rec: 0.023396, loss_semantic: 0.342738, loss_idmrf: 1.533525, loss_adv_gen: -38.768314
414
+ 2023-02-13 04:18:39,252 - INFO - [Train] step: 28799, loss_adv_disc: -1.913511
415
+ 2023-02-13 04:18:39,461 - INFO - [Train] step: 28799, loss_mpn: 0.005483, loss_rec: 0.022142, loss_semantic: 0.296328, loss_idmrf: 1.239522, loss_adv_gen: -41.124989
416
+ 2023-02-13 04:19:08,710 - INFO - [Train] step: 28899, loss_adv_disc: -1.888143
417
+ 2023-02-13 04:19:08,918 - INFO - [Train] step: 28899, loss_mpn: 0.006542, loss_rec: 0.025403, loss_semantic: 0.326764, loss_idmrf: 4.936932, loss_adv_gen: -59.378212
418
+ 2023-02-13 04:19:38,344 - INFO - [Train] step: 28999, loss_adv_disc: -1.588324
419
+ 2023-02-13 04:19:38,553 - INFO - [Train] step: 28999, loss_mpn: 0.006560, loss_rec: 0.025576, loss_semantic: 0.327476, loss_idmrf: 3.260027, loss_adv_gen: -48.005432
420
+ 2023-02-13 04:19:43,698 - INFO - [Eval] step: 28999, bce: 0.465755, psnr: 23.979197, ssim: 0.932822
421
+ 2023-02-13 04:20:13,994 - INFO - [Train] step: 29099, loss_adv_disc: -2.685792
422
+ 2023-02-13 04:20:14,203 - INFO - [Train] step: 29099, loss_mpn: 0.012146, loss_rec: 0.029994, loss_semantic: 0.365296, loss_idmrf: 2.682022, loss_adv_gen: -36.540726
423
+ 2023-02-13 04:20:43,452 - INFO - [Train] step: 29199, loss_adv_disc: -1.354690
424
+ 2023-02-13 04:20:43,661 - INFO - [Train] step: 29199, loss_mpn: 0.007733, loss_rec: 0.025824, loss_semantic: 0.352530, loss_idmrf: 1.492018, loss_adv_gen: -59.247963
425
+ 2023-02-13 04:21:13,086 - INFO - [Train] step: 29299, loss_adv_disc: -0.308071
426
+ 2023-02-13 04:21:13,295 - INFO - [Train] step: 29299, loss_mpn: 0.005732, loss_rec: 0.027066, loss_semantic: 0.429230, loss_idmrf: 5.198031, loss_adv_gen: -36.152596
427
+ 2023-02-13 04:21:42,523 - INFO - [Train] step: 29399, loss_adv_disc: -5.704453
428
+ 2023-02-13 04:21:42,733 - INFO - [Train] step: 29399, loss_mpn: 0.009742, loss_rec: 0.036058, loss_semantic: 0.422115, loss_idmrf: 3.425453, loss_adv_gen: -64.740868
429
+ 2023-02-13 04:22:11,979 - INFO - [Train] step: 29499, loss_adv_disc: -3.116785
430
+ 2023-02-13 04:22:12,187 - INFO - [Train] step: 29499, loss_mpn: 0.009215, loss_rec: 0.025835, loss_semantic: 0.381380, loss_idmrf: 4.609172, loss_adv_gen: -35.275311
431
+ 2023-02-13 04:22:41,608 - INFO - [Train] step: 29599, loss_adv_disc: -3.281649
432
+ 2023-02-13 04:22:41,817 - INFO - [Train] step: 29599, loss_mpn: 0.008651, loss_rec: 0.032746, loss_semantic: 0.376462, loss_idmrf: 1.623248, loss_adv_gen: -55.076706
433
+ 2023-02-13 04:23:11,065 - INFO - [Train] step: 29699, loss_adv_disc: -1.218207
434
+ 2023-02-13 04:23:11,274 - INFO - [Train] step: 29699, loss_mpn: 0.003264, loss_rec: 0.019132, loss_semantic: 0.286781, loss_idmrf: 2.779821, loss_adv_gen: -47.997040
435
+ 2023-02-13 04:23:40,521 - INFO - [Train] step: 29799, loss_adv_disc: -4.113495
436
+ 2023-02-13 04:23:40,730 - INFO - [Train] step: 29799, loss_mpn: 0.014174, loss_rec: 0.033425, loss_semantic: 0.392298, loss_idmrf: 2.298213, loss_adv_gen: -53.244530
437
+ 2023-02-13 04:24:09,972 - INFO - [Train] step: 29899, loss_adv_disc: -2.777640
438
+ 2023-02-13 04:24:10,181 - INFO - [Train] step: 29899, loss_mpn: 0.010283, loss_rec: 0.027365, loss_semantic: 0.370754, loss_idmrf: 2.673162, loss_adv_gen: -44.858421
439
+ 2023-02-13 04:24:39,617 - INFO - [Train] step: 29999, loss_adv_disc: -1.686746
440
+ 2023-02-13 04:24:39,826 - INFO - [Train] step: 29999, loss_mpn: 0.006264, loss_rec: 0.021137, loss_semantic: 0.292509, loss_idmrf: 2.118923, loss_adv_gen: -39.865761
441
+ 2023-02-13 04:24:44,947 - INFO - [Eval] step: 29999, bce: 0.492185, psnr: 23.945683, ssim: 0.932370
442
+ 2023-02-13 04:25:15,393 - INFO - [Train] step: 30099, loss_adv_disc: -2.197589
443
+ 2023-02-13 04:25:15,602 - INFO - [Train] step: 30099, loss_mpn: 0.014454, loss_rec: 0.025383, loss_semantic: 0.374786, loss_idmrf: 2.624727, loss_adv_gen: -40.343857
444
+ 2023-02-13 04:25:44,882 - INFO - [Train] step: 30199, loss_adv_disc: -1.674134
445
+ 2023-02-13 04:25:45,091 - INFO - [Train] step: 30199, loss_mpn: 0.006576, loss_rec: 0.023323, loss_semantic: 0.302643, loss_idmrf: 1.844227, loss_adv_gen: -39.347649
446
+ 2023-02-13 04:26:14,522 - INFO - [Train] step: 30299, loss_adv_disc: -2.467939
447
+ 2023-02-13 04:26:14,730 - INFO - [Train] step: 30299, loss_mpn: 0.004610, loss_rec: 0.019071, loss_semantic: 0.272741, loss_idmrf: 2.557956, loss_adv_gen: -35.477177
448
+ 2023-02-13 04:26:43,963 - INFO - [Train] step: 30399, loss_adv_disc: -3.106873
449
+ 2023-02-13 04:26:44,171 - INFO - [Train] step: 30399, loss_mpn: 0.011434, loss_rec: 0.034334, loss_semantic: 0.442513, loss_idmrf: 2.949084, loss_adv_gen: -48.490547
450
+ 2023-02-13 04:27:13,406 - INFO - [Train] step: 30499, loss_adv_disc: -3.356421
451
+ 2023-02-13 04:27:13,614 - INFO - [Train] step: 30499, loss_mpn: 0.006068, loss_rec: 0.027413, loss_semantic: 0.352242, loss_idmrf: 2.848948, loss_adv_gen: -59.479187
452
+ 2023-02-13 04:27:43,044 - INFO - [Train] step: 30599, loss_adv_disc: -1.141179
453
+ 2023-02-13 04:27:43,253 - INFO - [Train] step: 30599, loss_mpn: 0.004855, loss_rec: 0.017252, loss_semantic: 0.252036, loss_idmrf: 1.422145, loss_adv_gen: -30.667336
454
+ 2023-02-13 04:28:12,498 - INFO - [Train] step: 30699, loss_adv_disc: -2.391990
455
+ 2023-02-13 04:28:12,707 - INFO - [Train] step: 30699, loss_mpn: 0.008719, loss_rec: 0.027900, loss_semantic: 0.338170, loss_idmrf: 3.741982, loss_adv_gen: -48.991299
456
+ 2023-02-13 04:28:41,943 - INFO - [Train] step: 30799, loss_adv_disc: -2.743029
457
+ 2023-02-13 04:28:42,152 - INFO - [Train] step: 30799, loss_mpn: 0.007150, loss_rec: 0.029743, loss_semantic: 0.353363, loss_idmrf: 1.883271, loss_adv_gen: -57.676750
458
+ 2023-02-13 04:29:11,394 - INFO - [Train] step: 30899, loss_adv_disc: -1.434591
459
+ 2023-02-13 04:29:11,603 - INFO - [Train] step: 30899, loss_mpn: 0.004940, loss_rec: 0.021217, loss_semantic: 0.291477, loss_idmrf: 1.621642, loss_adv_gen: -61.386360
460
+ 2023-02-13 04:29:41,027 - INFO - [Train] step: 30999, loss_adv_disc: -3.122560
461
+ 2023-02-13 04:29:41,235 - INFO - [Train] step: 30999, loss_mpn: 0.007449, loss_rec: 0.035113, loss_semantic: 0.435334, loss_idmrf: 3.818401, loss_adv_gen: -52.961311
462
+ 2023-02-13 04:29:46,393 - INFO - [Eval] step: 30999, bce: 0.490164, psnr: 23.965448, ssim: 0.932622
463
+ 2023-02-13 04:30:16,716 - INFO - [Train] step: 31099, loss_adv_disc: -2.185221
464
+ 2023-02-13 04:30:16,924 - INFO - [Train] step: 31099, loss_mpn: 0.004554, loss_rec: 0.020886, loss_semantic: 0.288429, loss_idmrf: 1.497186, loss_adv_gen: -52.509830
465
+ 2023-02-13 04:30:46,170 - INFO - [Train] step: 31199, loss_adv_disc: -4.675382
466
+ 2023-02-13 04:30:46,378 - INFO - [Train] step: 31199, loss_mpn: 0.011436, loss_rec: 0.037177, loss_semantic: 0.422422, loss_idmrf: 3.885715, loss_adv_gen: -55.271225
467
+ 2023-02-13 04:31:15,811 - INFO - [Train] step: 31299, loss_adv_disc: -1.755943
468
+ 2023-02-13 04:31:16,020 - INFO - [Train] step: 31299, loss_mpn: 0.009999, loss_rec: 0.023714, loss_semantic: 0.357692, loss_idmrf: 2.204484, loss_adv_gen: -36.626755
469
+ 2023-02-13 04:31:45,265 - INFO - [Train] step: 31399, loss_adv_disc: -1.392416
470
+ 2023-02-13 04:31:45,474 - INFO - [Train] step: 31399, loss_mpn: 0.007405, loss_rec: 0.023420, loss_semantic: 0.318830, loss_idmrf: 2.317692, loss_adv_gen: -43.350834
471
+ 2023-02-13 04:32:14,705 - INFO - [Train] step: 31499, loss_adv_disc: -2.211383
472
+ 2023-02-13 04:32:14,914 - INFO - [Train] step: 31499, loss_mpn: 0.006516, loss_rec: 0.028322, loss_semantic: 0.381882, loss_idmrf: 1.678529, loss_adv_gen: -37.920712
473
+ 2023-02-13 04:32:44,356 - INFO - [Train] step: 31599, loss_adv_disc: -1.601540
474
+ 2023-02-13 04:32:44,564 - INFO - [Train] step: 31599, loss_mpn: 0.009201, loss_rec: 0.031845, loss_semantic: 0.375913, loss_idmrf: 1.841415, loss_adv_gen: -47.110973
475
+ 2023-02-13 04:33:13,811 - INFO - [Train] step: 31699, loss_adv_disc: -1.545524
476
+ 2023-02-13 04:33:14,019 - INFO - [Train] step: 31699, loss_mpn: 0.008448, loss_rec: 0.037653, loss_semantic: 0.368679, loss_idmrf: 2.739496, loss_adv_gen: -25.949142
477
+ 2023-02-13 04:33:43,266 - INFO - [Train] step: 31799, loss_adv_disc: -3.020606
478
+ 2023-02-13 04:33:43,475 - INFO - [Train] step: 31799, loss_mpn: 0.015079, loss_rec: 0.038613, loss_semantic: 0.437632, loss_idmrf: 2.923442, loss_adv_gen: -38.652390
479
+ 2023-02-13 04:34:12,732 - INFO - [Train] step: 31899, loss_adv_disc: -2.436349
480
+ 2023-02-13 04:34:12,940 - INFO - [Train] step: 31899, loss_mpn: 0.007766, loss_rec: 0.024789, loss_semantic: 0.337582, loss_idmrf: 2.340315, loss_adv_gen: -43.865879
481
+ 2023-02-13 04:34:42,373 - INFO - [Train] step: 31999, loss_adv_disc: -2.298278
482
+ 2023-02-13 04:34:42,582 - INFO - [Train] step: 31999, loss_mpn: 0.008034, loss_rec: 0.031308, loss_semantic: 0.375757, loss_idmrf: 2.158618, loss_adv_gen: -53.958466
483
+ 2023-02-13 04:34:47,725 - INFO - [Eval] step: 31999, bce: 0.511144, psnr: 23.987217, ssim: 0.932225
484
+ 2023-02-13 04:35:18,021 - INFO - [Train] step: 32099, loss_adv_disc: -2.084677
485
+ 2023-02-13 04:35:18,229 - INFO - [Train] step: 32099, loss_mpn: 0.009502, loss_rec: 0.025188, loss_semantic: 0.323963, loss_idmrf: 3.507452, loss_adv_gen: -46.396492
486
+ 2023-02-13 04:35:47,464 - INFO - [Train] step: 32199, loss_adv_disc: -4.015974
487
+ 2023-02-13 04:35:47,673 - INFO - [Train] step: 32199, loss_mpn: 0.009766, loss_rec: 0.030390, loss_semantic: 0.407730, loss_idmrf: 5.195652, loss_adv_gen: -59.783085
488
+ 2023-02-13 04:36:17,108 - INFO - [Train] step: 32299, loss_adv_disc: -3.174493
489
+ 2023-02-13 04:36:17,317 - INFO - [Train] step: 32299, loss_mpn: 0.009587, loss_rec: 0.022740, loss_semantic: 0.340636, loss_idmrf: 1.810525, loss_adv_gen: -43.734375
490
+ 2023-02-13 04:36:46,557 - INFO - [Train] step: 32399, loss_adv_disc: -2.961869
491
+ 2023-02-13 04:36:46,766 - INFO - [Train] step: 32399, loss_mpn: 0.006834, loss_rec: 0.026820, loss_semantic: 0.367303, loss_idmrf: 3.318485, loss_adv_gen: -48.750732
492
+ 2023-02-13 04:37:16,014 - INFO - [Train] step: 32499, loss_adv_disc: -1.361956
493
+ 2023-02-13 04:37:16,223 - INFO - [Train] step: 32499, loss_mpn: 0.003526, loss_rec: 0.021565, loss_semantic: 0.330520, loss_idmrf: 0.975429, loss_adv_gen: -30.753990
494
+ 2023-02-13 04:37:45,659 - INFO - [Train] step: 32599, loss_adv_disc: 0.066008
495
+ 2023-02-13 04:37:45,868 - INFO - [Train] step: 32599, loss_mpn: 0.011720, loss_rec: 0.032479, loss_semantic: 0.369994, loss_idmrf: 3.079811, loss_adv_gen: -62.844032
496
+ 2023-02-13 04:38:15,115 - INFO - [Train] step: 32699, loss_adv_disc: -2.147907
497
+ 2023-02-13 04:38:15,323 - INFO - [Train] step: 32699, loss_mpn: 0.010529, loss_rec: 0.030653, loss_semantic: 0.365616, loss_idmrf: 2.815894, loss_adv_gen: -38.179665
498
+ 2023-02-13 04:38:44,576 - INFO - [Train] step: 32799, loss_adv_disc: -3.022984
499
+ 2023-02-13 04:38:44,784 - INFO - [Train] step: 32799, loss_mpn: 0.007264, loss_rec: 0.024473, loss_semantic: 0.354572, loss_idmrf: 2.879258, loss_adv_gen: -47.123955
500
+ 2023-02-13 04:39:14,018 - INFO - [Train] step: 32899, loss_adv_disc: -2.134367
501
+ 2023-02-13 04:39:14,226 - INFO - [Train] step: 32899, loss_mpn: 0.010902, loss_rec: 0.024346, loss_semantic: 0.349439, loss_idmrf: 2.405934, loss_adv_gen: -49.264977
502
+ 2023-02-13 04:39:43,683 - INFO - [Train] step: 32999, loss_adv_disc: -1.867232
503
+ 2023-02-13 04:39:43,891 - INFO - [Train] step: 32999, loss_mpn: 0.007945, loss_rec: 0.030071, loss_semantic: 0.401522, loss_idmrf: 1.317394, loss_adv_gen: -43.877331
504
+ 2023-02-13 04:39:49,055 - INFO - [Eval] step: 32999, bce: 0.449431, psnr: 23.902346, ssim: 0.932984
505
+ 2023-02-13 04:40:19,349 - INFO - [Train] step: 33099, loss_adv_disc: -2.116894
506
+ 2023-02-13 04:40:19,558 - INFO - [Train] step: 33099, loss_mpn: 0.010758, loss_rec: 0.024110, loss_semantic: 0.369346, loss_idmrf: 1.759542, loss_adv_gen: -47.994667
507
+ 2023-02-13 04:40:48,792 - INFO - [Train] step: 33199, loss_adv_disc: -1.882688
508
+ 2023-02-13 04:40:49,001 - INFO - [Train] step: 33199, loss_mpn: 0.008150, loss_rec: 0.024921, loss_semantic: 0.324246, loss_idmrf: 3.030573, loss_adv_gen: -52.880379
509
+ 2023-02-13 04:41:18,442 - INFO - [Train] step: 33299, loss_adv_disc: -1.602988
510
+ 2023-02-13 04:41:18,651 - INFO - [Train] step: 33299, loss_mpn: 0.005471, loss_rec: 0.018395, loss_semantic: 0.298615, loss_idmrf: 5.933259, loss_adv_gen: -33.227077
511
+ 2023-02-13 04:41:47,905 - INFO - [Train] step: 33399, loss_adv_disc: -1.262914
512
+ 2023-02-13 04:41:48,113 - INFO - [Train] step: 33399, loss_mpn: 0.004389, loss_rec: 0.023651, loss_semantic: 0.292535, loss_idmrf: 1.633751, loss_adv_gen: -47.819061
513
+ 2023-02-13 04:42:17,374 - INFO - [Train] step: 33499, loss_adv_disc: -4.915746
514
+ 2023-02-13 04:42:17,583 - INFO - [Train] step: 33499, loss_mpn: 0.006723, loss_rec: 0.039050, loss_semantic: 0.405457, loss_idmrf: 2.034822, loss_adv_gen: -73.422531
515
+ 2023-02-13 04:42:47,032 - INFO - [Train] step: 33599, loss_adv_disc: -3.144320
516
+ 2023-02-13 04:42:47,241 - INFO - [Train] step: 33599, loss_mpn: 0.009663, loss_rec: 0.031112, loss_semantic: 0.353883, loss_idmrf: 2.664726, loss_adv_gen: -30.228935
517
+ 2023-02-13 04:43:16,501 - INFO - [Train] step: 33699, loss_adv_disc: -4.212373
518
+ 2023-02-13 04:43:16,710 - INFO - [Train] step: 33699, loss_mpn: 0.015043, loss_rec: 0.039189, loss_semantic: 0.461344, loss_idmrf: 2.188730, loss_adv_gen: -57.981586
519
+ 2023-02-13 04:43:45,968 - INFO - [Train] step: 33799, loss_adv_disc: -2.260951
520
+ 2023-02-13 04:43:46,177 - INFO - [Train] step: 33799, loss_mpn: 0.012627, loss_rec: 0.029967, loss_semantic: 0.422946, loss_idmrf: 1.669225, loss_adv_gen: -42.562790
521
+ 2023-02-13 04:44:15,423 - INFO - [Train] step: 33899, loss_adv_disc: -2.908595
522
+ 2023-02-13 04:44:15,631 - INFO - [Train] step: 33899, loss_mpn: 0.009963, loss_rec: 0.026809, loss_semantic: 0.331871, loss_idmrf: 4.898873, loss_adv_gen: -28.988209
523
+ 2023-02-13 04:44:45,067 - INFO - [Train] step: 33999, loss_adv_disc: -3.627460
524
+ 2023-02-13 04:44:45,276 - INFO - [Train] step: 33999, loss_mpn: 0.011132, loss_rec: 0.029245, loss_semantic: 0.385511, loss_idmrf: 2.419443, loss_adv_gen: -38.314835
525
+ 2023-02-13 04:44:50,427 - INFO - [Eval] step: 33999, bce: 0.493806, psnr: 24.006083, ssim: 0.933177
526
+ 2023-02-13 04:45:20,720 - INFO - [Train] step: 34099, loss_adv_disc: -2.766464
527
+ 2023-02-13 04:45:20,929 - INFO - [Train] step: 34099, loss_mpn: 0.012229, loss_rec: 0.027476, loss_semantic: 0.361568, loss_idmrf: 1.908972, loss_adv_gen: -34.426533
528
+ 2023-02-13 04:45:50,178 - INFO - [Train] step: 34199, loss_adv_disc: -1.794028
529
+ 2023-02-13 04:45:50,387 - INFO - [Train] step: 34199, loss_mpn: 0.011440, loss_rec: 0.023251, loss_semantic: 0.366893, loss_idmrf: 2.413965, loss_adv_gen: -36.843105
530
+ 2023-02-13 04:46:19,819 - INFO - [Train] step: 34299, loss_adv_disc: -3.664385
531
+ 2023-02-13 04:46:20,028 - INFO - [Train] step: 34299, loss_mpn: 0.005965, loss_rec: 0.025664, loss_semantic: 0.328096, loss_idmrf: 1.862936, loss_adv_gen: -62.822632
532
+ 2023-02-13 04:46:49,273 - INFO - [Train] step: 34399, loss_adv_disc: -4.225115
533
+ 2023-02-13 04:46:49,481 - INFO - [Train] step: 34399, loss_mpn: 0.005665, loss_rec: 0.026986, loss_semantic: 0.313127, loss_idmrf: 1.985018, loss_adv_gen: -42.136913
534
+ 2023-02-13 04:47:18,742 - INFO - [Train] step: 34499, loss_adv_disc: -2.175828
535
+ 2023-02-13 04:47:18,951 - INFO - [Train] step: 34499, loss_mpn: 0.009428, loss_rec: 0.028373, loss_semantic: 0.334881, loss_idmrf: 2.210392, loss_adv_gen: -56.427872
536
+ 2023-02-13 04:47:48,390 - INFO - [Train] step: 34599, loss_adv_disc: -2.655634
537
+ 2023-02-13 04:47:48,599 - INFO - [Train] step: 34599, loss_mpn: 0.007422, loss_rec: 0.026994, loss_semantic: 0.377276, loss_idmrf: 2.321023, loss_adv_gen: -45.277206
538
+ 2023-02-13 04:48:17,863 - INFO - [Train] step: 34699, loss_adv_disc: -2.667136
539
+ 2023-02-13 04:48:18,071 - INFO - [Train] step: 34699, loss_mpn: 0.012077, loss_rec: 0.026902, loss_semantic: 0.367508, loss_idmrf: 1.815201, loss_adv_gen: -30.165154
540
+ 2023-02-13 04:48:47,339 - INFO - [Train] step: 34799, loss_adv_disc: -0.941228
541
+ 2023-02-13 04:48:47,547 - INFO - [Train] step: 34799, loss_mpn: 0.008583, loss_rec: 0.029322, loss_semantic: 0.366778, loss_idmrf: 1.744337, loss_adv_gen: -32.191875
542
+ 2023-02-13 04:49:16,808 - INFO - [Train] step: 34899, loss_adv_disc: -2.496438
543
+ 2023-02-13 04:49:17,016 - INFO - [Train] step: 34899, loss_mpn: 0.008992, loss_rec: 0.033850, loss_semantic: 0.455653, loss_idmrf: 2.395776, loss_adv_gen: -40.789753
544
+ 2023-02-13 04:49:46,483 - INFO - [Train] step: 34999, loss_adv_disc: -3.898407
545
+ 2023-02-13 04:49:46,691 - INFO - [Train] step: 34999, loss_mpn: 0.007672, loss_rec: 0.029516, loss_semantic: 0.391788, loss_idmrf: 3.343580, loss_adv_gen: -49.658260
546
+ 2023-02-13 04:49:51,834 - INFO - [Eval] step: 34999, bce: 0.492687, psnr: 23.920872, ssim: 0.932854
547
+ 2023-02-13 04:50:22,250 - INFO - [Train] step: 35099, loss_adv_disc: -3.315386
548
+ 2023-02-13 04:50:22,458 - INFO - [Train] step: 35099, loss_mpn: 0.007049, loss_rec: 0.026398, loss_semantic: 0.327736, loss_idmrf: 2.569352, loss_adv_gen: -44.722576
549
+ 2023-02-13 04:50:51,704 - INFO - [Train] step: 35199, loss_adv_disc: -1.372233
550
+ 2023-02-13 04:50:51,912 - INFO - [Train] step: 35199, loss_mpn: 0.005760, loss_rec: 0.020659, loss_semantic: 0.290627, loss_idmrf: 1.633903, loss_adv_gen: -40.939323
551
+ 2023-02-13 04:51:21,346 - INFO - [Train] step: 35299, loss_adv_disc: -2.340883
552
+ 2023-02-13 04:51:21,555 - INFO - [Train] step: 35299, loss_mpn: 0.005978, loss_rec: 0.021761, loss_semantic: 0.300960, loss_idmrf: 2.090793, loss_adv_gen: -32.323677
553
+ 2023-02-13 04:51:50,793 - INFO - [Train] step: 35399, loss_adv_disc: -2.734210
554
+ 2023-02-13 04:51:51,001 - INFO - [Train] step: 35399, loss_mpn: 0.009728, loss_rec: 0.029348, loss_semantic: 0.389291, loss_idmrf: 3.652727, loss_adv_gen: -43.234627
555
+ 2023-02-13 04:52:20,240 - INFO - [Train] step: 35499, loss_adv_disc: -1.138059
556
+ 2023-02-13 04:52:20,449 - INFO - [Train] step: 35499, loss_mpn: 0.005424, loss_rec: 0.022387, loss_semantic: 0.280479, loss_idmrf: 4.068128, loss_adv_gen: -42.517578
557
+ 2023-02-13 04:52:49,873 - INFO - [Train] step: 35599, loss_adv_disc: -2.349144
558
+ 2023-02-13 04:52:50,082 - INFO - [Train] step: 35599, loss_mpn: 0.004034, loss_rec: 0.023441, loss_semantic: 0.325334, loss_idmrf: 1.199162, loss_adv_gen: -37.645535
559
+ 2023-02-13 04:53:19,322 - INFO - [Train] step: 35699, loss_adv_disc: -3.336832
560
+ 2023-02-13 04:53:19,531 - INFO - [Train] step: 35699, loss_mpn: 0.009035, loss_rec: 0.032002, loss_semantic: 0.351620, loss_idmrf: 2.143802, loss_adv_gen: -48.266075
561
+ 2023-02-13 04:53:48,771 - INFO - [Train] step: 35799, loss_adv_disc: -5.475993
562
+ 2023-02-13 04:53:48,981 - INFO - [Train] step: 35799, loss_mpn: 0.015121, loss_rec: 0.043468, loss_semantic: 0.448060, loss_idmrf: 4.249763, loss_adv_gen: -59.551453
563
+ 2023-02-13 04:54:18,420 - INFO - [Train] step: 35899, loss_adv_disc: -1.503309
564
+ 2023-02-13 04:54:18,629 - INFO - [Train] step: 35899, loss_mpn: 0.005271, loss_rec: 0.021715, loss_semantic: 0.305199, loss_idmrf: 3.841014, loss_adv_gen: -56.835781
565
+ 2023-02-13 04:54:47,871 - INFO - [Train] step: 35999, loss_adv_disc: -1.925685
566
+ 2023-02-13 04:54:48,079 - INFO - [Train] step: 35999, loss_mpn: 0.007125, loss_rec: 0.024240, loss_semantic: 0.358235, loss_idmrf: 2.990580, loss_adv_gen: -49.677704
567
+ 2023-02-13 04:54:53,246 - INFO - [Eval] step: 35999, bce: 0.451248, psnr: 24.089697, ssim: 0.933519
568
+ 2023-02-13 04:55:23,765 - INFO - [Train] step: 36099, loss_adv_disc: -2.582094
569
+ 2023-02-13 04:55:23,974 - INFO - [Train] step: 36099, loss_mpn: 0.010654, loss_rec: 0.032407, loss_semantic: 0.359470, loss_idmrf: 4.383611, loss_adv_gen: -51.658825
570
+ 2023-02-13 04:55:53,217 - INFO - [Train] step: 36199, loss_adv_disc: -1.860945
571
+ 2023-02-13 04:55:53,426 - INFO - [Train] step: 36199, loss_mpn: 0.005998, loss_rec: 0.023951, loss_semantic: 0.332017, loss_idmrf: 2.729879, loss_adv_gen: -52.860546
572
+ 2023-02-13 04:56:22,869 - INFO - [Train] step: 36299, loss_adv_disc: -3.733282
573
+ 2023-02-13 04:56:23,077 - INFO - [Train] step: 36299, loss_mpn: 0.005556, loss_rec: 0.029372, loss_semantic: 0.412161, loss_idmrf: 4.497084, loss_adv_gen: -80.765701
574
+ 2023-02-13 04:56:52,330 - INFO - [Train] step: 36399, loss_adv_disc: -1.141786
575
+ 2023-02-13 04:56:52,539 - INFO - [Train] step: 36399, loss_mpn: 0.005013, loss_rec: 0.021536, loss_semantic: 0.305248, loss_idmrf: 1.969661, loss_adv_gen: -45.550240
576
+ 2023-02-13 04:57:21,785 - INFO - [Train] step: 36499, loss_adv_disc: -1.790366
577
+ 2023-02-13 04:57:21,994 - INFO - [Train] step: 36499, loss_mpn: 0.005966, loss_rec: 0.025548, loss_semantic: 0.332869, loss_idmrf: 2.200314, loss_adv_gen: -37.936630
578
+ 2023-02-13 04:57:51,440 - INFO - [Train] step: 36599, loss_adv_disc: -3.809116
579
+ 2023-02-13 04:57:51,649 - INFO - [Train] step: 36599, loss_mpn: 0.005486, loss_rec: 0.024083, loss_semantic: 0.316460, loss_idmrf: 2.768273, loss_adv_gen: -49.244308
580
+ 2023-02-13 04:58:20,920 - INFO - [Train] step: 36699, loss_adv_disc: -2.374944
581
+ 2023-02-13 04:58:21,129 - INFO - [Train] step: 36699, loss_mpn: 0.005127, loss_rec: 0.021198, loss_semantic: 0.294804, loss_idmrf: 1.192759, loss_adv_gen: -24.119633
582
+ 2023-02-13 04:58:50,378 - INFO - [Train] step: 36799, loss_adv_disc: -3.080187
583
+ 2023-02-13 04:58:50,587 - INFO - [Train] step: 36799, loss_mpn: 0.009575, loss_rec: 0.025953, loss_semantic: 0.377960, loss_idmrf: 1.418940, loss_adv_gen: -53.199898
584
+ 2023-02-13 04:59:20,041 - INFO - [Train] step: 36899, loss_adv_disc: -1.956524
585
+ 2023-02-13 04:59:20,250 - INFO - [Train] step: 36899, loss_mpn: 0.004222, loss_rec: 0.018766, loss_semantic: 0.279596, loss_idmrf: 3.066632, loss_adv_gen: -43.155220
586
+ 2023-02-13 04:59:49,499 - INFO - [Train] step: 36999, loss_adv_disc: -2.218104
587
+ 2023-02-13 04:59:49,709 - INFO - [Train] step: 36999, loss_mpn: 0.006558, loss_rec: 0.021726, loss_semantic: 0.322024, loss_idmrf: 3.109234, loss_adv_gen: -44.358009
588
+ 2023-02-13 04:59:54,854 - INFO - [Eval] step: 36999, bce: 0.511608, psnr: 24.066105, ssim: 0.933602
589
+ 2023-02-13 05:00:25,154 - INFO - [Train] step: 37099, loss_adv_disc: -3.587424
590
+ 2023-02-13 05:00:25,363 - INFO - [Train] step: 37099, loss_mpn: 0.008115, loss_rec: 0.029436, loss_semantic: 0.385246, loss_idmrf: 1.943526, loss_adv_gen: -34.469177
591
+ 2023-02-13 05:00:54,625 - INFO - [Train] step: 37199, loss_adv_disc: -2.477960
592
+ 2023-02-13 05:00:54,833 - INFO - [Train] step: 37199, loss_mpn: 0.007001, loss_rec: 0.030382, loss_semantic: 0.373161, loss_idmrf: 7.979999, loss_adv_gen: -34.278618
593
+ 2023-02-13 05:01:24,262 - INFO - [Train] step: 37299, loss_adv_disc: -1.760101
594
+ 2023-02-13 05:01:24,471 - INFO - [Train] step: 37299, loss_mpn: 0.006662, loss_rec: 0.020768, loss_semantic: 0.294572, loss_idmrf: 2.288859, loss_adv_gen: -34.755966
595
+ 2023-02-13 05:01:53,723 - INFO - [Train] step: 37399, loss_adv_disc: -4.853773
596
+ 2023-02-13 05:01:53,932 - INFO - [Train] step: 37399, loss_mpn: 0.010881, loss_rec: 0.037399, loss_semantic: 0.408028, loss_idmrf: 5.442630, loss_adv_gen: -53.623989
597
+ 2023-02-13 05:02:23,179 - INFO - [Train] step: 37499, loss_adv_disc: -2.969615
598
+ 2023-02-13 05:02:23,387 - INFO - [Train] step: 37499, loss_mpn: 0.006965, loss_rec: 0.025584, loss_semantic: 0.342903, loss_idmrf: 2.896766, loss_adv_gen: -56.246437
599
+ 2023-02-13 05:02:52,832 - INFO - [Train] step: 37599, loss_adv_disc: -1.794725
600
+ 2023-02-13 05:02:53,041 - INFO - [Train] step: 37599, loss_mpn: 0.008978, loss_rec: 0.024542, loss_semantic: 0.356401, loss_idmrf: 1.502892, loss_adv_gen: -49.551048
601
+ 2023-02-13 05:03:22,289 - INFO - [Train] step: 37699, loss_adv_disc: -2.146141
602
+ 2023-02-13 05:03:22,497 - INFO - [Train] step: 37699, loss_mpn: 0.007144, loss_rec: 0.028536, loss_semantic: 0.323880, loss_idmrf: 2.761802, loss_adv_gen: -54.818451
603
+ 2023-02-13 05:03:51,758 - INFO - [Train] step: 37799, loss_adv_disc: -3.474980
604
+ 2023-02-13 05:03:51,966 - INFO - [Train] step: 37799, loss_mpn: 0.006934, loss_rec: 0.036860, loss_semantic: 0.379891, loss_idmrf: 2.068331, loss_adv_gen: -32.999737
605
+ 2023-02-13 05:04:21,416 - INFO - [Train] step: 37899, loss_adv_disc: -3.225597
606
+ 2023-02-13 05:04:21,625 - INFO - [Train] step: 37899, loss_mpn: 0.010186, loss_rec: 0.033910, loss_semantic: 0.345942, loss_idmrf: 3.703231, loss_adv_gen: -37.339417
607
+ 2023-02-13 05:04:50,873 - INFO - [Train] step: 37999, loss_adv_disc: -1.842927
608
+ 2023-02-13 05:04:51,082 - INFO - [Train] step: 37999, loss_mpn: 0.004289, loss_rec: 0.022910, loss_semantic: 0.307499, loss_idmrf: 0.953175, loss_adv_gen: -35.079369
609
+ 2023-02-13 05:04:56,221 - INFO - [Eval] step: 37999, bce: 0.550441, psnr: 23.929970, ssim: 0.932575
610
+ 2023-02-13 05:05:26,524 - INFO - [Train] step: 38099, loss_adv_disc: -3.334216
611
+ 2023-02-13 05:05:26,732 - INFO - [Train] step: 38099, loss_mpn: 0.010205, loss_rec: 0.032689, loss_semantic: 0.379243, loss_idmrf: 2.807850, loss_adv_gen: -46.448601
612
+ 2023-02-13 05:05:55,991 - INFO - [Train] step: 38199, loss_adv_disc: -4.326716
613
+ 2023-02-13 05:05:56,199 - INFO - [Train] step: 38199, loss_mpn: 0.009495, loss_rec: 0.031506, loss_semantic: 0.367149, loss_idmrf: 7.712746, loss_adv_gen: -44.628571
614
+ 2023-02-13 05:06:25,633 - INFO - [Train] step: 38299, loss_adv_disc: -4.521143
615
+ 2023-02-13 05:06:25,841 - INFO - [Train] step: 38299, loss_mpn: 0.010475, loss_rec: 0.033682, loss_semantic: 0.389523, loss_idmrf: 2.941085, loss_adv_gen: -46.803600
616
+ 2023-02-13 05:06:55,090 - INFO - [Train] step: 38399, loss_adv_disc: -3.862583
617
+ 2023-02-13 05:06:55,299 - INFO - [Train] step: 38399, loss_mpn: 0.008165, loss_rec: 0.033486, loss_semantic: 0.434651, loss_idmrf: 1.601594, loss_adv_gen: -46.221794
618
+ 2023-02-13 05:07:24,538 - INFO - [Train] step: 38499, loss_adv_disc: -2.403943
619
+ 2023-02-13 05:07:24,747 - INFO - [Train] step: 38499, loss_mpn: 0.007161, loss_rec: 0.027608, loss_semantic: 0.396924, loss_idmrf: 1.407646, loss_adv_gen: -39.244766
620
+ 2023-02-13 05:07:54,217 - INFO - [Train] step: 38599, loss_adv_disc: -3.030591
621
+ 2023-02-13 05:07:54,425 - INFO - [Train] step: 38599, loss_mpn: 0.005753, loss_rec: 0.027395, loss_semantic: 0.364165, loss_idmrf: 1.339588, loss_adv_gen: -38.628998
622
+ 2023-02-13 05:08:23,670 - INFO - [Train] step: 38699, loss_adv_disc: -3.780364
623
+ 2023-02-13 05:08:23,879 - INFO - [Train] step: 38699, loss_mpn: 0.009336, loss_rec: 0.031163, loss_semantic: 0.372193, loss_idmrf: 3.190570, loss_adv_gen: -35.400978
624
+ 2023-02-13 05:08:53,137 - INFO - [Train] step: 38799, loss_adv_disc: -3.990635
625
+ 2023-02-13 05:08:53,346 - INFO - [Train] step: 38799, loss_mpn: 0.011644, loss_rec: 0.034559, loss_semantic: 0.392621, loss_idmrf: 2.479134, loss_adv_gen: -49.558567
626
+ 2023-02-13 05:09:22,789 - INFO - [Train] step: 38899, loss_adv_disc: -1.296090
627
+ 2023-02-13 05:09:22,998 - INFO - [Train] step: 38899, loss_mpn: 0.002924, loss_rec: 0.020468, loss_semantic: 0.320249, loss_idmrf: 2.698936, loss_adv_gen: -34.255341
628
+ 2023-02-13 05:09:52,241 - INFO - [Train] step: 38999, loss_adv_disc: -1.709500
629
+ 2023-02-13 05:09:52,450 - INFO - [Train] step: 38999, loss_mpn: 0.007364, loss_rec: 0.024880, loss_semantic: 0.339524, loss_idmrf: 2.538348, loss_adv_gen: -32.774738
630
+ 2023-02-13 05:09:57,599 - INFO - [Eval] step: 38999, bce: 0.543258, psnr: 23.938129, ssim: 0.933121
631
+ 2023-02-13 05:10:27,906 - INFO - [Train] step: 39099, loss_adv_disc: -4.103148
632
+ 2023-02-13 05:10:28,115 - INFO - [Train] step: 39099, loss_mpn: 0.011113, loss_rec: 0.030247, loss_semantic: 0.417251, loss_idmrf: 1.839109, loss_adv_gen: -40.379929
633
+ 2023-02-13 05:10:57,376 - INFO - [Train] step: 39199, loss_adv_disc: -2.376977
634
+ 2023-02-13 05:10:57,585 - INFO - [Train] step: 39199, loss_mpn: 0.006472, loss_rec: 0.027119, loss_semantic: 0.307704, loss_idmrf: 2.029118, loss_adv_gen: -41.603371
635
+ 2023-02-13 05:11:27,026 - INFO - [Train] step: 39299, loss_adv_disc: -2.299066
636
+ 2023-02-13 05:11:27,236 - INFO - [Train] step: 39299, loss_mpn: 0.008503, loss_rec: 0.022436, loss_semantic: 0.329789, loss_idmrf: 1.689665, loss_adv_gen: -55.237625
637
+ 2023-02-13 05:11:56,492 - INFO - [Train] step: 39399, loss_adv_disc: -1.676680
638
+ 2023-02-13 05:11:56,701 - INFO - [Train] step: 39399, loss_mpn: 0.007318, loss_rec: 0.025969, loss_semantic: 0.338832, loss_idmrf: 2.227033, loss_adv_gen: -35.137005
639
+ 2023-02-13 05:12:25,968 - INFO - [Train] step: 39499, loss_adv_disc: -3.850043
640
+ 2023-02-13 05:12:26,177 - INFO - [Train] step: 39499, loss_mpn: 0.007011, loss_rec: 0.025364, loss_semantic: 0.338889, loss_idmrf: 2.077760, loss_adv_gen: -42.100559
641
+ 2023-02-13 05:12:55,617 - INFO - [Train] step: 39599, loss_adv_disc: -2.115081
642
+ 2023-02-13 05:12:55,827 - INFO - [Train] step: 39599, loss_mpn: 0.005659, loss_rec: 0.023901, loss_semantic: 0.306843, loss_idmrf: 1.831819, loss_adv_gen: -40.354424
643
+ 2023-02-13 05:13:25,069 - INFO - [Train] step: 39699, loss_adv_disc: -1.994590
644
+ 2023-02-13 05:13:25,278 - INFO - [Train] step: 39699, loss_mpn: 0.012201, loss_rec: 0.024626, loss_semantic: 0.370962, loss_idmrf: 1.804987, loss_adv_gen: -38.031910
645
+ 2023-02-13 05:13:54,538 - INFO - [Train] step: 39799, loss_adv_disc: -1.271324
646
+ 2023-02-13 05:13:54,747 - INFO - [Train] step: 39799, loss_mpn: 0.007453, loss_rec: 0.022167, loss_semantic: 0.320188, loss_idmrf: 2.214860, loss_adv_gen: -37.319176
647
+ 2023-02-13 05:14:24,190 - INFO - [Train] step: 39899, loss_adv_disc: -1.676649
648
+ 2023-02-13 05:14:24,399 - INFO - [Train] step: 39899, loss_mpn: 0.004491, loss_rec: 0.022268, loss_semantic: 0.293991, loss_idmrf: 2.233423, loss_adv_gen: -52.407726
649
+ 2023-02-13 05:14:53,659 - INFO - [Train] step: 39999, loss_adv_disc: -2.623639
650
+ 2023-02-13 05:14:53,867 - INFO - [Train] step: 39999, loss_mpn: 0.005937, loss_rec: 0.026519, loss_semantic: 0.336726, loss_idmrf: 1.825681, loss_adv_gen: -35.927650
651
+ 2023-02-13 05:14:59,026 - INFO - [Eval] step: 39999, bce: 0.525641, psnr: 23.983196, ssim: 0.932836
652
+ 2023-02-13 05:15:29,452 - INFO - [Train] step: 40099, loss_adv_disc: -2.439545
653
+ 2023-02-13 05:15:29,661 - INFO - [Train] step: 40099, loss_mpn: 0.003933, loss_rec: 0.025905, loss_semantic: 0.298822, loss_idmrf: 1.251885, loss_adv_gen: -25.356033
654
+ 2023-02-13 05:15:58,913 - INFO - [Train] step: 40199, loss_adv_disc: -2.549628
655
+ 2023-02-13 05:15:59,122 - INFO - [Train] step: 40199, loss_mpn: 0.003956, loss_rec: 0.025239, loss_semantic: 0.325746, loss_idmrf: 2.043177, loss_adv_gen: -53.137932
656
+ 2023-02-13 05:16:28,584 - INFO - [Train] step: 40299, loss_adv_disc: -2.381817
657
+ 2023-02-13 05:16:28,792 - INFO - [Train] step: 40299, loss_mpn: 0.005729, loss_rec: 0.019757, loss_semantic: 0.314869, loss_idmrf: 1.337579, loss_adv_gen: -36.087315
658
+ 2023-02-13 05:16:58,059 - INFO - [Train] step: 40399, loss_adv_disc: -1.973938
659
+ 2023-02-13 05:16:58,267 - INFO - [Train] step: 40399, loss_mpn: 0.005751, loss_rec: 0.021610, loss_semantic: 0.290376, loss_idmrf: 1.547544, loss_adv_gen: -47.893143
660
+ 2023-02-13 05:17:27,521 - INFO - [Train] step: 40499, loss_adv_disc: -1.127224
661
+ 2023-02-13 05:17:27,730 - INFO - [Train] step: 40499, loss_mpn: 0.008527, loss_rec: 0.023455, loss_semantic: 0.340019, loss_idmrf: 4.099547, loss_adv_gen: -28.373253
662
+ 2023-02-13 05:17:57,179 - INFO - [Train] step: 40599, loss_adv_disc: -1.630693
663
+ 2023-02-13 05:17:57,388 - INFO - [Train] step: 40599, loss_mpn: 0.004528, loss_rec: 0.022425, loss_semantic: 0.311950, loss_idmrf: 2.045025, loss_adv_gen: -42.663849
664
+ 2023-02-13 05:18:26,648 - INFO - [Train] step: 40699, loss_adv_disc: -4.531513
665
+ 2023-02-13 05:18:26,856 - INFO - [Train] step: 40699, loss_mpn: 0.006523, loss_rec: 0.032161, loss_semantic: 0.387854, loss_idmrf: 2.429147, loss_adv_gen: -54.620087
666
+ 2023-02-13 05:18:56,106 - INFO - [Train] step: 40799, loss_adv_disc: -1.792886
667
+ 2023-02-13 05:18:56,316 - INFO - [Train] step: 40799, loss_mpn: 0.004772, loss_rec: 0.020073, loss_semantic: 0.275349, loss_idmrf: 2.377422, loss_adv_gen: -30.988853
668
+ 2023-02-13 05:19:25,749 - INFO - [Train] step: 40899, loss_adv_disc: -2.539077
669
+ 2023-02-13 05:19:25,958 - INFO - [Train] step: 40899, loss_mpn: 0.006153, loss_rec: 0.024839, loss_semantic: 0.301111, loss_idmrf: 1.659654, loss_adv_gen: -32.603027
670
+ 2023-02-13 05:19:55,216 - INFO - [Train] step: 40999, loss_adv_disc: -2.101744
671
+ 2023-02-13 05:19:55,425 - INFO - [Train] step: 40999, loss_mpn: 0.006380, loss_rec: 0.018790, loss_semantic: 0.273095, loss_idmrf: 0.847124, loss_adv_gen: -29.596489
672
+ 2023-02-13 05:20:00,565 - INFO - [Eval] step: 40999, bce: 0.468785, psnr: 23.891504, ssim: 0.933626
673
+ 2023-02-13 05:20:30,853 - INFO - [Train] step: 41099, loss_adv_disc: -1.681476
674
+ 2023-02-13 05:20:31,061 - INFO - [Train] step: 41099, loss_mpn: 0.007705, loss_rec: 0.024250, loss_semantic: 0.294493, loss_idmrf: 2.501994, loss_adv_gen: -23.292068
675
+ 2023-02-13 05:21:00,315 - INFO - [Train] step: 41199, loss_adv_disc: -2.958653
676
+ 2023-02-13 05:21:00,523 - INFO - [Train] step: 41199, loss_mpn: 0.006746, loss_rec: 0.025472, loss_semantic: 0.317826, loss_idmrf: 1.403564, loss_adv_gen: -49.742279
677
+ 2023-02-13 05:21:29,965 - INFO - [Train] step: 41299, loss_adv_disc: -4.011607
678
+ 2023-02-13 05:21:30,174 - INFO - [Train] step: 41299, loss_mpn: 0.004297, loss_rec: 0.028533, loss_semantic: 0.317103, loss_idmrf: 2.485311, loss_adv_gen: -38.421883
679
+ 2023-02-13 05:21:59,436 - INFO - [Train] step: 41399, loss_adv_disc: -2.888487
680
+ 2023-02-13 05:21:59,645 - INFO - [Train] step: 41399, loss_mpn: 0.003805, loss_rec: 0.019754, loss_semantic: 0.258992, loss_idmrf: 1.701972, loss_adv_gen: -30.434830
downstream-shadow-removal/samples/step000999.png ADDED

Git LFS Details

  • SHA256: f1214759aa03f101533802175d817608281251805dd547d72753fda464607fb4
  • Pointer size: 132 Bytes
  • Size of remote file: 4.74 MB
downstream-shadow-removal/samples/step001999.png ADDED

Git LFS Details

  • SHA256: 82f18e6e0ec8bae05eed42a9d1f20985f9f71f85b95ec6700dd98cfd5ba90591
  • Pointer size: 132 Bytes
  • Size of remote file: 4.68 MB
downstream-shadow-removal/samples/step002999.png ADDED

Git LFS Details

  • SHA256: 6fa76994dc7f73c508c94f9f617d825a7b1281f7b01fed74111040a527346830
  • Pointer size: 132 Bytes
  • Size of remote file: 4.65 MB
downstream-shadow-removal/samples/step003999.png ADDED

Git LFS Details

  • SHA256: 1267b46cf8e3c267566852cff0dfcd51774cfabab7c0688e48bb38cdd9fb7f84
  • Pointer size: 132 Bytes
  • Size of remote file: 4.65 MB
downstream-shadow-removal/samples/step004999.png ADDED

Git LFS Details

  • SHA256: 576c911c91656a1b625286fb8206ac9131faa2609ab9314a946568804eecc1e0
  • Pointer size: 132 Bytes
  • Size of remote file: 4.7 MB
downstream-shadow-removal/samples/step005999.png ADDED

Git LFS Details

  • SHA256: 96aa1780f15a825333b5553c88aee426ba7cb820c2244813135d3ba9b65d944b
  • Pointer size: 132 Bytes
  • Size of remote file: 4.67 MB
downstream-shadow-removal/samples/step006999.png ADDED

Git LFS Details

  • SHA256: 8041666df7e11cc7e5e4a6661aff4d4850b84b8d0760dbd8d9f5a457f4ddc956
  • Pointer size: 132 Bytes
  • Size of remote file: 4.65 MB
downstream-shadow-removal/samples/step007999.png ADDED

Git LFS Details

  • SHA256: 2e48dd6e1f5ff2f3e4665aa09e771be290a62b893d8d8c598ff794c7d23fd590
  • Pointer size: 132 Bytes
  • Size of remote file: 4.64 MB
downstream-shadow-removal/samples/step008999.png ADDED

Git LFS Details

  • SHA256: 027609ac908b6a9b85c924191f6a3b810003b219cfbfdfe7014a05d16a5aaa5b
  • Pointer size: 132 Bytes
  • Size of remote file: 4.65 MB
downstream-shadow-removal/samples/step009999.png ADDED

Git LFS Details

  • SHA256: 73e7ebe106885f04570e75f8f08ffce10b1b262a230d606f22624bffcd3730d3
  • Pointer size: 132 Bytes
  • Size of remote file: 4.64 MB
downstream-shadow-removal/tensorboard/events.out.tfevents.1676208201.jason-system.49917.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:491cda9bfb4a46b18a438a1bbacd2e388cb63d31f930b3d473567adb85b850e9
3
+ size 3340690
downstream-shadow-removal/tensorboard/events.out.tfevents.1676256254.jason-system.54160.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:facf7f8557133fa989d8d1fa01dda6946b885368cbb6ca334dde150e7db61914
3
+ size 10672223
downstream-watermark-removal/ckpt/step099999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2af74105dabcebf32f3b700898727a4070b3d5afbe20a5e592d74c4495f0e132
3
+ size 160487111
downstream-watermark-removal/config-2023-02-12-14-28-38.yaml ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DATA:
2
+ DATAROOT: /data/LOGO/
3
+ IMG_SIZE: 256
4
+ NAME: LOGO_30K
5
+ DATALOADER:
6
+ BATCH_SIZE: 4
7
+ MICRO_BATCH: 0
8
+ NUM_WORKERS: 4
9
+ PIN_MEMORY: true
10
+ PREFETCH_FACTOR: 2
11
+ EVALUATE:
12
+ N_EVAL: 10000
13
+ MASK:
14
+ BRUSH_LENGTH_RATIO:
15
+ - 0.1
16
+ - 0.25
17
+ BRUSH_NUM: &id001
18
+ - 0
19
+ - 0
20
+ BRUSH_TURNS:
21
+ - 4
22
+ - 18
23
+ BRUSH_WIDTH_RATIO:
24
+ - 0.02
25
+ - 0.1
26
+ MASK_DIR: null
27
+ NOISE_DATASETS: []
28
+ RECT_LENGTH_RATIO:
29
+ - 0.2
30
+ - 0.8
31
+ RECT_NUM: *id001
32
+ SMOOTH_ITERS: 4
33
+ SMOOTH_KERNEL_SIZE: 15
34
+ SMOOTH_SIGMA: 4
35
+ MODEL:
36
+ MPN:
37
+ BASE_N_CHANNELS: 64
38
+ NECK_N_CHANNELS: 128
39
+ RIN:
40
+ BASE_N_CHANNELS: 32
41
+ NECK_N_CHANNELS: 128
42
+ WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
43
+ SAMPLE:
44
+ N_SAMPLES: 100
45
+ RANDOM: false
46
+ SAVE_DIR: ./samples/
47
+ SEED: 1234
48
+ TRAIN:
49
+ COEF_ADV: 0.001
50
+ COEF_IDMRF: 0.001
51
+ COEF_MPN: 2.0
52
+ COEF_REC: 1.4
53
+ COEF_SEMANTIC: 0.0001
54
+ EVAL_FREQ: 1000
55
+ OPTIM_DISC:
56
+ BETAS: &id002
57
+ - 0.5
58
+ - 0.9
59
+ LR: 1.0e-05
60
+ OPTIM_MPN:
61
+ BETAS: *id002
62
+ LR: 1.0e-05
63
+ OPTIM_RIN:
64
+ BETAS: *id002
65
+ LR: 1.0e-05
66
+ PRINT_FREQ: 100
67
+ RESUME: null
68
+ SAMPLE_FREQ: 1000
69
+ SAVE_FREQ: 5000
70
+ TRAIN_STEPS: 10000
downstream-watermark-removal/config-2023-02-12-15-41-28.yaml ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DATA:
2
+ DATAROOT: /data/LOGO/
3
+ IMG_SIZE: 256
4
+ NAME: LOGO_30K
5
+ DATALOADER:
6
+ BATCH_SIZE: 4
7
+ MICRO_BATCH: 0
8
+ NUM_WORKERS: 4
9
+ PIN_MEMORY: true
10
+ PREFETCH_FACTOR: 2
11
+ EVALUATE:
12
+ N_EVAL: 10000
13
+ MASK:
14
+ BRUSH_LENGTH_RATIO:
15
+ - 0.1
16
+ - 0.25
17
+ BRUSH_NUM: &id001
18
+ - 0
19
+ - 0
20
+ BRUSH_TURNS:
21
+ - 4
22
+ - 18
23
+ BRUSH_WIDTH_RATIO:
24
+ - 0.02
25
+ - 0.1
26
+ MASK_DIR: null
27
+ NOISE_DATASETS: []
28
+ RECT_LENGTH_RATIO:
29
+ - 0.2
30
+ - 0.8
31
+ RECT_NUM: *id001
32
+ SMOOTH_ITERS: 4
33
+ SMOOTH_KERNEL_SIZE: 15
34
+ SMOOTH_SIGMA: 4
35
+ MODEL:
36
+ MPN:
37
+ BASE_N_CHANNELS: 64
38
+ NECK_N_CHANNELS: 128
39
+ RIN:
40
+ BASE_N_CHANNELS: 32
41
+ NECK_N_CHANNELS: 128
42
+ WEIGHTS: ./runs/places-joint/ckpt/step149999.pt
43
+ SAMPLE:
44
+ N_SAMPLES: 100
45
+ RANDOM: false
46
+ SAVE_DIR: ./samples/
47
+ SEED: 1234
48
+ TRAIN:
49
+ COEF_ADV: 0.001
50
+ COEF_IDMRF: 0.001
51
+ COEF_MPN: 2.0
52
+ COEF_REC: 1.4
53
+ COEF_SEMANTIC: 0.0001
54
+ EVAL_FREQ: 1000
55
+ OPTIM_DISC:
56
+ BETAS: &id002
57
+ - 0.5
58
+ - 0.9
59
+ LR: 1.0e-05
60
+ OPTIM_MPN:
61
+ BETAS: *id002
62
+ LR: 1.0e-05
63
+ OPTIM_RIN:
64
+ BETAS: *id002
65
+ LR: 1.0e-05
66
+ PRINT_FREQ: 100
67
+ RESUME: latest
68
+ SAMPLE_FREQ: 1000
69
+ SAVE_FREQ: 5000
70
+ TRAIN_STEPS: 100000
downstream-watermark-removal/output-2023-02-12-14-28-38.log ADDED
@@ -0,0 +1,223 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-02-12 14:28:40,985 - INFO - Experiment directory: runs/downstream-watermark-removal
2
+ 2023-02-12 14:28:40,985 - INFO - Device: cuda
3
+ 2023-02-12 14:28:40,985 - INFO - Number of devices: 1
4
+ 2023-02-12 14:28:41,349 - INFO - Size of training set: 28352
5
+ 2023-02-12 14:28:41,349 - INFO - Size of validation set: 4051
6
+ 2023-02-12 14:28:41,349 - INFO - Batch size per device: 4
7
+ 2023-02-12 14:28:41,349 - INFO - Effective batch size: 4
8
+ 2023-02-12 14:28:42,908 - INFO - Successfully load mpn from ./runs/places-joint/ckpt/step149999.pt
9
+ 2023-02-12 14:28:42,913 - INFO - Successfully load rin from ./runs/places-joint/ckpt/step149999.pt
10
+ 2023-02-12 14:28:42,914 - INFO - Successfully load disc from ./runs/places-joint/ckpt/step149999.pt
11
+ 2023-02-12 14:28:42,916 - INFO - Successfully load pdisc from ./runs/places-joint/ckpt/step149999.pt
12
+ 2023-02-12 14:28:42,918 - INFO - Start training...
13
+ 2023-02-12 14:29:14,414 - INFO - [Train] step: 99, loss_adv_disc: 1.299828
14
+ 2023-02-12 14:29:14,621 - INFO - [Train] step: 99, loss_mpn: 0.037874, loss_rec: 0.041837, loss_semantic: 0.520840, loss_idmrf: 1.336823, loss_adv_gen: -128.098953
15
+ 2023-02-12 14:29:43,591 - INFO - [Train] step: 199, loss_adv_disc: -1.303024
16
+ 2023-02-12 14:29:43,800 - INFO - [Train] step: 199, loss_mpn: 0.036802, loss_rec: 0.040308, loss_semantic: 0.500949, loss_idmrf: 1.170398, loss_adv_gen: -188.719955
17
+ 2023-02-12 14:30:12,854 - INFO - [Train] step: 299, loss_adv_disc: 0.880611
18
+ 2023-02-12 14:30:13,062 - INFO - [Train] step: 299, loss_mpn: 0.022952, loss_rec: 0.032488, loss_semantic: 0.472941, loss_idmrf: 0.716021, loss_adv_gen: -164.942017
19
+ 2023-02-12 14:30:42,137 - INFO - [Train] step: 399, loss_adv_disc: 1.241273
20
+ 2023-02-12 14:30:42,346 - INFO - [Train] step: 399, loss_mpn: 0.020866, loss_rec: 0.041395, loss_semantic: 0.461653, loss_idmrf: 1.742797, loss_adv_gen: -94.545654
21
+ 2023-02-12 14:31:11,414 - INFO - [Train] step: 499, loss_adv_disc: -4.018159
22
+ 2023-02-12 14:31:11,624 - INFO - [Train] step: 499, loss_mpn: 0.042017, loss_rec: 0.038558, loss_semantic: 0.466815, loss_idmrf: 1.169562, loss_adv_gen: -143.637222
23
+ 2023-02-12 14:31:40,708 - INFO - [Train] step: 599, loss_adv_disc: 0.942124
24
+ 2023-02-12 14:31:40,916 - INFO - [Train] step: 599, loss_mpn: 0.012776, loss_rec: 0.029491, loss_semantic: 0.450419, loss_idmrf: 0.787979, loss_adv_gen: -97.693977
25
+ 2023-02-12 14:32:10,001 - INFO - [Train] step: 699, loss_adv_disc: 3.814938
26
+ 2023-02-12 14:32:10,211 - INFO - [Train] step: 699, loss_mpn: 0.028852, loss_rec: 0.043598, loss_semantic: 0.497849, loss_idmrf: 1.739692, loss_adv_gen: -155.648834
27
+ 2023-02-12 14:32:39,303 - INFO - [Train] step: 799, loss_adv_disc: -1.610433
28
+ 2023-02-12 14:32:39,512 - INFO - [Train] step: 799, loss_mpn: 0.029309, loss_rec: 0.045206, loss_semantic: 0.510565, loss_idmrf: 3.046881, loss_adv_gen: -96.266266
29
+ 2023-02-12 14:33:08,610 - INFO - [Train] step: 899, loss_adv_disc: 1.118233
30
+ 2023-02-12 14:33:08,819 - INFO - [Train] step: 899, loss_mpn: 0.012930, loss_rec: 0.022580, loss_semantic: 0.349494, loss_idmrf: 0.902996, loss_adv_gen: -88.348045
31
+ 2023-02-12 14:33:37,905 - INFO - [Train] step: 999, loss_adv_disc: 0.263744
32
+ 2023-02-12 14:33:38,114 - INFO - [Train] step: 999, loss_mpn: 0.023001, loss_rec: 0.043346, loss_semantic: 0.504102, loss_idmrf: 2.317935, loss_adv_gen: -102.413116
33
+ 2023-02-12 14:34:15,528 - INFO - [Eval] step: 999, bce: 0.338886, psnr: 26.837254, ssim: 0.944074
34
+ 2023-02-12 14:34:46,223 - INFO - [Train] step: 1099, loss_adv_disc: -2.614538
35
+ 2023-02-12 14:34:46,432 - INFO - [Train] step: 1099, loss_mpn: 0.049012, loss_rec: 0.041178, loss_semantic: 0.505731, loss_idmrf: 1.993219, loss_adv_gen: -61.293377
36
+ 2023-02-12 14:35:15,523 - INFO - [Train] step: 1199, loss_adv_disc: -3.433872
37
+ 2023-02-12 14:35:15,731 - INFO - [Train] step: 1199, loss_mpn: 0.033315, loss_rec: 0.036747, loss_semantic: 0.420346, loss_idmrf: 1.440817, loss_adv_gen: -119.410286
38
+ 2023-02-12 14:35:44,803 - INFO - [Train] step: 1299, loss_adv_disc: -1.207434
39
+ 2023-02-12 14:35:45,012 - INFO - [Train] step: 1299, loss_mpn: 0.015608, loss_rec: 0.038091, loss_semantic: 0.484080, loss_idmrf: 0.916523, loss_adv_gen: -110.312073
40
+ 2023-02-12 14:36:14,077 - INFO - [Train] step: 1399, loss_adv_disc: -5.074930
41
+ 2023-02-12 14:36:14,286 - INFO - [Train] step: 1399, loss_mpn: 0.040863, loss_rec: 0.052102, loss_semantic: 0.532582, loss_idmrf: 1.521286, loss_adv_gen: -111.052315
42
+ 2023-02-12 14:36:43,350 - INFO - [Train] step: 1499, loss_adv_disc: -1.698674
43
+ 2023-02-12 14:36:43,558 - INFO - [Train] step: 1499, loss_mpn: 0.019298, loss_rec: 0.017522, loss_semantic: 0.335085, loss_idmrf: 1.877787, loss_adv_gen: -124.831696
44
+ 2023-02-12 14:37:12,635 - INFO - [Train] step: 1599, loss_adv_disc: 0.078364
45
+ 2023-02-12 14:37:12,844 - INFO - [Train] step: 1599, loss_mpn: 0.032950, loss_rec: 0.033342, loss_semantic: 0.452769, loss_idmrf: 1.889498, loss_adv_gen: -102.988449
46
+ 2023-02-12 14:37:41,908 - INFO - [Train] step: 1699, loss_adv_disc: 3.190553
47
+ 2023-02-12 14:37:42,117 - INFO - [Train] step: 1699, loss_mpn: 0.024712, loss_rec: 0.037171, loss_semantic: 0.477129, loss_idmrf: 1.474838, loss_adv_gen: -107.239792
48
+ 2023-02-12 14:38:11,183 - INFO - [Train] step: 1799, loss_adv_disc: -0.911062
49
+ 2023-02-12 14:38:11,391 - INFO - [Train] step: 1799, loss_mpn: 0.018543, loss_rec: 0.032870, loss_semantic: 0.416096, loss_idmrf: 2.942292, loss_adv_gen: -97.946312
50
+ 2023-02-12 14:38:40,450 - INFO - [Train] step: 1899, loss_adv_disc: -2.992178
51
+ 2023-02-12 14:38:40,658 - INFO - [Train] step: 1899, loss_mpn: 0.044249, loss_rec: 0.053050, loss_semantic: 0.493338, loss_idmrf: 1.900263, loss_adv_gen: -102.013123
52
+ 2023-02-12 14:39:09,716 - INFO - [Train] step: 1999, loss_adv_disc: 1.107126
53
+ 2023-02-12 14:39:09,925 - INFO - [Train] step: 1999, loss_mpn: 0.049591, loss_rec: 0.039920, loss_semantic: 0.413558, loss_idmrf: 1.118119, loss_adv_gen: -138.656616
54
+ 2023-02-12 14:39:47,107 - INFO - [Eval] step: 1999, bce: 0.373344, psnr: 27.043365, ssim: 0.946821
55
+ 2023-02-12 14:40:17,430 - INFO - [Train] step: 2099, loss_adv_disc: -2.116822
56
+ 2023-02-12 14:40:17,638 - INFO - [Train] step: 2099, loss_mpn: 0.022930, loss_rec: 0.023332, loss_semantic: 0.383731, loss_idmrf: 1.122571, loss_adv_gen: -93.001617
57
+ 2023-02-12 14:40:46,705 - INFO - [Train] step: 2199, loss_adv_disc: 1.641629
58
+ 2023-02-12 14:40:46,913 - INFO - [Train] step: 2199, loss_mpn: 0.015344, loss_rec: 0.028494, loss_semantic: 0.344031, loss_idmrf: 1.670086, loss_adv_gen: -85.842712
59
+ 2023-02-12 14:41:15,969 - INFO - [Train] step: 2299, loss_adv_disc: -1.291130
60
+ 2023-02-12 14:41:16,177 - INFO - [Train] step: 2299, loss_mpn: 0.013970, loss_rec: 0.028013, loss_semantic: 0.418684, loss_idmrf: 1.546552, loss_adv_gen: -87.390312
61
+ 2023-02-12 14:41:45,234 - INFO - [Train] step: 2399, loss_adv_disc: 2.109378
62
+ 2023-02-12 14:41:45,443 - INFO - [Train] step: 2399, loss_mpn: 0.038823, loss_rec: 0.049957, loss_semantic: 0.547463, loss_idmrf: 1.906920, loss_adv_gen: -140.577087
63
+ 2023-02-12 14:42:14,492 - INFO - [Train] step: 2499, loss_adv_disc: -2.110101
64
+ 2023-02-12 14:42:14,700 - INFO - [Train] step: 2499, loss_mpn: 0.028055, loss_rec: 0.036691, loss_semantic: 0.471377, loss_idmrf: 1.584231, loss_adv_gen: -66.208374
65
+ 2023-02-12 14:42:43,743 - INFO - [Train] step: 2599, loss_adv_disc: 1.007918
66
+ 2023-02-12 14:42:43,951 - INFO - [Train] step: 2599, loss_mpn: 0.011720, loss_rec: 0.039426, loss_semantic: 0.476948, loss_idmrf: 1.416666, loss_adv_gen: -108.188354
67
+ 2023-02-12 14:43:12,995 - INFO - [Train] step: 2699, loss_adv_disc: 0.309885
68
+ 2023-02-12 14:43:13,203 - INFO - [Train] step: 2699, loss_mpn: 0.043115, loss_rec: 0.048996, loss_semantic: 0.506540, loss_idmrf: 1.823579, loss_adv_gen: -79.640060
69
+ 2023-02-12 14:43:42,259 - INFO - [Train] step: 2799, loss_adv_disc: 1.769973
70
+ 2023-02-12 14:43:42,470 - INFO - [Train] step: 2799, loss_mpn: 0.029426, loss_rec: 0.040913, loss_semantic: 0.455353, loss_idmrf: 1.369126, loss_adv_gen: -82.098961
71
+ 2023-02-12 14:44:11,518 - INFO - [Train] step: 2899, loss_adv_disc: -2.826468
72
+ 2023-02-12 14:44:11,727 - INFO - [Train] step: 2899, loss_mpn: 0.024427, loss_rec: 0.036017, loss_semantic: 0.450512, loss_idmrf: 1.423253, loss_adv_gen: -119.713104
73
+ 2023-02-12 14:44:40,782 - INFO - [Train] step: 2999, loss_adv_disc: -0.511095
74
+ 2023-02-12 14:44:40,992 - INFO - [Train] step: 2999, loss_mpn: 0.019680, loss_rec: 0.028514, loss_semantic: 0.381640, loss_idmrf: 1.594770, loss_adv_gen: -121.859711
75
+ 2023-02-12 14:45:18,178 - INFO - [Eval] step: 2999, bce: 0.325142, psnr: 27.303087, ssim: 0.947815
76
+ 2023-02-12 14:45:48,502 - INFO - [Train] step: 3099, loss_adv_disc: 0.043045
77
+ 2023-02-12 14:45:48,711 - INFO - [Train] step: 3099, loss_mpn: 0.016258, loss_rec: 0.021658, loss_semantic: 0.378096, loss_idmrf: 1.504422, loss_adv_gen: -198.545273
78
+ 2023-02-12 14:46:17,756 - INFO - [Train] step: 3199, loss_adv_disc: -1.038017
79
+ 2023-02-12 14:46:17,965 - INFO - [Train] step: 3199, loss_mpn: 0.012933, loss_rec: 0.026565, loss_semantic: 0.426411, loss_idmrf: 1.404650, loss_adv_gen: -96.168388
80
+ 2023-02-12 14:46:47,018 - INFO - [Train] step: 3299, loss_adv_disc: 0.046620
81
+ 2023-02-12 14:46:47,226 - INFO - [Train] step: 3299, loss_mpn: 0.011906, loss_rec: 0.014265, loss_semantic: 0.269762, loss_idmrf: 0.559204, loss_adv_gen: -99.662216
82
+ 2023-02-12 14:47:16,291 - INFO - [Train] step: 3399, loss_adv_disc: -0.021643
83
+ 2023-02-12 14:47:16,499 - INFO - [Train] step: 3399, loss_mpn: 0.017540, loss_rec: 0.029993, loss_semantic: 0.392757, loss_idmrf: 1.623783, loss_adv_gen: -72.550812
84
+ 2023-02-12 14:47:45,556 - INFO - [Train] step: 3499, loss_adv_disc: 0.026034
85
+ 2023-02-12 14:47:45,764 - INFO - [Train] step: 3499, loss_mpn: 0.024654, loss_rec: 0.025382, loss_semantic: 0.367133, loss_idmrf: 1.421693, loss_adv_gen: -106.288788
86
+ 2023-02-12 14:48:14,815 - INFO - [Train] step: 3599, loss_adv_disc: -0.349758
87
+ 2023-02-12 14:48:15,023 - INFO - [Train] step: 3599, loss_mpn: 0.032561, loss_rec: 0.033780, loss_semantic: 0.448922, loss_idmrf: 1.408511, loss_adv_gen: -133.010773
88
+ 2023-02-12 14:48:44,080 - INFO - [Train] step: 3699, loss_adv_disc: -3.845224
89
+ 2023-02-12 14:48:44,288 - INFO - [Train] step: 3699, loss_mpn: 0.046263, loss_rec: 0.035030, loss_semantic: 0.449073, loss_idmrf: 1.693964, loss_adv_gen: -101.824783
90
+ 2023-02-12 14:49:13,351 - INFO - [Train] step: 3799, loss_adv_disc: 0.691240
91
+ 2023-02-12 14:49:13,559 - INFO - [Train] step: 3799, loss_mpn: 0.033147, loss_rec: 0.032454, loss_semantic: 0.412963, loss_idmrf: 1.078437, loss_adv_gen: -133.369232
92
+ 2023-02-12 14:49:42,612 - INFO - [Train] step: 3899, loss_adv_disc: -1.092527
93
+ 2023-02-12 14:49:42,820 - INFO - [Train] step: 3899, loss_mpn: 0.030574, loss_rec: 0.035086, loss_semantic: 0.450594, loss_idmrf: 1.633320, loss_adv_gen: -157.005127
94
+ 2023-02-12 14:50:11,886 - INFO - [Train] step: 3999, loss_adv_disc: 0.675812
95
+ 2023-02-12 14:50:12,095 - INFO - [Train] step: 3999, loss_mpn: 0.021050, loss_rec: 0.033965, loss_semantic: 0.449186, loss_idmrf: 1.248645, loss_adv_gen: -97.598892
96
+ 2023-02-12 14:50:49,252 - INFO - [Eval] step: 3999, bce: 0.346579, psnr: 27.542393, ssim: 0.949793
97
+ 2023-02-12 14:51:19,568 - INFO - [Train] step: 4099, loss_adv_disc: 0.768379
98
+ 2023-02-12 14:51:19,777 - INFO - [Train] step: 4099, loss_mpn: 0.013409, loss_rec: 0.029262, loss_semantic: 0.393055, loss_idmrf: 1.478358, loss_adv_gen: -109.265968
99
+ 2023-02-12 14:51:48,839 - INFO - [Train] step: 4199, loss_adv_disc: -0.045594
100
+ 2023-02-12 14:51:49,047 - INFO - [Train] step: 4199, loss_mpn: 0.024171, loss_rec: 0.017396, loss_semantic: 0.286983, loss_idmrf: 0.789566, loss_adv_gen: -148.535095
101
+ 2023-02-12 14:52:18,101 - INFO - [Train] step: 4299, loss_adv_disc: -0.430947
102
+ 2023-02-12 14:52:18,310 - INFO - [Train] step: 4299, loss_mpn: 0.014691, loss_rec: 0.021662, loss_semantic: 0.326292, loss_idmrf: 0.946436, loss_adv_gen: -91.638748
103
+ 2023-02-12 14:52:47,360 - INFO - [Train] step: 4399, loss_adv_disc: -0.410857
104
+ 2023-02-12 14:52:47,568 - INFO - [Train] step: 4399, loss_mpn: 0.011707, loss_rec: 0.021423, loss_semantic: 0.334674, loss_idmrf: 0.605388, loss_adv_gen: -137.055222
105
+ 2023-02-12 14:53:16,621 - INFO - [Train] step: 4499, loss_adv_disc: -0.533930
106
+ 2023-02-12 14:53:16,829 - INFO - [Train] step: 4499, loss_mpn: 0.012462, loss_rec: 0.036778, loss_semantic: 0.428064, loss_idmrf: 1.709253, loss_adv_gen: -201.091476
107
+ 2023-02-12 14:53:45,873 - INFO - [Train] step: 4599, loss_adv_disc: -1.762612
108
+ 2023-02-12 14:53:46,082 - INFO - [Train] step: 4599, loss_mpn: 0.026616, loss_rec: 0.052115, loss_semantic: 0.480837, loss_idmrf: 1.606442, loss_adv_gen: -97.646194
109
+ 2023-02-12 14:54:15,132 - INFO - [Train] step: 4699, loss_adv_disc: -0.790706
110
+ 2023-02-12 14:54:15,340 - INFO - [Train] step: 4699, loss_mpn: 0.018837, loss_rec: 0.028844, loss_semantic: 0.392761, loss_idmrf: 1.836121, loss_adv_gen: -98.939705
111
+ 2023-02-12 14:54:44,391 - INFO - [Train] step: 4799, loss_adv_disc: -0.965508
112
+ 2023-02-12 14:54:44,601 - INFO - [Train] step: 4799, loss_mpn: 0.032487, loss_rec: 0.035681, loss_semantic: 0.437911, loss_idmrf: 1.694472, loss_adv_gen: -81.396034
113
+ 2023-02-12 14:55:13,651 - INFO - [Train] step: 4899, loss_adv_disc: -1.041402
114
+ 2023-02-12 14:55:13,860 - INFO - [Train] step: 4899, loss_mpn: 0.028312, loss_rec: 0.028561, loss_semantic: 0.396117, loss_idmrf: 1.250476, loss_adv_gen: -97.100685
115
+ 2023-02-12 14:55:42,913 - INFO - [Train] step: 4999, loss_adv_disc: 0.711361
116
+ 2023-02-12 14:55:43,122 - INFO - [Train] step: 4999, loss_mpn: 0.021475, loss_rec: 0.026028, loss_semantic: 0.379468, loss_idmrf: 1.322870, loss_adv_gen: -115.331512
117
+ 2023-02-12 14:56:20,330 - INFO - [Eval] step: 4999, bce: 0.351477, psnr: 27.780710, ssim: 0.950434
118
+ 2023-02-12 14:56:50,783 - INFO - [Train] step: 5099, loss_adv_disc: -2.985783
119
+ 2023-02-12 14:56:50,991 - INFO - [Train] step: 5099, loss_mpn: 0.019004, loss_rec: 0.041138, loss_semantic: 0.508738, loss_idmrf: 0.819962, loss_adv_gen: -182.782623
120
+ 2023-02-12 14:57:20,046 - INFO - [Train] step: 5199, loss_adv_disc: -1.645151
121
+ 2023-02-12 14:57:20,256 - INFO - [Train] step: 5199, loss_mpn: 0.016225, loss_rec: 0.025021, loss_semantic: 0.370635, loss_idmrf: 0.914956, loss_adv_gen: -103.670837
122
+ 2023-02-12 14:57:49,304 - INFO - [Train] step: 5299, loss_adv_disc: -0.446883
123
+ 2023-02-12 14:57:49,512 - INFO - [Train] step: 5299, loss_mpn: 0.025473, loss_rec: 0.035687, loss_semantic: 0.464093, loss_idmrf: 1.667971, loss_adv_gen: -153.179428
124
+ 2023-02-12 14:58:18,560 - INFO - [Train] step: 5399, loss_adv_disc: 0.041118
125
+ 2023-02-12 14:58:18,767 - INFO - [Train] step: 5399, loss_mpn: 0.022868, loss_rec: 0.027386, loss_semantic: 0.379084, loss_idmrf: 1.039361, loss_adv_gen: -118.277161
126
+ 2023-02-12 14:58:47,819 - INFO - [Train] step: 5499, loss_adv_disc: -0.695383
127
+ 2023-02-12 14:58:48,028 - INFO - [Train] step: 5499, loss_mpn: 0.017166, loss_rec: 0.023030, loss_semantic: 0.343284, loss_idmrf: 1.867031, loss_adv_gen: -132.320648
128
+ 2023-02-12 14:59:17,075 - INFO - [Train] step: 5599, loss_adv_disc: -0.396028
129
+ 2023-02-12 14:59:17,283 - INFO - [Train] step: 5599, loss_mpn: 0.016136, loss_rec: 0.024782, loss_semantic: 0.387950, loss_idmrf: 1.010486, loss_adv_gen: -124.445305
130
+ 2023-02-12 14:59:46,331 - INFO - [Train] step: 5699, loss_adv_disc: -2.364127
131
+ 2023-02-12 14:59:46,539 - INFO - [Train] step: 5699, loss_mpn: 0.022420, loss_rec: 0.027926, loss_semantic: 0.412665, loss_idmrf: 1.334894, loss_adv_gen: -62.740250
132
+ 2023-02-12 15:00:15,585 - INFO - [Train] step: 5799, loss_adv_disc: -0.426686
133
+ 2023-02-12 15:00:15,793 - INFO - [Train] step: 5799, loss_mpn: 0.020284, loss_rec: 0.029912, loss_semantic: 0.416979, loss_idmrf: 2.073402, loss_adv_gen: -149.003220
134
+ 2023-02-12 15:00:44,848 - INFO - [Train] step: 5899, loss_adv_disc: -0.774378
135
+ 2023-02-12 15:00:45,057 - INFO - [Train] step: 5899, loss_mpn: 0.013465, loss_rec: 0.031138, loss_semantic: 0.415459, loss_idmrf: 0.909214, loss_adv_gen: -113.032188
136
+ 2023-02-12 15:01:14,101 - INFO - [Train] step: 5999, loss_adv_disc: -0.389311
137
+ 2023-02-12 15:01:14,312 - INFO - [Train] step: 5999, loss_mpn: 0.039259, loss_rec: 0.035546, loss_semantic: 0.422057, loss_idmrf: 1.271151, loss_adv_gen: -117.240402
138
+ 2023-02-12 15:01:51,501 - INFO - [Eval] step: 5999, bce: 0.340868, psnr: 27.871330, ssim: 0.950925
139
+ 2023-02-12 15:02:21,828 - INFO - [Train] step: 6099, loss_adv_disc: -0.167378
140
+ 2023-02-12 15:02:22,037 - INFO - [Train] step: 6099, loss_mpn: 0.016576, loss_rec: 0.036952, loss_semantic: 0.467681, loss_idmrf: 1.342181, loss_adv_gen: -104.226501
141
+ 2023-02-12 15:02:51,091 - INFO - [Train] step: 6199, loss_adv_disc: -0.683909
142
+ 2023-02-12 15:02:51,300 - INFO - [Train] step: 6199, loss_mpn: 0.018274, loss_rec: 0.032154, loss_semantic: 0.425463, loss_idmrf: 0.816652, loss_adv_gen: -86.657425
143
+ 2023-02-12 15:03:20,351 - INFO - [Train] step: 6299, loss_adv_disc: -1.493726
144
+ 2023-02-12 15:03:20,560 - INFO - [Train] step: 6299, loss_mpn: 0.023145, loss_rec: 0.033411, loss_semantic: 0.432057, loss_idmrf: 0.897825, loss_adv_gen: -139.223816
145
+ 2023-02-12 15:03:49,616 - INFO - [Train] step: 6399, loss_adv_disc: -0.277531
146
+ 2023-02-12 15:03:49,824 - INFO - [Train] step: 6399, loss_mpn: 0.017375, loss_rec: 0.029168, loss_semantic: 0.402245, loss_idmrf: 0.622658, loss_adv_gen: -142.869843
147
+ 2023-02-12 15:04:18,892 - INFO - [Train] step: 6499, loss_adv_disc: -0.235311
148
+ 2023-02-12 15:04:19,101 - INFO - [Train] step: 6499, loss_mpn: 0.012269, loss_rec: 0.022133, loss_semantic: 0.319729, loss_idmrf: 1.260930, loss_adv_gen: -171.630920
149
+ 2023-02-12 15:04:48,160 - INFO - [Train] step: 6599, loss_adv_disc: -0.469994
150
+ 2023-02-12 15:04:48,370 - INFO - [Train] step: 6599, loss_mpn: 0.008784, loss_rec: 0.019133, loss_semantic: 0.306887, loss_idmrf: 0.801923, loss_adv_gen: -64.075722
151
+ 2023-02-12 15:05:17,426 - INFO - [Train] step: 6699, loss_adv_disc: -1.466449
152
+ 2023-02-12 15:05:17,634 - INFO - [Train] step: 6699, loss_mpn: 0.021287, loss_rec: 0.035671, loss_semantic: 0.454560, loss_idmrf: 1.030252, loss_adv_gen: -120.588272
153
+ 2023-02-12 15:05:46,695 - INFO - [Train] step: 6799, loss_adv_disc: 0.515251
154
+ 2023-02-12 15:05:46,904 - INFO - [Train] step: 6799, loss_mpn: 0.013280, loss_rec: 0.027942, loss_semantic: 0.360047, loss_idmrf: 1.654778, loss_adv_gen: -180.273392
155
+ 2023-02-12 15:06:15,963 - INFO - [Train] step: 6899, loss_adv_disc: -0.489713
156
+ 2023-02-12 15:06:16,171 - INFO - [Train] step: 6899, loss_mpn: 0.018902, loss_rec: 0.038339, loss_semantic: 0.442717, loss_idmrf: 2.431321, loss_adv_gen: -115.624718
157
+ 2023-02-12 15:06:45,239 - INFO - [Train] step: 6999, loss_adv_disc: -1.769455
158
+ 2023-02-12 15:06:45,447 - INFO - [Train] step: 6999, loss_mpn: 0.019381, loss_rec: 0.028560, loss_semantic: 0.383751, loss_idmrf: 1.163507, loss_adv_gen: -153.030273
159
+ 2023-02-12 15:07:22,635 - INFO - [Eval] step: 6999, bce: 0.306903, psnr: 28.013441, ssim: 0.951795
160
+ 2023-02-12 15:07:53,174 - INFO - [Train] step: 7099, loss_adv_disc: 1.659884
161
+ 2023-02-12 15:07:53,383 - INFO - [Train] step: 7099, loss_mpn: 0.028287, loss_rec: 0.037931, loss_semantic: 0.468431, loss_idmrf: 1.497301, loss_adv_gen: -132.254868
162
+ 2023-02-12 15:08:22,426 - INFO - [Train] step: 7199, loss_adv_disc: 0.472226
163
+ 2023-02-12 15:08:22,636 - INFO - [Train] step: 7199, loss_mpn: 0.013172, loss_rec: 0.024137, loss_semantic: 0.364234, loss_idmrf: 1.162045, loss_adv_gen: -132.014572
164
+ 2023-02-12 15:08:51,699 - INFO - [Train] step: 7299, loss_adv_disc: -0.581740
165
+ 2023-02-12 15:08:51,908 - INFO - [Train] step: 7299, loss_mpn: 0.012158, loss_rec: 0.031973, loss_semantic: 0.362457, loss_idmrf: 1.268194, loss_adv_gen: -137.127380
166
+ 2023-02-12 15:09:20,969 - INFO - [Train] step: 7399, loss_adv_disc: -0.978511
167
+ 2023-02-12 15:09:21,178 - INFO - [Train] step: 7399, loss_mpn: 0.022330, loss_rec: 0.024040, loss_semantic: 0.309038, loss_idmrf: 0.792379, loss_adv_gen: -106.908493
168
+ 2023-02-12 15:09:50,228 - INFO - [Train] step: 7499, loss_adv_disc: -0.073679
169
+ 2023-02-12 15:09:50,436 - INFO - [Train] step: 7499, loss_mpn: 0.018063, loss_rec: 0.022740, loss_semantic: 0.319768, loss_idmrf: 1.349909, loss_adv_gen: -165.113327
170
+ 2023-02-12 15:10:19,481 - INFO - [Train] step: 7599, loss_adv_disc: 0.049674
171
+ 2023-02-12 15:10:19,689 - INFO - [Train] step: 7599, loss_mpn: 0.010805, loss_rec: 0.025964, loss_semantic: 0.350963, loss_idmrf: 0.961297, loss_adv_gen: -113.559540
172
+ 2023-02-12 15:10:48,747 - INFO - [Train] step: 7699, loss_adv_disc: -0.608275
173
+ 2023-02-12 15:10:48,955 - INFO - [Train] step: 7699, loss_mpn: 0.016553, loss_rec: 0.027370, loss_semantic: 0.380454, loss_idmrf: 1.702928, loss_adv_gen: -108.241295
174
+ 2023-02-12 15:11:18,013 - INFO - [Train] step: 7799, loss_adv_disc: 1.413847
175
+ 2023-02-12 15:11:18,221 - INFO - [Train] step: 7799, loss_mpn: 0.021216, loss_rec: 0.033353, loss_semantic: 0.390046, loss_idmrf: 1.304851, loss_adv_gen: -118.235794
176
+ 2023-02-12 15:11:47,280 - INFO - [Train] step: 7899, loss_adv_disc: -0.615377
177
+ 2023-02-12 15:11:47,489 - INFO - [Train] step: 7899, loss_mpn: 0.016149, loss_rec: 0.024272, loss_semantic: 0.358250, loss_idmrf: 0.852178, loss_adv_gen: -133.236221
178
+ 2023-02-12 15:12:16,555 - INFO - [Train] step: 7999, loss_adv_disc: -0.062408
179
+ 2023-02-12 15:12:16,764 - INFO - [Train] step: 7999, loss_mpn: 0.013659, loss_rec: 0.026106, loss_semantic: 0.386249, loss_idmrf: 1.021867, loss_adv_gen: -147.716049
180
+ 2023-02-12 15:12:53,951 - INFO - [Eval] step: 7999, bce: 0.288697, psnr: 28.163181, ssim: 0.952470
181
+ 2023-02-12 15:13:24,273 - INFO - [Train] step: 8099, loss_adv_disc: -1.120922
182
+ 2023-02-12 15:13:24,481 - INFO - [Train] step: 8099, loss_mpn: 0.004955, loss_rec: 0.020915, loss_semantic: 0.322940, loss_idmrf: 1.106715, loss_adv_gen: -151.821106
183
+ 2023-02-12 15:13:53,543 - INFO - [Train] step: 8199, loss_adv_disc: 0.029411
184
+ 2023-02-12 15:13:53,753 - INFO - [Train] step: 8199, loss_mpn: 0.026964, loss_rec: 0.028958, loss_semantic: 0.445343, loss_idmrf: 1.209411, loss_adv_gen: -102.843552
185
+ 2023-02-12 15:14:22,805 - INFO - [Train] step: 8299, loss_adv_disc: -4.315375
186
+ 2023-02-12 15:14:23,012 - INFO - [Train] step: 8299, loss_mpn: 0.019549, loss_rec: 0.034941, loss_semantic: 0.462220, loss_idmrf: 0.890771, loss_adv_gen: -154.584610
187
+ 2023-02-12 15:14:52,068 - INFO - [Train] step: 8399, loss_adv_disc: -1.995814
188
+ 2023-02-12 15:14:52,276 - INFO - [Train] step: 8399, loss_mpn: 0.012006, loss_rec: 0.034098, loss_semantic: 0.396413, loss_idmrf: 0.811635, loss_adv_gen: -96.562088
189
+ 2023-02-12 15:15:21,323 - INFO - [Train] step: 8499, loss_adv_disc: 0.820693
190
+ 2023-02-12 15:15:21,532 - INFO - [Train] step: 8499, loss_mpn: 0.013124, loss_rec: 0.029301, loss_semantic: 0.393588, loss_idmrf: 1.013259, loss_adv_gen: -89.326385
191
+ 2023-02-12 15:15:50,592 - INFO - [Train] step: 8599, loss_adv_disc: 0.393147
192
+ 2023-02-12 15:15:50,800 - INFO - [Train] step: 8599, loss_mpn: 0.011680, loss_rec: 0.017792, loss_semantic: 0.300831, loss_idmrf: 0.998240, loss_adv_gen: -184.119385
193
+ 2023-02-12 15:16:19,842 - INFO - [Train] step: 8699, loss_adv_disc: -0.531485
194
+ 2023-02-12 15:16:20,051 - INFO - [Train] step: 8699, loss_mpn: 0.015221, loss_rec: 0.025195, loss_semantic: 0.341359, loss_idmrf: 1.062933, loss_adv_gen: -110.207268
195
+ 2023-02-12 15:16:49,123 - INFO - [Train] step: 8799, loss_adv_disc: -0.491042
196
+ 2023-02-12 15:16:49,332 - INFO - [Train] step: 8799, loss_mpn: 0.010106, loss_rec: 0.024656, loss_semantic: 0.360912, loss_idmrf: 0.767988, loss_adv_gen: -151.354294
197
+ 2023-02-12 15:17:18,385 - INFO - [Train] step: 8899, loss_adv_disc: -0.419860
198
+ 2023-02-12 15:17:18,594 - INFO - [Train] step: 8899, loss_mpn: 0.013824, loss_rec: 0.030028, loss_semantic: 0.417085, loss_idmrf: 1.085957, loss_adv_gen: -207.431549
199
+ 2023-02-12 15:17:47,642 - INFO - [Train] step: 8999, loss_adv_disc: -0.628583
200
+ 2023-02-12 15:17:47,850 - INFO - [Train] step: 8999, loss_mpn: 0.024702, loss_rec: 0.037819, loss_semantic: 0.416768, loss_idmrf: 2.331512, loss_adv_gen: -108.836723
201
+ 2023-02-12 15:18:25,047 - INFO - [Eval] step: 8999, bce: 0.302010, psnr: 28.254166, ssim: 0.952631
202
+ 2023-02-12 15:18:55,405 - INFO - [Train] step: 9099, loss_adv_disc: 0.632391
203
+ 2023-02-12 15:18:55,616 - INFO - [Train] step: 9099, loss_mpn: 0.014931, loss_rec: 0.024709, loss_semantic: 0.325625, loss_idmrf: 1.853235, loss_adv_gen: -115.922569
204
+ 2023-02-12 15:19:24,673 - INFO - [Train] step: 9199, loss_adv_disc: -1.815902
205
+ 2023-02-12 15:19:24,881 - INFO - [Train] step: 9199, loss_mpn: 0.016712, loss_rec: 0.022784, loss_semantic: 0.345748, loss_idmrf: 1.113174, loss_adv_gen: -51.955341
206
+ 2023-02-12 15:19:53,938 - INFO - [Train] step: 9299, loss_adv_disc: -0.156092
207
+ 2023-02-12 15:19:54,147 - INFO - [Train] step: 9299, loss_mpn: 0.011803, loss_rec: 0.023657, loss_semantic: 0.339971, loss_idmrf: 1.318946, loss_adv_gen: -127.398438
208
+ 2023-02-12 15:20:23,205 - INFO - [Train] step: 9399, loss_adv_disc: -0.345600
209
+ 2023-02-12 15:20:23,415 - INFO - [Train] step: 9399, loss_mpn: 0.036090, loss_rec: 0.027977, loss_semantic: 0.409671, loss_idmrf: 1.271996, loss_adv_gen: -96.280853
210
+ 2023-02-12 15:20:52,463 - INFO - [Train] step: 9499, loss_adv_disc: -0.154040
211
+ 2023-02-12 15:20:52,671 - INFO - [Train] step: 9499, loss_mpn: 0.008822, loss_rec: 0.023087, loss_semantic: 0.333166, loss_idmrf: 1.134625, loss_adv_gen: -172.469971
212
+ 2023-02-12 15:21:21,728 - INFO - [Train] step: 9599, loss_adv_disc: -2.609488
213
+ 2023-02-12 15:21:21,937 - INFO - [Train] step: 9599, loss_mpn: 0.018622, loss_rec: 0.031683, loss_semantic: 0.426192, loss_idmrf: 0.870782, loss_adv_gen: -63.231483
214
+ 2023-02-12 15:21:50,997 - INFO - [Train] step: 9699, loss_adv_disc: -1.726371
215
+ 2023-02-12 15:21:51,205 - INFO - [Train] step: 9699, loss_mpn: 0.016741, loss_rec: 0.030167, loss_semantic: 0.398618, loss_idmrf: 1.257529, loss_adv_gen: -196.563248
216
+ 2023-02-12 15:22:20,258 - INFO - [Train] step: 9799, loss_adv_disc: -2.347822
217
+ 2023-02-12 15:22:20,466 - INFO - [Train] step: 9799, loss_mpn: 0.021154, loss_rec: 0.028975, loss_semantic: 0.403203, loss_idmrf: 0.936794, loss_adv_gen: -92.555832
218
+ 2023-02-12 15:22:49,519 - INFO - [Train] step: 9899, loss_adv_disc: -0.370060
219
+ 2023-02-12 15:22:49,728 - INFO - [Train] step: 9899, loss_mpn: 0.013475, loss_rec: 0.026137, loss_semantic: 0.385814, loss_idmrf: 0.893062, loss_adv_gen: -135.118530
220
+ 2023-02-12 15:23:18,785 - INFO - [Train] step: 9999, loss_adv_disc: -1.345161
221
+ 2023-02-12 15:23:18,993 - INFO - [Train] step: 9999, loss_mpn: 0.022469, loss_rec: 0.028983, loss_semantic: 0.421686, loss_idmrf: 1.735447, loss_adv_gen: -143.855499
222
+ 2023-02-12 15:23:56,181 - INFO - [Eval] step: 9999, bce: 0.237329, psnr: 28.375944, ssim: 0.953470
223
+ 2023-02-12 15:23:57,541 - INFO - End of training
downstream-watermark-removal/output-2023-02-12-15-41-28.log ADDED
The diff for this file is too large to render. See raw diff
 
downstream-watermark-removal/samples/step009999.png ADDED

Git LFS Details

  • SHA256: 4da650e215f32583d010618439f2250f048c589606777a81331f5782e92f8aa7
  • Pointer size: 132 Bytes
  • Size of remote file: 4.21 MB
downstream-watermark-removal/samples/step099999.png ADDED

Git LFS Details

  • SHA256: 9f1c9f7d02a43c2a6c9e511ffb9b25c33bc8f9da7a523a547be9d2cf3e1ab4be
  • Pointer size: 132 Bytes
  • Size of remote file: 4.21 MB
downstream-watermark-removal/tensorboard/events.out.tfevents.1676212121.jason-system.50864.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72f66be949f0e94261734eec3cbad480c18474ca8baa123a2a45c385766c716d
3
+ size 3340690
downstream-watermark-removal/tensorboard/events.out.tfevents.1676216490.jason-system.51652.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e75c5e9ee010d9336b6c412ca566ae68faaedcece11d58e810a0f37f1a197be9
3
+ size 30574858
joint_ffhq_brush_realnoise/ckpt/step149999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
3
+ size 425
joint_ffhq_brush_realnoise/ckpt/step149999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d809ab838d9255922d407abef2cba547971c2a927491a17be7c50c2a330c4197
3
+ size 53562757
joint_ffhq_brush_realnoise/ckpt/step149999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48843df6a56026b7b32b5df254a7920bfd82ec587d071398a38d2e620fd29b5d
3
+ size 106908613
joint_ffhq_brush_realnoise/config-2023-03-20-11-52-44.yaml ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ dataroot: /amax/xyf/data/FFHQ/
3
+ img_size: 256
4
+ name: FFHQ
5
+ dataloader:
6
+ micro_batch: 0
7
+ num_workers: 4
8
+ pin_memory: true
9
+ prefetch_factor: 2
10
+ mask:
11
+ mask_type: brush
12
+ noise_type: real
13
+ real_dataset:
14
+ - dataroot: /amax/xyf/data/CelebA-HQ/
15
+ img_size: 256
16
+ name: CelebA-HQ
17
+ - dataroot: /amax/xyf/data/ImageNet/
18
+ img_size: 256
19
+ name: ImageNet
20
+ smooth_type: iterative_gaussian
21
+ model:
22
+ mpn:
23
+ base_n_channels: 64
24
+ neck_n_channels: 128
25
+ rin:
26
+ base_n_channels: 32
27
+ neck_n_channels: 128
28
+ seed: 1234
29
+ train:
30
+ batch_size: 4
31
+ coef_adv: 0.001
32
+ coef_idmrf: 0.001
33
+ coef_mpn: 2.0
34
+ coef_rec: 1.4
35
+ coef_semantic: 0.0001
36
+ eval_freq: 1000
37
+ n_steps: 150000
38
+ optim_disc:
39
+ betas:
40
+ - 0.5
41
+ - 0.9
42
+ lr: 1.0e-05
43
+ optim_mpn:
44
+ betas:
45
+ - 0.5
46
+ - 0.9
47
+ lr: 1.0e-05
48
+ optim_rin:
49
+ betas:
50
+ - 0.5
51
+ - 0.9
52
+ lr: 1.0e-05
53
+ pretrained: ./runs/separate_ffhq_brush_realnoise/ckpt/step079999/model.pt
54
+ print_freq: 100
55
+ resume: null
56
+ sample_freq: 1000
57
+ save_freq: 5000
joint_ffhq_brush_realnoise/output-2023-03-20-11-52-44.log ADDED
The diff for this file is too large to render. See raw diff
 
joint_ffhq_brush_realnoise/samples/step149999.png ADDED

Git LFS Details

  • SHA256: 890e8b4c0999852a6e792efd4611a3a4ff9e9d307467b4bcfccf0d32e3e09662
  • Pointer size: 132 Bytes
  • Size of remote file: 5.04 MB
joint_ffhq_brush_realnoise/tensorboard/events.out.tfevents.1679284370.admin.cluster.local.35162.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6ba10d8f4626a47316c7d9f6174a6335bb323bbb5add80852704d3b7589d7b6
3
+ size 74423266
joint_ffhq_maskdir_realnoise/ckpt/step149999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
3
+ size 425
joint_ffhq_maskdir_realnoise/ckpt/step149999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70bbc5fd9861e9f8898e85a15932d529b39c3e72d9dd85671754d09250ff1063
3
+ size 53562757
joint_ffhq_maskdir_realnoise/ckpt/step149999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4bc75ed20245fd2b74310ee4f6d2f5a66e52a25cd348184ace9d1a8215b8155
3
+ size 106908613
joint_ffhq_maskdir_realnoise/config-2023-03-20-11-55-31.yaml ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ dataroot: /amax/xyf/data/FFHQ/
3
+ img_size: 256
4
+ name: FFHQ
5
+ dataloader:
6
+ micro_batch: 0
7
+ num_workers: 4
8
+ pin_memory: true
9
+ prefetch_factor: 2
10
+ mask:
11
+ dir_invert_color: true
12
+ dir_path: /amax/xyf/data/NVIDIAIrregularMaskDataset/train/
13
+ mask_type: dir
14
+ noise_type: real
15
+ real_dataset:
16
+ - dataroot: /amax/xyf/data/CelebA-HQ/
17
+ img_size: 256
18
+ name: CelebA-HQ
19
+ - dataroot: /amax/xyf/data/ImageNet/
20
+ img_size: 256
21
+ name: ImageNet
22
+ smooth_kernel_size: 13
23
+ smooth_sigma: 3
24
+ smooth_type: gaussian
25
+ model:
26
+ mpn:
27
+ base_n_channels: 64
28
+ neck_n_channels: 128
29
+ rin:
30
+ base_n_channels: 32
31
+ neck_n_channels: 128
32
+ seed: 1234
33
+ train:
34
+ batch_size: 4
35
+ coef_adv: 0.001
36
+ coef_idmrf: 0.001
37
+ coef_mpn: 2.0
38
+ coef_rec: 1.4
39
+ coef_semantic: 0.0001
40
+ eval_freq: 1000
41
+ n_steps: 150000
42
+ optim_disc:
43
+ betas:
44
+ - 0.5
45
+ - 0.9
46
+ lr: 1.0e-05
47
+ optim_mpn:
48
+ betas:
49
+ - 0.5
50
+ - 0.9
51
+ lr: 1.0e-05
52
+ optim_rin:
53
+ betas:
54
+ - 0.5
55
+ - 0.9
56
+ lr: 1.0e-05
57
+ pretrained: ./runs/separate_ffhq_maskdir_realnoise/ckpt/step079999/model.pt
58
+ print_freq: 100
59
+ resume: null
60
+ sample_freq: 1000
61
+ save_freq: 5000
joint_ffhq_maskdir_realnoise/output-2023-03-20-11-55-31.log ADDED
The diff for this file is too large to render. See raw diff
 
joint_ffhq_maskdir_realnoise/samples/step149999.png ADDED

Git LFS Details

  • SHA256: ae0633e510b1e0ac513f6ba542cd9aea4127e99e535028d968666793347d965c
  • Pointer size: 132 Bytes
  • Size of remote file: 4.79 MB
joint_ffhq_maskdir_realnoise/tensorboard/events.out.tfevents.1679284533.admin.cluster.local.727.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ada1bcfab3f9ee5ff8c747daa086e98cdc3d3a368d73e703a3a0ff7a1c3c6b50
3
+ size 74423266
joint_imagenet_brush_realnoise/ckpt/step149999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
3
+ size 425
joint_imagenet_brush_realnoise/ckpt/step149999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:248ca4c17d6ef0894d98d0bbf7556c025362489ba27800b2b175e45707e4b604
3
+ size 53562757
joint_imagenet_brush_realnoise/ckpt/step149999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09d31cf2963d5ace5747cc410bbfac3c66b5263fd75ee3850f7d332f2ba1343e
3
+ size 106908613
joint_imagenet_brush_realnoise/config-2023-03-22-15-20-53.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ dataroot: /amax/xyf/data/ImageNet/
3
+ img_size: 256
4
+ name: ImageNet
5
+ dataloader:
6
+ micro_batch: 0
7
+ num_workers: 4
8
+ pin_memory: true
9
+ prefetch_factor: 2
10
+ mask:
11
+ mask_type: brush
12
+ noise_type: real
13
+ real_dataset:
14
+ - dataroot: /amax/xyf/data/Places365/
15
+ img_size: 256
16
+ name: Places365
17
+ smooth_type: iterative_gaussian
18
+ model:
19
+ mpn:
20
+ base_n_channels: 64
21
+ neck_n_channels: 128
22
+ rin:
23
+ base_n_channels: 32
24
+ neck_n_channels: 128
25
+ seed: 1234
26
+ train:
27
+ batch_size: 4
28
+ coef_adv: 0.001
29
+ coef_idmrf: 0.001
30
+ coef_mpn: 2.0
31
+ coef_rec: 1.4
32
+ coef_semantic: 0.0001
33
+ eval_freq: 1000
34
+ n_steps: 150000
35
+ optim_disc:
36
+ betas:
37
+ - 0.5
38
+ - 0.9
39
+ lr: 1.0e-05
40
+ optim_mpn:
41
+ betas:
42
+ - 0.5
43
+ - 0.9
44
+ lr: 1.0e-05
45
+ optim_rin:
46
+ betas:
47
+ - 0.5
48
+ - 0.9
49
+ lr: 1.0e-05
50
+ pretrained: ./runs/separate_imagenet_brush_realnoise/ckpt/step149999/model.pt
51
+ print_freq: 100
52
+ resume: null
53
+ sample_freq: 1000
54
+ save_freq: 5000
joint_imagenet_brush_realnoise/output-2023-03-22-15-20-53.log ADDED
The diff for this file is too large to render. See raw diff
 
joint_imagenet_brush_realnoise/samples/step149999.png ADDED

Git LFS Details

  • SHA256: 18f9bc5e45bef8f05d80db275a56586e2f9b2b93b603b5958dd46cba9aa56003
  • Pointer size: 132 Bytes
  • Size of remote file: 5.72 MB
joint_imagenet_brush_realnoise/tensorboard/events.out.tfevents.1679469675.admin.cluster.local.8096.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfbcea7719488dd19a69e053cbd84bd0e8bcc89471b67219ebf9fd2541691c98
3
+ size 74430750
joint_imagenet_maskdir_realnoise/ckpt/step149999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d73f91e6f9f0e8222c08d532226feb041d885e7aed885e0536b91a6887793879
3
+ size 425
joint_imagenet_maskdir_realnoise/ckpt/step149999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02f470d4182208054de483d2729e56f03904a2e64526f9fb61362dc7a69add86
3
+ size 53562757