Hannes Kuchelmeister commited on
Commit
400abb9
1 Parent(s): d056ddb

add simple conv model with relu and batchnorm

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. configs/experiment/focusConvReLU_MSE_150.yaml +40 -0
  2. configs/model/focusConvReLU_150.yaml +13 -0
  3. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/config.yaml +81 -0
  4. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/hydra.yaml +170 -0
  5. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/overrides.yaml +1 -0
  6. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/config.yaml +81 -0
  7. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/hydra.yaml +170 -0
  8. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/overrides.yaml +1 -0
  9. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/config.yaml +81 -0
  10. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/hydra.yaml +170 -0
  11. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/overrides.yaml +1 -0
  12. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/config.yaml +81 -0
  13. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/hydra.yaml +170 -0
  14. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/overrides.yaml +1 -0
  15. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/epoch_096.ckpt +3 -0
  16. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/last.ckpt +3 -0
  17. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/hparams.yaml +55 -0
  18. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/metrics.csv +202 -0
  19. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652279809.c49d1a46d86b.1.0 +3 -0
  20. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652280195.c49d1a46d86b.1.1 +3 -0
  21. logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/hparams.yaml +55 -0
  22. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/meta.yaml +15 -0
  23. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/epoch +201 -0
  24. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/loss +1 -0
  25. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/mae +1 -0
  26. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/loss +100 -0
  27. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/mae +100 -0
  28. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/loss +100 -0
  29. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae +100 -0
  30. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae_best +100 -0
  31. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/_target_ +1 -0
  32. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/min_delta +1 -0
  33. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/mode +1 -0
  34. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/monitor +1 -0
  35. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/patience +1 -0
  36. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/_target_ +1 -0
  37. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/auto_insert_metric_name +1 -0
  38. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/dirpath +1 -0
  39. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/filename +1 -0
  40. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/mode +1 -0
  41. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/monitor +1 -0
  42. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_last +1 -0
  43. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_top_k +1 -0
  44. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/verbose +1 -0
  45. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/_target_ +1 -0
  46. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/max_depth +1 -0
  47. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/rich_progress_bar/_target_ +1 -0
  48. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/_target_ +1 -0
  49. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/augmentation +1 -0
  50. logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/batch_size +1 -0
configs/experiment/focusConvReLU_MSE_150.yaml ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # @package _global_
2
+
3
+ # to execute this experiment run:
4
+ # python train.py experiment=example
5
+
6
+ defaults:
7
+ - override /datamodule: focus150.yaml
8
+ - override /model: focusConvReLU_150.yaml
9
+ - override /callbacks: default.yaml
10
+ - override /logger: many_loggers
11
+ - override /trainer: default.yaml
12
+
13
+ # all parameters below will be merged with parameters from default configurations set above
14
+ # this allows you to overwrite only specified parameters
15
+
16
+ # name of the run determines folder name in logs
17
+ name: "focusConvReLU_MSE_150"
18
+ seed: 12345
19
+
20
+ trainer:
21
+ min_epochs: 1
22
+ max_epochs: 100
23
+
24
+ model:
25
+ image_size: 150
26
+ pool_size: 2
27
+ conv1_size: 3
28
+ conv1_channels: 9
29
+ conv2_size: 7
30
+ conv2_channels: 6
31
+ lin1_size: 32
32
+ lin2_size: 72
33
+ output_size: 1
34
+ lr: 0.001
35
+ weight_decay: 0.0005
36
+
37
+ datamodule:
38
+ batch_size: 64
39
+ augmentation: True
40
+
configs/model/focusConvReLU_150.yaml ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
2
+
3
+ image_size: 150
4
+ pool_size: 2
5
+ conv1_size: 5
6
+ conv1_channels: 6
7
+ conv2_size: 5
8
+ conv2_channels: 16
9
+ lin1_size: 100
10
+ lin2_size: 80
11
+ output_size: 1
12
+ lr: 0.001
13
+ weight_decay: 0.0005
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/config.yaml ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: 12345
8
+ name: focusConvReLU_MSE_150
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_train_file: ${data_dir}/focus150/train_metadata.csv
13
+ csv_val_file: ${data_dir}/focus150/validation_metadata.csv
14
+ csv_test_file: ${data_dir}/focus150/test_metadata.csv
15
+ batch_size: 64
16
+ num_workers: 0
17
+ pin_memory: false
18
+ augmentation: true
19
+ model:
20
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
21
+ image_size: 150
22
+ pool_size: 2
23
+ conv1_size: 3
24
+ conv1_channels: 9
25
+ conv2_size: 7
26
+ conv2_channels: 6
27
+ lin1_size: 32
28
+ lin2_size: 72
29
+ output_size: 1
30
+ lr: 0.001
31
+ weight_decay: 0.0005
32
+ callbacks:
33
+ model_checkpoint:
34
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
35
+ monitor: val/mae
36
+ mode: min
37
+ save_top_k: 1
38
+ save_last: true
39
+ verbose: false
40
+ dirpath: checkpoints/
41
+ filename: epoch_{epoch:03d}
42
+ auto_insert_metric_name: false
43
+ early_stopping:
44
+ _target_: pytorch_lightning.callbacks.EarlyStopping
45
+ monitor: val/mae
46
+ mode: min
47
+ patience: 100
48
+ min_delta: 0
49
+ model_summary:
50
+ _target_: pytorch_lightning.callbacks.RichModelSummary
51
+ max_depth: -1
52
+ rich_progress_bar:
53
+ _target_: pytorch_lightning.callbacks.RichProgressBar
54
+ logger:
55
+ csv:
56
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
57
+ save_dir: .
58
+ name: csv/
59
+ prefix: ''
60
+ mlflow:
61
+ _target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
62
+ experiment_name: ${name}
63
+ tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
64
+ tags: null
65
+ save_dir: ./mlruns
66
+ prefix: ''
67
+ artifact_location: null
68
+ tensorboard:
69
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
70
+ save_dir: tensorboard/
71
+ name: null
72
+ version: ${name}
73
+ log_graph: false
74
+ default_hp_metric: true
75
+ prefix: ''
76
+ trainer:
77
+ _target_: pytorch_lightning.Trainer
78
+ gpus: 1
79
+ min_epochs: 1
80
+ max_epochs: 100
81
+ resume_from_checkpoint: null
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - experiment=focusConvReLU_MSE_150
121
+ job:
122
+ name: train
123
+ override_dirname: experiment=focusConvReLU_MSE_150
124
+ id: ???
125
+ num: ???
126
+ config_name: train.yaml
127
+ env_set: {}
128
+ env_copy: []
129
+ config:
130
+ override_dirname:
131
+ kv_sep: '='
132
+ item_sep: ','
133
+ exclude_keys: []
134
+ runtime:
135
+ version: 1.1.2
136
+ cwd: /usr/src/app
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /usr/src/app/configs
142
+ schema: file
143
+ provider: main
144
+ - path: hydra_plugins.hydra_colorlog.conf
145
+ schema: pkg
146
+ provider: hydra-colorlog
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ choices:
151
+ local: default.yaml
152
+ hparams_search: null
153
+ debug: null
154
+ experiment: focusConvReLU_MSE_150
155
+ log_dir: default.yaml
156
+ trainer: default.yaml
157
+ logger: many_loggers
158
+ callbacks: default.yaml
159
+ model: focusConvReLU_150.yaml
160
+ datamodule: focus150.yaml
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: colorlog
164
+ hydra/hydra_logging: colorlog
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=focusConvReLU_MSE_150
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/config.yaml ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: 12345
8
+ name: focusConvReLU_MSE_150
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_train_file: ${data_dir}/focus150/train_metadata.csv
13
+ csv_val_file: ${data_dir}/focus150/validation_metadata.csv
14
+ csv_test_file: ${data_dir}/focus150/test_metadata.csv
15
+ batch_size: 64
16
+ num_workers: 0
17
+ pin_memory: false
18
+ augmentation: true
19
+ model:
20
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
21
+ image_size: 150
22
+ pool_size: 2
23
+ conv1_size: 3
24
+ conv1_channels: 9
25
+ conv2_size: 7
26
+ conv2_channels: 6
27
+ lin1_size: 32
28
+ lin2_size: 72
29
+ output_size: 1
30
+ lr: 0.001
31
+ weight_decay: 0.0005
32
+ callbacks:
33
+ model_checkpoint:
34
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
35
+ monitor: val/mae
36
+ mode: min
37
+ save_top_k: 1
38
+ save_last: true
39
+ verbose: false
40
+ dirpath: checkpoints/
41
+ filename: epoch_{epoch:03d}
42
+ auto_insert_metric_name: false
43
+ early_stopping:
44
+ _target_: pytorch_lightning.callbacks.EarlyStopping
45
+ monitor: val/mae
46
+ mode: min
47
+ patience: 100
48
+ min_delta: 0
49
+ model_summary:
50
+ _target_: pytorch_lightning.callbacks.RichModelSummary
51
+ max_depth: -1
52
+ rich_progress_bar:
53
+ _target_: pytorch_lightning.callbacks.RichProgressBar
54
+ logger:
55
+ csv:
56
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
57
+ save_dir: .
58
+ name: csv/
59
+ prefix: ''
60
+ mlflow:
61
+ _target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
62
+ experiment_name: ${name}
63
+ tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
64
+ tags: null
65
+ save_dir: ./mlruns
66
+ prefix: ''
67
+ artifact_location: null
68
+ tensorboard:
69
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
70
+ save_dir: tensorboard/
71
+ name: null
72
+ version: ${name}
73
+ log_graph: false
74
+ default_hp_metric: true
75
+ prefix: ''
76
+ trainer:
77
+ _target_: pytorch_lightning.Trainer
78
+ gpus: 1
79
+ min_epochs: 1
80
+ max_epochs: 100
81
+ resume_from_checkpoint: null
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - experiment=focusConvReLU_MSE_150.yaml
121
+ job:
122
+ name: train
123
+ override_dirname: experiment=focusConvReLU_MSE_150.yaml
124
+ id: ???
125
+ num: ???
126
+ config_name: train.yaml
127
+ env_set: {}
128
+ env_copy: []
129
+ config:
130
+ override_dirname:
131
+ kv_sep: '='
132
+ item_sep: ','
133
+ exclude_keys: []
134
+ runtime:
135
+ version: 1.1.2
136
+ cwd: /usr/src/app
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /usr/src/app/configs
142
+ schema: file
143
+ provider: main
144
+ - path: hydra_plugins.hydra_colorlog.conf
145
+ schema: pkg
146
+ provider: hydra-colorlog
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ choices:
151
+ local: default.yaml
152
+ hparams_search: null
153
+ debug: null
154
+ experiment: focusConvReLU_MSE_150.yaml
155
+ log_dir: default.yaml
156
+ trainer: default.yaml
157
+ logger: many_loggers
158
+ callbacks: default.yaml
159
+ model: focusConvReLU_150.yaml
160
+ datamodule: focus150.yaml
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: colorlog
164
+ hydra/hydra_logging: colorlog
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=focusConvReLU_MSE_150.yaml
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/config.yaml ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: 12345
8
+ name: focusConvReLU_MSE_150
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_train_file: ${data_dir}/focus150/train_metadata.csv
13
+ csv_val_file: ${data_dir}/focus150/validation_metadata.csv
14
+ csv_test_file: ${data_dir}/focus150/test_metadata.csv
15
+ batch_size: 64
16
+ num_workers: 0
17
+ pin_memory: false
18
+ augmentation: true
19
+ model:
20
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
21
+ image_size: 150
22
+ pool_size: 2
23
+ conv1_size: 3
24
+ conv1_channels: 9
25
+ conv2_size: 7
26
+ conv2_channels: 6
27
+ lin1_size: 32
28
+ lin2_size: 72
29
+ output_size: 1
30
+ lr: 0.001
31
+ weight_decay: 0.0005
32
+ callbacks:
33
+ model_checkpoint:
34
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
35
+ monitor: val/mae
36
+ mode: min
37
+ save_top_k: 1
38
+ save_last: true
39
+ verbose: false
40
+ dirpath: checkpoints/
41
+ filename: epoch_{epoch:03d}
42
+ auto_insert_metric_name: false
43
+ early_stopping:
44
+ _target_: pytorch_lightning.callbacks.EarlyStopping
45
+ monitor: val/mae
46
+ mode: min
47
+ patience: 100
48
+ min_delta: 0
49
+ model_summary:
50
+ _target_: pytorch_lightning.callbacks.RichModelSummary
51
+ max_depth: -1
52
+ rich_progress_bar:
53
+ _target_: pytorch_lightning.callbacks.RichProgressBar
54
+ logger:
55
+ csv:
56
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
57
+ save_dir: .
58
+ name: csv/
59
+ prefix: ''
60
+ mlflow:
61
+ _target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
62
+ experiment_name: ${name}
63
+ tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
64
+ tags: null
65
+ save_dir: ./mlruns
66
+ prefix: ''
67
+ artifact_location: null
68
+ tensorboard:
69
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
70
+ save_dir: tensorboard/
71
+ name: null
72
+ version: ${name}
73
+ log_graph: false
74
+ default_hp_metric: true
75
+ prefix: ''
76
+ trainer:
77
+ _target_: pytorch_lightning.Trainer
78
+ gpus: 1
79
+ min_epochs: 1
80
+ max_epochs: 100
81
+ resume_from_checkpoint: null
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - experiment=focusConvReLU_MSE_150.yaml
121
+ job:
122
+ name: train
123
+ override_dirname: experiment=focusConvReLU_MSE_150.yaml
124
+ id: ???
125
+ num: ???
126
+ config_name: train.yaml
127
+ env_set: {}
128
+ env_copy: []
129
+ config:
130
+ override_dirname:
131
+ kv_sep: '='
132
+ item_sep: ','
133
+ exclude_keys: []
134
+ runtime:
135
+ version: 1.1.2
136
+ cwd: /usr/src/app
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /usr/src/app/configs
142
+ schema: file
143
+ provider: main
144
+ - path: hydra_plugins.hydra_colorlog.conf
145
+ schema: pkg
146
+ provider: hydra-colorlog
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ choices:
151
+ local: default.yaml
152
+ hparams_search: null
153
+ debug: null
154
+ experiment: focusConvReLU_MSE_150.yaml
155
+ log_dir: default.yaml
156
+ trainer: default.yaml
157
+ logger: many_loggers
158
+ callbacks: default.yaml
159
+ model: focusConvReLU_150.yaml
160
+ datamodule: focus150.yaml
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: colorlog
164
+ hydra/hydra_logging: colorlog
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=focusConvReLU_MSE_150.yaml
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/config.yaml ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: 12345
8
+ name: focusConvReLU_MSE_150
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_train_file: ${data_dir}/focus150/train_metadata.csv
13
+ csv_val_file: ${data_dir}/focus150/validation_metadata.csv
14
+ csv_test_file: ${data_dir}/focus150/test_metadata.csv
15
+ batch_size: 64
16
+ num_workers: 0
17
+ pin_memory: false
18
+ augmentation: true
19
+ model:
20
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
21
+ image_size: 150
22
+ pool_size: 2
23
+ conv1_size: 3
24
+ conv1_channels: 9
25
+ conv2_size: 7
26
+ conv2_channels: 6
27
+ lin1_size: 32
28
+ lin2_size: 72
29
+ output_size: 1
30
+ lr: 0.001
31
+ weight_decay: 0.0005
32
+ callbacks:
33
+ model_checkpoint:
34
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
35
+ monitor: val/mae
36
+ mode: min
37
+ save_top_k: 1
38
+ save_last: true
39
+ verbose: false
40
+ dirpath: checkpoints/
41
+ filename: epoch_{epoch:03d}
42
+ auto_insert_metric_name: false
43
+ early_stopping:
44
+ _target_: pytorch_lightning.callbacks.EarlyStopping
45
+ monitor: val/mae
46
+ mode: min
47
+ patience: 100
48
+ min_delta: 0
49
+ model_summary:
50
+ _target_: pytorch_lightning.callbacks.RichModelSummary
51
+ max_depth: -1
52
+ rich_progress_bar:
53
+ _target_: pytorch_lightning.callbacks.RichProgressBar
54
+ logger:
55
+ csv:
56
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
57
+ save_dir: .
58
+ name: csv/
59
+ prefix: ''
60
+ mlflow:
61
+ _target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
62
+ experiment_name: ${name}
63
+ tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
64
+ tags: null
65
+ save_dir: ./mlruns
66
+ prefix: ''
67
+ artifact_location: null
68
+ tensorboard:
69
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
70
+ save_dir: tensorboard/
71
+ name: null
72
+ version: ${name}
73
+ log_graph: false
74
+ default_hp_metric: true
75
+ prefix: ''
76
+ trainer:
77
+ _target_: pytorch_lightning.Trainer
78
+ gpus: 1
79
+ min_epochs: 1
80
+ max_epochs: 100
81
+ resume_from_checkpoint: null
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - experiment=focusConvReLU_MSE_150.yaml
121
+ job:
122
+ name: train
123
+ override_dirname: experiment=focusConvReLU_MSE_150.yaml
124
+ id: ???
125
+ num: ???
126
+ config_name: train.yaml
127
+ env_set: {}
128
+ env_copy: []
129
+ config:
130
+ override_dirname:
131
+ kv_sep: '='
132
+ item_sep: ','
133
+ exclude_keys: []
134
+ runtime:
135
+ version: 1.1.2
136
+ cwd: /usr/src/app
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /usr/src/app/configs
142
+ schema: file
143
+ provider: main
144
+ - path: hydra_plugins.hydra_colorlog.conf
145
+ schema: pkg
146
+ provider: hydra-colorlog
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ choices:
151
+ local: default.yaml
152
+ hparams_search: null
153
+ debug: null
154
+ experiment: focusConvReLU_MSE_150.yaml
155
+ log_dir: default.yaml
156
+ trainer: default.yaml
157
+ logger: many_loggers
158
+ callbacks: default.yaml
159
+ model: focusConvReLU_150.yaml
160
+ datamodule: focus150.yaml
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: colorlog
164
+ hydra/hydra_logging: colorlog
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=focusConvReLU_MSE_150.yaml
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/epoch_096.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3e9159964f903b3063b0a2170eb2baa7d138b219689f46620567a0b821cbf98
3
+ size 2612879
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/last.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d046d8e08e9ab7ff46290261d0f10a025a04c4c43ff48de76df4a91d77ee0f6
3
+ size 2612879
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 1
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
9
+ image_size: 150
10
+ pool_size: 2
11
+ conv1_size: 3
12
+ conv1_channels: 9
13
+ conv2_size: 7
14
+ conv2_channels: 6
15
+ lin1_size: 32
16
+ lin2_size: 72
17
+ output_size: 1
18
+ lr: 0.001
19
+ weight_decay: 0.0005
20
+ datamodule:
21
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
22
+ data_dir: /usr/src/app/data/focus150
23
+ csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
24
+ csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
25
+ csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
26
+ batch_size: 64
27
+ num_workers: 0
28
+ pin_memory: false
29
+ augmentation: true
30
+ seed: 12345
31
+ callbacks:
32
+ model_checkpoint:
33
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
34
+ monitor: val/mae
35
+ mode: min
36
+ save_top_k: 1
37
+ save_last: true
38
+ verbose: false
39
+ dirpath: checkpoints/
40
+ filename: epoch_{epoch:03d}
41
+ auto_insert_metric_name: false
42
+ early_stopping:
43
+ _target_: pytorch_lightning.callbacks.EarlyStopping
44
+ monitor: val/mae
45
+ mode: min
46
+ patience: 100
47
+ min_delta: 0
48
+ model_summary:
49
+ _target_: pytorch_lightning.callbacks.RichModelSummary
50
+ max_depth: -1
51
+ rich_progress_bar:
52
+ _target_: pytorch_lightning.callbacks.RichProgressBar
53
+ model/params/total: 216231
54
+ model/params/trainable: 216231
55
+ model/params/non_trainable: 0
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 0.0010634043719619513,0.02607358619570732,0.02607358619570732,0,47,,,,
3
+ ,,,1,47,0.015712032094597816,0.07011217623949051,,
4
+ 0.0008477172814309597,0.023356162011623383,0.023356162011623383,1,95,,,,
5
+ ,,,2,95,0.0006998287281021476,0.021280117332935333,,
6
+ 0.0006586583331227303,0.02111729420721531,0.02111729420721531,2,143,,,,
7
+ ,,,3,143,0.0006262350943870842,0.020297976210713387,,
8
+ 0.0006095922435633838,0.020395344123244286,0.020395344123244286,3,191,,,,
9
+ ,,,4,191,0.0006112757255323231,0.020039746537804604,,
10
+ 0.000579743820708245,0.02008357085287571,0.02008357085287571,4,239,,,,
11
+ ,,,5,239,0.0005892604240216315,0.019787773489952087,,
12
+ 0.000539788044989109,0.019109351560473442,0.019109351560473442,5,287,,,,
13
+ ,,,6,287,0.0004991526948288083,0.018337024375796318,,
14
+ 0.0005701080663129687,0.020062454044818878,0.019109351560473442,6,335,,,,
15
+ ,,,7,335,0.00046538031892850995,0.017815634608268738,,
16
+ 0.0005549467750824988,0.018724771216511726,0.018724771216511726,7,383,,,,
17
+ ,,,8,383,0.0004903865046799183,0.017987234517931938,,
18
+ 0.0005555949755944312,0.01976374350488186,0.018724771216511726,8,431,,,,
19
+ ,,,9,431,0.0004449250118341297,0.017149975523352623,,
20
+ 0.0004789295489899814,0.018088672310113907,0.018088672310113907,9,479,,,,
21
+ ,,,10,479,0.0004655240918509662,0.01764928176999092,,
22
+ 0.0005822633975185454,0.02009824477136135,0.018088672310113907,10,527,,,,
23
+ ,,,11,527,0.00044383766362443566,0.017262466251850128,,
24
+ 0.00048447673907503486,0.017433589324355125,0.017433589324355125,11,575,,,,
25
+ ,,,12,575,0.0004356526769697666,0.017195068299770355,,
26
+ 0.00046181678771972656,0.018075214698910713,0.017433589324355125,12,623,,,,
27
+ ,,,13,623,0.00041521640378050506,0.01676551066339016,,
28
+ 0.00041472347220405936,0.016927972435951233,0.016927972435951233,13,671,,,,
29
+ ,,,14,671,0.0003966042131651193,0.016418082639575005,,
30
+ 0.0005150702781975269,0.01795283891260624,0.016927972435951233,14,719,,,,
31
+ ,,,15,719,0.0004069829301442951,0.016572223976254463,,
32
+ 0.00043978836038149893,0.017694905400276184,0.016927972435951233,15,767,,,,
33
+ ,,,16,767,0.0004001975175924599,0.016431566327810287,,
34
+ 0.0004524872056208551,0.017309539020061493,0.016927972435951233,16,815,,,,
35
+ ,,,17,815,0.000411732355132699,0.016624923795461655,,
36
+ 0.0004895858000963926,0.017853353172540665,0.016927972435951233,17,863,,,,
37
+ ,,,18,863,0.0003962284536100924,0.016285978257656097,,
38
+ 0.00048630539095029235,0.017320722341537476,0.016927972435951233,18,911,,,,
39
+ ,,,19,911,0.0003755573998205364,0.015845887362957,,
40
+ 0.00045636019785888493,0.017117636278271675,0.016927972435951233,19,959,,,,
41
+ ,,,20,959,0.00040089385583996773,0.016440516337752342,,
42
+ 0.00042457759263925254,0.01685478165745735,0.01685478165745735,20,1007,,,,
43
+ ,,,21,1007,0.00037927881930954754,0.015909109264612198,,
44
+ 0.0005504836444742978,0.01946953497827053,0.01685478165745735,21,1055,,,,
45
+ ,,,22,1055,0.00037079042522236705,0.015598192811012268,,
46
+ 0.0005828228895552456,0.019640492275357246,0.01685478165745735,22,1103,,,,
47
+ ,,,23,1103,0.0003636741894297302,0.015478463843464851,,
48
+ 0.000536997162271291,0.01877320185303688,0.01685478165745735,23,1151,,,,
49
+ ,,,24,1151,0.0003473551769275218,0.014973259530961514,,
50
+ 0.0004296917177271098,0.016905883327126503,0.01685478165745735,24,1199,,,,
51
+ ,,,25,1199,0.0003502711479086429,0.014955243095755577,,
52
+ 0.000369079876691103,0.015234655700623989,0.015234655700623989,25,1247,,,,
53
+ ,,,26,1247,0.0003374406951479614,0.014734053052961826,,
54
+ 0.0005687376251444221,0.019466156139969826,0.015234655700623989,26,1295,,,,
55
+ ,,,27,1295,0.0003077377623412758,0.013931847177445889,,
56
+ 0.0003877778071910143,0.01577661745250225,0.015234655700623989,27,1343,,,,
57
+ ,,,28,1343,0.00030867080204188824,0.013890136033296585,,
58
+ 0.0007391179096885026,0.021662525832653046,0.015234655700623989,28,1391,,,,
59
+ ,,,29,1391,0.0003139588516205549,0.013945831917226315,,
60
+ 0.0002891223703045398,0.013201462104916573,0.013201462104916573,29,1439,,,,
61
+ ,,,30,1439,0.00029595510568469763,0.013475497253239155,,
62
+ 0.00136762042529881,0.03221268951892853,0.013201462104916573,30,1487,,,,
63
+ ,,,31,1487,0.0002615793200675398,0.012566673569381237,,
64
+ 0.0005622635362669826,0.018861766904592514,0.013201462104916573,31,1535,,,,
65
+ ,,,32,1535,0.0002453128108754754,0.012109467759728432,,
66
+ 0.0002613224496599287,0.012555716559290886,0.012555716559290886,32,1583,,,,
67
+ ,,,33,1583,0.0002475597430020571,0.01209949143230915,,
68
+ 0.00026621491997502744,0.01243590097874403,0.01243590097874403,33,1631,,,,
69
+ ,,,34,1631,0.00021561248286161572,0.011168932542204857,,
70
+ 0.0003965417272411287,0.015400874428451061,0.01243590097874403,34,1679,,,,
71
+ ,,,35,1679,0.0002476295630913228,0.012050420045852661,,
72
+ 0.0006301697576418519,0.019847549498081207,0.01243590097874403,35,1727,,,,
73
+ ,,,36,1727,0.00022967555560171604,0.011594988405704498,,
74
+ 0.0004612059856299311,0.01633538119494915,0.01243590097874403,36,1775,,,,
75
+ ,,,37,1775,0.0002333874726900831,0.011586652137339115,,
76
+ 0.00025420173187740147,0.012520545162260532,0.01243590097874403,37,1823,,,,
77
+ ,,,38,1823,0.00023667479399591684,0.01181836612522602,,
78
+ 0.0003383368020877242,0.015009582974016666,0.01243590097874403,38,1871,,,,
79
+ ,,,39,1871,0.0001987472496693954,0.010663148015737534,,
80
+ 0.0003864422906190157,0.016114825382828712,0.01243590097874403,39,1919,,,,
81
+ ,,,40,1919,0.00019275647355243564,0.010579877533018589,,
82
+ 0.00022850626555737108,0.011395555920898914,0.011395555920898914,40,1967,,,,
83
+ ,,,41,1967,0.00020515960932243615,0.010919580236077309,,
84
+ 0.00036217307206243277,0.015877339988946915,0.011395555920898914,41,2015,,,,
85
+ ,,,42,2015,0.00020770113042090088,0.010937352664768696,,
86
+ 0.00026963415439240634,0.012874774634838104,0.011395555920898914,42,2063,,,,
87
+ ,,,43,2063,0.00020995769591536373,0.010980397462844849,,
88
+ 0.0003081822069361806,0.01374065037816763,0.011395555920898914,43,2111,,,,
89
+ ,,,44,2111,0.0001873430737759918,0.010383875109255314,,
90
+ 0.0008588534547016025,0.02455970272421837,0.011395555920898914,44,2159,,,,
91
+ ,,,45,2159,0.0001887517428258434,0.010248820297420025,,
92
+ 0.00025588576681911945,0.012964467518031597,0.011395555920898914,45,2207,,,,
93
+ ,,,46,2207,0.00017797888722270727,0.009944379329681396,,
94
+ 0.0012111393734812737,0.030586877837777138,0.011395555920898914,46,2255,,,,
95
+ ,,,47,2255,0.00017462909454479814,0.009952726773917675,,
96
+ 0.0005991487414576113,0.01970861665904522,0.011395555920898914,47,2303,,,,
97
+ ,,,48,2303,0.00018134394485969096,0.010111919604241848,,
98
+ 0.00027215175214223564,0.012582853436470032,0.011395555920898914,48,2351,,,,
99
+ ,,,49,2351,0.0002078687393805012,0.010890776291489601,,
100
+ 0.0002326530375285074,0.011895284987986088,0.011395555920898914,49,2399,,,,
101
+ ,,,50,2399,0.00016190602036658674,0.009552615694701672,,
102
+ 0.00034161240910179913,0.014539946801960468,0.011395555920898914,50,2447,,,,
103
+ ,,,51,2447,0.0001574503257870674,0.009343606419861317,,
104
+ 0.0002047311281785369,0.010513107292354107,0.010513107292354107,51,2495,,,,
105
+ ,,,52,2495,0.00017259498417843133,0.00977871473878622,,
106
+ 0.0002189291553804651,0.011585352011024952,0.010513107292354107,52,2543,,,,
107
+ ,,,53,2543,0.00016215378127526492,0.009527333080768585,,
108
+ 0.0008305492228828371,0.025230778381228447,0.010513107292354107,53,2591,,,,
109
+ ,,,54,2591,0.00017192678933497518,0.009975402615964413,,
110
+ 0.0003245118132326752,0.01448488887399435,0.010513107292354107,54,2639,,,,
111
+ ,,,55,2639,0.00014025233394932002,0.008768484927713871,,
112
+ 0.00043747638119384646,0.016850745305418968,0.010513107292354107,55,2687,,,,
113
+ ,,,56,2687,0.00014388358977157623,0.008977909572422504,,
114
+ 0.0003352710627950728,0.014916492626070976,0.010513107292354107,56,2735,,,,
115
+ ,,,57,2735,0.00013614425552077591,0.008777343668043613,,
116
+ 0.00017761132039595395,0.009772807359695435,0.009772807359695435,57,2783,,,,
117
+ ,,,58,2783,0.00014166149776428938,0.008911573328077793,,
118
+ 0.0002451064356137067,0.012413904070854187,0.009772807359695435,58,2831,,,,
119
+ ,,,59,2831,0.00014307026867754757,0.008944649249315262,,
120
+ 0.00045808948925696313,0.017842639237642288,0.009772807359695435,59,2879,,,,
121
+ ,,,60,2879,0.00012790328764822334,0.008465489372611046,,
122
+ 0.00016568403225392103,0.009426051750779152,0.009426051750779152,60,2927,,,,
123
+ ,,,61,2927,0.00013250892516225576,0.008481234312057495,,
124
+ 0.0002005849382840097,0.011247269809246063,0.009426051750779152,61,2975,,,,
125
+ ,,,62,2975,0.00012225286627653986,0.008217735216021538,,
126
+ 0.0004337916907388717,0.017680322751402855,0.009426051750779152,62,3023,,,,
127
+ ,,,63,3023,0.00013184384442865849,0.008489103056490421,,
128
+ 0.00043202925007790327,0.016938434913754463,0.009426051750779152,63,3071,,,,
129
+ ,,,64,3071,0.0001139832820626907,0.00794004462659359,,
130
+ 0.0006070298259146512,0.020868053659796715,0.009426051750779152,64,3119,,,,
131
+ ,,,65,3119,0.00012551495456136763,0.00839218683540821,,
132
+ 0.00125686835963279,0.03198228031396866,0.009426051750779152,65,3167,,,,
133
+ ,,,66,3167,0.00013683686847798526,0.008686883375048637,,
134
+ 0.0003458949795458466,0.01481217984110117,0.009426051750779152,66,3215,,,,
135
+ ,,,67,3215,0.0001252585934707895,0.008334343321621418,,
136
+ 0.00023582404537592083,0.01217477023601532,0.009426051750779152,67,3263,,,,
137
+ ,,,68,3263,0.0001234338415088132,0.008174673654139042,,
138
+ 0.00043027440551668406,0.0175019484013319,0.009426051750779152,68,3311,,,,
139
+ ,,,69,3311,0.0001386790390824899,0.008791987784206867,,
140
+ 0.00018619702314026654,0.01059352234005928,0.009426051750779152,69,3359,,,,
141
+ ,,,70,3359,0.000133068097056821,0.008551678620278835,,
142
+ 0.0004644407599698752,0.017586909234523773,0.009426051750779152,70,3407,,,,
143
+ ,,,71,3407,0.0001232567592523992,0.008277542889118195,,
144
+ 0.00025989525602199137,0.013158340007066727,0.009426051750779152,71,3455,,,,
145
+ ,,,72,3455,0.00010765787737909704,0.007683832198381424,,
146
+ 0.000271034223260358,0.013016403652727604,0.009426051750779152,72,3503,,,,
147
+ ,,,73,3503,0.00010420197213534266,0.007514127530157566,,
148
+ 0.00015595319564454257,0.00941098015755415,0.00941098015755415,73,3551,,,,
149
+ ,,,74,3551,0.00011743942741304636,0.008030056953430176,,
150
+ 0.00018362083937972784,0.010173797607421875,0.00941098015755415,74,3599,,,,
151
+ ,,,75,3599,0.00013075751485303044,0.008549727499485016,,
152
+ 0.0001723619207041338,0.010034985840320587,0.00941098015755415,75,3647,,,,
153
+ ,,,76,3647,0.00012383297143969685,0.008334793150424957,,
154
+ 0.00026339536998420954,0.01230852585285902,0.00941098015755415,76,3695,,,,
155
+ ,,,77,3695,0.00012997408339288086,0.00849801953881979,,
156
+ 0.0022478969767689705,0.043913356959819794,0.00941098015755415,77,3743,,,,
157
+ ,,,78,3743,0.0001171917756437324,0.00794607400894165,,
158
+ 0.00043688094592653215,0.01787818782031536,0.00941098015755415,78,3791,,,,
159
+ ,,,79,3791,0.0001384491624776274,0.008685767650604248,,
160
+ 0.000253666948992759,0.012860489077866077,0.00941098015755415,79,3839,,,,
161
+ ,,,80,3839,0.00011071155313402414,0.007812082301825285,,
162
+ 0.0007027353858575225,0.02276649698615074,0.00941098015755415,80,3887,,,,
163
+ ,,,81,3887,0.00012242942466400564,0.008264751173555851,,
164
+ 0.00036985089536756277,0.015199720859527588,0.00941098015755415,81,3935,,,,
165
+ ,,,82,3935,0.00012077093560947105,0.008274727500975132,,
166
+ 0.00042630801908671856,0.016393927857279778,0.00941098015755415,82,3983,,,,
167
+ ,,,83,3983,0.00011708390229614452,0.008091365918517113,,
168
+ 0.00018512520182412118,0.010693005286157131,0.00941098015755415,83,4031,,,,
169
+ ,,,84,4031,0.00013511900033336133,0.008610659278929234,,
170
+ 0.0006041005253791809,0.02147362008690834,0.00941098015755415,84,4079,,,,
171
+ ,,,85,4079,0.00010951827425742522,0.007766522467136383,,
172
+ 0.0005739088519476354,0.019695602357387543,0.00941098015755415,85,4127,,,,
173
+ ,,,86,4127,0.00011793604062404484,0.008045573718845844,,
174
+ 0.00020085058349650353,0.010959284380078316,0.00941098015755415,86,4175,,,,
175
+ ,,,87,4175,0.00011460384848760441,0.007821462117135525,,
176
+ 0.00035394675796851516,0.014944094233214855,0.00941098015755415,87,4223,,,,
177
+ ,,,88,4223,0.00012705886911135167,0.008322164416313171,,
178
+ 0.001467306399717927,0.03422271087765694,0.00941098015755415,88,4271,,,,
179
+ ,,,89,4271,0.00011695512512233108,0.008126244880259037,,
180
+ 0.0003908797516487539,0.015736214816570282,0.00941098015755415,89,4319,,,,
181
+ ,,,90,4319,0.00011259406892349944,0.007857552729547024,,
182
+ 0.0001805946958484128,0.009996251203119755,0.00941098015755415,90,4367,,,,
183
+ ,,,91,4367,0.00011745867959689349,0.008035344071686268,,
184
+ 0.00035339672467671335,0.015099494718015194,0.00941098015755415,91,4415,,,,
185
+ ,,,92,4415,0.00013093209418002516,0.008491233922541142,,
186
+ 0.00040678223012946546,0.015847915783524513,0.00941098015755415,92,4463,,,,
187
+ ,,,93,4463,0.00013953317829873413,0.008733570575714111,,
188
+ 0.00018399060354568064,0.010453518480062485,0.00941098015755415,93,4511,,,,
189
+ ,,,94,4511,0.00012743366823997349,0.008426659740507603,,
190
+ 0.0005399276851676404,0.018864858895540237,0.00941098015755415,94,4559,,,,
191
+ ,,,95,4559,0.00011808565614046529,0.008046725764870644,,
192
+ 0.00029979468672536314,0.01469734963029623,0.00941098015755415,95,4607,,,,
193
+ ,,,96,4607,0.00015584049106109887,0.009322610683739185,,
194
+ 0.00013096585462335497,0.00869279820472002,0.00869279820472002,96,4655,,,,
195
+ ,,,97,4655,0.00013836825382895768,0.008852057158946991,,
196
+ 0.00021977233700454235,0.011555126868188381,0.00869279820472002,97,4703,,,,
197
+ ,,,98,4703,0.00012112926197005436,0.008111957460641861,,
198
+ 0.00024704227689653635,0.01307331956923008,0.00869279820472002,98,4751,,,,
199
+ ,,,99,4751,0.00012230231368448585,0.008170234970748425,,
200
+ 0.0003437100094743073,0.01433614082634449,0.00869279820472002,99,4799,,,,
201
+ ,,,100,4799,0.00011741576599888504,0.008008270524442196,,
202
+ ,,,96,4800,,,0.00014260406896937639,0.008947640657424927
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652279809.c49d1a46d86b.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c87544f737c7a20d4e44d813742d70c0b9acd226f353912b880bbdbf298495a
3
+ size 36870
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652280195.c49d1a46d86b.1.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bdc9d2c64d369ea43ddafc00d36574863f5c53b08e1a856563afc76f942d11c
3
+ size 179
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/hparams.yaml ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 1
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_conv_module.FocusConvReLULitModule
9
+ image_size: 150
10
+ pool_size: 2
11
+ conv1_size: 3
12
+ conv1_channels: 9
13
+ conv2_size: 7
14
+ conv2_channels: 6
15
+ lin1_size: 32
16
+ lin2_size: 72
17
+ output_size: 1
18
+ lr: 0.001
19
+ weight_decay: 0.0005
20
+ datamodule:
21
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
22
+ data_dir: /usr/src/app/data/focus150
23
+ csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
24
+ csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
25
+ csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
26
+ batch_size: 64
27
+ num_workers: 0
28
+ pin_memory: false
29
+ augmentation: true
30
+ seed: 12345
31
+ callbacks:
32
+ model_checkpoint:
33
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
34
+ monitor: val/mae
35
+ mode: min
36
+ save_top_k: 1
37
+ save_last: true
38
+ verbose: false
39
+ dirpath: checkpoints/
40
+ filename: epoch_{epoch:03d}
41
+ auto_insert_metric_name: false
42
+ early_stopping:
43
+ _target_: pytorch_lightning.callbacks.EarlyStopping
44
+ monitor: val/mae
45
+ mode: min
46
+ patience: 100
47
+ min_delta: 0
48
+ model_summary:
49
+ _target_: pytorch_lightning.callbacks.RichModelSummary
50
+ max_depth: -1
51
+ rich_progress_bar:
52
+ _target_: pytorch_lightning.callbacks.RichProgressBar
53
+ model/params/total: 216231
54
+ model/params/trainable: 216231
55
+ model/params/non_trainable: 0
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/meta.yaml ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ artifact_uri: /usr/src/app/logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/artifacts
2
+ end_time: 1652280195396
3
+ entry_point_name: ''
4
+ experiment_id: '11'
5
+ lifecycle_stage: active
6
+ name: ''
7
+ run_id: fc9d4c9c610e4ae9a51aeec7e544684b
8
+ run_uuid: fc9d4c9c610e4ae9a51aeec7e544684b
9
+ source_name: ''
10
+ source_type: 4
11
+ source_version: ''
12
+ start_time: 1652279809740
13
+ status: 3
14
+ tags: []
15
+ user_id: unknown
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/epoch ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1652279820704 0 47
2
+ 1652279820749 1 47
3
+ 1652279824357 1 95
4
+ 1652279824405 2 95
5
+ 1652279828044 2 143
6
+ 1652279828096 3 143
7
+ 1652279831676 3 191
8
+ 1652279831720 4 191
9
+ 1652279835446 4 239
10
+ 1652279835495 5 239
11
+ 1652279839271 5 287
12
+ 1652279839319 6 287
13
+ 1652279843079 6 335
14
+ 1652279843119 7 335
15
+ 1652279846830 7 383
16
+ 1652279846878 8 383
17
+ 1652279850614 8 431
18
+ 1652279850653 9 431
19
+ 1652279854390 9 479
20
+ 1652279854434 10 479
21
+ 1652279858241 10 527
22
+ 1652279858281 11 527
23
+ 1652279862079 11 575
24
+ 1652279862126 12 575
25
+ 1652279865967 12 623
26
+ 1652279866012 13 623
27
+ 1652279869769 13 671
28
+ 1652279869817 14 671
29
+ 1652279873629 14 719
30
+ 1652279873668 15 719
31
+ 1652279877456 15 767
32
+ 1652279877494 16 767
33
+ 1652279881261 16 815
34
+ 1652279881296 17 815
35
+ 1652279885044 17 863
36
+ 1652279885082 18 863
37
+ 1652279888829 18 911
38
+ 1652279888864 19 911
39
+ 1652279892527 19 959
40
+ 1652279892566 20 959
41
+ 1652279896237 20 1007
42
+ 1652279896290 21 1007
43
+ 1652279900002 21 1055
44
+ 1652279900037 22 1055
45
+ 1652279903680 22 1103
46
+ 1652279903717 23 1103
47
+ 1652279907516 23 1151
48
+ 1652279907559 24 1151
49
+ 1652279911251 24 1199
50
+ 1652279911367 25 1199
51
+ 1652279915084 25 1247
52
+ 1652279915129 26 1247
53
+ 1652279918947 26 1295
54
+ 1652279918986 27 1295
55
+ 1652279922767 27 1343
56
+ 1652279922806 28 1343
57
+ 1652279926483 28 1391
58
+ 1652279926523 29 1391
59
+ 1652279930329 29 1439
60
+ 1652279930377 30 1439
61
+ 1652279934121 30 1487
62
+ 1652279934160 31 1487
63
+ 1652279937833 31 1535
64
+ 1652279937872 32 1535
65
+ 1652279941623 32 1583
66
+ 1652279941677 33 1583
67
+ 1652279945268 33 1631
68
+ 1652279945317 34 1631
69
+ 1652279949005 34 1679
70
+ 1652279949040 35 1679
71
+ 1652279952772 35 1727
72
+ 1652279952812 36 1727
73
+ 1652279956496 36 1775
74
+ 1652279956535 37 1775
75
+ 1652279960298 37 1823
76
+ 1652279960344 38 1823
77
+ 1652279964073 38 1871
78
+ 1652279964110 39 1871
79
+ 1652279967844 39 1919
80
+ 1652279967883 40 1919
81
+ 1652279971446 40 1967
82
+ 1652279971491 41 1967
83
+ 1652279975126 41 2015
84
+ 1652279975165 42 2015
85
+ 1652279978811 42 2063
86
+ 1652279978850 43 2063
87
+ 1652279982539 43 2111
88
+ 1652279982578 44 2111
89
+ 1652279986321 44 2159
90
+ 1652279986365 45 2159
91
+ 1652279990134 45 2207
92
+ 1652279990170 46 2207
93
+ 1652279993983 46 2255
94
+ 1652279994023 47 2255
95
+ 1652279997788 47 2303
96
+ 1652279997827 48 2303
97
+ 1652280001609 48 2351
98
+ 1652280001649 49 2351
99
+ 1652280005406 49 2399
100
+ 1652280005492 50 2399
101
+ 1652280009144 50 2447
102
+ 1652280009187 51 2447
103
+ 1652280012811 51 2495
104
+ 1652280012862 52 2495
105
+ 1652280016584 52 2543
106
+ 1652280016623 53 2543
107
+ 1652280020312 53 2591
108
+ 1652280020351 54 2591
109
+ 1652280024138 54 2639
110
+ 1652280024177 55 2639
111
+ 1652280027939 55 2687
112
+ 1652280027981 56 2687
113
+ 1652280031799 56 2735
114
+ 1652280031839 57 2735
115
+ 1652280035376 57 2783
116
+ 1652280035425 58 2783
117
+ 1652280039147 58 2831
118
+ 1652280039190 59 2831
119
+ 1652280043008 59 2879
120
+ 1652280043047 60 2879
121
+ 1652280046735 60 2927
122
+ 1652280046780 61 2927
123
+ 1652280050532 61 2975
124
+ 1652280050567 62 2975
125
+ 1652280054457 62 3023
126
+ 1652280054496 63 3023
127
+ 1652280058279 63 3071
128
+ 1652280058318 64 3071
129
+ 1652280062047 64 3119
130
+ 1652280062086 65 3119
131
+ 1652280065810 65 3167
132
+ 1652280065848 66 3167
133
+ 1652280069610 66 3215
134
+ 1652280069646 67 3215
135
+ 1652280073394 67 3263
136
+ 1652280073440 68 3263
137
+ 1652280077179 68 3311
138
+ 1652280077219 69 3311
139
+ 1652280081010 69 3359
140
+ 1652280081049 70 3359
141
+ 1652280084783 70 3407
142
+ 1652280084823 71 3407
143
+ 1652280088575 71 3455
144
+ 1652280088615 72 3455
145
+ 1652280092368 72 3503
146
+ 1652280092408 73 3503
147
+ 1652280096159 73 3551
148
+ 1652280096210 74 3551
149
+ 1652280099894 74 3599
150
+ 1652280099978 75 3599
151
+ 1652280103580 75 3647
152
+ 1652280103622 76 3647
153
+ 1652280107354 76 3695
154
+ 1652280107393 77 3695
155
+ 1652280111162 77 3743
156
+ 1652280111203 78 3743
157
+ 1652280114987 78 3791
158
+ 1652280115112 79 3791
159
+ 1652280118870 79 3839
160
+ 1652280118906 80 3839
161
+ 1652280122629 80 3887
162
+ 1652280122669 81 3887
163
+ 1652280126475 81 3935
164
+ 1652280126515 82 3935
165
+ 1652280130198 82 3983
166
+ 1652280130237 83 3983
167
+ 1652280134070 83 4031
168
+ 1652280134111 84 4031
169
+ 1652280137810 84 4079
170
+ 1652280137849 85 4079
171
+ 1652280141642 85 4127
172
+ 1652280141678 86 4127
173
+ 1652280145396 86 4175
174
+ 1652280145432 87 4175
175
+ 1652280149183 87 4223
176
+ 1652280149222 88 4223
177
+ 1652280152927 88 4271
178
+ 1652280152969 89 4271
179
+ 1652280156721 89 4319
180
+ 1652280156762 90 4319
181
+ 1652280160468 90 4367
182
+ 1652280160504 91 4367
183
+ 1652280164338 91 4415
184
+ 1652280164378 92 4415
185
+ 1652280168035 92 4463
186
+ 1652280168071 93 4463
187
+ 1652280171787 93 4511
188
+ 1652280171823 94 4511
189
+ 1652280175551 94 4559
190
+ 1652280175591 95 4559
191
+ 1652280179385 95 4607
192
+ 1652280179426 96 4607
193
+ 1652280183124 96 4655
194
+ 1652280183168 97 4655
195
+ 1652280186940 97 4703
196
+ 1652280186980 98 4703
197
+ 1652280190746 98 4751
198
+ 1652280190782 99 4751
199
+ 1652280194552 99 4799
200
+ 1652280194640 100 4799
201
+ 1652280195352 96 4800
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/loss ADDED
@@ -0,0 +1 @@
 
 
1
+ 1652280195352 0.00014260406896937639 4800
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/mae ADDED
@@ -0,0 +1 @@
 
 
1
+ 1652280195352 0.008947640657424927 4800
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/loss ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1652279820749 0.015712032094597816 47
2
+ 1652279824405 0.0006998287281021476 95
3
+ 1652279828096 0.0006262350943870842 143
4
+ 1652279831720 0.0006112757255323231 191
5
+ 1652279835495 0.0005892604240216315 239
6
+ 1652279839319 0.0004991526948288083 287
7
+ 1652279843119 0.00046538031892850995 335
8
+ 1652279846878 0.0004903865046799183 383
9
+ 1652279850653 0.0004449250118341297 431
10
+ 1652279854434 0.0004655240918509662 479
11
+ 1652279858281 0.00044383766362443566 527
12
+ 1652279862126 0.0004356526769697666 575
13
+ 1652279866012 0.00041521640378050506 623
14
+ 1652279869817 0.0003966042131651193 671
15
+ 1652279873668 0.0004069829301442951 719
16
+ 1652279877494 0.0004001975175924599 767
17
+ 1652279881296 0.000411732355132699 815
18
+ 1652279885082 0.0003962284536100924 863
19
+ 1652279888864 0.0003755573998205364 911
20
+ 1652279892566 0.00040089385583996773 959
21
+ 1652279896290 0.00037927881930954754 1007
22
+ 1652279900037 0.00037079042522236705 1055
23
+ 1652279903717 0.0003636741894297302 1103
24
+ 1652279907559 0.0003473551769275218 1151
25
+ 1652279911367 0.0003502711479086429 1199
26
+ 1652279915129 0.0003374406951479614 1247
27
+ 1652279918986 0.0003077377623412758 1295
28
+ 1652279922806 0.00030867080204188824 1343
29
+ 1652279926523 0.0003139588516205549 1391
30
+ 1652279930377 0.00029595510568469763 1439
31
+ 1652279934160 0.0002615793200675398 1487
32
+ 1652279937872 0.0002453128108754754 1535
33
+ 1652279941677 0.0002475597430020571 1583
34
+ 1652279945317 0.00021561248286161572 1631
35
+ 1652279949040 0.0002476295630913228 1679
36
+ 1652279952812 0.00022967555560171604 1727
37
+ 1652279956535 0.0002333874726900831 1775
38
+ 1652279960344 0.00023667479399591684 1823
39
+ 1652279964110 0.0001987472496693954 1871
40
+ 1652279967883 0.00019275647355243564 1919
41
+ 1652279971491 0.00020515960932243615 1967
42
+ 1652279975165 0.00020770113042090088 2015
43
+ 1652279978850 0.00020995769591536373 2063
44
+ 1652279982578 0.0001873430737759918 2111
45
+ 1652279986365 0.0001887517428258434 2159
46
+ 1652279990170 0.00017797888722270727 2207
47
+ 1652279994023 0.00017462909454479814 2255
48
+ 1652279997827 0.00018134394485969096 2303
49
+ 1652280001649 0.0002078687393805012 2351
50
+ 1652280005492 0.00016190602036658674 2399
51
+ 1652280009187 0.0001574503257870674 2447
52
+ 1652280012862 0.00017259498417843133 2495
53
+ 1652280016623 0.00016215378127526492 2543
54
+ 1652280020351 0.00017192678933497518 2591
55
+ 1652280024177 0.00014025233394932002 2639
56
+ 1652280027981 0.00014388358977157623 2687
57
+ 1652280031839 0.00013614425552077591 2735
58
+ 1652280035425 0.00014166149776428938 2783
59
+ 1652280039190 0.00014307026867754757 2831
60
+ 1652280043047 0.00012790328764822334 2879
61
+ 1652280046780 0.00013250892516225576 2927
62
+ 1652280050567 0.00012225286627653986 2975
63
+ 1652280054496 0.00013184384442865849 3023
64
+ 1652280058318 0.0001139832820626907 3071
65
+ 1652280062086 0.00012551495456136763 3119
66
+ 1652280065848 0.00013683686847798526 3167
67
+ 1652280069646 0.0001252585934707895 3215
68
+ 1652280073440 0.0001234338415088132 3263
69
+ 1652280077219 0.0001386790390824899 3311
70
+ 1652280081049 0.000133068097056821 3359
71
+ 1652280084823 0.0001232567592523992 3407
72
+ 1652280088615 0.00010765787737909704 3455
73
+ 1652280092408 0.00010420197213534266 3503
74
+ 1652280096210 0.00011743942741304636 3551
75
+ 1652280099978 0.00013075751485303044 3599
76
+ 1652280103622 0.00012383297143969685 3647
77
+ 1652280107393 0.00012997408339288086 3695
78
+ 1652280111203 0.0001171917756437324 3743
79
+ 1652280115112 0.0001384491624776274 3791
80
+ 1652280118906 0.00011071155313402414 3839
81
+ 1652280122669 0.00012242942466400564 3887
82
+ 1652280126515 0.00012077093560947105 3935
83
+ 1652280130237 0.00011708390229614452 3983
84
+ 1652280134111 0.00013511900033336133 4031
85
+ 1652280137849 0.00010951827425742522 4079
86
+ 1652280141678 0.00011793604062404484 4127
87
+ 1652280145432 0.00011460384848760441 4175
88
+ 1652280149222 0.00012705886911135167 4223
89
+ 1652280152969 0.00011695512512233108 4271
90
+ 1652280156762 0.00011259406892349944 4319
91
+ 1652280160504 0.00011745867959689349 4367
92
+ 1652280164378 0.00013093209418002516 4415
93
+ 1652280168071 0.00013953317829873413 4463
94
+ 1652280171823 0.00012743366823997349 4511
95
+ 1652280175591 0.00011808565614046529 4559
96
+ 1652280179426 0.00015584049106109887 4607
97
+ 1652280183168 0.00013836825382895768 4655
98
+ 1652280186980 0.00012112926197005436 4703
99
+ 1652280190782 0.00012230231368448585 4751
100
+ 1652280194640 0.00011741576599888504 4799
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/mae ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1652279820749 0.07011217623949051 47
2
+ 1652279824405 0.021280117332935333 95
3
+ 1652279828096 0.020297976210713387 143
4
+ 1652279831720 0.020039746537804604 191
5
+ 1652279835495 0.019787773489952087 239
6
+ 1652279839319 0.018337024375796318 287
7
+ 1652279843119 0.017815634608268738 335
8
+ 1652279846878 0.017987234517931938 383
9
+ 1652279850653 0.017149975523352623 431
10
+ 1652279854434 0.01764928176999092 479
11
+ 1652279858281 0.017262466251850128 527
12
+ 1652279862126 0.017195068299770355 575
13
+ 1652279866012 0.01676551066339016 623
14
+ 1652279869817 0.016418082639575005 671
15
+ 1652279873668 0.016572223976254463 719
16
+ 1652279877494 0.016431566327810287 767
17
+ 1652279881296 0.016624923795461655 815
18
+ 1652279885082 0.016285978257656097 863
19
+ 1652279888864 0.015845887362957 911
20
+ 1652279892566 0.016440516337752342 959
21
+ 1652279896290 0.015909109264612198 1007
22
+ 1652279900037 0.015598192811012268 1055
23
+ 1652279903717 0.015478463843464851 1103
24
+ 1652279907559 0.014973259530961514 1151
25
+ 1652279911367 0.014955243095755577 1199
26
+ 1652279915129 0.014734053052961826 1247
27
+ 1652279918986 0.013931847177445889 1295
28
+ 1652279922806 0.013890136033296585 1343
29
+ 1652279926523 0.013945831917226315 1391
30
+ 1652279930377 0.013475497253239155 1439
31
+ 1652279934160 0.012566673569381237 1487
32
+ 1652279937872 0.012109467759728432 1535
33
+ 1652279941677 0.01209949143230915 1583
34
+ 1652279945317 0.011168932542204857 1631
35
+ 1652279949040 0.012050420045852661 1679
36
+ 1652279952812 0.011594988405704498 1727
37
+ 1652279956535 0.011586652137339115 1775
38
+ 1652279960344 0.01181836612522602 1823
39
+ 1652279964110 0.010663148015737534 1871
40
+ 1652279967883 0.010579877533018589 1919
41
+ 1652279971491 0.010919580236077309 1967
42
+ 1652279975165 0.010937352664768696 2015
43
+ 1652279978850 0.010980397462844849 2063
44
+ 1652279982578 0.010383875109255314 2111
45
+ 1652279986365 0.010248820297420025 2159
46
+ 1652279990170 0.009944379329681396 2207
47
+ 1652279994023 0.009952726773917675 2255
48
+ 1652279997827 0.010111919604241848 2303
49
+ 1652280001649 0.010890776291489601 2351
50
+ 1652280005492 0.009552615694701672 2399
51
+ 1652280009187 0.009343606419861317 2447
52
+ 1652280012862 0.00977871473878622 2495
53
+ 1652280016623 0.009527333080768585 2543
54
+ 1652280020351 0.009975402615964413 2591
55
+ 1652280024177 0.008768484927713871 2639
56
+ 1652280027981 0.008977909572422504 2687
57
+ 1652280031839 0.008777343668043613 2735
58
+ 1652280035425 0.008911573328077793 2783
59
+ 1652280039190 0.008944649249315262 2831
60
+ 1652280043047 0.008465489372611046 2879
61
+ 1652280046780 0.008481234312057495 2927
62
+ 1652280050567 0.008217735216021538 2975
63
+ 1652280054496 0.008489103056490421 3023
64
+ 1652280058318 0.00794004462659359 3071
65
+ 1652280062086 0.00839218683540821 3119
66
+ 1652280065848 0.008686883375048637 3167
67
+ 1652280069646 0.008334343321621418 3215
68
+ 1652280073440 0.008174673654139042 3263
69
+ 1652280077219 0.008791987784206867 3311
70
+ 1652280081049 0.008551678620278835 3359
71
+ 1652280084823 0.008277542889118195 3407
72
+ 1652280088615 0.007683832198381424 3455
73
+ 1652280092408 0.007514127530157566 3503
74
+ 1652280096210 0.008030056953430176 3551
75
+ 1652280099978 0.008549727499485016 3599
76
+ 1652280103622 0.008334793150424957 3647
77
+ 1652280107393 0.00849801953881979 3695
78
+ 1652280111203 0.00794607400894165 3743
79
+ 1652280115112 0.008685767650604248 3791
80
+ 1652280118906 0.007812082301825285 3839
81
+ 1652280122669 0.008264751173555851 3887
82
+ 1652280126515 0.008274727500975132 3935
83
+ 1652280130237 0.008091365918517113 3983
84
+ 1652280134111 0.008610659278929234 4031
85
+ 1652280137849 0.007766522467136383 4079
86
+ 1652280141678 0.008045573718845844 4127
87
+ 1652280145432 0.007821462117135525 4175
88
+ 1652280149222 0.008322164416313171 4223
89
+ 1652280152969 0.008126244880259037 4271
90
+ 1652280156762 0.007857552729547024 4319
91
+ 1652280160504 0.008035344071686268 4367
92
+ 1652280164378 0.008491233922541142 4415
93
+ 1652280168071 0.008733570575714111 4463
94
+ 1652280171823 0.008426659740507603 4511
95
+ 1652280175591 0.008046725764870644 4559
96
+ 1652280179426 0.009322610683739185 4607
97
+ 1652280183168 0.008852057158946991 4655
98
+ 1652280186980 0.008111957460641861 4703
99
+ 1652280190782 0.008170234970748425 4751
100
+ 1652280194640 0.008008270524442196 4799
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/loss ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1652279820704 0.0010634043719619513 47
2
+ 1652279824357 0.0008477172814309597 95
3
+ 1652279828044 0.0006586583331227303 143
4
+ 1652279831676 0.0006095922435633838 191
5
+ 1652279835446 0.000579743820708245 239
6
+ 1652279839271 0.000539788044989109 287
7
+ 1652279843079 0.0005701080663129687 335
8
+ 1652279846830 0.0005549467750824988 383
9
+ 1652279850614 0.0005555949755944312 431
10
+ 1652279854390 0.0004789295489899814 479
11
+ 1652279858241 0.0005822633975185454 527
12
+ 1652279862079 0.00048447673907503486 575
13
+ 1652279865967 0.00046181678771972656 623
14
+ 1652279869769 0.00041472347220405936 671
15
+ 1652279873629 0.0005150702781975269 719
16
+ 1652279877456 0.00043978836038149893 767
17
+ 1652279881261 0.0004524872056208551 815
18
+ 1652279885044 0.0004895858000963926 863
19
+ 1652279888829 0.00048630539095029235 911
20
+ 1652279892527 0.00045636019785888493 959
21
+ 1652279896237 0.00042457759263925254 1007
22
+ 1652279900002 0.0005504836444742978 1055
23
+ 1652279903680 0.0005828228895552456 1103
24
+ 1652279907516 0.000536997162271291 1151
25
+ 1652279911251 0.0004296917177271098 1199
26
+ 1652279915084 0.000369079876691103 1247
27
+ 1652279918947 0.0005687376251444221 1295
28
+ 1652279922767 0.0003877778071910143 1343
29
+ 1652279926483 0.0007391179096885026 1391
30
+ 1652279930329 0.0002891223703045398 1439
31
+ 1652279934121 0.00136762042529881 1487
32
+ 1652279937833 0.0005622635362669826 1535
33
+ 1652279941623 0.0002613224496599287 1583
34
+ 1652279945268 0.00026621491997502744 1631
35
+ 1652279949005 0.0003965417272411287 1679
36
+ 1652279952772 0.0006301697576418519 1727
37
+ 1652279956496 0.0004612059856299311 1775
38
+ 1652279960298 0.00025420173187740147 1823
39
+ 1652279964073 0.0003383368020877242 1871
40
+ 1652279967844 0.0003864422906190157 1919
41
+ 1652279971446 0.00022850626555737108 1967
42
+ 1652279975126 0.00036217307206243277 2015
43
+ 1652279978811 0.00026963415439240634 2063
44
+ 1652279982539 0.0003081822069361806 2111
45
+ 1652279986321 0.0008588534547016025 2159
46
+ 1652279990134 0.00025588576681911945 2207
47
+ 1652279993983 0.0012111393734812737 2255
48
+ 1652279997788 0.0005991487414576113 2303
49
+ 1652280001609 0.00027215175214223564 2351
50
+ 1652280005406 0.0002326530375285074 2399
51
+ 1652280009144 0.00034161240910179913 2447
52
+ 1652280012811 0.0002047311281785369 2495
53
+ 1652280016584 0.0002189291553804651 2543
54
+ 1652280020312 0.0008305492228828371 2591
55
+ 1652280024138 0.0003245118132326752 2639
56
+ 1652280027939 0.00043747638119384646 2687
57
+ 1652280031799 0.0003352710627950728 2735
58
+ 1652280035376 0.00017761132039595395 2783
59
+ 1652280039147 0.0002451064356137067 2831
60
+ 1652280043008 0.00045808948925696313 2879
61
+ 1652280046735 0.00016568403225392103 2927
62
+ 1652280050532 0.0002005849382840097 2975
63
+ 1652280054457 0.0004337916907388717 3023
64
+ 1652280058279 0.00043202925007790327 3071
65
+ 1652280062047 0.0006070298259146512 3119
66
+ 1652280065810 0.00125686835963279 3167
67
+ 1652280069610 0.0003458949795458466 3215
68
+ 1652280073394 0.00023582404537592083 3263
69
+ 1652280077179 0.00043027440551668406 3311
70
+ 1652280081010 0.00018619702314026654 3359
71
+ 1652280084783 0.0004644407599698752 3407
72
+ 1652280088575 0.00025989525602199137 3455
73
+ 1652280092368 0.000271034223260358 3503
74
+ 1652280096159 0.00015595319564454257 3551
75
+ 1652280099894 0.00018362083937972784 3599
76
+ 1652280103580 0.0001723619207041338 3647
77
+ 1652280107354 0.00026339536998420954 3695
78
+ 1652280111162 0.0022478969767689705 3743
79
+ 1652280114987 0.00043688094592653215 3791
80
+ 1652280118870 0.000253666948992759 3839
81
+ 1652280122629 0.0007027353858575225 3887
82
+ 1652280126475 0.00036985089536756277 3935
83
+ 1652280130198 0.00042630801908671856 3983
84
+ 1652280134070 0.00018512520182412118 4031
85
+ 1652280137810 0.0006041005253791809 4079
86
+ 1652280141642 0.0005739088519476354 4127
87
+ 1652280145396 0.00020085058349650353 4175
88
+ 1652280149183 0.00035394675796851516 4223
89
+ 1652280152927 0.001467306399717927 4271
90
+ 1652280156721 0.0003908797516487539 4319
91
+ 1652280160468 0.0001805946958484128 4367
92
+ 1652280164338 0.00035339672467671335 4415
93
+ 1652280168035 0.00040678223012946546 4463
94
+ 1652280171787 0.00018399060354568064 4511
95
+ 1652280175551 0.0005399276851676404 4559
96
+ 1652280179385 0.00029979468672536314 4607
97
+ 1652280183124 0.00013096585462335497 4655
98
+ 1652280186940 0.00021977233700454235 4703
99
+ 1652280190746 0.00024704227689653635 4751
100
+ 1652280194552 0.0003437100094743073 4799
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1652279820704 0.02607358619570732 47
2
+ 1652279824357 0.023356162011623383 95
3
+ 1652279828044 0.02111729420721531 143
4
+ 1652279831676 0.020395344123244286 191
5
+ 1652279835446 0.02008357085287571 239
6
+ 1652279839271 0.019109351560473442 287
7
+ 1652279843079 0.020062454044818878 335
8
+ 1652279846830 0.018724771216511726 383
9
+ 1652279850614 0.01976374350488186 431
10
+ 1652279854390 0.018088672310113907 479
11
+ 1652279858241 0.02009824477136135 527
12
+ 1652279862079 0.017433589324355125 575
13
+ 1652279865967 0.018075214698910713 623
14
+ 1652279869769 0.016927972435951233 671
15
+ 1652279873629 0.01795283891260624 719
16
+ 1652279877456 0.017694905400276184 767
17
+ 1652279881261 0.017309539020061493 815
18
+ 1652279885044 0.017853353172540665 863
19
+ 1652279888829 0.017320722341537476 911
20
+ 1652279892527 0.017117636278271675 959
21
+ 1652279896237 0.01685478165745735 1007
22
+ 1652279900002 0.01946953497827053 1055
23
+ 1652279903680 0.019640492275357246 1103
24
+ 1652279907516 0.01877320185303688 1151
25
+ 1652279911251 0.016905883327126503 1199
26
+ 1652279915084 0.015234655700623989 1247
27
+ 1652279918947 0.019466156139969826 1295
28
+ 1652279922767 0.01577661745250225 1343
29
+ 1652279926483 0.021662525832653046 1391
30
+ 1652279930329 0.013201462104916573 1439
31
+ 1652279934121 0.03221268951892853 1487
32
+ 1652279937833 0.018861766904592514 1535
33
+ 1652279941623 0.012555716559290886 1583
34
+ 1652279945268 0.01243590097874403 1631
35
+ 1652279949005 0.015400874428451061 1679
36
+ 1652279952772 0.019847549498081207 1727
37
+ 1652279956496 0.01633538119494915 1775
38
+ 1652279960298 0.012520545162260532 1823
39
+ 1652279964073 0.015009582974016666 1871
40
+ 1652279967844 0.016114825382828712 1919
41
+ 1652279971446 0.011395555920898914 1967
42
+ 1652279975126 0.015877339988946915 2015
43
+ 1652279978811 0.012874774634838104 2063
44
+ 1652279982539 0.01374065037816763 2111
45
+ 1652279986321 0.02455970272421837 2159
46
+ 1652279990134 0.012964467518031597 2207
47
+ 1652279993983 0.030586877837777138 2255
48
+ 1652279997788 0.01970861665904522 2303
49
+ 1652280001609 0.012582853436470032 2351
50
+ 1652280005406 0.011895284987986088 2399
51
+ 1652280009144 0.014539946801960468 2447
52
+ 1652280012811 0.010513107292354107 2495
53
+ 1652280016584 0.011585352011024952 2543
54
+ 1652280020312 0.025230778381228447 2591
55
+ 1652280024138 0.01448488887399435 2639
56
+ 1652280027939 0.016850745305418968 2687
57
+ 1652280031799 0.014916492626070976 2735
58
+ 1652280035376 0.009772807359695435 2783
59
+ 1652280039147 0.012413904070854187 2831
60
+ 1652280043008 0.017842639237642288 2879
61
+ 1652280046735 0.009426051750779152 2927
62
+ 1652280050532 0.011247269809246063 2975
63
+ 1652280054457 0.017680322751402855 3023
64
+ 1652280058279 0.016938434913754463 3071
65
+ 1652280062047 0.020868053659796715 3119
66
+ 1652280065810 0.03198228031396866 3167
67
+ 1652280069610 0.01481217984110117 3215
68
+ 1652280073394 0.01217477023601532 3263
69
+ 1652280077179 0.0175019484013319 3311
70
+ 1652280081010 0.01059352234005928 3359
71
+ 1652280084783 0.017586909234523773 3407
72
+ 1652280088575 0.013158340007066727 3455
73
+ 1652280092368 0.013016403652727604 3503
74
+ 1652280096159 0.00941098015755415 3551
75
+ 1652280099894 0.010173797607421875 3599
76
+ 1652280103580 0.010034985840320587 3647
77
+ 1652280107354 0.01230852585285902 3695
78
+ 1652280111162 0.043913356959819794 3743
79
+ 1652280114987 0.01787818782031536 3791
80
+ 1652280118870 0.012860489077866077 3839
81
+ 1652280122629 0.02276649698615074 3887
82
+ 1652280126475 0.015199720859527588 3935
83
+ 1652280130198 0.016393927857279778 3983
84
+ 1652280134070 0.010693005286157131 4031
85
+ 1652280137810 0.02147362008690834 4079
86
+ 1652280141642 0.019695602357387543 4127
87
+ 1652280145396 0.010959284380078316 4175
88
+ 1652280149183 0.014944094233214855 4223
89
+ 1652280152927 0.03422271087765694 4271
90
+ 1652280156721 0.015736214816570282 4319
91
+ 1652280160468 0.009996251203119755 4367
92
+ 1652280164338 0.015099494718015194 4415
93
+ 1652280168035 0.015847915783524513 4463
94
+ 1652280171787 0.010453518480062485 4511
95
+ 1652280175551 0.018864858895540237 4559
96
+ 1652280179385 0.01469734963029623 4607
97
+ 1652280183124 0.00869279820472002 4655
98
+ 1652280186940 0.011555126868188381 4703
99
+ 1652280190746 0.01307331956923008 4751
100
+ 1652280194552 0.01433614082634449 4799
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae_best ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1652279820704 0.02607358619570732 47
2
+ 1652279824357 0.023356162011623383 95
3
+ 1652279828044 0.02111729420721531 143
4
+ 1652279831676 0.020395344123244286 191
5
+ 1652279835446 0.02008357085287571 239
6
+ 1652279839271 0.019109351560473442 287
7
+ 1652279843079 0.019109351560473442 335
8
+ 1652279846830 0.018724771216511726 383
9
+ 1652279850614 0.018724771216511726 431
10
+ 1652279854390 0.018088672310113907 479
11
+ 1652279858241 0.018088672310113907 527
12
+ 1652279862079 0.017433589324355125 575
13
+ 1652279865967 0.017433589324355125 623
14
+ 1652279869769 0.016927972435951233 671
15
+ 1652279873629 0.016927972435951233 719
16
+ 1652279877456 0.016927972435951233 767
17
+ 1652279881261 0.016927972435951233 815
18
+ 1652279885044 0.016927972435951233 863
19
+ 1652279888829 0.016927972435951233 911
20
+ 1652279892527 0.016927972435951233 959
21
+ 1652279896237 0.01685478165745735 1007
22
+ 1652279900002 0.01685478165745735 1055
23
+ 1652279903680 0.01685478165745735 1103
24
+ 1652279907516 0.01685478165745735 1151
25
+ 1652279911251 0.01685478165745735 1199
26
+ 1652279915084 0.015234655700623989 1247
27
+ 1652279918947 0.015234655700623989 1295
28
+ 1652279922767 0.015234655700623989 1343
29
+ 1652279926483 0.015234655700623989 1391
30
+ 1652279930329 0.013201462104916573 1439
31
+ 1652279934121 0.013201462104916573 1487
32
+ 1652279937833 0.013201462104916573 1535
33
+ 1652279941623 0.012555716559290886 1583
34
+ 1652279945268 0.01243590097874403 1631
35
+ 1652279949005 0.01243590097874403 1679
36
+ 1652279952772 0.01243590097874403 1727
37
+ 1652279956496 0.01243590097874403 1775
38
+ 1652279960298 0.01243590097874403 1823
39
+ 1652279964073 0.01243590097874403 1871
40
+ 1652279967844 0.01243590097874403 1919
41
+ 1652279971446 0.011395555920898914 1967
42
+ 1652279975126 0.011395555920898914 2015
43
+ 1652279978811 0.011395555920898914 2063
44
+ 1652279982539 0.011395555920898914 2111
45
+ 1652279986321 0.011395555920898914 2159
46
+ 1652279990134 0.011395555920898914 2207
47
+ 1652279993983 0.011395555920898914 2255
48
+ 1652279997788 0.011395555920898914 2303
49
+ 1652280001609 0.011395555920898914 2351
50
+ 1652280005406 0.011395555920898914 2399
51
+ 1652280009144 0.011395555920898914 2447
52
+ 1652280012811 0.010513107292354107 2495
53
+ 1652280016584 0.010513107292354107 2543
54
+ 1652280020312 0.010513107292354107 2591
55
+ 1652280024138 0.010513107292354107 2639
56
+ 1652280027939 0.010513107292354107 2687
57
+ 1652280031799 0.010513107292354107 2735
58
+ 1652280035376 0.009772807359695435 2783
59
+ 1652280039147 0.009772807359695435 2831
60
+ 1652280043008 0.009772807359695435 2879
61
+ 1652280046735 0.009426051750779152 2927
62
+ 1652280050532 0.009426051750779152 2975
63
+ 1652280054457 0.009426051750779152 3023
64
+ 1652280058279 0.009426051750779152 3071
65
+ 1652280062047 0.009426051750779152 3119
66
+ 1652280065810 0.009426051750779152 3167
67
+ 1652280069610 0.009426051750779152 3215
68
+ 1652280073394 0.009426051750779152 3263
69
+ 1652280077179 0.009426051750779152 3311
70
+ 1652280081010 0.009426051750779152 3359
71
+ 1652280084783 0.009426051750779152 3407
72
+ 1652280088575 0.009426051750779152 3455
73
+ 1652280092368 0.009426051750779152 3503
74
+ 1652280096159 0.00941098015755415 3551
75
+ 1652280099894 0.00941098015755415 3599
76
+ 1652280103580 0.00941098015755415 3647
77
+ 1652280107354 0.00941098015755415 3695
78
+ 1652280111162 0.00941098015755415 3743
79
+ 1652280114987 0.00941098015755415 3791
80
+ 1652280118870 0.00941098015755415 3839
81
+ 1652280122629 0.00941098015755415 3887
82
+ 1652280126475 0.00941098015755415 3935
83
+ 1652280130198 0.00941098015755415 3983
84
+ 1652280134070 0.00941098015755415 4031
85
+ 1652280137810 0.00941098015755415 4079
86
+ 1652280141642 0.00941098015755415 4127
87
+ 1652280145396 0.00941098015755415 4175
88
+ 1652280149183 0.00941098015755415 4223
89
+ 1652280152927 0.00941098015755415 4271
90
+ 1652280156721 0.00941098015755415 4319
91
+ 1652280160468 0.00941098015755415 4367
92
+ 1652280164338 0.00941098015755415 4415
93
+ 1652280168035 0.00941098015755415 4463
94
+ 1652280171787 0.00941098015755415 4511
95
+ 1652280175551 0.00941098015755415 4559
96
+ 1652280179385 0.00941098015755415 4607
97
+ 1652280183124 0.00869279820472002 4655
98
+ 1652280186940 0.00869279820472002 4703
99
+ 1652280190746 0.00869279820472002 4751
100
+ 1652280194552 0.00869279820472002 4799
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.EarlyStopping
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/min_delta ADDED
@@ -0,0 +1 @@
 
 
1
+ 0
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/mode ADDED
@@ -0,0 +1 @@
 
 
1
+ min
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/monitor ADDED
@@ -0,0 +1 @@
 
 
1
+ val/mae
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/patience ADDED
@@ -0,0 +1 @@
 
 
1
+ 100
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.ModelCheckpoint
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/auto_insert_metric_name ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/dirpath ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoints/
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/filename ADDED
@@ -0,0 +1 @@
 
 
1
+ epoch_{epoch:03d}
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/mode ADDED
@@ -0,0 +1 @@
 
 
1
+ min
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/monitor ADDED
@@ -0,0 +1 @@
 
 
1
+ val/mae
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_last ADDED
@@ -0,0 +1 @@
 
 
1
+ True
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_top_k ADDED
@@ -0,0 +1 @@
 
 
1
+ 1
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/verbose ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.RichModelSummary
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/max_depth ADDED
@@ -0,0 +1 @@
 
 
1
+ -1
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/rich_progress_bar/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.RichProgressBar
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ src.datamodules.focus_datamodule.FocusDataModule
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/augmentation ADDED
@@ -0,0 +1 @@
 
 
1
+ True
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/batch_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 64