Hannes Kuchelmeister commited on
Commit
1a50bbb
1 Parent(s): eac0701

remove all runs due to results being not useful (error in loss function)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/config.yaml +0 -66
  2. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/hydra.yaml +0 -232
  3. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/overrides.yaml +0 -6
  4. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393430.e02c2cd1d8ab.1.0 +0 -3
  5. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393675.e02c2cd1d8ab.1.1 +0 -3
  6. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/hparams.yaml +0 -52
  7. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/config.yaml +0 -66
  8. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/hydra.yaml +0 -232
  9. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/overrides.yaml +0 -6
  10. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646393676.e02c2cd1d8ab.1.2 +0 -3
  11. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646394071.e02c2cd1d8ab.1.3 +0 -3
  12. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/hparams.yaml +0 -52
  13. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/config.yaml +0 -66
  14. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/hydra.yaml +0 -232
  15. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/overrides.yaml +0 -6
  16. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646399561.e02c2cd1d8ab.1.19 +0 -3
  17. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646400250.e02c2cd1d8ab.1.20 +0 -3
  18. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/hparams.yaml +0 -52
  19. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/config.yaml +0 -66
  20. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/hydra.yaml +0 -232
  21. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/overrides.yaml +0 -6
  22. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400251.e02c2cd1d8ab.1.21 +0 -3
  23. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400340.e02c2cd1d8ab.1.22 +0 -3
  24. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/hparams.yaml +0 -52
  25. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/config.yaml +0 -66
  26. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/hydra.yaml +0 -232
  27. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/overrides.yaml +0 -6
  28. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400341.e02c2cd1d8ab.1.23 +0 -3
  29. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.24 +0 -3
  30. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/hparams.yaml +0 -52
  31. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/config.yaml +0 -66
  32. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/hydra.yaml +0 -232
  33. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/overrides.yaml +0 -6
  34. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.25 +0 -3
  35. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400569.e02c2cd1d8ab.1.26 +0 -3
  36. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/hparams.yaml +0 -52
  37. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/config.yaml +0 -66
  38. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/hydra.yaml +0 -232
  39. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/overrides.yaml +0 -6
  40. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646400570.e02c2cd1d8ab.1.27 +0 -3
  41. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646403523.e02c2cd1d8ab.1.28 +0 -3
  42. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/hparams.yaml +0 -52
  43. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/config.yaml +0 -66
  44. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/hydra.yaml +0 -232
  45. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/overrides.yaml +0 -6
  46. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403524.e02c2cd1d8ab.1.29 +0 -3
  47. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403650.e02c2cd1d8ab.1.30 +0 -3
  48. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/hparams.yaml +0 -52
  49. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/config.yaml +0 -66
  50. logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/hydra.yaml +0 -232
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 32
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 128
24
- lin2_size: 32
25
- lin3_size: 32
26
- output_size: 1
27
- lr: 0.04099159968275264
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=32
177
- - model.lr=0.04099159968275264
178
- - model.lin1_size=128
179
- - model.lin2_size=32
180
- - model.lin3_size=32
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=32,hparams_search=focus150_optuna,model.lin1_size=128,model.lin2_size=32,model.lin3_size=32,model.lr=0.04099159968275264
185
- id: '0'
186
- num: 0
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=32
2
- - model.lr=0.04099159968275264
3
- - model.lin1_size=128
4
- - model.lin2_size=32
5
- - model.lin3_size=32
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393430.e02c2cd1d8ab.1.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6afc0f43b3c1bd7933f2a0e2d56b93f74680505b96c42a15f8970c35864b322
3
- size 36391
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393675.e02c2cd1d8ab.1.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7356d4d877d5517b1738d3fbb03153b84e6a6c09d506bbabbdffcebca279aba
3
- size 179
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 128
11
- lin2_size: 32
12
- lin3_size: 32
13
- output_size: 1
14
- lr: 0.04099159968275264
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 32
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 8645729
51
- model/params/trainable: 8645729
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 64
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 256
24
- lin2_size: 128
25
- lin3_size: 1024
26
- output_size: 1
27
- lr: 0.15818542131777294
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=64
177
- - model.lr=0.15818542131777294
178
- - model.lin1_size=256
179
- - model.lin2_size=128
180
- - model.lin3_size=1024
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=1024,model.lr=0.15818542131777294
185
- id: '1'
186
- num: 1
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=64
2
- - model.lr=0.15818542131777294
3
- - model.lin1_size=256
4
- - model.lin2_size=128
5
- - model.lin3_size=1024
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646393676.e02c2cd1d8ab.1.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f6997fead061ad1b7ca1a54e9df1c6e59c656a41f5642597c9791fd1f5944a9
3
- size 36328
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646394071.e02c2cd1d8ab.1.3 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:98d4647ce1487bed60e00a1b0991c7f1a1d6298d661b2af6befa36b542c3d805
3
- size 179
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 256
11
- lin2_size: 128
12
- lin3_size: 1024
13
- output_size: 1
14
- lr: 0.15818542131777294
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 64
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 17449089
51
- model/params/trainable: 17449089
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 128
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 512
24
- lin2_size: 512
25
- lin3_size: 256
26
- output_size: 1
27
- lr: 0.07148210917089061
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=128
177
- - model.lr=0.07148210917089061
178
- - model.lin1_size=512
179
- - model.lin2_size=512
180
- - model.lin3_size=256
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=512,model.lin2_size=512,model.lin3_size=256,model.lr=0.07148210917089061
185
- id: '10'
186
- num: 10
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=128
2
- - model.lr=0.07148210917089061
3
- - model.lin1_size=512
4
- - model.lin2_size=512
5
- - model.lin3_size=256
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646399561.e02c2cd1d8ab.1.19 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3093c010acb1da620f30c23572d0cfd10dfd6b38f5b81062c7b070360f3d326e
3
- size 36230
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646400250.e02c2cd1d8ab.1.20 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5dcb14bc03802b02af910756b125403e6a53e54513795ee022d8e68555492ce9
3
- size 176
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 512
11
- lin2_size: 512
12
- lin3_size: 256
13
- output_size: 1
14
- lr: 0.07148210917089061
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 128
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 34957313
51
- model/params/trainable: 34957313
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 64
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 32
24
- lin2_size: 256
25
- lin3_size: 128
26
- output_size: 1
27
- lr: 0.11779269530261958
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=64
177
- - model.lr=0.11779269530261958
178
- - model.lin1_size=32
179
- - model.lin2_size=256
180
- - model.lin3_size=128
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=256,model.lin3_size=128,model.lr=0.11779269530261958
185
- id: '11'
186
- num: 11
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=64
2
- - model.lr=0.11779269530261958
3
- - model.lin1_size=32
4
- - model.lin2_size=256
5
- - model.lin3_size=128
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400251.e02c2cd1d8ab.1.21 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:99dcbe2276408d1477fed434497dfedc2e1c55e37deff1fec0513c0df7336762
3
- size 36328
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400340.e02c2cd1d8ab.1.22 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:303733f02185c2365f205d30101dd2f463231152f5186c6366958d57f32db557
3
- size 179
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 32
11
- lin2_size: 256
12
- lin3_size: 128
13
- output_size: 1
14
- lr: 0.11779269530261958
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 64
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 2202337
51
- model/params/trainable: 2202337
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 64
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 32
24
- lin2_size: 256
25
- lin3_size: 128
26
- output_size: 1
27
- lr: 0.13355237942960652
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=64
177
- - model.lr=0.13355237942960652
178
- - model.lin1_size=32
179
- - model.lin2_size=256
180
- - model.lin3_size=128
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=256,model.lin3_size=128,model.lr=0.13355237942960652
185
- id: '12'
186
- num: 12
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=64
2
- - model.lr=0.13355237942960652
3
- - model.lin1_size=32
4
- - model.lin2_size=256
5
- - model.lin3_size=128
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400341.e02c2cd1d8ab.1.23 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8869683dbbc16e523ec4daec9634a5844e47c2d42cb5e48fcde03b47f057c09b
3
- size 36328
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.24 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1989b8df8bbec51baec3199e84b0bff5cb0e9f00f01c15396087aeede34c5e25
3
- size 176
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 32
11
- lin2_size: 256
12
- lin3_size: 128
13
- output_size: 1
14
- lr: 0.13355237942960652
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 64
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 2202337
51
- model/params/trainable: 2202337
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 128
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 32
24
- lin2_size: 512
25
- lin3_size: 2048
26
- output_size: 1
27
- lr: 0.07220274494059103
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=128
177
- - model.lr=0.07220274494059103
178
- - model.lin1_size=32
179
- - model.lin2_size=512
180
- - model.lin3_size=2048
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=512,model.lin3_size=2048,model.lr=0.07220274494059103
185
- id: '13'
186
- num: 13
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=128
2
- - model.lr=0.07220274494059103
3
- - model.lin1_size=32
4
- - model.lin2_size=512
5
- - model.lin3_size=2048
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.25 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d966126e50625b723d1aa3ea74101b1e0012a80c8ce55c9b9a7b9b0fc7e838c4
3
- size 36230
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400569.e02c2cd1d8ab.1.26 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d80b675e81b6ba40bda511fec56d0a2eb745dccb51a04e4d428b5d9216b44714
3
- size 179
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 32
11
- lin2_size: 512
12
- lin3_size: 2048
13
- output_size: 1
14
- lr: 0.07220274494059103
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 128
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 3234785
51
- model/params/trainable: 3234785
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 128
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 2048
24
- lin2_size: 512
25
- lin3_size: 128
26
- output_size: 1
27
- lr: 0.08335231197967785
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=128
177
- - model.lr=0.08335231197967785
178
- - model.lin1_size=2048
179
- - model.lin2_size=512
180
- - model.lin3_size=128
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=2048,model.lin2_size=512,model.lin3_size=128,model.lr=0.08335231197967785
185
- id: '14'
186
- num: 14
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=128
2
- - model.lr=0.08335231197967785
3
- - model.lin1_size=2048
4
- - model.lin2_size=512
5
- - model.lin3_size=128
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646400570.e02c2cd1d8ab.1.27 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d05f9112c18daa456c3ae5551f39f14bd694958d48a3292ebee05fefe8a256a
3
- size 36230
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646403523.e02c2cd1d8ab.1.28 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:70fb6b94c38eb4ba4538834d1114cf2f128d7d9f729d09e10798d95e0a6ae839
3
- size 179
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 2048
11
- lin2_size: 512
12
- lin3_size: 128
13
- output_size: 1
14
- lr: 0.08335231197967785
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 128
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 139362305
51
- model/params/trainable: 139362305
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 64
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 64
24
- lin2_size: 256
25
- lin3_size: 128
26
- output_size: 1
27
- lr: 0.04131276528182434
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=64
177
- - model.lr=0.04131276528182434
178
- - model.lin1_size=64
179
- - model.lin2_size=256
180
- - model.lin3_size=128
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=64,model.lin2_size=256,model.lin3_size=128,model.lr=0.04131276528182434
185
- id: '15'
186
- num: 15
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/overrides.yaml DELETED
@@ -1,6 +0,0 @@
1
- - datamodule.batch_size=64
2
- - model.lr=0.04131276528182434
3
- - model.lin1_size=64
4
- - model.lin2_size=256
5
- - model.lin3_size=128
6
- - hparams_search=focus150_optuna
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403524.e02c2cd1d8ab.1.29 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f5f63724068b3fe2d9530ced3aa79622ad58b3a5c1e48f29b09a3c23fd909d6
3
- size 36328
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403650.e02c2cd1d8ab.1.30 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c94ff7b41ccf935cdb8bcd338c81f331607912868cfeecc00906530eb326bae2
3
- size 179
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/hparams.yaml DELETED
@@ -1,52 +0,0 @@
1
- trainer:
2
- _target_: pytorch_lightning.Trainer
3
- gpus: 1
4
- min_epochs: 1
5
- max_epochs: 100
6
- resume_from_checkpoint: null
7
- model:
8
- _target_: src.models.focus_module.FocusLitModule
9
- input_size: 67500
10
- lin1_size: 64
11
- lin2_size: 256
12
- lin3_size: 128
13
- output_size: 1
14
- lr: 0.04131276528182434
15
- weight_decay: 0.0005
16
- datamodule:
17
- _target_: src.datamodules.focus_datamodule.FocusDataModule
18
- data_dir: /usr/src/app/data/focus150
19
- csv_file: /usr/src/app/data/focus150/metadata.csv
20
- batch_size: 64
21
- train_val_test_split_percentage:
22
- - 0.7
23
- - 0.15
24
- - 0.15
25
- num_workers: 0
26
- pin_memory: false
27
- seed: null
28
- callbacks:
29
- model_checkpoint:
30
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
- monitor: val/mae
32
- mode: min
33
- save_top_k: 1
34
- save_last: true
35
- verbose: false
36
- dirpath: checkpoints/
37
- filename: epoch_{epoch:03d}
38
- auto_insert_metric_name: false
39
- early_stopping:
40
- _target_: pytorch_lightning.callbacks.EarlyStopping
41
- monitor: val/mae
42
- mode: min
43
- patience: 100
44
- min_delta: 0
45
- model_summary:
46
- _target_: pytorch_lightning.callbacks.RichModelSummary
47
- max_depth: -1
48
- rich_progress_bar:
49
- _target_: pytorch_lightning.callbacks.RichProgressBar
50
- model/params/total: 4370625
51
- model/params/trainable: 4370625
52
- model/params/non_trainable: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/config.yaml DELETED
@@ -1,66 +0,0 @@
1
- original_work_dir: ${hydra:runtime.cwd}
2
- data_dir: ${original_work_dir}/data
3
- print_config: true
4
- ignore_warnings: true
5
- train: true
6
- test: true
7
- seed: null
8
- name: focus150_optuna
9
- datamodule:
10
- _target_: src.datamodules.focus_datamodule.FocusDataModule
11
- data_dir: ${data_dir}/focus150
12
- csv_file: ${data_dir}/focus150/metadata.csv
13
- batch_size: 128
14
- train_val_test_split_percentage:
15
- - 0.7
16
- - 0.15
17
- - 0.15
18
- num_workers: 0
19
- pin_memory: false
20
- model:
21
- _target_: src.models.focus_module.FocusLitModule
22
- input_size: 67500
23
- lin1_size: 512
24
- lin2_size: 32
25
- lin3_size: 2048
26
- output_size: 1
27
- lr: 0.12875356688119352
28
- weight_decay: 0.0005
29
- callbacks:
30
- model_checkpoint:
31
- _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
- monitor: val/mae
33
- mode: min
34
- save_top_k: 1
35
- save_last: true
36
- verbose: false
37
- dirpath: checkpoints/
38
- filename: epoch_{epoch:03d}
39
- auto_insert_metric_name: false
40
- early_stopping:
41
- _target_: pytorch_lightning.callbacks.EarlyStopping
42
- monitor: val/mae
43
- mode: min
44
- patience: 100
45
- min_delta: 0
46
- model_summary:
47
- _target_: pytorch_lightning.callbacks.RichModelSummary
48
- max_depth: -1
49
- rich_progress_bar:
50
- _target_: pytorch_lightning.callbacks.RichProgressBar
51
- logger:
52
- tensorboard:
53
- _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
54
- save_dir: tensorboard/
55
- name: null
56
- version: ${name}
57
- log_graph: false
58
- default_hp_metric: true
59
- prefix: ''
60
- trainer:
61
- _target_: pytorch_lightning.Trainer
62
- gpus: 1
63
- min_epochs: 1
64
- max_epochs: 100
65
- resume_from_checkpoint: null
66
- optimized_metric: val/mae_best
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/hydra.yaml DELETED
@@ -1,232 +0,0 @@
1
- hydra:
2
- run:
3
- dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
- sweep:
5
- dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- sampler:
11
- _target_: optuna.samplers.TPESampler
12
- seed: 12345
13
- consider_prior: true
14
- prior_weight: 1.0
15
- consider_magic_clip: true
16
- consider_endpoints: false
17
- n_startup_trials: 10
18
- n_ei_candidates: 24
19
- multivariate: false
20
- warn_independent_sampling: true
21
- _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
- direction: minimize
23
- storage: null
24
- study_name: focus150_hyperparameter_search
25
- n_trials: 25
26
- n_jobs: 1
27
- search_space:
28
- datamodule.batch_size:
29
- type: categorical
30
- choices:
31
- - 32
32
- - 64
33
- - 128
34
- model.lr:
35
- type: float
36
- low: 0.0001
37
- high: 0.2
38
- model.lin1_size:
39
- type: categorical
40
- choices:
41
- - 32
42
- - 64
43
- - 128
44
- - 256
45
- - 512
46
- - 1024
47
- - 2048
48
- model.lin2_size:
49
- type: categorical
50
- choices:
51
- - 32
52
- - 64
53
- - 128
54
- - 256
55
- - 512
56
- - 1024
57
- - 2048
58
- model.lin3_size:
59
- type: categorical
60
- choices:
61
- - 32
62
- - 64
63
- - 128
64
- - 256
65
- - 512
66
- - 1024
67
- - 2048
68
- help:
69
- app_name: ${hydra.job.name}
70
- header: '${hydra.help.app_name} is powered by Hydra.
71
-
72
- '
73
- footer: 'Powered by Hydra (https://hydra.cc)
74
-
75
- Use --hydra-help to view Hydra specific help
76
-
77
- '
78
- template: '${hydra.help.header}
79
-
80
- == Configuration groups ==
81
-
82
- Compose your configuration from those groups (group=option)
83
-
84
-
85
- $APP_CONFIG_GROUPS
86
-
87
-
88
- == Config ==
89
-
90
- Override anything in the config (foo.bar=value)
91
-
92
-
93
- $CONFIG
94
-
95
-
96
- ${hydra.help.footer}
97
-
98
- '
99
- hydra_help:
100
- template: 'Hydra (${hydra.runtime.version})
101
-
102
- See https://hydra.cc for more info.
103
-
104
-
105
- == Flags ==
106
-
107
- $FLAGS_HELP
108
-
109
-
110
- == Configuration groups ==
111
-
112
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
113
- to command line)
114
-
115
-
116
- $HYDRA_CONFIG_GROUPS
117
-
118
-
119
- Use ''--cfg hydra'' to Show the Hydra config.
120
-
121
- '
122
- hydra_help: ???
123
- hydra_logging:
124
- version: 1
125
- formatters:
126
- colorlog:
127
- (): colorlog.ColoredFormatter
128
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
129
- handlers:
130
- console:
131
- class: logging.StreamHandler
132
- formatter: colorlog
133
- stream: ext://sys.stdout
134
- root:
135
- level: INFO
136
- handlers:
137
- - console
138
- disable_existing_loggers: false
139
- job_logging:
140
- version: 1
141
- formatters:
142
- simple:
143
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
144
- colorlog:
145
- (): colorlog.ColoredFormatter
146
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
147
- - %(message)s'
148
- log_colors:
149
- DEBUG: purple
150
- INFO: green
151
- WARNING: yellow
152
- ERROR: red
153
- CRITICAL: red
154
- handlers:
155
- console:
156
- class: logging.StreamHandler
157
- formatter: colorlog
158
- stream: ext://sys.stdout
159
- file:
160
- class: logging.FileHandler
161
- formatter: simple
162
- filename: ${hydra.job.name}.log
163
- root:
164
- level: INFO
165
- handlers:
166
- - console
167
- - file
168
- disable_existing_loggers: false
169
- env: {}
170
- searchpath: []
171
- callbacks: {}
172
- output_subdir: .hydra
173
- overrides:
174
- hydra: []
175
- task:
176
- - datamodule.batch_size=128
177
- - model.lr=0.12875356688119352
178
- - model.lin1_size=512
179
- - model.lin2_size=32
180
- - model.lin3_size=2048
181
- - hparams_search=focus150_optuna
182
- job:
183
- name: train
184
- override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=512,model.lin2_size=32,model.lin3_size=2048,model.lr=0.12875356688119352
185
- id: '16'
186
- num: 16
187
- config_name: train.yaml
188
- env_set: {}
189
- env_copy: []
190
- config:
191
- override_dirname:
192
- kv_sep: '='
193
- item_sep: ','
194
- exclude_keys: []
195
- runtime:
196
- version: 1.1.1
197
- cwd: /usr/src/app
198
- config_sources:
199
- - path: hydra.conf
200
- schema: pkg
201
- provider: hydra
202
- - path: /usr/src/app/configs
203
- schema: file
204
- provider: main
205
- - path: hydra_plugins.hydra_colorlog.conf
206
- schema: pkg
207
- provider: hydra-colorlog
208
- - path: ''
209
- schema: structured
210
- provider: schema
211
- choices:
212
- local: default.yaml
213
- hparams_search: focus150_optuna
214
- debug: null
215
- experiment: null
216
- log_dir: default.yaml
217
- trainer: long.yaml
218
- logger: tensorboard
219
- callbacks: default.yaml
220
- model: focus150.yaml
221
- datamodule: focus150.yaml
222
- hydra/env: default
223
- hydra/callbacks: null
224
- hydra/job_logging: colorlog
225
- hydra/hydra_logging: colorlog
226
- hydra/hydra_help: default
227
- hydra/help: default
228
- hydra/sweeper: optuna
229
- hydra/sweeper/sampler: tpe
230
- hydra/launcher: basic
231
- hydra/output: default
232
- verbose: false