diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fde4c59acb39619445daffac88d5c5b808419ca3 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 128 + lin2_size: 32 + lin3_size: 32 + output_size: 1 + lr: 0.04099159968275264 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..715e3ccf1d9d889704d1c60460518384b0f47e08 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.04099159968275264 + - model.lin1_size=128 + - model.lin2_size=32 + - model.lin3_size=32 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus150_optuna,model.lin1_size=128,model.lin2_size=32,model.lin3_size=32,model.lr=0.04099159968275264 + id: '0' + num: 0 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6197a0dd92e0c051edb233941f3017e4a6ed16cf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.04099159968275264 +- model.lin1_size=128 +- model.lin2_size=32 +- model.lin3_size=32 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/checkpoints/epoch_035.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/checkpoints/epoch_035.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..8617e25bacfbb9a5702eec16436e8c22690d4f5f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/checkpoints/epoch_035.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76789e42b047a8b83c49e2852bfa0fd7081d2978df6085698bc85bb6e1cc00f4 +size 103767494 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..1035ad74ee68ad7e24652408430a31a2714e842b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:832c39971ca0f6ca5c53721d2a6d32125c4893b34a74187f9c7dfd55ce6ee706 +size 103767494 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393430.e02c2cd1d8ab.1.0 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393430.e02c2cd1d8ab.1.0 new file mode 100644 index 0000000000000000000000000000000000000000..00af08c00b7642f6a5f2ef325976e69363213a85 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393430.e02c2cd1d8ab.1.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6afc0f43b3c1bd7933f2a0e2d56b93f74680505b96c42a15f8970c35864b322 +size 36391 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393675.e02c2cd1d8ab.1.1 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393675.e02c2cd1d8ab.1.1 new file mode 100644 index 0000000000000000000000000000000000000000..d2fd28a62e561be336dc68abaa8c7a55a90c8748 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/events.out.tfevents.1646393675.e02c2cd1d8ab.1.1 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7356d4d877d5517b1738d3fbb03153b84e6a6c09d506bbabbdffcebca279aba +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f4544e60bd2420f2eed4f783b6a78561e41daab0 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/0/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 128 + lin2_size: 32 + lin3_size: 32 + output_size: 1 + lr: 0.04099159968275264 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 8645729 +model/params/trainable: 8645729 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8b26197a938727026805bdea06de8b40f92f375b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 128 + lin3_size: 1024 + output_size: 1 + lr: 0.15818542131777294 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fdae6e8f780273770bf7ea101619ee5d8b533c5d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.15818542131777294 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=1024 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=1024,model.lr=0.15818542131777294 + id: '1' + num: 1 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bf861272fea7531129612005eec28f4d2afddeba --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.15818542131777294 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=1024 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/checkpoints/epoch_034.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/checkpoints/epoch_034.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..3ac0f8c735d24842cf00f503adf716da8c8ffac4 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/checkpoints/epoch_034.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e34abece5594593d3ffa53372127f90d33e229d5a9f219a846ff27518046079 +size 209417606 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d06df60e97c56d846cb1de2915909f2eb2ac0e0b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca746e527d3f3dcad2032b106265a9447cdfcc3d1116dfb87cdb1b9ac4f9a6e1 +size 209417606 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646393676.e02c2cd1d8ab.1.2 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646393676.e02c2cd1d8ab.1.2 new file mode 100644 index 0000000000000000000000000000000000000000..67e65e17647622cda6907dfaa2b8f907a54431cb --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646393676.e02c2cd1d8ab.1.2 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f6997fead061ad1b7ca1a54e9df1c6e59c656a41f5642597c9791fd1f5944a9 +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646394071.e02c2cd1d8ab.1.3 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646394071.e02c2cd1d8ab.1.3 new file mode 100644 index 0000000000000000000000000000000000000000..322ee69378a983dcd3d3878abd6a4098e22e55fb --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/events.out.tfevents.1646394071.e02c2cd1d8ab.1.3 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98d4647ce1487bed60e00a1b0991c7f1a1d6298d661b2af6befa36b542c3d805 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4b57efa7c8a96ea4a8bbcf9fc65804c3de295623 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/1/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 128 + lin3_size: 1024 + output_size: 1 + lr: 0.15818542131777294 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 17449089 +model/params/trainable: 17449089 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a20715827808d8435572343a8dd9857b33006af3 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 512 + lin2_size: 512 + lin3_size: 256 + output_size: 1 + lr: 0.07148210917089061 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fff3b69d4b079dd852d56bd33780e10239206ab2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.07148210917089061 + - model.lin1_size=512 + - model.lin2_size=512 + - model.lin3_size=256 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=512,model.lin2_size=512,model.lin3_size=256,model.lr=0.07148210917089061 + id: '10' + num: 10 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..541f5d678e1ea61b998a6a115ff228c2212742b8 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.07148210917089061 +- model.lin1_size=512 +- model.lin2_size=512 +- model.lin3_size=256 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/checkpoints/epoch_025.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/checkpoints/epoch_025.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..90fbf7727c5a0eb4e8cd2978d7aa7805579be9c3 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/checkpoints/epoch_025.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09d4d96a74552a96e77c6a11f07a7ea188656829292744dea5e019ddc1dfa609 +size 419515270 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..7951b969c01c0c041d5f8cb4c1ffe2fc3edac568 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fab35171fe59ee61b1ad70a681d6b47f618d582a2ef0115844af98a23c5a81e5 +size 419515334 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646399561.e02c2cd1d8ab.1.19 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646399561.e02c2cd1d8ab.1.19 new file mode 100644 index 0000000000000000000000000000000000000000..f9e7a2da2d9fda4c3c1f60e023f32f7b0cb5cf31 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646399561.e02c2cd1d8ab.1.19 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3093c010acb1da620f30c23572d0cfd10dfd6b38f5b81062c7b070360f3d326e +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646400250.e02c2cd1d8ab.1.20 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646400250.e02c2cd1d8ab.1.20 new file mode 100644 index 0000000000000000000000000000000000000000..0b85f32975581d9ac30aa379f28b7adb26176c36 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/events.out.tfevents.1646400250.e02c2cd1d8ab.1.20 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5dcb14bc03802b02af910756b125403e6a53e54513795ee022d8e68555492ce9 +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..076362de15b834c13134567b3669d6070d66be92 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/10/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 512 + lin2_size: 512 + lin3_size: 256 + output_size: 1 + lr: 0.07148210917089061 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 34957313 +model/params/trainable: 34957313 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2e0272a616e916212359d7c42e3563372de03fdf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.11779269530261958 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..63444dbaa998c43c485764ee7d04d40825096220 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.11779269530261958 + - model.lin1_size=32 + - model.lin2_size=256 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=256,model.lin3_size=128,model.lr=0.11779269530261958 + id: '11' + num: 11 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ede484811399ec5eeea7108e635e8a576c26d7bf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.11779269530261958 +- model.lin1_size=32 +- model.lin2_size=256 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/checkpoints/epoch_031.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/checkpoints/epoch_031.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..730b19dc02c847df52b054a8ae71c318600ffc23 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/checkpoints/epoch_031.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8367ecd00cc20d3b26d0f249b42f75fa463b7b3d9040cc3ccfbb454ed48d646d +size 26448582 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..2b2bb836f2d533809d11b427b71340bba4a564bf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77483463b312329b851048316f308712aeafcefa2d5f62f6f724644649c9e3c6 +size 26448582 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400251.e02c2cd1d8ab.1.21 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400251.e02c2cd1d8ab.1.21 new file mode 100644 index 0000000000000000000000000000000000000000..0178a29532b63350e949956076bd8f7b881b3257 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400251.e02c2cd1d8ab.1.21 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99dcbe2276408d1477fed434497dfedc2e1c55e37deff1fec0513c0df7336762 +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400340.e02c2cd1d8ab.1.22 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400340.e02c2cd1d8ab.1.22 new file mode 100644 index 0000000000000000000000000000000000000000..e739ea626482525ebc77d43c1d9c317928e77078 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/events.out.tfevents.1646400340.e02c2cd1d8ab.1.22 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:303733f02185c2365f205d30101dd2f463231152f5186c6366958d57f32db557 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d4eedd084f748fe296501ec33688a43e5f92be98 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/11/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.11779269530261958 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2202337 +model/params/trainable: 2202337 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d7bd1545b950232669f5b784f8de1117856cad81 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.13355237942960652 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ded6a2ed35aa4489c36ddaccf659621a707e9b8c --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.13355237942960652 + - model.lin1_size=32 + - model.lin2_size=256 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=256,model.lin3_size=128,model.lr=0.13355237942960652 + id: '12' + num: 12 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5c0afa3c3f6c2a5d1d21a6495a1b576a240dfaee --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.13355237942960652 +- model.lin1_size=32 +- model.lin2_size=256 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/checkpoints/epoch_007.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/checkpoints/epoch_007.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..55b3ca3c7c53313cfd69107ae1777c69e3f847b2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/checkpoints/epoch_007.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b82068c1b9535f88c097f28c442c2d4110e75e6f7072b1d69084110f1a7e474 +size 26448582 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..da76bd1dcf638e60c48a06546beb2adad4a59aaf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae3e95298d9f03ab9e09e5baba944412a19c8b90157954567c56adb105548b9b +size 26448582 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400341.e02c2cd1d8ab.1.23 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400341.e02c2cd1d8ab.1.23 new file mode 100644 index 0000000000000000000000000000000000000000..090403885bf1d644c2b387e2d20fb9278f43a9d2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400341.e02c2cd1d8ab.1.23 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8869683dbbc16e523ec4daec9634a5844e47c2d42cb5e48fcde03b47f057c09b +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.24 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.24 new file mode 100644 index 0000000000000000000000000000000000000000..25c396e69b6119da7c0a24a068ea1af8a35e3403 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.24 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1989b8df8bbec51baec3199e84b0bff5cb0e9f00f01c15396087aeede34c5e25 +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cac11de4b74455cc56530092fe7f92bf073d11a4 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/12/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.13355237942960652 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2202337 +model/params/trainable: 2202337 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1eec09f804a4c4eca2bb52b7358c692cda272c70 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.07220274494059103 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3fe9f5647874cad3053e5483e0ef04a21ad2ea6d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.07220274494059103 + - model.lin1_size=32 + - model.lin2_size=512 + - model.lin3_size=2048 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=512,model.lin3_size=2048,model.lr=0.07220274494059103 + id: '13' + num: 13 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3cb10a485c1bdd9ebbd6f6f1ae6aba733c3a2922 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.07220274494059103 +- model.lin1_size=32 +- model.lin2_size=512 +- model.lin3_size=2048 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/checkpoints/epoch_033.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/checkpoints/epoch_033.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..b75eaaff6545c4c0dd46a4be418eea92a9bb19d2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/checkpoints/epoch_033.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8499970279989031c9b81c0e4dd15d0ec69e022381fa5cf103e88e27457a57ab +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..79a804d90fb247cf37b0870a5acc9e439ee2b024 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed0715b6a5eaafc0dbcb06bfabc11417d06d14605ecbcb01fe27bc46f344b8e5 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.25 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.25 new file mode 100644 index 0000000000000000000000000000000000000000..1c24746ed44666bf591197593b10dab0d7af8a5b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400436.e02c2cd1d8ab.1.25 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d966126e50625b723d1aa3ea74101b1e0012a80c8ce55c9b9a7b9b0fc7e838c4 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400569.e02c2cd1d8ab.1.26 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400569.e02c2cd1d8ab.1.26 new file mode 100644 index 0000000000000000000000000000000000000000..8d08099fa878cf9156422f6318f3b69900890d44 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/events.out.tfevents.1646400569.e02c2cd1d8ab.1.26 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d80b675e81b6ba40bda511fec56d0a2eb745dccb51a04e4d428b5d9216b44714 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..78c58b5b7948825fa801029fb5a8bcbad7ae7ee8 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/13/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.07220274494059103 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 3234785 +model/params/trainable: 3234785 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b7ec538d75059274645444dc280a7b6accd56ef4 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 2048 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.08335231197967785 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f5b4097bccf0bea79e78ee8fbb0070fa085be3be --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.08335231197967785 + - model.lin1_size=2048 + - model.lin2_size=512 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=2048,model.lin2_size=512,model.lin3_size=128,model.lr=0.08335231197967785 + id: '14' + num: 14 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2294f80bd01d35b645bfff6940062f98da89e0a6 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.08335231197967785 +- model.lin1_size=2048 +- model.lin2_size=512 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/checkpoints/epoch_038.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/checkpoints/epoch_038.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..fadb2136cee91490389eaa9dbe78184bb9deb8fe --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/checkpoints/epoch_038.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec87ae32d869dc5eaabfa04750de022bbc646b724cc444552a1cfb36237d51c5 +size 1672386438 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..ef7b3c5ee39d509da16954fdc38b1351f787f771 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b01626d458b96bce2f165a7f7829dbc174304385c35bab41cf4df88fd47d6192 +size 1672386438 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646400570.e02c2cd1d8ab.1.27 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646400570.e02c2cd1d8ab.1.27 new file mode 100644 index 0000000000000000000000000000000000000000..2bf81f667ff8b95527ebafbe28224d7492f684e5 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646400570.e02c2cd1d8ab.1.27 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d05f9112c18daa456c3ae5551f39f14bd694958d48a3292ebee05fefe8a256a +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646403523.e02c2cd1d8ab.1.28 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646403523.e02c2cd1d8ab.1.28 new file mode 100644 index 0000000000000000000000000000000000000000..049d2800e6e413bdcdf1c3835c8209bbd9b8d47a --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/events.out.tfevents.1646403523.e02c2cd1d8ab.1.28 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70fb6b94c38eb4ba4538834d1114cf2f128d7d9f729d09e10798d95e0a6ae839 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b7fb25af231f8551decf950cd81c1f8df41ee74e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/14/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 2048 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.08335231197967785 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 139362305 +model/params/trainable: 139362305 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ef03b92bbfb81828690b4006989477fa80135f81 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 64 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.04131276528182434 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9cc8da3e8b17b91c2920edaa62ff017580db612d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.04131276528182434 + - model.lin1_size=64 + - model.lin2_size=256 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=64,model.lin2_size=256,model.lin3_size=128,model.lr=0.04131276528182434 + id: '15' + num: 15 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f163c74709f5c16afe75ceeb6fa36cb1f610f468 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.04131276528182434 +- model.lin1_size=64 +- model.lin2_size=256 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/checkpoints/epoch_029.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/checkpoints/epoch_029.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..8dcdf9311a3bbda598f8c7e39d8f99a3151cdc14 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/checkpoints/epoch_029.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18868f2af702ad7fc65db66eb08b43b021832ccc80506b91639df2254fab66eb +size 52468294 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..67425cc026e09c8900475cf9bfd93d829e225994 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:027eaf4a1ccd9f4c7cecfa233dc4749964fbb8d2d42e39c8fd927b3f53866682 +size 52468294 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403524.e02c2cd1d8ab.1.29 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403524.e02c2cd1d8ab.1.29 new file mode 100644 index 0000000000000000000000000000000000000000..0902cb3c23f65e50b5fe29de62c71fe29d086c72 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403524.e02c2cd1d8ab.1.29 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f5f63724068b3fe2d9530ced3aa79622ad58b3a5c1e48f29b09a3c23fd909d6 +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403650.e02c2cd1d8ab.1.30 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403650.e02c2cd1d8ab.1.30 new file mode 100644 index 0000000000000000000000000000000000000000..7908c3b371056edf345d404a002249c2c3c23ad7 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/events.out.tfevents.1646403650.e02c2cd1d8ab.1.30 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c94ff7b41ccf935cdb8bcd338c81f331607912868cfeecc00906530eb326bae2 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c02697d10a25311be46c1e1051f830e616bc6ff8 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/15/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 64 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.04131276528182434 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4370625 +model/params/trainable: 4370625 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..26726a5cace2ab3133e10b77a6fc87781c40d6d8 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 512 + lin2_size: 32 + lin3_size: 2048 + output_size: 1 + lr: 0.12875356688119352 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d91133bb377250afb7972d13707053835140f55e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.12875356688119352 + - model.lin1_size=512 + - model.lin2_size=32 + - model.lin3_size=2048 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=512,model.lin2_size=32,model.lin3_size=2048,model.lr=0.12875356688119352 + id: '16' + num: 16 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a56c047670dab179f00866a6a2842bb87bb494e5 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.12875356688119352 +- model.lin1_size=512 +- model.lin2_size=32 +- model.lin3_size=2048 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/checkpoints/epoch_056.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/checkpoints/epoch_056.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..c2855c5f3c5cfb74be9d99e27fb0ee2deca295d2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/checkpoints/epoch_056.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dcbaf143337d0d4a3539740a34270d377e6f2fcf514897a5f61d0554f114161 +size 415858950 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..17d3a52c685f33ae14859b0316a6c139ac8d329e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d1fe68fb29f552fb26f5a2180d53deae1c0724bd743ec98145c95e568879c47 +size 415858950 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/events.out.tfevents.1646403651.e02c2cd1d8ab.1.31 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/events.out.tfevents.1646403651.e02c2cd1d8ab.1.31 new file mode 100644 index 0000000000000000000000000000000000000000..b881d58b9fa569586bf0ee67574c81de0d80ae59 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/events.out.tfevents.1646403651.e02c2cd1d8ab.1.31 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e3fb02310887740ec064030d6b249654ad975400ab5b5ceba4d38b0ee1b6ea3 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/events.out.tfevents.1646404342.e02c2cd1d8ab.1.32 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/events.out.tfevents.1646404342.e02c2cd1d8ab.1.32 new file mode 100644 index 0000000000000000000000000000000000000000..bf58a84412bcc867f04d484172127122917203e4 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/events.out.tfevents.1646404342.e02c2cd1d8ab.1.32 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8b4de9909627583877d45362c233458c8f4b80e76b7e77ef396370e50b91534 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cfdab4f0eb1d26e6444ecbfea7db5878702fd047 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/16/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 512 + lin2_size: 32 + lin3_size: 2048 + output_size: 1 + lr: 0.12875356688119352 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 34651745 +model/params/trainable: 34651745 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9e4ca456b451eddc802568de8aad1cc7bcacfe74 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 128 + lin2_size: 128 + lin3_size: 256 + output_size: 1 + lr: 0.14899729967070047 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c151d7cda8f6eabdaafe4cfac7326b0b569ad668 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.14899729967070047 + - model.lin1_size=128 + - model.lin2_size=128 + - model.lin3_size=256 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=128,model.lin2_size=128,model.lin3_size=256,model.lr=0.14899729967070047 + id: '17' + num: 17 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b2e611127ad5ea1a45c6ac58a252e8eb2c917f53 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.14899729967070047 +- model.lin1_size=128 +- model.lin2_size=128 +- model.lin3_size=256 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/checkpoints/epoch_019.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/checkpoints/epoch_019.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..eb4d6917bdade88c3711b95ed28aff2bfcedd73f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/checkpoints/epoch_019.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4473b3dc9b0c7bf71bb1bb069213afd7dac803045e6fc2674f63098cf0d38b10 +size 104312646 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..dad4af1e8c928e47feea2e14e00eaa28e73e28c0 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b9e86ddaff1729904d0889cd5ce47834b42c62a67e6e3e82e9cef1d35f559cc +size 104312646 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/events.out.tfevents.1646404343.e02c2cd1d8ab.1.33 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/events.out.tfevents.1646404343.e02c2cd1d8ab.1.33 new file mode 100644 index 0000000000000000000000000000000000000000..656f1af3e640fb66a52aed90b0739eb0b810db12 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/events.out.tfevents.1646404343.e02c2cd1d8ab.1.33 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97506cfb4e0ae653dbabeb25cea691b2f0655ebe32d218adccf52b5197fab773 +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/events.out.tfevents.1646404545.e02c2cd1d8ab.1.34 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/events.out.tfevents.1646404545.e02c2cd1d8ab.1.34 new file mode 100644 index 0000000000000000000000000000000000000000..b481f284490e889101f9ce7f8a7bb1b63d879671 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/events.out.tfevents.1646404545.e02c2cd1d8ab.1.34 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afced4017f1373a3653b1761c052269f3f1d03c097337d2da52b1e4d4a6c2966 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..165f74b9e8f8f61ed7485a71f4cd28241e1326b9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/17/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 128 + lin2_size: 128 + lin3_size: 256 + output_size: 1 + lr: 0.14899729967070047 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 8690945 +model/params/trainable: 8690945 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6d011a6604810d54b514f1074237a8a5bbfc64c5 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 64 + lin3_size: 512 + output_size: 1 + lr: 0.043609714610201814 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..563160bbafd521df519360dae0080c5840a2da47 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.043609714610201814 + - model.lin1_size=32 + - model.lin2_size=64 + - model.lin3_size=512 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=64,model.lin3_size=512,model.lr=0.043609714610201814 + id: '18' + num: 18 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..59a439321a0352c6cd37e679955a66e772c14a5e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.043609714610201814 +- model.lin1_size=32 +- model.lin2_size=64 +- model.lin3_size=512 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/checkpoints/epoch_073.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/checkpoints/epoch_073.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..daab38a9b7e079e74796b085ece4deacff2d8f49 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/checkpoints/epoch_073.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c03e422c2828d9612444fcabb98171d81ba1da970f43d85d5ff7169334986806 +size 26387910 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..a227f1c0785ffb4102721b3eb23c5ef34b630ca4 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d62300c7e90eab32d76efe75c2214ecc88f19f75134732f295239910e0a1ab2b +size 26387910 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/events.out.tfevents.1646404545.e02c2cd1d8ab.1.35 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/events.out.tfevents.1646404545.e02c2cd1d8ab.1.35 new file mode 100644 index 0000000000000000000000000000000000000000..5c5906e3daeb050166be67e7024109a46e4362c2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/events.out.tfevents.1646404545.e02c2cd1d8ab.1.35 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdf99a4fb785f3090d8c8179b170f3427025514d68a12c8f15067e0737b1f6be +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/events.out.tfevents.1646404634.e02c2cd1d8ab.1.36 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/events.out.tfevents.1646404634.e02c2cd1d8ab.1.36 new file mode 100644 index 0000000000000000000000000000000000000000..cb74994a5653a69764eba6ff8dfd0f5360069ba5 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/events.out.tfevents.1646404634.e02c2cd1d8ab.1.36 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb29ec2c809dae3307355c80996de3564685ebac48c7cf6461a5fb222e24f6bb +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..804e6c95dce57c8f23a0e83d95e79b8b8a422672 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/18/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 64 + lin3_size: 512 + output_size: 1 + lr: 0.043609714610201814 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2197153 +model/params/trainable: 2197153 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bd57d2cd0e79545012af134335ce16164acabe98 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 256 + lin3_size: 64 + output_size: 1 + lr: 0.001567877853839779 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..888daf694c6e629b322129a1280da6e965ab7779 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.001567877853839779 + - model.lin1_size=256 + - model.lin2_size=256 + - model.lin3_size=64 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=256,model.lin3_size=64,model.lr=0.001567877853839779 + id: '19' + num: 19 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a9cd0c0e6b6bf86e5d79748b29e25697b0491954 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.001567877853839779 +- model.lin1_size=256 +- model.lin2_size=256 +- model.lin3_size=64 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/checkpoints/epoch_072.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/checkpoints/epoch_072.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..226cd6e5ca9b937d5e9f65cb3c4d5bb39b995377 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/checkpoints/epoch_072.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:671b07040e1782dd1201e05f0275b412eabb6ccd6941f23a390ab6c09e76def4 +size 208386438 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..1bae39d786dba02fc143162dd49a8c944762210e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29fbed0a8e6963ebf3dbf1a8c52119ffc5c8c15f45231fa5144291143d04c116 +size 208386438 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/events.out.tfevents.1646404634.e02c2cd1d8ab.1.37 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/events.out.tfevents.1646404634.e02c2cd1d8ab.1.37 new file mode 100644 index 0000000000000000000000000000000000000000..da1037a009c64f712f3098c80d0cb140bf4c8aaa --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/events.out.tfevents.1646404634.e02c2cd1d8ab.1.37 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5958ec8e708e2fdfef161a1f2273d10d87ca900d3520b33d5ea09c43e7143d4f +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/events.out.tfevents.1646405000.e02c2cd1d8ab.1.38 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/events.out.tfevents.1646405000.e02c2cd1d8ab.1.38 new file mode 100644 index 0000000000000000000000000000000000000000..1f6846f8796eebd7a5ee36fd95d14a50d33e8d81 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/events.out.tfevents.1646405000.e02c2cd1d8ab.1.38 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfc2baf70998d8d1e06c39fdab8c99bf2e12dd8b2534d89ba3a2ad1041c6ec63 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..361e35a16ab44f5015a4937b6c4ba53f4488c64a --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/19/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 256 + lin3_size: 64 + output_size: 1 + lr: 0.001567877853839779 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 17363713 +model/params/trainable: 17363713 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b86d9956384c074ad1339dde78791281eda7bf52 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 1024 + lin2_size: 1024 + lin3_size: 32 + output_size: 1 + lr: 0.18564129276815303 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ff4f26ed677689f24ed5f5f08508f672907344e1 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.18564129276815303 + - model.lin1_size=1024 + - model.lin2_size=1024 + - model.lin3_size=32 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=1024,model.lin2_size=1024,model.lin3_size=32,model.lr=0.18564129276815303 + id: '2' + num: 2 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2cd996711399ee9365e63d051ef83f2f472a81a9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.18564129276815303 +- model.lin1_size=1024 +- model.lin2_size=1024 +- model.lin3_size=32 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/checkpoints/epoch_031.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/checkpoints/epoch_031.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..924972c542b5baa15972c32b3f4cee49d665e2cc --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/checkpoints/epoch_031.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8795b7fc7181caca1fcf1d52796f26de17ea91f90b6a4fcd694ee86134212b47 +size 842525318 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..fcdc78e86fd3a1272aba944ad0b85cfbbe618bfe --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:587ca293f7910be0515e8a2cf1ccf8de7293fd61d19e267235904d9180e30ac6 +size 842525318 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/events.out.tfevents.1646394072.e02c2cd1d8ab.1.4 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/events.out.tfevents.1646394072.e02c2cd1d8ab.1.4 new file mode 100644 index 0000000000000000000000000000000000000000..75e8f3e905dd9ca7bdf6c94965e4fbf016be7c7f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/events.out.tfevents.1646394072.e02c2cd1d8ab.1.4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c34e0c728abd186eb1aeb89fe5f2aa8a7cd7a254872619697e68b1fe7a6a8366 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/events.out.tfevents.1646395466.e02c2cd1d8ab.1.5 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/events.out.tfevents.1646395466.e02c2cd1d8ab.1.5 new file mode 100644 index 0000000000000000000000000000000000000000..9b47f699620a8cdb82849464488ec01c78f1053f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/events.out.tfevents.1646395466.e02c2cd1d8ab.1.5 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bee9769a533de761bebaaadd5f504bf0495ffb62d6e9fdfb3f529adbef197a5 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1a5029aa98f9ed2c4c98169f7e5215208cdc8bd7 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/2/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 1024 + lin2_size: 1024 + lin3_size: 32 + output_size: 1 + lr: 0.18564129276815303 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 70207617 +model/params/trainable: 70207617 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0a3572cf12533afa6f43fe211c59d7cf8497dab1 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.10487184499754953 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..71d6cc2f4dbd9dd1e8a33ee6e6239d0165b90510 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.10487184499754953 + - model.lin1_size=256 + - model.lin2_size=512 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=512,model.lin3_size=128,model.lr=0.10487184499754953 + id: '20' + num: 20 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3b21b02da9a7e8eedb93e2ad888e2d2714093653 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.10487184499754953 +- model.lin1_size=256 +- model.lin2_size=512 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/checkpoints/epoch_029.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/checkpoints/epoch_029.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d2710cce7e4d8de7d1448c75cc4ddc726b6242ea --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/checkpoints/epoch_029.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26436b1871e1df08139a09a0fa13f5ca29f8b1f42bb7004d0250ea721ceb25c3 +size 209777542 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..fec766ea41d7a832456fe060a3206d8636ed6753 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2b3fd309cc53ee207fcdfcdb53783557db8cadae0f3bbc45cd6d9625b1c11b3 +size 209777542 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/events.out.tfevents.1646405000.e02c2cd1d8ab.1.39 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/events.out.tfevents.1646405000.e02c2cd1d8ab.1.39 new file mode 100644 index 0000000000000000000000000000000000000000..f39acb2e4de27f0d0f5e035dad3b32fe14214b5e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/events.out.tfevents.1646405000.e02c2cd1d8ab.1.39 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2d5337aa51ed1ede8c56b34128c7d230e53588b9224435f813296090386ea65 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/events.out.tfevents.1646405357.e02c2cd1d8ab.1.40 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/events.out.tfevents.1646405357.e02c2cd1d8ab.1.40 new file mode 100644 index 0000000000000000000000000000000000000000..dd1d975c78ac714edf5049caf31852f712388b0c --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/events.out.tfevents.1646405357.e02c2cd1d8ab.1.40 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01611ac9a425f8161978c24080350542409046ffc631e757eaa3e470414b0c5f +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..38fb78fd84c321cab16899b95ab3e9f572cfbe0b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/20/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.10487184499754953 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 17479425 +model/params/trainable: 17479425 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d97168e5f62d1c647f724e82e3d62b8bc72f22c2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.06607730400973591 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca6ee4a049a586572e201d95b215db1ab4d02297 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.06607730400973591 + - model.lin1_size=32 + - model.lin2_size=512 + - model.lin3_size=2048 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=512,model.lin3_size=2048,model.lr=0.06607730400973591 + id: '21' + num: 21 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..86fd533d2e9e0e55e591cc60e4fda1ad66fa9445 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.06607730400973591 +- model.lin1_size=32 +- model.lin2_size=512 +- model.lin3_size=2048 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/checkpoints/epoch_013.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/checkpoints/epoch_013.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..edf0558b29a15193d6094999392065b8a4e36332 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/checkpoints/epoch_013.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fd37a1f421fedbfd842f579bd588528fabd7ec5730c16578b1dbfb1d0834935 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..85ce8495587a18ecef2af1b605835a38fcd5f753 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2545cf3108123f77df8939ada04fc138c3492ce37086cfac35e95d63ac73c674 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/events.out.tfevents.1646405358.e02c2cd1d8ab.1.41 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/events.out.tfevents.1646405358.e02c2cd1d8ab.1.41 new file mode 100644 index 0000000000000000000000000000000000000000..d65346de926ee47491748fdf9b095decb9340fbb --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/events.out.tfevents.1646405358.e02c2cd1d8ab.1.41 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44e2ccd76c98d8edd177ade6cf95074f7ddf4da8659bea0f1a22820394886396 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/events.out.tfevents.1646405459.e02c2cd1d8ab.1.42 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/events.out.tfevents.1646405459.e02c2cd1d8ab.1.42 new file mode 100644 index 0000000000000000000000000000000000000000..b46c1184c43df5c5e6d590a7e98f305bc9178fb9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/events.out.tfevents.1646405459.e02c2cd1d8ab.1.42 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd7fdf3b887575718dffd63122fb819f8844216bedac480f643b3a0e123a48bd +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5796cefb294aeb9a8294e3f335d8505deccd9d7c --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/21/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.06607730400973591 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 3234785 +model/params/trainable: 3234785 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5027e1defac5ff1545860d665a9e65b322582b79 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.056028463603035976 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..99100751e98d2bf279c8e077a655f57ccb616af9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.056028463603035976 + - model.lin1_size=32 + - model.lin2_size=512 + - model.lin3_size=2048 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=512,model.lin3_size=2048,model.lr=0.056028463603035976 + id: '22' + num: 22 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..21081b524a159307b1144d3cff470a0f26957fd3 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.056028463603035976 +- model.lin1_size=32 +- model.lin2_size=512 +- model.lin3_size=2048 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/checkpoints/epoch_035.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/checkpoints/epoch_035.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..14cba2babd3cec5973f3b0e88ca0e9dacc2c05b2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/checkpoints/epoch_035.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f75b6ba795600cda33e3cb29af3875dee353a33c20627fb7088f98dcb665194 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..7ed214b5bb1f43bdbbb9dbecfbb073d2cc9c7eb1 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:164a3ecfcfd35efef7c2903aa5f0569c4c419d3721fa5b81a244ee70b61dbcf2 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/events.out.tfevents.1646405459.e02c2cd1d8ab.1.43 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/events.out.tfevents.1646405459.e02c2cd1d8ab.1.43 new file mode 100644 index 0000000000000000000000000000000000000000..a7c8bb2036b8e9185282c67c22ed887384fa7783 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/events.out.tfevents.1646405459.e02c2cd1d8ab.1.43 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d2715198412ce895618c69707fe67c189cbbcf776fc0920af7b44935e9b7787 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/events.out.tfevents.1646405569.e02c2cd1d8ab.1.44 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/events.out.tfevents.1646405569.e02c2cd1d8ab.1.44 new file mode 100644 index 0000000000000000000000000000000000000000..7e03fedad6d4056cc5d65fa782ba23cf1b4dac3c --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/events.out.tfevents.1646405569.e02c2cd1d8ab.1.44 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:600bb6948ae0d30dbf47f46abea46d7ab97494e226026e34e47ecb2c1b2eaaf3 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..35cc8e2f235d4c8432e9895e83a8f02f6321a6e8 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/22/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.056028463603035976 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 3234785 +model/params/trainable: 3234785 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6b31409f7d38946dd5130397d8006730f99e0a8f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.08775537240022759 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..12d9d9c137a37bbf2e37f8045c017d108450044b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.08775537240022759 + - model.lin1_size=32 + - model.lin2_size=512 + - model.lin3_size=2048 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=512,model.lin3_size=2048,model.lr=0.08775537240022759 + id: '23' + num: 23 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7a5078338968a0d7323d532bd9b681d8d89b164e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.08775537240022759 +- model.lin1_size=32 +- model.lin2_size=512 +- model.lin3_size=2048 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/checkpoints/epoch_034.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/checkpoints/epoch_034.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..395c0f72c9b298dcff0f30ece3433b3d5862e568 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/checkpoints/epoch_034.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2f7b38d6bc7ba769730ee00c379807905457c5e3f876904a8d9a30f17f5c4f6 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..47d85ab4a43bec090ff7d9332f1005f73e3f0f33 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2085d579d29160e94d069b6a741f683bfafda1044223f6671fbe08f855d78a1 +size 38855430 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/events.out.tfevents.1646405569.e02c2cd1d8ab.1.45 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/events.out.tfevents.1646405569.e02c2cd1d8ab.1.45 new file mode 100644 index 0000000000000000000000000000000000000000..70b42cd2fbe3bdf552ad89b6ef71a4bb12f85038 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/events.out.tfevents.1646405569.e02c2cd1d8ab.1.45 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc3e059585bac6abacd063c05eaa2784fdae37e3a20219824ebcc723e6e72b28 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/events.out.tfevents.1646405675.e02c2cd1d8ab.1.46 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/events.out.tfevents.1646405675.e02c2cd1d8ab.1.46 new file mode 100644 index 0000000000000000000000000000000000000000..cd5021fa04c88437adf8f2b099171f45be94b372 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/events.out.tfevents.1646405675.e02c2cd1d8ab.1.46 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a79a40037b60f1e09a3415b4751de20764ecf3f02feb027e6475a183eb894bc4 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..46d2017243fd68ee851cbd077c51e46b7c64bff6 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/23/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 2048 + output_size: 1 + lr: 0.08775537240022759 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 3234785 +model/params/trainable: 3234785 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c3a59174d2ee045a12294ae7c0c7d2d9da22a28b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.02016131295868198 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..65353df476d8aca58f00335b30e4fa3656e809ab --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.02016131295868198 + - model.lin1_size=32 + - model.lin2_size=512 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=512,model.lin3_size=128,model.lr=0.02016131295868198 + id: '24' + num: 24 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a5dbb03d4d8499f7701e14e1cc96559d5ed1234 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.02016131295868198 +- model.lin1_size=32 +- model.lin2_size=512 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/checkpoints/epoch_037.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/checkpoints/epoch_037.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..ee67e745a4041911c7188f746f384393c0fcb0e7 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/checkpoints/epoch_037.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ec3fa8bbd841e354294dba8eb58d637a12acbe2ba03f74ea099e9ae39387c12 +size 26951366 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..f65df32fe19a793458215a873cb344abf06e4712 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00fd40355c82e80fe5a23e7834789fb8036f55408d405cc86bd984eea816dcbe +size 26951366 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/events.out.tfevents.1646405675.e02c2cd1d8ab.1.47 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/events.out.tfevents.1646405675.e02c2cd1d8ab.1.47 new file mode 100644 index 0000000000000000000000000000000000000000..8e02652e8dcadf70b625b01dec52da2f4595b97f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/events.out.tfevents.1646405675.e02c2cd1d8ab.1.47 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bec5c438e457e168ad3f7823da89c93d604482dcf32d6096ad27af4720c091b3 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/events.out.tfevents.1646405764.e02c2cd1d8ab.1.48 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/events.out.tfevents.1646405764.e02c2cd1d8ab.1.48 new file mode 100644 index 0000000000000000000000000000000000000000..d86e3e3c745c0c69e65b0b02528cee16497c19be --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/events.out.tfevents.1646405764.e02c2cd1d8ab.1.48 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf4462cb6112df5cd3c1e882b38f2ffa5fcd4413f4fc9c0b0430243c997e1af4 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..227c8f13081aa904ea6e9467fef6d76383ab4c0a --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/24/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.02016131295868198 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2244065 +model/params/trainable: 2244065 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8e005e16f22513a8f49f4b775704a7a208023536 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 64 + lin2_size: 2048 + lin3_size: 32 + output_size: 1 + lr: 0.09908496566082269 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a863176a56b4c854828af27def815fdab79fa0fc --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.09908496566082269 + - model.lin1_size=64 + - model.lin2_size=2048 + - model.lin3_size=32 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=64,model.lin2_size=2048,model.lin3_size=32,model.lr=0.09908496566082269 + id: '3' + num: 3 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9152e8dafce336f286de49a9bccfd2e968842d62 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.09908496566082269 +- model.lin1_size=64 +- model.lin2_size=2048 +- model.lin3_size=32 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/checkpoints/epoch_005.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/checkpoints/epoch_005.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..445823b5cf44b5fecdf5541151724c086ba2160f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/checkpoints/epoch_005.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2d6142028d686581f04940663d6f65dc92f16d1074c8110d9f8884e38ef4218 +size 54311238 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d24b01ddedd0214b541e6458e54127f539f5eb39 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da819a7baa00fbdc7df0c2dbe31d0f28d0c4ad3771e72149fb4088a8dc7f093f +size 54311302 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/events.out.tfevents.1646395467.e02c2cd1d8ab.1.6 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/events.out.tfevents.1646395467.e02c2cd1d8ab.1.6 new file mode 100644 index 0000000000000000000000000000000000000000..8449e411f7ea3fadd37eecf0a00449edebfa5aec --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/events.out.tfevents.1646395467.e02c2cd1d8ab.1.6 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55c7397c7291ce0088bcbf730d2aa5a106ad926b48004259de701431a5045dda +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/events.out.tfevents.1646395590.e02c2cd1d8ab.1.7 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/events.out.tfevents.1646395590.e02c2cd1d8ab.1.7 new file mode 100644 index 0000000000000000000000000000000000000000..9b3f1cc0e49e0f7c0df2abc9352e3d5adee92eef --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/events.out.tfevents.1646395590.e02c2cd1d8ab.1.7 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:442436893b711d7e9a5fc87ef734e4d44b9b13981a5e761a3f9a38baf7c8f0c7 +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c79a2902967f2db2f6fb4463867ccfee8428df58 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/3/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 64 + lin2_size: 2048 + lin3_size: 32 + output_size: 1 + lr: 0.09908496566082269 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4523073 +model/params/trainable: 4523073 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bdc6203b66976d810491f2ee960b3fa8de9a8518 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.017335998714038146 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..42eed7bb07090b7c2cd73f3bedda248fd013a7d9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.017335998714038146 + - model.lin1_size=256 + - model.lin2_size=512 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=512,model.lin3_size=128,model.lr=0.017335998714038146 + id: '4' + num: 4 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8f33fc86c58ce62a55248a48b18f5caaecff70c2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.017335998714038146 +- model.lin1_size=256 +- model.lin2_size=512 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/checkpoints/epoch_049.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/checkpoints/epoch_049.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..a41486557736a7bbd590788ef533c8473353bf9a --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/checkpoints/epoch_049.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cd490a5afbacda40c5653588bb4c28fa165cc5021b7fd7aad740943ebe9d091 +size 209777542 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..fb43bddf073c58a38fba4818092b234dffb550aa --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d74eb6848221433c5cdc0a8c7e0df414292f7c8967bc45081ce193ef2b08d3e +size 209777542 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/events.out.tfevents.1646395591.e02c2cd1d8ab.1.8 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/events.out.tfevents.1646395591.e02c2cd1d8ab.1.8 new file mode 100644 index 0000000000000000000000000000000000000000..e796c14ffd14dd0ef529e7c20ec325f5cfaaa6a3 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/events.out.tfevents.1646395591.e02c2cd1d8ab.1.8 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be6cc73ff538dd62f651a7e41561ee28e33ef53dcb65700a18cfd5b95e6f9883 +size 36230 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/events.out.tfevents.1646395963.e02c2cd1d8ab.1.9 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/events.out.tfevents.1646395963.e02c2cd1d8ab.1.9 new file mode 100644 index 0000000000000000000000000000000000000000..c8792bce3b2c43fbee70673be97077dba06f1b19 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/events.out.tfevents.1646395963.e02c2cd1d8ab.1.9 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55a8d3a8912a6733a5fc596237fb86eee66f490377a49c73d8055fa62b61c07d +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..71f80903be9e03934b00a9d8f5b4321d26e5294d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/4/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.017335998714038146 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 17479425 +model/params/trainable: 17479425 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..717955cbbb5ef7fedba41403543d03f0dfacab19 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 1024 + lin2_size: 64 + lin3_size: 128 + output_size: 1 + lr: 0.01956889375661914 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..21e8eadce94dd2e9a7026cffe29a88c24baa793d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.01956889375661914 + - model.lin1_size=1024 + - model.lin2_size=64 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus150_optuna,model.lin1_size=1024,model.lin2_size=64,model.lin3_size=128,model.lr=0.01956889375661914 + id: '5' + num: 5 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1820cc83c3c8dafdbadfc7e6db54865eed9498ed --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.01956889375661914 +- model.lin1_size=1024 +- model.lin2_size=64 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/checkpoints/epoch_031.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/checkpoints/epoch_031.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..8782aafd105c80a32763937bcc55d38e89213f72 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/checkpoints/epoch_031.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75497f34fa2635317bad2eec59123833278ba574575cefb6e3013816a431db02 +size 830396998 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..9d00b37bfdee76647a774c99bf58668eef322c2a --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:813ba3e8cbd5b4e4e99d77aae5919dd0735f2161a4823834470b33dadb4cbeaa +size 830396998 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/events.out.tfevents.1646395964.e02c2cd1d8ab.1.10 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/events.out.tfevents.1646395964.e02c2cd1d8ab.1.10 new file mode 100644 index 0000000000000000000000000000000000000000..c03e4c8461d1a026e5e4ff354f6e2dbdc9389ddc --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/events.out.tfevents.1646395964.e02c2cd1d8ab.1.10 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1911f58aa0db426008b232a79bd09392121e82cd0d5c9d444b8030312735b4d7 +size 36391 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/events.out.tfevents.1646397283.e02c2cd1d8ab.1.11 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/events.out.tfevents.1646397283.e02c2cd1d8ab.1.11 new file mode 100644 index 0000000000000000000000000000000000000000..56374ab3ebff48da9da87a4faf77e5de176a12c2 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/events.out.tfevents.1646397283.e02c2cd1d8ab.1.11 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24f326a62f538804abac4b6bba124b8bd8989c9bbefc389cfa0d9857efe1f66b +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..302f39e5a5d80dd5b59b6029acb22ed3bb80c67e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/5/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 1024 + lin2_size: 64 + lin3_size: 128 + output_size: 1 + lr: 0.01956889375661914 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 69197505 +model/params/trainable: 69197505 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..72f4c880fd3e22250b10b14904741beb00e5cfd6 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 512 + output_size: 1 + lr: 0.012981602213938178 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..86eb2e44ae8ddb0809297f54b2894e0f9b800cb9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.012981602213938178 + - model.lin1_size=256 + - model.lin2_size=1024 + - model.lin3_size=512 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=1024,model.lin3_size=512,model.lr=0.012981602213938178 + id: '6' + num: 6 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b2842d12e03335fa4977f51a61ffe796b5025981 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.012981602213938178 +- model.lin1_size=256 +- model.lin2_size=1024 +- model.lin3_size=512 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/checkpoints/epoch_031.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/checkpoints/epoch_031.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..60168b5a30307d0e4fa1868cb7d5402db8a352d1 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/checkpoints/epoch_031.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b012ac07d921941e5d47d29628a5aaa6507707f41fbcc4d8613aaaf26627aa5f +size 216899526 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d7cdf3c173029e1ac05203ebe6c1ad1e8772f274 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74b16e5e6ad1ed60ed11a6d0ed283d6b2957d4f488932bc36af6398d82b01347 +size 216899526 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/events.out.tfevents.1646397284.e02c2cd1d8ab.1.12 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/events.out.tfevents.1646397284.e02c2cd1d8ab.1.12 new file mode 100644 index 0000000000000000000000000000000000000000..c6aaebbde8a7567a04f944c9af83708b325d8478 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/events.out.tfevents.1646397284.e02c2cd1d8ab.1.12 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb42be304fc7d0e8b29e863c71d1d034dcad5abee842ca10ddeda2796aacbbd6 +size 36391 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/events.out.tfevents.1646397655.e02c2cd1d8ab.1.13 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/events.out.tfevents.1646397655.e02c2cd1d8ab.1.13 new file mode 100644 index 0000000000000000000000000000000000000000..2e3da30bbbbd34ad7dd27974d5c268782de2defb --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/events.out.tfevents.1646397655.e02c2cd1d8ab.1.13 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7db2525437ef7e1b2e85ed68ad7b62972be94a6274aff849058f5dede2ff87c4 +size 179 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3bad09a1a6a01b85e16af0d032fbd6fc039a9766 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/6/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 512 + output_size: 1 + lr: 0.012981602213938178 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 18072321 +model/params/trainable: 18072321 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe13be443deaee7ce21645dd7008d11ddca43f9e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 64 + output_size: 1 + lr: 0.19115771252214422 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a4212410cd1d53f79e8056d48c6cb36d1c28d61c --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.19115771252214422 + - model.lin1_size=256 + - model.lin2_size=1024 + - model.lin3_size=64 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=256,model.lin2_size=1024,model.lin3_size=64,model.lr=0.19115771252214422 + id: '7' + num: 7 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2ab5874034b7f4b9939c4cea0e1acfaf2e0f9b0f --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.19115771252214422 +- model.lin1_size=256 +- model.lin2_size=1024 +- model.lin3_size=64 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/checkpoints/epoch_003.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/checkpoints/epoch_003.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..72bfb4f8f53f0a23613d44380ac6cd54bf341e45 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/checkpoints/epoch_003.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:423867df2c74fde2e800c29d9cf4b872057410bcb16c29898548b126edf35f0b +size 211369350 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..8d2526ac8d789102a0c9e83127f6228952321b2d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a31adbbb479ca2f878955ae678393fb8dd059eac3d7233f50ee41b37f49ce52f +size 211369350 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/events.out.tfevents.1646397655.e02c2cd1d8ab.1.14 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/events.out.tfevents.1646397655.e02c2cd1d8ab.1.14 new file mode 100644 index 0000000000000000000000000000000000000000..d164354f95103d82947f757b562753cc2e966186 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/events.out.tfevents.1646397655.e02c2cd1d8ab.1.14 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:601ac743b860712bb8c0ef3a7beec02378a70ac644cfb14e23c4acbc2023b740 +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/events.out.tfevents.1646398015.e02c2cd1d8ab.1.15 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/events.out.tfevents.1646398015.e02c2cd1d8ab.1.15 new file mode 100644 index 0000000000000000000000000000000000000000..3779be976236280f2229ac22368b3175fb2e42bf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/events.out.tfevents.1646398015.e02c2cd1d8ab.1.15 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69c8fa5843aa51e544e34d0f60bc7c83101eb0bfe6ef14bbbbf834541856c215 +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6866b52ce97e2e40da4c4845f2874e29b9628a03 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/7/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 64 + output_size: 1 + lr: 0.19115771252214422 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 17611777 +model/params/trainable: 17611777 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..42b7e91c9183b71fb317bc730c493bad1edda75e --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 2048 + lin2_size: 2048 + lin3_size: 1024 + output_size: 1 + lr: 0.16654057333719488 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0eccb8c3882edc4829d31807e2af264b3a0e656a --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.16654057333719488 + - model.lin1_size=2048 + - model.lin2_size=2048 + - model.lin3_size=1024 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus150_optuna,model.lin1_size=2048,model.lin2_size=2048,model.lin3_size=1024,model.lr=0.16654057333719488 + id: '8' + num: 8 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7427231a47b9dc49d62f38f2b3b6ae13198a8922 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.16654057333719488 +- model.lin1_size=2048 +- model.lin2_size=2048 +- model.lin3_size=1024 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/checkpoints/epoch_016.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/checkpoints/epoch_016.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d069a312959f74893b7c586d3f5df333b40eb21b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/checkpoints/epoch_016.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c11ac48b0b034d2bc5bca48c110b0a198003597c7677c13e47d0caaf484c567 +size 1734632326 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..c025eeeaf05acdb96e6e79b87394625285c35e60 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14ead4687e0bd4352c1fbcb64235f66b7a2151fc97ed3b6664e5fa3587b3cfd3 +size 1734632390 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/tensorboard/focus150_optuna/events.out.tfevents.1646398016.e02c2cd1d8ab.1.16 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/tensorboard/focus150_optuna/events.out.tfevents.1646398016.e02c2cd1d8ab.1.16 new file mode 100644 index 0000000000000000000000000000000000000000..391c1c6f2c53888662dc4e26fd08e9740030dd58 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/tensorboard/focus150_optuna/events.out.tfevents.1646398016.e02c2cd1d8ab.1.16 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fc762015da01ec8b033fb91cc4f8af5562007186b30c24509af934b58103f08 +size 21255 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9f39f731eadadf4d189d20fe0f42dfbad5ce53e4 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/8/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 2048 + lin2_size: 2048 + lin3_size: 1024 + output_size: 1 + lr: 0.16654057333719488 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 144547841 +model/params/trainable: 144547841 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/config.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..eae454583a092a9bdec10b625d731f003e5ebcbf --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus150_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus150 + csv_file: ${data_dir}/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.1335867765111329 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/hydra.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b175cb45c4ee62bbf28fc3f54e1be42842011ff7 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus150_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.1335867765111329 + - model.lin1_size=32 + - model.lin2_size=256 + - model.lin3_size=128 + - hparams_search=focus150_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus150_optuna,model.lin1_size=32,model.lin2_size=256,model.lin3_size=128,model.lr=0.1335867765111329 + id: '9' + num: 9 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus150_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus150.yaml + datamodule: focus150.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/overrides.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e28efaca087db98526a3c331250de5f9093a7fe9 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.1335867765111329 +- model.lin1_size=32 +- model.lin2_size=256 +- model.lin3_size=128 +- hparams_search=focus150_optuna diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/checkpoints/epoch_003.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/checkpoints/epoch_003.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..0bfb31469632b43c651bb7474963c9680d30fd8b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/checkpoints/epoch_003.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8dc9374628f461da7eecda844324476d6df20a77c9a5f6e0228204ad1b657c2 +size 26448582 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/checkpoints/last.ckpt b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..9502771d92ca383be56163cb906ec50c4eb5441d --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e3e83ea21ce84921ad3f3f347a23cd75c89519742e36f66fe43e2360ad94db2 +size 26448582 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/events.out.tfevents.1646399475.e02c2cd1d8ab.1.17 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/events.out.tfevents.1646399475.e02c2cd1d8ab.1.17 new file mode 100644 index 0000000000000000000000000000000000000000..dc640f5565be81702791c0b196439a7dc2c0452b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/events.out.tfevents.1646399475.e02c2cd1d8ab.1.17 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b114d5a6b9b6c7746bf453fb3957dddf45b38cd396c5124a3dfd572ca1089ba9 +size 36328 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/events.out.tfevents.1646399560.e02c2cd1d8ab.1.18 b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/events.out.tfevents.1646399560.e02c2cd1d8ab.1.18 new file mode 100644 index 0000000000000000000000000000000000000000..5edf20cd039481b7d99dec5e2d4a1c236f0965e1 --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/events.out.tfevents.1646399560.e02c2cd1d8ab.1.18 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:608caecb3347e1778663016f2a217be4e3c5978be967c10938184bae5bf8ece2 +size 176 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/hparams.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d869ba65826a8b1e5133067883cd1a2c02ad971b --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/9/tensorboard/focus150_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 67500 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.1335867765111329 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus150 + csv_file: /usr/src/app/data/focus150/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2202337 +model/params/trainable: 2202337 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/optimization_results.yaml b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/optimization_results.yaml new file mode 100644 index 0000000000000000000000000000000000000000..92c1cbff17f8eb2bf3133ccef4871d2bd6b999ec --- /dev/null +++ b/logs/experiments/multiruns/focus150_optuna/2022-03-04_11-30-27/optimization_results.yaml @@ -0,0 +1,8 @@ +name: optuna +best_params: + datamodule.batch_size: 128 + model.lr: 0.017335998714038146 + model.lin1_size: 256 + model.lin2_size: 512 + model.lin3_size: 128 +best_value: 1.6476445198059082 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e25785e760b7cc37df98de092890f932b49ce4de --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 128 + lin2_size: 32 + lin3_size: 32 + output_size: 1 + lr: 0.04099159968275264 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a529e55a88f4c28b366701058ad34a9614f6209 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.04099159968275264 + - model.lin1_size=128 + - model.lin2_size=32 + - model.lin3_size=32 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=128,model.lin2_size=32,model.lin3_size=32,model.lr=0.04099159968275264 + id: '0' + num: 0 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..25d3c574525d7eb3bbe963120f350077487d04da --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.04099159968275264 +- model.lin1_size=128 +- model.lin2_size=32 +- model.lin3_size=32 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/checkpoints/epoch_025.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/checkpoints/epoch_025.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..7eed248803b82374761b1acdcb2ffa45d972d9a0 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/checkpoints/epoch_025.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03c441232f95b5b0e7a8eabe147370b44b414df8cf71e72364263cafc2a1e7a4 +size 26007430 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d57a18c6ea0ea4c0496e62a0aa10761e5a088d10 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:410df08e859231a4b02a8e85b4c340ea946513fe835aed5bd68a5a70d2b6e788 +size 26007430 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/events.out.tfevents.1646230713.82c8ad9a8c00.1.0 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/events.out.tfevents.1646230713.82c8ad9a8c00.1.0 new file mode 100644 index 0000000000000000000000000000000000000000..0283cb4ba8e08c9a49d3193087f1c2f404f2ef3a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/events.out.tfevents.1646230713.82c8ad9a8c00.1.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3fce43531cf37c0aa80c897797f368e0f5554cd0f12d6030b428bf411983b65 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/events.out.tfevents.1646230787.82c8ad9a8c00.1.1 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/events.out.tfevents.1646230787.82c8ad9a8c00.1.1 new file mode 100644 index 0000000000000000000000000000000000000000..1a2c44b34898bb0dcfb21e63eb1a71ac3a5eec3b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/events.out.tfevents.1646230787.82c8ad9a8c00.1.1 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2639527c404e715e18ba15b2cd35f09779f85f6e5476a93e0f5ac9b82dd40457 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..023ec3490749689406832f9d337f448141231c55 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/0/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 128 + lin2_size: 32 + lin3_size: 32 + output_size: 1 + lr: 0.04099159968275264 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2165729 +model/params/trainable: 2165729 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..41a454801b9d7b08e8d5c169f79237f50fed34f5 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 1024 + output_size: 1 + lr: 0.15818542131777294 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..68fa56c0685fd075e8e5f01cd8114cf9ad857cda --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.15818542131777294 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=1024 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=1024,model.lr=0.15818542131777294 + id: '1' + num: 1 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..dfc88b7353c8e6db65adedb8ee5b184c73413102 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.15818542131777294 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=1024 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/checkpoints/epoch_019.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/checkpoints/epoch_019.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..c6dfafc833bc61df41d2cdc920b1c3083fe45766 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/checkpoints/epoch_019.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae8109867097dcdbbbe20ac496b61f48919ea3c67387bff98578fb2e1bba858d +size 53897542 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..42c625fb821f9fedc171bf8017cc1719073c275c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d80a718c99769bfdf856f0b340799c7feeb6c4df87592052d5759e768b530a5f +size 53897606 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/events.out.tfevents.1646230788.82c8ad9a8c00.1.2 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/events.out.tfevents.1646230788.82c8ad9a8c00.1.2 new file mode 100644 index 0000000000000000000000000000000000000000..ab3b6da16a051801371897d19d6a8857de2bef68 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/events.out.tfevents.1646230788.82c8ad9a8c00.1.2 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:981fc0df76a30b785e4e4223341814f7fd652afebcafc03ea8f04ca7960cb56d +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/events.out.tfevents.1646230904.82c8ad9a8c00.1.3 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/events.out.tfevents.1646230904.82c8ad9a8c00.1.3 new file mode 100644 index 0000000000000000000000000000000000000000..5ccea00fab803a24afd09b919d848eedd71b17a8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/events.out.tfevents.1646230904.82c8ad9a8c00.1.3 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1babf837cb821e2e775bc3ef5199a09e427ddc477dfe1674c8d54e90e79b1b9 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b61d7bc29ff51a1c28ff44fb5c60d920d215b6dd --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/1/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 1024 + output_size: 1 + lr: 0.15818542131777294 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4489089 +model/params/trainable: 4489089 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..582e2a8191f6e207c4ef53bf378e7f0be1e772e6 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 512 + lin2_size: 1024 + lin3_size: 512 + output_size: 1 + lr: 0.07148210917089061 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3f7bfc89384a2638d001e7f6fd149fd6ff3f865c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.07148210917089061 + - model.lin1_size=512 + - model.lin2_size=1024 + - model.lin3_size=512 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=512,model.lin2_size=1024,model.lin3_size=512,model.lr=0.07148210917089061 + id: '10' + num: 10 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..105e44a0fb8438061eede3b61bb796429157d20d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.07148210917089061 +- model.lin1_size=512 +- model.lin2_size=1024 +- model.lin3_size=512 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/checkpoints/epoch_002.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/checkpoints/epoch_002.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..400af84da5aeca2055aa48aa84bf1a113fea3fa8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/checkpoints/epoch_002.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:067a9aa76ec028083e5ba4263f3e964caadb7e7cd8d3152069e06c2836144295 +size 116376454 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..7b5f5d270e3de53b055d45d5633afe21dca5ac41 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:666e543ae2099dbb8209d962757250ef81457f30c9d837fed6dbdc99c001bdf5 +size 116376454 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/events.out.tfevents.1646232812.82c8ad9a8c00.1.20 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/events.out.tfevents.1646232812.82c8ad9a8c00.1.20 new file mode 100644 index 0000000000000000000000000000000000000000..d804f84e0e99fdc658b84780cbc10f3697f1854b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/events.out.tfevents.1646232812.82c8ad9a8c00.1.20 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c29e0380bfd83adc767d7778fa2d1df242321ff922d9c16ce622e1ad9ad17a6 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/events.out.tfevents.1646233005.82c8ad9a8c00.1.21 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/events.out.tfevents.1646233005.82c8ad9a8c00.1.21 new file mode 100644 index 0000000000000000000000000000000000000000..c3ba5c5c3fea9bee2a6024b38c53f5216420d5c8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/events.out.tfevents.1646233005.82c8ad9a8c00.1.21 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92e0f22061792f0889cfe1f4a706181b15388e9f95988cb6e57773788c15808b +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..97869144c9106b6ef7c3cb44672f5aaeff98f5fe --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/10/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 512 + lin2_size: 1024 + lin3_size: 512 + output_size: 1 + lr: 0.07148210917089061 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 9695233 +model/params/trainable: 9695233 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..74fe0b3762ec2dbcc54e63ed09b1ab6ef7ec7a2b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 512 + output_size: 1 + lr: 0.10955577193802624 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8607264a5e4f3f6053fdbe1101d2cfed64c3292d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.10955577193802624 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=512 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=512,model.lr=0.10955577193802624 + id: '11' + num: 11 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..77fcfc3545878848c6da2c429a5f4f2550ad5e90 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.10955577193802624 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=512 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/checkpoints/epoch_011.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/checkpoints/epoch_011.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..aaca86f3f4992b7de17d96fcd7b8be98562c5dc8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/checkpoints/epoch_011.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d7af0e87a741594dfacd03e05467d68a16042f24c9a20139cf7be59778e0e3f +size 53082438 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..f8853c7bcc674e0df22d442031f25c67a106be7d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc767f835ef42b362db7e0be53b108534681309195f44ccd9435c2b9ecf196c4 +size 53082502 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/events.out.tfevents.1646233006.82c8ad9a8c00.1.22 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/events.out.tfevents.1646233006.82c8ad9a8c00.1.22 new file mode 100644 index 0000000000000000000000000000000000000000..43775fe288e93491d9e4976681670bdcf1b5069c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/events.out.tfevents.1646233006.82c8ad9a8c00.1.22 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebe3acccbbe19a8cda32bce673aaf33e07838465e7b1034496a52df627ffe84f +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/events.out.tfevents.1646233109.82c8ad9a8c00.1.23 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/events.out.tfevents.1646233109.82c8ad9a8c00.1.23 new file mode 100644 index 0000000000000000000000000000000000000000..0d1d51acdf13638a02d2e0c53c10657b4cb5a0a8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/events.out.tfevents.1646233109.82c8ad9a8c00.1.23 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c517fa7ce717843f63b789db59557534b1851d58308f9dcd5f5f4f43acb28885 +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0e63bfee6c42aad98ced67892ca9bd5aaf1f6c92 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/11/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 512 + output_size: 1 + lr: 0.10955577193802624 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4421505 +model/params/trainable: 4421505 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9959815576dd7b272daba2fbc3a20d89a0af1725 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.14872784644674936 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5043e48746de4f77d530b2b0711a5f63dbca91f3 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.14872784644674936 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=2048 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=2048,model.lr=0.14872784644674936 + id: '12' + num: 12 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f3269ec3470006a5e16ab34f507c87e1ef6a6968 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.14872784644674936 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=2048 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/checkpoints/epoch_062.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/checkpoints/epoch_062.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..a198acc08f6e7beed6122c8fe48de632e4ec9b07 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/checkpoints/epoch_062.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a907ddc2266391b2fb23dbd328ff3ff505662607eb75dd457993e4677a3f76f0 +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..188051021395166177afff05f2806258922d06ea --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3e4a4e7059e467bb91f30ba2a92838e7873c049f86d506218b13c75abf0b97f +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/events.out.tfevents.1646233109.82c8ad9a8c00.1.24 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/events.out.tfevents.1646233109.82c8ad9a8c00.1.24 new file mode 100644 index 0000000000000000000000000000000000000000..db2df583064af24038acc3ea825c87cfd1d3ffaa --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/events.out.tfevents.1646233109.82c8ad9a8c00.1.24 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e7106283fb88ad835dd94f90fc9234914d584b9d9cd4a8cd74ff26fb23716a7 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/events.out.tfevents.1646233222.82c8ad9a8c00.1.25 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/events.out.tfevents.1646233222.82c8ad9a8c00.1.25 new file mode 100644 index 0000000000000000000000000000000000000000..83feae37c3bd0c44728baad34773bbed506b0d2d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/events.out.tfevents.1646233222.82c8ad9a8c00.1.25 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3c2d138224e9f6e2fd215ad6fcbc07fb5827f596e9efc038b1add6be8e3f59f +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..97e3943f661552b66cb735a9020b3c6fd45bee69 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/12/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.14872784644674936 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4624257 +model/params/trainable: 4624257 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b84009187c34023b2647d20ece2ed70920c65f61 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 256 + output_size: 1 + lr: 0.061614158561522185 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7bcea816005bacc86a4b1a64706d25220ae0d519 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.061614158561522185 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=256 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=256,model.lr=0.061614158561522185 + id: '13' + num: 13 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0c9fc6ae52d7d8b72a3dae46ff4d17a667f29358 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.061614158561522185 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=256 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/checkpoints/epoch_001.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/checkpoints/epoch_001.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..01a09a961ed33d67f1433c8b6e0fbf071066c807 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/checkpoints/epoch_001.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1335bedcedad7097211a37f32bb9952682f72a70fe9a426735c7d3e91b96587 +size 52674886 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..aa6bce105eca37ecc9d0fb0f3e3254269e9d5192 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d5ef4eb4e8c3197452b74d20be4cbd69ac67623e0e56a06444cf144f66d99a6 +size 52674950 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/events.out.tfevents.1646233222.82c8ad9a8c00.1.26 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/events.out.tfevents.1646233222.82c8ad9a8c00.1.26 new file mode 100644 index 0000000000000000000000000000000000000000..f09df6905aeb10bbfa60c20971f6585ccd93c369 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/events.out.tfevents.1646233222.82c8ad9a8c00.1.26 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da66fcb3b35c8517a4a8cb3eed2f37b28b0a156f4d7d525e0fa7d0ab627a1332 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/events.out.tfevents.1646233336.82c8ad9a8c00.1.27 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/events.out.tfevents.1646233336.82c8ad9a8c00.1.27 new file mode 100644 index 0000000000000000000000000000000000000000..0c860a6abc65df618c45e6f6090448f2e7d7926e --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/events.out.tfevents.1646233336.82c8ad9a8c00.1.27 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64cc9ec8d415e1a0d6e2064f55125987d2392cc024bad5738fe56350794aee19 +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..de47581378a24330d70409dd723c0f44bfad7d29 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/13/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 256 + output_size: 1 + lr: 0.061614158561522185 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4387713 +model/params/trainable: 4387713 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..30a19080212578b0b1ce30890b685871c857d6a7 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 128 + lin2_size: 512 + lin3_size: 1024 + output_size: 1 + lr: 0.11620493578986595 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9e8f514614f51ee21ea914fd92d7e3bd63877637 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.11620493578986595 + - model.lin1_size=128 + - model.lin2_size=512 + - model.lin3_size=1024 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=128,model.lin2_size=512,model.lin3_size=1024,model.lr=0.11620493578986595 + id: '14' + num: 14 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..eae2dbb3a8ef6df194e7c72c1afe05fc6505af6a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.11620493578986595 +- model.lin1_size=128 +- model.lin2_size=512 +- model.lin3_size=1024 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/checkpoints/epoch_022.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/checkpoints/epoch_022.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..661817f04002e22be3953401b143213a35fe91dd --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/checkpoints/epoch_022.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd457c83c421ccde25623c5509f762d5e260d97b7a6b2bca86a2e2cf060be6c5 +size 33100678 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..5a3f9a895c9f9788fc57b5e5e6380c44c5876f07 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c041350fd85800ab9d4b54ec4407030ac49897721a08ed69b606edc642e7a924 +size 33100678 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/events.out.tfevents.1646233336.82c8ad9a8c00.1.28 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/events.out.tfevents.1646233336.82c8ad9a8c00.1.28 new file mode 100644 index 0000000000000000000000000000000000000000..5f64a78e20e819432702137a68a5b40c62895b4c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/events.out.tfevents.1646233336.82c8ad9a8c00.1.28 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58883f7006a954d147508a95adc08aef5f9f0f8510e95ecbee6b6f60ddeaed3e +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/events.out.tfevents.1646233413.82c8ad9a8c00.1.29 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/events.out.tfevents.1646233413.82c8ad9a8c00.1.29 new file mode 100644 index 0000000000000000000000000000000000000000..ee2b4784b9e99c53eff0a975410652bd2c3b810c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/events.out.tfevents.1646233413.82c8ad9a8c00.1.29 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f7bd41c9ea77b2e8a4944ed65f2c22aa4502069072a08cf32c389efd4689793 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3188db4dbc4dc7fba4a63579cb641dc1ba2a9f28 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/14/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 128 + lin2_size: 512 + lin3_size: 1024 + output_size: 1 + lr: 0.11620493578986595 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 2755841 +model/params/trainable: 2755841 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..afde0fc6da1cb3fc81281c442e37a71e2b09d470 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 64 + lin2_size: 256 + lin3_size: 512 + output_size: 1 + lr: 0.0004139807643157769 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6d73a748228cca002f0270eeb5902b4c9835a00c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.0004139807643157769 + - model.lin1_size=64 + - model.lin2_size=256 + - model.lin3_size=512 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=64,model.lin2_size=256,model.lin3_size=512,model.lr=0.0004139807643157769 + id: '15' + num: 15 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0a0e1b93c7bb2c2136e6d7c6af1996458095d626 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.0004139807643157769 +- model.lin1_size=64 +- model.lin2_size=256 +- model.lin3_size=512 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/checkpoints/epoch_079.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/checkpoints/epoch_079.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..0a841ada04536a67208cd29b7bb01d17186122ca --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/checkpoints/epoch_079.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03865ba7b4a1709b8bc93d7325a86721c413fb76c04083efee207bdb660544bd +size 14789510 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..dff133c584bbc0fee71151bf6a68d7305424d4c5 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43c8d3082b99a44d64688d9eb7c7a01034c616257b587c5c764a5ae12698b0f8 +size 14789510 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/events.out.tfevents.1646233413.82c8ad9a8c00.1.30 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/events.out.tfevents.1646233413.82c8ad9a8c00.1.30 new file mode 100644 index 0000000000000000000000000000000000000000..4a216e95bc2831628bc756f69ab67fe03f7cda1d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/events.out.tfevents.1646233413.82c8ad9a8c00.1.30 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd18e832a582508c779bd3c02ba73d0532507489a51ad8653fedef29370a244f +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/events.out.tfevents.1646233469.82c8ad9a8c00.1.31 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/events.out.tfevents.1646233469.82c8ad9a8c00.1.31 new file mode 100644 index 0000000000000000000000000000000000000000..03ab42ac6b09b1bacb679847d395a7ce6846f65b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/events.out.tfevents.1646233469.82c8ad9a8c00.1.31 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f908c8d852c71f901489dc7996899e86f00e1c254759f799e65381dc7cb8de84 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4529f719f5a9055dd91484045013f1d2c1b8ecf6 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/15/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 64 + lin2_size: 256 + lin3_size: 512 + output_size: 1 + lr: 0.0004139807643157769 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 1230465 +model/params/trainable: 1230465 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..707a94584ec9376165c7da1971e35a66211ad9a9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 512 + lin2_size: 32 + lin3_size: 1024 + output_size: 1 + lr: 0.08117903431854909 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4724aceb6befb19b0302431df82c938e23a32896 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.08117903431854909 + - model.lin1_size=512 + - model.lin2_size=32 + - model.lin3_size=1024 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=512,model.lin2_size=32,model.lin3_size=1024,model.lr=0.08117903431854909 + id: '16' + num: 16 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ad36d96844a36eb925b206a9952542c4998d79f6 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.08117903431854909 +- model.lin1_size=512 +- model.lin2_size=32 +- model.lin3_size=1024 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/checkpoints/epoch_095.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/checkpoints/epoch_095.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..8c2c4e3a01fc380c44cb04dafd291185d9ece46f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/checkpoints/epoch_095.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5366bd93f41504f31715f74b64adea6a0b8dfdaee2e80801097efb41684049e5 +size 104368390 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..2ec5cf57bf89f5f78070ff7d83fb09a857295245 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e4d9a11180180a2e140d557ee59dcba67db4f3e25c8e32579f4ea676c3040fe +size 104368390 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/events.out.tfevents.1646233470.82c8ad9a8c00.1.32 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/events.out.tfevents.1646233470.82c8ad9a8c00.1.32 new file mode 100644 index 0000000000000000000000000000000000000000..ad156d026e88cd45aa52eb5ff8e6a4eccc5bfd8b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/events.out.tfevents.1646233470.82c8ad9a8c00.1.32 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d544d54440905049c2515f747ab4112abcd662667b95c5e0446216be2eb5e34 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/events.out.tfevents.1646233654.82c8ad9a8c00.1.33 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/events.out.tfevents.1646233654.82c8ad9a8c00.1.33 new file mode 100644 index 0000000000000000000000000000000000000000..980adb894ac4c417f5d729d7101be939cf5d3d1e --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/events.out.tfevents.1646233654.82c8ad9a8c00.1.33 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0988bce22333e171a063ba58278487bc33d83efe4f8dcbde5490fde2c1f26990 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9ea364d1488c1e914c14d9b6111c7949067ce7c9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/16/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 512 + lin2_size: 32 + lin3_size: 1024 + output_size: 1 + lr: 0.08117903431854909 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 8694881 +model/params/trainable: 8694881 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e88ab2972fb77d6efb0ea9b4ea3af18bd792acba --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 2048 + lin2_size: 64 + lin3_size: 2048 + output_size: 1 + lr: 0.15496012629324551 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..754cb958c1c7ecaebbca6e2d6aecee48d5afd583 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.15496012629324551 + - model.lin1_size=2048 + - model.lin2_size=64 + - model.lin3_size=2048 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=2048,model.lin2_size=64,model.lin3_size=2048,model.lr=0.15496012629324551 + id: '17' + num: 17 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..78c07eb823adfa83b8dd5592db4ff7f91d4fb6c4 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.15496012629324551 +- model.lin1_size=2048 +- model.lin2_size=64 +- model.lin3_size=2048 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/checkpoints/epoch_021.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/checkpoints/epoch_021.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..c00c7707f6c1e45ee4fef715705fb59e8942d5c7 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/checkpoints/epoch_021.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fa4ebe8b024c3e74a8c678408a00d18f6e8bc485c0bf947274dc7a6523f484b +size 418090630 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..435609c5128302208fd3e439645698dd8ac0ad2d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b789de9d909c9fde19b3397df2eeb8777eb2fb894689b988004c4015fcf7c9c4 +size 418090630 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/events.out.tfevents.1646233654.82c8ad9a8c00.1.34 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/events.out.tfevents.1646233654.82c8ad9a8c00.1.34 new file mode 100644 index 0000000000000000000000000000000000000000..5a4fc0e07f4868ff272b9f269343bf0426eb6377 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/events.out.tfevents.1646233654.82c8ad9a8c00.1.34 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93c300e003a245b11f092c95b8b23aacdeedb453cafff3f2a41cfa552b839aef +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/events.out.tfevents.1646234311.82c8ad9a8c00.1.35 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/events.out.tfevents.1646234311.82c8ad9a8c00.1.35 new file mode 100644 index 0000000000000000000000000000000000000000..878151dd4cdfd3b941e23bd8e6585e055bd5a5f8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/events.out.tfevents.1646234311.82c8ad9a8c00.1.35 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b014fc7c4bbc2a0d7e5cab80eb38850377960610891bf245af41f100965c0284 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fb74bc80c7ec20a096ab544dea823f1523f18ec0 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/17/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 2048 + lin2_size: 64 + lin3_size: 2048 + output_size: 1 + lr: 0.15496012629324551 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 34836673 +model/params/trainable: 34836673 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2d7a33a8acb716f7b39fe86b169909d0371ec1a2 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 32 + lin2_size: 128 + lin3_size: 64 + output_size: 1 + lr: 0.05150073483440863 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7499b2590fd783f552ad56f38ff1f62d311d73f9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.05150073483440863 + - model.lin1_size=32 + - model.lin2_size=128 + - model.lin3_size=64 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=32,model.lin2_size=128,model.lin3_size=64,model.lr=0.05150073483440863 + id: '18' + num: 18 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3498781715082e48c88ff84e457a9ac5c5f5148c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.05150073483440863 +- model.lin1_size=32 +- model.lin2_size=128 +- model.lin3_size=64 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/checkpoints/epoch_005.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/checkpoints/epoch_005.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..5b8efc5ade967f7d927de8643a885eca240d2e1d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/checkpoints/epoch_005.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e666b4d8b8873c5d503467e827ff688d1b1f6ee030350900b320b7f6e2781c39 +size 6655238 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..7e66791e7d1d069e777e27b30388fbe21ee86561 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68d2791fcb6db9ff32135e6d372bac4d02dd0006a46e40ce18105f07da44d7db +size 6655302 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/events.out.tfevents.1646234312.82c8ad9a8c00.1.36 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/events.out.tfevents.1646234312.82c8ad9a8c00.1.36 new file mode 100644 index 0000000000000000000000000000000000000000..07a9ac703da1bbd8950e4b10d2d1f57f4503fea8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/events.out.tfevents.1646234312.82c8ad9a8c00.1.36 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1bbceef244f19d22fd11ff0929b4ac555f8fa16853a327d5467c8bd7a940653 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/events.out.tfevents.1646234360.82c8ad9a8c00.1.37 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/events.out.tfevents.1646234360.82c8ad9a8c00.1.37 new file mode 100644 index 0000000000000000000000000000000000000000..82a9ee1572b334157adc77f9c186bf3b695ccf25 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/events.out.tfevents.1646234360.82c8ad9a8c00.1.37 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:149a2de8073dd48b2d4b1bb6f209fe2ec26696a2f196f59042f4fa90b077015a +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cc0476b11f22cf20124dcafb7024881363f0d631 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/18/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 32 + lin2_size: 128 + lin3_size: 64 + output_size: 1 + lr: 0.05150073483440863 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 553025 +model/params/trainable: 553025 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3522c47d4658447862b7120773142f16841568ab --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 256 + output_size: 1 + lr: 0.09059799301091165 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f211d4f8977681fc4f0000c4d8ce58b3d3a49240 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.09059799301091165 + - model.lin1_size=256 + - model.lin2_size=1024 + - model.lin3_size=256 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=1024,model.lin3_size=256,model.lr=0.09059799301091165 + id: '19' + num: 19 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..91b17c5a44d766d4ff3bb1d27b26ab9894414170 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.09059799301091165 +- model.lin1_size=256 +- model.lin2_size=1024 +- model.lin3_size=256 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/checkpoints/epoch_006.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/checkpoints/epoch_006.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d26944549b3a95e2c240dbada6c92b27ea0fdc0f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/checkpoints/epoch_006.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:640b76fbea863c8d89f66f4e167c1e263cd021896b1c07c372d5e37cb46b69c2 +size 58219398 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..b8714ce0d4d03887d35658c3d5f82c6fa3f3c30f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f130a56313251ae8993e825c7db058922b1a5c015f1cc0d8f4fbc83f8c0dca65 +size 58219398 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/events.out.tfevents.1646234361.82c8ad9a8c00.1.38 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/events.out.tfevents.1646234361.82c8ad9a8c00.1.38 new file mode 100644 index 0000000000000000000000000000000000000000..2cbc3f2b10d950eb4cc95f5cc5041d80f6330435 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/events.out.tfevents.1646234361.82c8ad9a8c00.1.38 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74928ec594e759f37420fcb8b26ea55058cfe587790a3e749cf837e15d03706f +size 36224 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/events.out.tfevents.1646234472.82c8ad9a8c00.1.39 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/events.out.tfevents.1646234472.82c8ad9a8c00.1.39 new file mode 100644 index 0000000000000000000000000000000000000000..0654d0ef7594ef5218cef277b406c13d73a8064d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/events.out.tfevents.1646234472.82c8ad9a8c00.1.39 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b5cce2020e8c92c6f3b08def9a376fe119c905015bf0bd327deb97d15a0ebc9 +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0f7c58d449a67142a5d06171860f317191a6a994 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/19/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 256 + output_size: 1 + lr: 0.09059799301091165 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4849153 +model/params/trainable: 4849153 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e07723b5600bf9aeb451455f5013ecdff30c9ae8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 1024 + lin2_size: 1024 + lin3_size: 32 + output_size: 1 + lr: 0.18564129276815303 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0c620eeb9dcc8ebd2ba5cf6b8944511334f7d306 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.18564129276815303 + - model.lin1_size=1024 + - model.lin2_size=1024 + - model.lin3_size=32 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus_optuna,model.lin1_size=1024,model.lin2_size=1024,model.lin3_size=32,model.lr=0.18564129276815303 + id: '2' + num: 2 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cd9ab7740e40fcb2bb69f403893c68ce8a482681 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.18564129276815303 +- model.lin1_size=1024 +- model.lin2_size=1024 +- model.lin3_size=32 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/checkpoints/epoch_041.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/checkpoints/epoch_041.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..c3bd9df8213f08cf99f44690f7a1a82cdc6df1c4 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/checkpoints/epoch_041.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be50287cf41bab2a44dbb38393076543b2f8297d395821065c40a7493a70057f +size 220445254 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..9331b098d95bcf51136408581a63fabef7368f36 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f08b6f9a3069816962a64d04a451aab29a2a0938b9c6d83f02c7ab3597d7cd2 +size 220445318 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/events.out.tfevents.1646230905.82c8ad9a8c00.1.4 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/events.out.tfevents.1646230905.82c8ad9a8c00.1.4 new file mode 100644 index 0000000000000000000000000000000000000000..ce4b7e641b977cb473caec85e0af1b12bc0f88c3 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/events.out.tfevents.1646230905.82c8ad9a8c00.1.4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0d33693318f2e7f71633368d20f0e709ca374a5ac444b7804cdc7e77ee7e2aa +size 36224 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/events.out.tfevents.1646231265.82c8ad9a8c00.1.5 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/events.out.tfevents.1646231265.82c8ad9a8c00.1.5 new file mode 100644 index 0000000000000000000000000000000000000000..a9641e4fa4d964b34a4acd1e86efbf5b793cbd61 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/events.out.tfevents.1646231265.82c8ad9a8c00.1.5 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91050ee23a19716c24a20d7ca640325eaf50e2cc448c54c0e3a9f0ec23e20813 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e7441f246488a7758e59d6dad701ef8bc03d2137 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/2/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 1024 + lin2_size: 1024 + lin3_size: 32 + output_size: 1 + lr: 0.18564129276815303 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 18367617 +model/params/trainable: 18367617 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..484e2947574a12a0ab3195a3fb21e0c0d458b3fd --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 1024 + output_size: 1 + lr: 0.1316071233536351 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..23682877e7f856b3cfc584b27123f6ba0e4c8c67 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.1316071233536351 + - model.lin1_size=256 + - model.lin2_size=1024 + - model.lin3_size=1024 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=1024,model.lin3_size=1024,model.lr=0.1316071233536351 + id: '20' + num: 20 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6320ab8dc068e4257f1aa0f348e32d6da7cd80f1 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.1316071233536351 +- model.lin1_size=256 +- model.lin2_size=1024 +- model.lin3_size=1024 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/checkpoints/epoch_038.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/checkpoints/epoch_038.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..d1835005d8b1d8e5ff844cb21e60d4f1e811418f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/checkpoints/epoch_038.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c1b364c157da3b3fec285b3137e79acf32c2a0f7aac4feba56ca9797b4dc35a +size 67699590 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..a3161de9125d6b7522b8fe61205c83527dca01e0 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66e2d978d4e624f59646528699f9a7294a64127400e79e5c51fc28b26983fff6 +size 67699590 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/events.out.tfevents.1646234472.82c8ad9a8c00.1.40 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/events.out.tfevents.1646234472.82c8ad9a8c00.1.40 new file mode 100644 index 0000000000000000000000000000000000000000..1cf99f682508069de863ef405253dc618668e56a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/events.out.tfevents.1646234472.82c8ad9a8c00.1.40 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1159082ec8b4fc08fb6711c0e416a15385b4b8909ede72131c556be47210410 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/events.out.tfevents.1646234602.82c8ad9a8c00.1.41 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/events.out.tfevents.1646234602.82c8ad9a8c00.1.41 new file mode 100644 index 0000000000000000000000000000000000000000..dcab74206f777a1b478ea62ec4092cbce4a26b3a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/events.out.tfevents.1646234602.82c8ad9a8c00.1.41 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1a8aaf0d7a174c78f77d091a2258c064f9fdefd26520bb9850d94bb97273eab +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..13be7ce9f32e6d2995915c518801263c19d2c3e0 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/20/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 1024 + output_size: 1 + lr: 0.1316071233536351 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 5638657 +model/params/trainable: 5638657 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f4af8c0082937a0bb953a9e1810109674321b669 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.1589506088094987 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cb5353976107e435ebe53cd4887553d7bf83d863 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.1589506088094987 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=2048 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=2048,model.lr=0.1589506088094987 + id: '21' + num: 21 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..637c8e6397b900c75bfccab7dc5f9a04959e0450 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.1589506088094987 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=2048 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/checkpoints/epoch_064.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/checkpoints/epoch_064.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..34aaaeee523e017e8d8bf9fd08e7d9ac62a094db --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/checkpoints/epoch_064.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac1fb49b6d940bf39f60f331b063b515f5316916f4a1fa3aae85bdb47504ee53 +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..125a1a52f8e4a7ac4a6a06c57b6c146253c878d3 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc7951f602eaad7b03931ec80ab564f4f8df81eeb9ae480aba420422502bc6a6 +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/events.out.tfevents.1646234603.82c8ad9a8c00.1.42 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/events.out.tfevents.1646234603.82c8ad9a8c00.1.42 new file mode 100644 index 0000000000000000000000000000000000000000..dd8eb388fe0ff8e37678eeb683389baaeba08211 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/events.out.tfevents.1646234603.82c8ad9a8c00.1.42 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0e6eb09078250dadb96344c5d8bb3a8c2aebba2a0872ef758e3b48ff1a93727 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/events.out.tfevents.1646234713.82c8ad9a8c00.1.43 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/events.out.tfevents.1646234713.82c8ad9a8c00.1.43 new file mode 100644 index 0000000000000000000000000000000000000000..4757f43f4655b52f2f7462af6bec0204c63e53f4 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/events.out.tfevents.1646234713.82c8ad9a8c00.1.43 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07774c842a69b7509f772bd47e10b2888b75d04d1f16f5f33582930d96944694 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..984b4fbde73b88fcba7ed7eacb83e185b03b7526 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/21/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.1589506088094987 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4624257 +model/params/trainable: 4624257 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f6449003da3ee3668f6589ab2bfd01c66bb04738 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.13350487078732648 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8e6a6cc6e2ddd7c13d0ce4344af5c9514e46d3cb --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.13350487078732648 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=2048 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=2048,model.lr=0.13350487078732648 + id: '22' + num: 22 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..53e7b5b23de221d8d6b3938ab04868cb56b12f0f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.13350487078732648 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=2048 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/checkpoints/epoch_027.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/checkpoints/epoch_027.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..e017f30b6c47749f8c4c7db8d9a45aca30787fe4 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/checkpoints/epoch_027.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e4d3ef07947dec227356d82c6dc807f213761cf33287dc26ca3e4093723dfb6 +size 55527750 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..6046c263e9798233eb44810a29e4ccee64b3d5c6 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dec18b8473e45929ac1a1f78d96e4db6ff85e01a9df03106a71f9fe141d1fcd7 +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/events.out.tfevents.1646234713.82c8ad9a8c00.1.44 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/events.out.tfevents.1646234713.82c8ad9a8c00.1.44 new file mode 100644 index 0000000000000000000000000000000000000000..12ce56e0025c69dff7b9bc3efd5cc11d30bb631a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/events.out.tfevents.1646234713.82c8ad9a8c00.1.44 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:941bfd5dbc7d7b36c23790d603416fd61be9cc57e803787754abd2b937f9a205 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/events.out.tfevents.1646234822.82c8ad9a8c00.1.45 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/events.out.tfevents.1646234822.82c8ad9a8c00.1.45 new file mode 100644 index 0000000000000000000000000000000000000000..e5a73a29b34e8a66bab94bdcaa3f89c82da982b1 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/events.out.tfevents.1646234822.82c8ad9a8c00.1.45 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb806bb1ae00cea4ca4a91fd202d53b72e340851517b88df5ad92f5170f1fbbb +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..73cf43c65ac8440d027edca37ba102cb60dcc4dc --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/22/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.13350487078732648 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4624257 +model/params/trainable: 4624257 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..166f86ebff7bba02409021cac20209dbf1deac35 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.17588782769447536 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e3bc971624ca0d2be99c38081bac1767605ca3ad --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.17588782769447536 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=2048 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=2048,model.lr=0.17588782769447536 + id: '23' + num: 23 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..29b1e2b0e0efe6692818314b90573d3cd9616bf9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.17588782769447536 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=2048 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/checkpoints/epoch_084.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/checkpoints/epoch_084.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..6b4bf1e8ed2c8138623ff93e8ae5ab8cb43c8caf --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/checkpoints/epoch_084.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba3607986ec6cfe63cdc39fbb5e64a46223b8b54d77fddfdbd5c648b318b3585 +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..463ac23b8f10a329dd89c99a166c8046ae9798f4 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:280fb7ce69338b09c9a7a7e8df9726f0efb9f7619dda404916f15f8468fc2fe7 +size 55527814 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/events.out.tfevents.1646234822.82c8ad9a8c00.1.46 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/events.out.tfevents.1646234822.82c8ad9a8c00.1.46 new file mode 100644 index 0000000000000000000000000000000000000000..64ce0da452d537953ec725bf08cd923a23ec44fa --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/events.out.tfevents.1646234822.82c8ad9a8c00.1.46 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d6ae6498f240521f1a16610d542c69e5f744fe7de3947127323227887588e92 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/events.out.tfevents.1646234936.82c8ad9a8c00.1.47 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/events.out.tfevents.1646234936.82c8ad9a8c00.1.47 new file mode 100644 index 0000000000000000000000000000000000000000..891972f26fe63d9fafc790f37918efca842ba42b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/events.out.tfevents.1646234936.82c8ad9a8c00.1.47 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff45673ebc8cd8e05014a0db8a6d8437e79a6348179024e1969c0781d831db45 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6d00829247b355a0a500dcb01419c0e4aa906817 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/23/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 2048 + output_size: 1 + lr: 0.17588782769447536 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4624257 +model/params/trainable: 4624257 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e0079e683ecb09f7ff7775f8798bc558e8bfc9fa --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 512 + output_size: 1 + lr: 0.17504576036895456 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..53edc1ddcd7e9fb2f8f3e6f328d748de41d6600c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.17504576036895456 + - model.lin1_size=256 + - model.lin2_size=128 + - model.lin3_size=512 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=512,model.lr=0.17504576036895456 + id: '24' + num: 24 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d1917117f3902e3351bbd8191d544409a264174f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.17504576036895456 +- model.lin1_size=256 +- model.lin2_size=128 +- model.lin3_size=512 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/checkpoints/epoch_039.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/checkpoints/epoch_039.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..688be3b7122d7f3fa34bdde3202f4061025817af --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/checkpoints/epoch_039.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:062dd9937e18e97c95ba8e3e9f977930db2438f6e1b30f0b32f98ceafc425701 +size 53082502 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..811176650ecca6afc111ae686fe2faa65f85acc3 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46e6f3c4dcb0957abb8f6c4b1a5dc103fff9b238a1867b0410e89459927a22c3 +size 53082502 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/events.out.tfevents.1646234936.82c8ad9a8c00.1.48 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/events.out.tfevents.1646234936.82c8ad9a8c00.1.48 new file mode 100644 index 0000000000000000000000000000000000000000..8bf725e184b00820095698592fa3982d776f6fdf --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/events.out.tfevents.1646234936.82c8ad9a8c00.1.48 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34cff3283459349c1cc6e26d9f4e21fc8fc8156b925063d70f0bb55a5e29a3b4 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/events.out.tfevents.1646235046.82c8ad9a8c00.1.49 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/events.out.tfevents.1646235046.82c8ad9a8c00.1.49 new file mode 100644 index 0000000000000000000000000000000000000000..f58b24046103727ff6d680e54452a68f7afcf421 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/events.out.tfevents.1646235046.82c8ad9a8c00.1.49 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d32046666eb56137e540a2a15bbbf1bd05d1eceb6ba1e5106aedcb6dd7313ab8 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..777a3bf18f50a0942ce574499b4c827880dbd1cd --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/24/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 128 + lin3_size: 512 + output_size: 1 + lr: 0.17504576036895456 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4421505 +model/params/trainable: 4421505 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..dcbc19c4491973db6e60f75fa8c9ea060ce31642 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 64 + lin2_size: 2048 + lin3_size: 32 + output_size: 1 + lr: 0.09908496566082269 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..91f497ab99362ac5c69e55d9510b098806067686 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.09908496566082269 + - model.lin1_size=64 + - model.lin2_size=2048 + - model.lin3_size=32 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus_optuna,model.lin1_size=64,model.lin2_size=2048,model.lin3_size=32,model.lr=0.09908496566082269 + id: '3' + num: 3 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..89cc56548549b372ea4fd76e689b5284f16c5737 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.09908496566082269 +- model.lin1_size=64 +- model.lin2_size=2048 +- model.lin3_size=32 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/checkpoints/epoch_003.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/checkpoints/epoch_003.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..a6170e05086e71500fdd4889aa1efb1b995ca760 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/checkpoints/epoch_003.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f56a6d79010a5cf44249f93a07c19c31a795fefba213460b838ddaf44f1b5ca4 +size 15431238 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..4ef301a7ef868e963e9f4a4959905a082f38e365 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:668d31c37bbb25cd64999d076395a0a14a3f1295e2273195cb867f37567d5ead +size 15431238 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/events.out.tfevents.1646231265.82c8ad9a8c00.1.6 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/events.out.tfevents.1646231265.82c8ad9a8c00.1.6 new file mode 100644 index 0000000000000000000000000000000000000000..d1f96d6c73b133471576b46483333f0dd77d97db --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/events.out.tfevents.1646231265.82c8ad9a8c00.1.6 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:687fd8657e4ee664cbcc200b3d5081c7ca34f9fdb6bafa72ff91f951681cccf5 +size 36224 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/events.out.tfevents.1646231315.82c8ad9a8c00.1.7 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/events.out.tfevents.1646231315.82c8ad9a8c00.1.7 new file mode 100644 index 0000000000000000000000000000000000000000..47c7487b96cfcbd4c448ebde721655bf66f9f4c8 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/events.out.tfevents.1646231315.82c8ad9a8c00.1.7 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e0c49a81092e24dd7f11b0f0e933f3cec0fa9a7a4d2e6fdfa6327bfe2bf8588 +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3145f556c6d4209ca008c819d51ee2cf88242e65 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/3/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 64 + lin2_size: 2048 + lin3_size: 32 + output_size: 1 + lr: 0.09908496566082269 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 1283073 +model/params/trainable: 1283073 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9535fbdb0168876b4af381c1615d973a4f45963e --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.017335998714038146 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1d244cae30cf5974683c65c904955bce73bda8d9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=128 + - model.lr=0.017335998714038146 + - model.lin1_size=256 + - model.lin2_size=512 + - model.lin3_size=128 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=128,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=512,model.lin3_size=128,model.lr=0.017335998714038146 + id: '4' + num: 4 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..57858ced2a704891a6082b0b9b41d987cc45349a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=128 +- model.lr=0.017335998714038146 +- model.lin1_size=256 +- model.lin2_size=512 +- model.lin3_size=128 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/checkpoints/epoch_090.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/checkpoints/epoch_090.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..b581ddc0e66b232c42e66a19a2a345816058cb12 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/checkpoints/epoch_090.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a1f772022d3593da2ca930ac4844d995b2a253ca50d6bc4e0858a78fd5ebcd6 +size 54257542 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..47567ac6132f675c364c8e32677c186c059fbed2 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee66460b4e83f75d5d425fe4a21e4bcd202c4e127e0adba7739266436e3d0824 +size 54257542 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/events.out.tfevents.1646231315.82c8ad9a8c00.1.8 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/events.out.tfevents.1646231315.82c8ad9a8c00.1.8 new file mode 100644 index 0000000000000000000000000000000000000000..4f8398ad4e675c4c2c68588345b8ba20c3855263 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/events.out.tfevents.1646231315.82c8ad9a8c00.1.8 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c75e3bf6d269af1ce44d3a0c8e5dfe35c4e221f9c4beb0ee65a301d215551479 +size 36224 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/events.out.tfevents.1646231425.82c8ad9a8c00.1.9 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/events.out.tfevents.1646231425.82c8ad9a8c00.1.9 new file mode 100644 index 0000000000000000000000000000000000000000..11b754a7211ad6168720fa27b4abed698f3fb883 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/events.out.tfevents.1646231425.82c8ad9a8c00.1.9 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25ca661d0e5368659d0fa70d03afddf7cdb8611f1aa1ea5d94a5d312181632cc +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0c3c00d0a0b932c3c2fbf80a3bbf22d3d595109e --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/4/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 512 + lin3_size: 128 + output_size: 1 + lr: 0.017335998714038146 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 128 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4519425 +model/params/trainable: 4519425 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2fe7850cac652d24e7729ffc0584f165be063482 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 1024 + lin2_size: 64 + lin3_size: 128 + output_size: 1 + lr: 0.01956889375661914 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5896940cadba03cfadc93566af3bab5b1e002ab7 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.01956889375661914 + - model.lin1_size=1024 + - model.lin2_size=64 + - model.lin3_size=128 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=1024,model.lin2_size=64,model.lin3_size=128,model.lr=0.01956889375661914 + id: '5' + num: 5 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f8592164e531c12932f7791447614a3832d004b4 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.01956889375661914 +- model.lin1_size=1024 +- model.lin2_size=64 +- model.lin3_size=128 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/checkpoints/epoch_029.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/checkpoints/epoch_029.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..f8f969ff830a24dfb6a4af1e03a568ee3bb8b3e7 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/checkpoints/epoch_029.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8a1e96aab9c16e63a03c1dc19e41a7132b794a61c412fcbf23fdfb2f786b5ac +size 208316998 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..3de33c32ab80774bb30b4fa82991959e3248ac74 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97a1da96cf1e6f65c09a0078198f9fadec6ebfaea80092ea6e64f29edc640bd9 +size 208316998 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/events.out.tfevents.1646231425.82c8ad9a8c00.1.10 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/events.out.tfevents.1646231425.82c8ad9a8c00.1.10 new file mode 100644 index 0000000000000000000000000000000000000000..db8c7a1a691b8b064744184b445ababbfb54606b --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/events.out.tfevents.1646231425.82c8ad9a8c00.1.10 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ecb800217c41dfea93429ef3423deec8fee19d654f2d1126f8fdb4aa414e483 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/events.out.tfevents.1646231770.82c8ad9a8c00.1.11 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/events.out.tfevents.1646231770.82c8ad9a8c00.1.11 new file mode 100644 index 0000000000000000000000000000000000000000..16c006178e2543c1bcc92ed401711b1f00ceef5e --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/events.out.tfevents.1646231770.82c8ad9a8c00.1.11 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e2e0527a9a7dbb7a6be7bbe942889f7284e587e226e153534dc2a240f36797e +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..357182e5934652a6107e4a59d54d67da8691275d --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/5/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 1024 + lin2_size: 64 + lin3_size: 128 + output_size: 1 + lr: 0.01956889375661914 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 17357505 +model/params/trainable: 17357505 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..197fc284c30a15a8899556b32730c85b5da8080c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 512 + output_size: 1 + lr: 0.012981602213938178 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..df7d78174a965ad834b97b8af25cc70c96d09e77 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.012981602213938178 + - model.lin1_size=256 + - model.lin2_size=1024 + - model.lin3_size=512 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=1024,model.lin3_size=512,model.lr=0.012981602213938178 + id: '6' + num: 6 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..261c16e78c5266f47bdd87ad9687adab725e8539 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.012981602213938178 +- model.lin1_size=256 +- model.lin2_size=1024 +- model.lin3_size=512 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/checkpoints/epoch_015.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/checkpoints/epoch_015.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..ca27c3cf415778e69bd0a543634935127bd019c1 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/checkpoints/epoch_015.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db97a04d450160a929842add8b44e42d67b72e9156a4e60b86da1500cc70bb0a +size 61379462 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..ade313982974774145047d4581e3ead9abed5463 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2d4688413eea92111420edadf06dd03590d72271bd061829f0306b27514462b +size 61379462 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/events.out.tfevents.1646231770.82c8ad9a8c00.1.12 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/events.out.tfevents.1646231770.82c8ad9a8c00.1.12 new file mode 100644 index 0000000000000000000000000000000000000000..33c2e00c31da51766c750e4c3410282ce7f65ee3 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/events.out.tfevents.1646231770.82c8ad9a8c00.1.12 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b9bb71d80a9d8f1b72d917ae003f9720acbbd2169162744c25ac464b14c8048 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/events.out.tfevents.1646231900.82c8ad9a8c00.1.13 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/events.out.tfevents.1646231900.82c8ad9a8c00.1.13 new file mode 100644 index 0000000000000000000000000000000000000000..92ee3fd5c8122ca4bff985f15a7311ba39ae8a35 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/events.out.tfevents.1646231900.82c8ad9a8c00.1.13 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:761b4cba86a351ddd607b66bdc1581f9554fb30a9a95b43cb3944c63d5b5dbca +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..dbbdb547deb3d6b36a306bfaeaac4b6b6d89a206 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/6/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 512 + output_size: 1 + lr: 0.012981602213938178 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 5112321 +model/params/trainable: 5112321 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..97fd9ff52d72b3dc654d5b7c71109f0e7a5fe0dc --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 64 + output_size: 1 + lr: 0.19115771252214422 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..58af2c6364612e0438fc5a375d1e61bad3cefa1c --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.19115771252214422 + - model.lin1_size=256 + - model.lin2_size=1024 + - model.lin3_size=64 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=256,model.lin2_size=1024,model.lin3_size=64,model.lr=0.19115771252214422 + id: '7' + num: 7 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e3354b0254f5a67e1b131e4f140ef0c1a7215bf5 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.19115771252214422 +- model.lin1_size=256 +- model.lin2_size=1024 +- model.lin3_size=64 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/checkpoints/epoch_001.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/checkpoints/epoch_001.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..e877ccb86c8700267d2ff711c732785727012a4a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/checkpoints/epoch_001.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71b26271967ede56474baf756db28cd30d43635eb1fde497cd706b5a229997a3 +size 55849286 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..eb4030564fc5542d0cf19e90bfc0bbe4f54f76d9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f605cd18f6b25aa2e1b7a5a26f873e4ce42f7f03e290ae22bd2066753e733a5d +size 55849350 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/events.out.tfevents.1646231900.82c8ad9a8c00.1.14 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/events.out.tfevents.1646231900.82c8ad9a8c00.1.14 new file mode 100644 index 0000000000000000000000000000000000000000..a6de79169920a157ee21185fc7b8a0eec413d891 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/events.out.tfevents.1646231900.82c8ad9a8c00.1.14 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48a1f29736f2467118ba3ad7b9f86b01f0e91335e9ffbdf30ee6b513b5d3fab4 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/events.out.tfevents.1646232006.82c8ad9a8c00.1.15 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/events.out.tfevents.1646232006.82c8ad9a8c00.1.15 new file mode 100644 index 0000000000000000000000000000000000000000..85daa0ffaa1bb2ab987f38155df8595e49760b3a --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/events.out.tfevents.1646232006.82c8ad9a8c00.1.15 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a438c709e6b1cc416f5b03513e37e5a874f6429d4ba74e58a38357cd5d964544 +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ecb0325543213ec0a9502838b8b9c14dac543186 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/7/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 256 + lin2_size: 1024 + lin3_size: 64 + output_size: 1 + lr: 0.19115771252214422 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 4651777 +model/params/trainable: 4651777 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..455713fd2a2c24031d8f49b93835a7b143bf2df7 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 2048 + lin2_size: 2048 + lin3_size: 1024 + output_size: 1 + lr: 0.16654057333719488 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b5863bf69dd0929b5bf53dd75e995d7f8b94e006 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=32 + - model.lr=0.16654057333719488 + - model.lin1_size=2048 + - model.lin2_size=2048 + - model.lin3_size=1024 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=32,hparams_search=focus_optuna,model.lin1_size=2048,model.lin2_size=2048,model.lin3_size=1024,model.lr=0.16654057333719488 + id: '8' + num: 8 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..514f296d4f745e5882f86e5641aee4d9fb3ef271 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=32 +- model.lr=0.16654057333719488 +- model.lin1_size=2048 +- model.lin2_size=2048 +- model.lin3_size=1024 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/checkpoints/epoch_011.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/checkpoints/epoch_011.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..0d15704d70d31c3a40bfa390d47d447bd0447b81 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/checkpoints/epoch_011.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39889dc0990d1cfe948b5a375fc4094692d5e6d4fdd93646b2ea5b8ffc3f80c4 +size 490472326 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..475c0ca6334d308e32febc079c87a5ff2584a5c0 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15d5edc12d406a5b0ffc7003639e676e59b586dfeffcc037678152d563a9ec22 +size 490472326 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/events.out.tfevents.1646232007.82c8ad9a8c00.1.16 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/events.out.tfevents.1646232007.82c8ad9a8c00.1.16 new file mode 100644 index 0000000000000000000000000000000000000000..ee702c458af0238b5926e166a81f49c7699995f9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/events.out.tfevents.1646232007.82c8ad9a8c00.1.16 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ea99bbb3f37bd0ea5d577496aa384d23e3eee51d38113107155563703214fd1 +size 36385 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/events.out.tfevents.1646232767.82c8ad9a8c00.1.17 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/events.out.tfevents.1646232767.82c8ad9a8c00.1.17 new file mode 100644 index 0000000000000000000000000000000000000000..0ee803f75ecc58f082ae3bdf98d4a3de5cf400df --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/events.out.tfevents.1646232767.82c8ad9a8c00.1.17 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c63750af4c0ce51552e8d8a9800555ea9b1849b6d81cd1c4119d6460d9ab8a65 +size 179 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b7002d8bad36279bf660dc693f75e809cb8518e0 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/8/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 2048 + lin2_size: 2048 + lin3_size: 1024 + output_size: 1 + lr: 0.16654057333719488 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 32 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 40867841 +model/params/trainable: 40867841 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/config.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8bb5f2db0d6962359bd63d72f01b1b8f85d2900f --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/config.yaml @@ -0,0 +1,66 @@ +original_work_dir: ${hydra:runtime.cwd} +data_dir: ${original_work_dir}/data +print_config: true +ignore_warnings: true +train: true +test: true +seed: null +name: focus_optuna +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: ${data_dir}/focus + csv_file: ${data_dir}/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.1335867765111329 + weight_decay: 0.0005 +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +logger: + tensorboard: + _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger + save_dir: tensorboard/ + name: null + version: ${name} + log_graph: false + default_hp_metric: true + prefix: '' +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +optimized_metric: val/mae_best diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/hydra.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4df7a90a417eca93b5c4081dc7ed937a913abbe9 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/hydra.yaml @@ -0,0 +1,232 @@ +hydra: + run: + dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + sweep: + dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + sampler: + _target_: optuna.samplers.TPESampler + seed: 12345 + consider_prior: true + prior_weight: 1.0 + consider_magic_clip: true + consider_endpoints: false + n_startup_trials: 10 + n_ei_candidates: 24 + multivariate: false + warn_independent_sampling: true + _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper + direction: minimize + storage: null + study_name: focus_hyperparameter_search + n_trials: 25 + n_jobs: 1 + search_space: + datamodule.batch_size: + type: categorical + choices: + - 32 + - 64 + - 128 + model.lr: + type: float + low: 0.0001 + high: 0.2 + model.lin1_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin2_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + model.lin3_size: + type: categorical + choices: + - 32 + - 64 + - 128 + - 256 + - 512 + - 1024 + - 2048 + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: [] + task: + - datamodule.batch_size=64 + - model.lr=0.1335867765111329 + - model.lin1_size=32 + - model.lin2_size=256 + - model.lin3_size=128 + - hparams_search=focus_optuna + job: + name: train + override_dirname: datamodule.batch_size=64,hparams_search=focus_optuna,model.lin1_size=32,model.lin2_size=256,model.lin3_size=128,model.lr=0.1335867765111329 + id: '9' + num: 9 + config_name: train.yaml + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.1.1 + cwd: /usr/src/app + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /usr/src/app/configs + schema: file + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: '' + schema: structured + provider: schema + choices: + local: default.yaml + hparams_search: focus_optuna + debug: null + experiment: null + log_dir: default.yaml + trainer: long.yaml + logger: tensorboard + callbacks: default.yaml + model: focus.yaml + datamodule: focus.yaml + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: optuna + hydra/sweeper/sampler: tpe + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/overrides.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..46cbc16fc675a78e0d97d05a389d6d2c855d7a72 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/.hydra/overrides.yaml @@ -0,0 +1,6 @@ +- datamodule.batch_size=64 +- model.lr=0.1335867765111329 +- model.lin1_size=32 +- model.lin2_size=256 +- model.lin3_size=128 +- hparams_search=focus_optuna diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/checkpoints/epoch_003.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/checkpoints/epoch_003.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..4d5be1d4b2467e498320f70fc3fe357e76e801b5 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/checkpoints/epoch_003.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebeeed7cc286370fa6f4697a312a3ac994af57c448e89be8290d77185ecc82d8 +size 7008582 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/checkpoints/last.ckpt b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/checkpoints/last.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..eaa1a903cafee8a2e066d5700fb80804994a7f72 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/checkpoints/last.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20f9b4e900a4769171905fbacda3b0cccb942cc163d5e0859d308ac979d8663a +size 7008582 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/events.out.tfevents.1646232767.82c8ad9a8c00.1.18 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/events.out.tfevents.1646232767.82c8ad9a8c00.1.18 new file mode 100644 index 0000000000000000000000000000000000000000..777c537449cb3140aed340c3e26a4b0c0a4af4ff --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/events.out.tfevents.1646232767.82c8ad9a8c00.1.18 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d850f356c8262dfb6a2b3aafa49fbbcfe66a846e6cfff2c62ae26aabc9cac175 +size 36322 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/events.out.tfevents.1646232812.82c8ad9a8c00.1.19 b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/events.out.tfevents.1646232812.82c8ad9a8c00.1.19 new file mode 100644 index 0000000000000000000000000000000000000000..f7affcc3d004857e9efe70caafdbb9527b075a48 --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/events.out.tfevents.1646232812.82c8ad9a8c00.1.19 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d01b59234a3b8f44e21d2489beb92d0a5f91a96aa60bc82ea040b59fced74cdc +size 176 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/hparams.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/hparams.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ab8682d4fd04905e60f729e7f97830cea8043fbb --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/9/tensorboard/focus_optuna/hparams.yaml @@ -0,0 +1,52 @@ +trainer: + _target_: pytorch_lightning.Trainer + gpus: 1 + min_epochs: 1 + max_epochs: 100 + resume_from_checkpoint: null +model: + _target_: src.models.focus_module.FocusLitModule + input_size: 16875 + lin1_size: 32 + lin2_size: 256 + lin3_size: 128 + output_size: 1 + lr: 0.1335867765111329 + weight_decay: 0.0005 +datamodule: + _target_: src.datamodules.focus_datamodule.FocusDataModule + data_dir: /usr/src/app/data/focus + csv_file: /usr/src/app/data/focus/metadata.csv + batch_size: 64 + train_val_test_split_percentage: + - 0.7 + - 0.15 + - 0.15 + num_workers: 0 + pin_memory: false +seed: null +callbacks: + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + monitor: val/mae + mode: min + save_top_k: 1 + save_last: true + verbose: false + dirpath: checkpoints/ + filename: epoch_{epoch:03d} + auto_insert_metric_name: false + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: val/mae + mode: min + patience: 100 + min_delta: 0 + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + max_depth: -1 + rich_progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar +model/params/total: 582337 +model/params/trainable: 582337 +model/params/non_trainable: 0 diff --git a/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/optimization_results.yaml b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/optimization_results.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d9c5aae3f115241e85833f2f3b628a15cc471bdf --- /dev/null +++ b/logs/experiments/multiruns/focus_optuna/2022-03-02_14-18-30/optimization_results.yaml @@ -0,0 +1,8 @@ +name: optuna +best_params: + datamodule.batch_size: 32 + model.lr: 0.012981602213938178 + model.lin1_size: 256 + model.lin2_size: 1024 + model.lin3_size: 512 +best_value: 1.5875822305679321