Hannes Kuchelmeister
commited on
Commit
•
400abb9
1
Parent(s):
d056ddb
add simple conv model with relu and batchnorm
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- configs/experiment/focusConvReLU_MSE_150.yaml +40 -0
- configs/model/focusConvReLU_150.yaml +13 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/config.yaml +81 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/hydra.yaml +170 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/overrides.yaml +1 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/config.yaml +81 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/hydra.yaml +170 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/overrides.yaml +1 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/config.yaml +81 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/hydra.yaml +170 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/overrides.yaml +1 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/config.yaml +81 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/hydra.yaml +170 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/overrides.yaml +1 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/epoch_096.ckpt +3 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/last.ckpt +3 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/hparams.yaml +55 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/metrics.csv +202 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652279809.c49d1a46d86b.1.0 +3 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652280195.c49d1a46d86b.1.1 +3 -0
- logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/hparams.yaml +55 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/meta.yaml +15 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/epoch +201 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/loss +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/mae +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/loss +100 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/mae +100 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/loss +100 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae +100 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae_best +100 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/_target_ +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/min_delta +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/mode +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/monitor +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/patience +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/_target_ +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/auto_insert_metric_name +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/dirpath +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/filename +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/mode +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/monitor +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_last +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_top_k +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/verbose +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/_target_ +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/max_depth +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/rich_progress_bar/_target_ +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/_target_ +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/augmentation +1 -0
- logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/batch_size +1 -0
configs/experiment/focusConvReLU_MSE_150.yaml
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# @package _global_
|
2 |
+
|
3 |
+
# to execute this experiment run:
|
4 |
+
# python train.py experiment=example
|
5 |
+
|
6 |
+
defaults:
|
7 |
+
- override /datamodule: focus150.yaml
|
8 |
+
- override /model: focusConvReLU_150.yaml
|
9 |
+
- override /callbacks: default.yaml
|
10 |
+
- override /logger: many_loggers
|
11 |
+
- override /trainer: default.yaml
|
12 |
+
|
13 |
+
# all parameters below will be merged with parameters from default configurations set above
|
14 |
+
# this allows you to overwrite only specified parameters
|
15 |
+
|
16 |
+
# name of the run determines folder name in logs
|
17 |
+
name: "focusConvReLU_MSE_150"
|
18 |
+
seed: 12345
|
19 |
+
|
20 |
+
trainer:
|
21 |
+
min_epochs: 1
|
22 |
+
max_epochs: 100
|
23 |
+
|
24 |
+
model:
|
25 |
+
image_size: 150
|
26 |
+
pool_size: 2
|
27 |
+
conv1_size: 3
|
28 |
+
conv1_channels: 9
|
29 |
+
conv2_size: 7
|
30 |
+
conv2_channels: 6
|
31 |
+
lin1_size: 32
|
32 |
+
lin2_size: 72
|
33 |
+
output_size: 1
|
34 |
+
lr: 0.001
|
35 |
+
weight_decay: 0.0005
|
36 |
+
|
37 |
+
datamodule:
|
38 |
+
batch_size: 64
|
39 |
+
augmentation: True
|
40 |
+
|
configs/model/focusConvReLU_150.yaml
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
2 |
+
|
3 |
+
image_size: 150
|
4 |
+
pool_size: 2
|
5 |
+
conv1_size: 5
|
6 |
+
conv1_channels: 6
|
7 |
+
conv2_size: 5
|
8 |
+
conv2_channels: 16
|
9 |
+
lin1_size: 100
|
10 |
+
lin2_size: 80
|
11 |
+
output_size: 1
|
12 |
+
lr: 0.001
|
13 |
+
weight_decay: 0.0005
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/config.yaml
ADDED
@@ -0,0 +1,81 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
original_work_dir: ${hydra:runtime.cwd}
|
2 |
+
data_dir: ${original_work_dir}/data
|
3 |
+
print_config: true
|
4 |
+
ignore_warnings: true
|
5 |
+
train: true
|
6 |
+
test: true
|
7 |
+
seed: 12345
|
8 |
+
name: focusConvReLU_MSE_150
|
9 |
+
datamodule:
|
10 |
+
_target_: src.datamodules.focus_datamodule.FocusDataModule
|
11 |
+
data_dir: ${data_dir}/focus150
|
12 |
+
csv_train_file: ${data_dir}/focus150/train_metadata.csv
|
13 |
+
csv_val_file: ${data_dir}/focus150/validation_metadata.csv
|
14 |
+
csv_test_file: ${data_dir}/focus150/test_metadata.csv
|
15 |
+
batch_size: 64
|
16 |
+
num_workers: 0
|
17 |
+
pin_memory: false
|
18 |
+
augmentation: true
|
19 |
+
model:
|
20 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
21 |
+
image_size: 150
|
22 |
+
pool_size: 2
|
23 |
+
conv1_size: 3
|
24 |
+
conv1_channels: 9
|
25 |
+
conv2_size: 7
|
26 |
+
conv2_channels: 6
|
27 |
+
lin1_size: 32
|
28 |
+
lin2_size: 72
|
29 |
+
output_size: 1
|
30 |
+
lr: 0.001
|
31 |
+
weight_decay: 0.0005
|
32 |
+
callbacks:
|
33 |
+
model_checkpoint:
|
34 |
+
_target_: pytorch_lightning.callbacks.ModelCheckpoint
|
35 |
+
monitor: val/mae
|
36 |
+
mode: min
|
37 |
+
save_top_k: 1
|
38 |
+
save_last: true
|
39 |
+
verbose: false
|
40 |
+
dirpath: checkpoints/
|
41 |
+
filename: epoch_{epoch:03d}
|
42 |
+
auto_insert_metric_name: false
|
43 |
+
early_stopping:
|
44 |
+
_target_: pytorch_lightning.callbacks.EarlyStopping
|
45 |
+
monitor: val/mae
|
46 |
+
mode: min
|
47 |
+
patience: 100
|
48 |
+
min_delta: 0
|
49 |
+
model_summary:
|
50 |
+
_target_: pytorch_lightning.callbacks.RichModelSummary
|
51 |
+
max_depth: -1
|
52 |
+
rich_progress_bar:
|
53 |
+
_target_: pytorch_lightning.callbacks.RichProgressBar
|
54 |
+
logger:
|
55 |
+
csv:
|
56 |
+
_target_: pytorch_lightning.loggers.csv_logs.CSVLogger
|
57 |
+
save_dir: .
|
58 |
+
name: csv/
|
59 |
+
prefix: ''
|
60 |
+
mlflow:
|
61 |
+
_target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
|
62 |
+
experiment_name: ${name}
|
63 |
+
tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
|
64 |
+
tags: null
|
65 |
+
save_dir: ./mlruns
|
66 |
+
prefix: ''
|
67 |
+
artifact_location: null
|
68 |
+
tensorboard:
|
69 |
+
_target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
|
70 |
+
save_dir: tensorboard/
|
71 |
+
name: null
|
72 |
+
version: ${name}
|
73 |
+
log_graph: false
|
74 |
+
default_hp_metric: true
|
75 |
+
prefix: ''
|
76 |
+
trainer:
|
77 |
+
_target_: pytorch_lightning.Trainer
|
78 |
+
gpus: 1
|
79 |
+
min_epochs: 1
|
80 |
+
max_epochs: 100
|
81 |
+
resume_from_checkpoint: null
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/hydra.yaml
ADDED
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
hydra:
|
2 |
+
run:
|
3 |
+
dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
4 |
+
sweep:
|
5 |
+
dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
6 |
+
subdir: ${hydra.job.num}
|
7 |
+
launcher:
|
8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
9 |
+
sweeper:
|
10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
11 |
+
max_batch_size: null
|
12 |
+
help:
|
13 |
+
app_name: ${hydra.job.name}
|
14 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
15 |
+
|
16 |
+
'
|
17 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
18 |
+
|
19 |
+
Use --hydra-help to view Hydra specific help
|
20 |
+
|
21 |
+
'
|
22 |
+
template: '${hydra.help.header}
|
23 |
+
|
24 |
+
== Configuration groups ==
|
25 |
+
|
26 |
+
Compose your configuration from those groups (group=option)
|
27 |
+
|
28 |
+
|
29 |
+
$APP_CONFIG_GROUPS
|
30 |
+
|
31 |
+
|
32 |
+
== Config ==
|
33 |
+
|
34 |
+
Override anything in the config (foo.bar=value)
|
35 |
+
|
36 |
+
|
37 |
+
$CONFIG
|
38 |
+
|
39 |
+
|
40 |
+
${hydra.help.footer}
|
41 |
+
|
42 |
+
'
|
43 |
+
hydra_help:
|
44 |
+
template: 'Hydra (${hydra.runtime.version})
|
45 |
+
|
46 |
+
See https://hydra.cc for more info.
|
47 |
+
|
48 |
+
|
49 |
+
== Flags ==
|
50 |
+
|
51 |
+
$FLAGS_HELP
|
52 |
+
|
53 |
+
|
54 |
+
== Configuration groups ==
|
55 |
+
|
56 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
57 |
+
to command line)
|
58 |
+
|
59 |
+
|
60 |
+
$HYDRA_CONFIG_GROUPS
|
61 |
+
|
62 |
+
|
63 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
64 |
+
|
65 |
+
'
|
66 |
+
hydra_help: ???
|
67 |
+
hydra_logging:
|
68 |
+
version: 1
|
69 |
+
formatters:
|
70 |
+
colorlog:
|
71 |
+
(): colorlog.ColoredFormatter
|
72 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
73 |
+
handlers:
|
74 |
+
console:
|
75 |
+
class: logging.StreamHandler
|
76 |
+
formatter: colorlog
|
77 |
+
stream: ext://sys.stdout
|
78 |
+
root:
|
79 |
+
level: INFO
|
80 |
+
handlers:
|
81 |
+
- console
|
82 |
+
disable_existing_loggers: false
|
83 |
+
job_logging:
|
84 |
+
version: 1
|
85 |
+
formatters:
|
86 |
+
simple:
|
87 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
88 |
+
colorlog:
|
89 |
+
(): colorlog.ColoredFormatter
|
90 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
91 |
+
- %(message)s'
|
92 |
+
log_colors:
|
93 |
+
DEBUG: purple
|
94 |
+
INFO: green
|
95 |
+
WARNING: yellow
|
96 |
+
ERROR: red
|
97 |
+
CRITICAL: red
|
98 |
+
handlers:
|
99 |
+
console:
|
100 |
+
class: logging.StreamHandler
|
101 |
+
formatter: colorlog
|
102 |
+
stream: ext://sys.stdout
|
103 |
+
file:
|
104 |
+
class: logging.FileHandler
|
105 |
+
formatter: simple
|
106 |
+
filename: ${hydra.job.name}.log
|
107 |
+
root:
|
108 |
+
level: INFO
|
109 |
+
handlers:
|
110 |
+
- console
|
111 |
+
- file
|
112 |
+
disable_existing_loggers: false
|
113 |
+
env: {}
|
114 |
+
searchpath: []
|
115 |
+
callbacks: {}
|
116 |
+
output_subdir: .hydra
|
117 |
+
overrides:
|
118 |
+
hydra: []
|
119 |
+
task:
|
120 |
+
- experiment=focusConvReLU_MSE_150
|
121 |
+
job:
|
122 |
+
name: train
|
123 |
+
override_dirname: experiment=focusConvReLU_MSE_150
|
124 |
+
id: ???
|
125 |
+
num: ???
|
126 |
+
config_name: train.yaml
|
127 |
+
env_set: {}
|
128 |
+
env_copy: []
|
129 |
+
config:
|
130 |
+
override_dirname:
|
131 |
+
kv_sep: '='
|
132 |
+
item_sep: ','
|
133 |
+
exclude_keys: []
|
134 |
+
runtime:
|
135 |
+
version: 1.1.2
|
136 |
+
cwd: /usr/src/app
|
137 |
+
config_sources:
|
138 |
+
- path: hydra.conf
|
139 |
+
schema: pkg
|
140 |
+
provider: hydra
|
141 |
+
- path: /usr/src/app/configs
|
142 |
+
schema: file
|
143 |
+
provider: main
|
144 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
145 |
+
schema: pkg
|
146 |
+
provider: hydra-colorlog
|
147 |
+
- path: ''
|
148 |
+
schema: structured
|
149 |
+
provider: schema
|
150 |
+
choices:
|
151 |
+
local: default.yaml
|
152 |
+
hparams_search: null
|
153 |
+
debug: null
|
154 |
+
experiment: focusConvReLU_MSE_150
|
155 |
+
log_dir: default.yaml
|
156 |
+
trainer: default.yaml
|
157 |
+
logger: many_loggers
|
158 |
+
callbacks: default.yaml
|
159 |
+
model: focusConvReLU_150.yaml
|
160 |
+
datamodule: focus150.yaml
|
161 |
+
hydra/env: default
|
162 |
+
hydra/callbacks: null
|
163 |
+
hydra/job_logging: colorlog
|
164 |
+
hydra/hydra_logging: colorlog
|
165 |
+
hydra/hydra_help: default
|
166 |
+
hydra/help: default
|
167 |
+
hydra/sweeper: basic
|
168 |
+
hydra/launcher: basic
|
169 |
+
hydra/output: default
|
170 |
+
verbose: false
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-32-57/.hydra/overrides.yaml
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
- experiment=focusConvReLU_MSE_150
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/config.yaml
ADDED
@@ -0,0 +1,81 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
original_work_dir: ${hydra:runtime.cwd}
|
2 |
+
data_dir: ${original_work_dir}/data
|
3 |
+
print_config: true
|
4 |
+
ignore_warnings: true
|
5 |
+
train: true
|
6 |
+
test: true
|
7 |
+
seed: 12345
|
8 |
+
name: focusConvReLU_MSE_150
|
9 |
+
datamodule:
|
10 |
+
_target_: src.datamodules.focus_datamodule.FocusDataModule
|
11 |
+
data_dir: ${data_dir}/focus150
|
12 |
+
csv_train_file: ${data_dir}/focus150/train_metadata.csv
|
13 |
+
csv_val_file: ${data_dir}/focus150/validation_metadata.csv
|
14 |
+
csv_test_file: ${data_dir}/focus150/test_metadata.csv
|
15 |
+
batch_size: 64
|
16 |
+
num_workers: 0
|
17 |
+
pin_memory: false
|
18 |
+
augmentation: true
|
19 |
+
model:
|
20 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
21 |
+
image_size: 150
|
22 |
+
pool_size: 2
|
23 |
+
conv1_size: 3
|
24 |
+
conv1_channels: 9
|
25 |
+
conv2_size: 7
|
26 |
+
conv2_channels: 6
|
27 |
+
lin1_size: 32
|
28 |
+
lin2_size: 72
|
29 |
+
output_size: 1
|
30 |
+
lr: 0.001
|
31 |
+
weight_decay: 0.0005
|
32 |
+
callbacks:
|
33 |
+
model_checkpoint:
|
34 |
+
_target_: pytorch_lightning.callbacks.ModelCheckpoint
|
35 |
+
monitor: val/mae
|
36 |
+
mode: min
|
37 |
+
save_top_k: 1
|
38 |
+
save_last: true
|
39 |
+
verbose: false
|
40 |
+
dirpath: checkpoints/
|
41 |
+
filename: epoch_{epoch:03d}
|
42 |
+
auto_insert_metric_name: false
|
43 |
+
early_stopping:
|
44 |
+
_target_: pytorch_lightning.callbacks.EarlyStopping
|
45 |
+
monitor: val/mae
|
46 |
+
mode: min
|
47 |
+
patience: 100
|
48 |
+
min_delta: 0
|
49 |
+
model_summary:
|
50 |
+
_target_: pytorch_lightning.callbacks.RichModelSummary
|
51 |
+
max_depth: -1
|
52 |
+
rich_progress_bar:
|
53 |
+
_target_: pytorch_lightning.callbacks.RichProgressBar
|
54 |
+
logger:
|
55 |
+
csv:
|
56 |
+
_target_: pytorch_lightning.loggers.csv_logs.CSVLogger
|
57 |
+
save_dir: .
|
58 |
+
name: csv/
|
59 |
+
prefix: ''
|
60 |
+
mlflow:
|
61 |
+
_target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
|
62 |
+
experiment_name: ${name}
|
63 |
+
tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
|
64 |
+
tags: null
|
65 |
+
save_dir: ./mlruns
|
66 |
+
prefix: ''
|
67 |
+
artifact_location: null
|
68 |
+
tensorboard:
|
69 |
+
_target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
|
70 |
+
save_dir: tensorboard/
|
71 |
+
name: null
|
72 |
+
version: ${name}
|
73 |
+
log_graph: false
|
74 |
+
default_hp_metric: true
|
75 |
+
prefix: ''
|
76 |
+
trainer:
|
77 |
+
_target_: pytorch_lightning.Trainer
|
78 |
+
gpus: 1
|
79 |
+
min_epochs: 1
|
80 |
+
max_epochs: 100
|
81 |
+
resume_from_checkpoint: null
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/hydra.yaml
ADDED
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
hydra:
|
2 |
+
run:
|
3 |
+
dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
4 |
+
sweep:
|
5 |
+
dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
6 |
+
subdir: ${hydra.job.num}
|
7 |
+
launcher:
|
8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
9 |
+
sweeper:
|
10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
11 |
+
max_batch_size: null
|
12 |
+
help:
|
13 |
+
app_name: ${hydra.job.name}
|
14 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
15 |
+
|
16 |
+
'
|
17 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
18 |
+
|
19 |
+
Use --hydra-help to view Hydra specific help
|
20 |
+
|
21 |
+
'
|
22 |
+
template: '${hydra.help.header}
|
23 |
+
|
24 |
+
== Configuration groups ==
|
25 |
+
|
26 |
+
Compose your configuration from those groups (group=option)
|
27 |
+
|
28 |
+
|
29 |
+
$APP_CONFIG_GROUPS
|
30 |
+
|
31 |
+
|
32 |
+
== Config ==
|
33 |
+
|
34 |
+
Override anything in the config (foo.bar=value)
|
35 |
+
|
36 |
+
|
37 |
+
$CONFIG
|
38 |
+
|
39 |
+
|
40 |
+
${hydra.help.footer}
|
41 |
+
|
42 |
+
'
|
43 |
+
hydra_help:
|
44 |
+
template: 'Hydra (${hydra.runtime.version})
|
45 |
+
|
46 |
+
See https://hydra.cc for more info.
|
47 |
+
|
48 |
+
|
49 |
+
== Flags ==
|
50 |
+
|
51 |
+
$FLAGS_HELP
|
52 |
+
|
53 |
+
|
54 |
+
== Configuration groups ==
|
55 |
+
|
56 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
57 |
+
to command line)
|
58 |
+
|
59 |
+
|
60 |
+
$HYDRA_CONFIG_GROUPS
|
61 |
+
|
62 |
+
|
63 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
64 |
+
|
65 |
+
'
|
66 |
+
hydra_help: ???
|
67 |
+
hydra_logging:
|
68 |
+
version: 1
|
69 |
+
formatters:
|
70 |
+
colorlog:
|
71 |
+
(): colorlog.ColoredFormatter
|
72 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
73 |
+
handlers:
|
74 |
+
console:
|
75 |
+
class: logging.StreamHandler
|
76 |
+
formatter: colorlog
|
77 |
+
stream: ext://sys.stdout
|
78 |
+
root:
|
79 |
+
level: INFO
|
80 |
+
handlers:
|
81 |
+
- console
|
82 |
+
disable_existing_loggers: false
|
83 |
+
job_logging:
|
84 |
+
version: 1
|
85 |
+
formatters:
|
86 |
+
simple:
|
87 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
88 |
+
colorlog:
|
89 |
+
(): colorlog.ColoredFormatter
|
90 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
91 |
+
- %(message)s'
|
92 |
+
log_colors:
|
93 |
+
DEBUG: purple
|
94 |
+
INFO: green
|
95 |
+
WARNING: yellow
|
96 |
+
ERROR: red
|
97 |
+
CRITICAL: red
|
98 |
+
handlers:
|
99 |
+
console:
|
100 |
+
class: logging.StreamHandler
|
101 |
+
formatter: colorlog
|
102 |
+
stream: ext://sys.stdout
|
103 |
+
file:
|
104 |
+
class: logging.FileHandler
|
105 |
+
formatter: simple
|
106 |
+
filename: ${hydra.job.name}.log
|
107 |
+
root:
|
108 |
+
level: INFO
|
109 |
+
handlers:
|
110 |
+
- console
|
111 |
+
- file
|
112 |
+
disable_existing_loggers: false
|
113 |
+
env: {}
|
114 |
+
searchpath: []
|
115 |
+
callbacks: {}
|
116 |
+
output_subdir: .hydra
|
117 |
+
overrides:
|
118 |
+
hydra: []
|
119 |
+
task:
|
120 |
+
- experiment=focusConvReLU_MSE_150.yaml
|
121 |
+
job:
|
122 |
+
name: train
|
123 |
+
override_dirname: experiment=focusConvReLU_MSE_150.yaml
|
124 |
+
id: ???
|
125 |
+
num: ???
|
126 |
+
config_name: train.yaml
|
127 |
+
env_set: {}
|
128 |
+
env_copy: []
|
129 |
+
config:
|
130 |
+
override_dirname:
|
131 |
+
kv_sep: '='
|
132 |
+
item_sep: ','
|
133 |
+
exclude_keys: []
|
134 |
+
runtime:
|
135 |
+
version: 1.1.2
|
136 |
+
cwd: /usr/src/app
|
137 |
+
config_sources:
|
138 |
+
- path: hydra.conf
|
139 |
+
schema: pkg
|
140 |
+
provider: hydra
|
141 |
+
- path: /usr/src/app/configs
|
142 |
+
schema: file
|
143 |
+
provider: main
|
144 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
145 |
+
schema: pkg
|
146 |
+
provider: hydra-colorlog
|
147 |
+
- path: ''
|
148 |
+
schema: structured
|
149 |
+
provider: schema
|
150 |
+
choices:
|
151 |
+
local: default.yaml
|
152 |
+
hparams_search: null
|
153 |
+
debug: null
|
154 |
+
experiment: focusConvReLU_MSE_150.yaml
|
155 |
+
log_dir: default.yaml
|
156 |
+
trainer: default.yaml
|
157 |
+
logger: many_loggers
|
158 |
+
callbacks: default.yaml
|
159 |
+
model: focusConvReLU_150.yaml
|
160 |
+
datamodule: focus150.yaml
|
161 |
+
hydra/env: default
|
162 |
+
hydra/callbacks: null
|
163 |
+
hydra/job_logging: colorlog
|
164 |
+
hydra/hydra_logging: colorlog
|
165 |
+
hydra/hydra_help: default
|
166 |
+
hydra/help: default
|
167 |
+
hydra/sweeper: basic
|
168 |
+
hydra/launcher: basic
|
169 |
+
hydra/output: default
|
170 |
+
verbose: false
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-33-41/.hydra/overrides.yaml
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
- experiment=focusConvReLU_MSE_150.yaml
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/config.yaml
ADDED
@@ -0,0 +1,81 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
original_work_dir: ${hydra:runtime.cwd}
|
2 |
+
data_dir: ${original_work_dir}/data
|
3 |
+
print_config: true
|
4 |
+
ignore_warnings: true
|
5 |
+
train: true
|
6 |
+
test: true
|
7 |
+
seed: 12345
|
8 |
+
name: focusConvReLU_MSE_150
|
9 |
+
datamodule:
|
10 |
+
_target_: src.datamodules.focus_datamodule.FocusDataModule
|
11 |
+
data_dir: ${data_dir}/focus150
|
12 |
+
csv_train_file: ${data_dir}/focus150/train_metadata.csv
|
13 |
+
csv_val_file: ${data_dir}/focus150/validation_metadata.csv
|
14 |
+
csv_test_file: ${data_dir}/focus150/test_metadata.csv
|
15 |
+
batch_size: 64
|
16 |
+
num_workers: 0
|
17 |
+
pin_memory: false
|
18 |
+
augmentation: true
|
19 |
+
model:
|
20 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
21 |
+
image_size: 150
|
22 |
+
pool_size: 2
|
23 |
+
conv1_size: 3
|
24 |
+
conv1_channels: 9
|
25 |
+
conv2_size: 7
|
26 |
+
conv2_channels: 6
|
27 |
+
lin1_size: 32
|
28 |
+
lin2_size: 72
|
29 |
+
output_size: 1
|
30 |
+
lr: 0.001
|
31 |
+
weight_decay: 0.0005
|
32 |
+
callbacks:
|
33 |
+
model_checkpoint:
|
34 |
+
_target_: pytorch_lightning.callbacks.ModelCheckpoint
|
35 |
+
monitor: val/mae
|
36 |
+
mode: min
|
37 |
+
save_top_k: 1
|
38 |
+
save_last: true
|
39 |
+
verbose: false
|
40 |
+
dirpath: checkpoints/
|
41 |
+
filename: epoch_{epoch:03d}
|
42 |
+
auto_insert_metric_name: false
|
43 |
+
early_stopping:
|
44 |
+
_target_: pytorch_lightning.callbacks.EarlyStopping
|
45 |
+
monitor: val/mae
|
46 |
+
mode: min
|
47 |
+
patience: 100
|
48 |
+
min_delta: 0
|
49 |
+
model_summary:
|
50 |
+
_target_: pytorch_lightning.callbacks.RichModelSummary
|
51 |
+
max_depth: -1
|
52 |
+
rich_progress_bar:
|
53 |
+
_target_: pytorch_lightning.callbacks.RichProgressBar
|
54 |
+
logger:
|
55 |
+
csv:
|
56 |
+
_target_: pytorch_lightning.loggers.csv_logs.CSVLogger
|
57 |
+
save_dir: .
|
58 |
+
name: csv/
|
59 |
+
prefix: ''
|
60 |
+
mlflow:
|
61 |
+
_target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
|
62 |
+
experiment_name: ${name}
|
63 |
+
tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
|
64 |
+
tags: null
|
65 |
+
save_dir: ./mlruns
|
66 |
+
prefix: ''
|
67 |
+
artifact_location: null
|
68 |
+
tensorboard:
|
69 |
+
_target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
|
70 |
+
save_dir: tensorboard/
|
71 |
+
name: null
|
72 |
+
version: ${name}
|
73 |
+
log_graph: false
|
74 |
+
default_hp_metric: true
|
75 |
+
prefix: ''
|
76 |
+
trainer:
|
77 |
+
_target_: pytorch_lightning.Trainer
|
78 |
+
gpus: 1
|
79 |
+
min_epochs: 1
|
80 |
+
max_epochs: 100
|
81 |
+
resume_from_checkpoint: null
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/hydra.yaml
ADDED
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
hydra:
|
2 |
+
run:
|
3 |
+
dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
4 |
+
sweep:
|
5 |
+
dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
6 |
+
subdir: ${hydra.job.num}
|
7 |
+
launcher:
|
8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
9 |
+
sweeper:
|
10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
11 |
+
max_batch_size: null
|
12 |
+
help:
|
13 |
+
app_name: ${hydra.job.name}
|
14 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
15 |
+
|
16 |
+
'
|
17 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
18 |
+
|
19 |
+
Use --hydra-help to view Hydra specific help
|
20 |
+
|
21 |
+
'
|
22 |
+
template: '${hydra.help.header}
|
23 |
+
|
24 |
+
== Configuration groups ==
|
25 |
+
|
26 |
+
Compose your configuration from those groups (group=option)
|
27 |
+
|
28 |
+
|
29 |
+
$APP_CONFIG_GROUPS
|
30 |
+
|
31 |
+
|
32 |
+
== Config ==
|
33 |
+
|
34 |
+
Override anything in the config (foo.bar=value)
|
35 |
+
|
36 |
+
|
37 |
+
$CONFIG
|
38 |
+
|
39 |
+
|
40 |
+
${hydra.help.footer}
|
41 |
+
|
42 |
+
'
|
43 |
+
hydra_help:
|
44 |
+
template: 'Hydra (${hydra.runtime.version})
|
45 |
+
|
46 |
+
See https://hydra.cc for more info.
|
47 |
+
|
48 |
+
|
49 |
+
== Flags ==
|
50 |
+
|
51 |
+
$FLAGS_HELP
|
52 |
+
|
53 |
+
|
54 |
+
== Configuration groups ==
|
55 |
+
|
56 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
57 |
+
to command line)
|
58 |
+
|
59 |
+
|
60 |
+
$HYDRA_CONFIG_GROUPS
|
61 |
+
|
62 |
+
|
63 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
64 |
+
|
65 |
+
'
|
66 |
+
hydra_help: ???
|
67 |
+
hydra_logging:
|
68 |
+
version: 1
|
69 |
+
formatters:
|
70 |
+
colorlog:
|
71 |
+
(): colorlog.ColoredFormatter
|
72 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
73 |
+
handlers:
|
74 |
+
console:
|
75 |
+
class: logging.StreamHandler
|
76 |
+
formatter: colorlog
|
77 |
+
stream: ext://sys.stdout
|
78 |
+
root:
|
79 |
+
level: INFO
|
80 |
+
handlers:
|
81 |
+
- console
|
82 |
+
disable_existing_loggers: false
|
83 |
+
job_logging:
|
84 |
+
version: 1
|
85 |
+
formatters:
|
86 |
+
simple:
|
87 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
88 |
+
colorlog:
|
89 |
+
(): colorlog.ColoredFormatter
|
90 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
91 |
+
- %(message)s'
|
92 |
+
log_colors:
|
93 |
+
DEBUG: purple
|
94 |
+
INFO: green
|
95 |
+
WARNING: yellow
|
96 |
+
ERROR: red
|
97 |
+
CRITICAL: red
|
98 |
+
handlers:
|
99 |
+
console:
|
100 |
+
class: logging.StreamHandler
|
101 |
+
formatter: colorlog
|
102 |
+
stream: ext://sys.stdout
|
103 |
+
file:
|
104 |
+
class: logging.FileHandler
|
105 |
+
formatter: simple
|
106 |
+
filename: ${hydra.job.name}.log
|
107 |
+
root:
|
108 |
+
level: INFO
|
109 |
+
handlers:
|
110 |
+
- console
|
111 |
+
- file
|
112 |
+
disable_existing_loggers: false
|
113 |
+
env: {}
|
114 |
+
searchpath: []
|
115 |
+
callbacks: {}
|
116 |
+
output_subdir: .hydra
|
117 |
+
overrides:
|
118 |
+
hydra: []
|
119 |
+
task:
|
120 |
+
- experiment=focusConvReLU_MSE_150.yaml
|
121 |
+
job:
|
122 |
+
name: train
|
123 |
+
override_dirname: experiment=focusConvReLU_MSE_150.yaml
|
124 |
+
id: ???
|
125 |
+
num: ???
|
126 |
+
config_name: train.yaml
|
127 |
+
env_set: {}
|
128 |
+
env_copy: []
|
129 |
+
config:
|
130 |
+
override_dirname:
|
131 |
+
kv_sep: '='
|
132 |
+
item_sep: ','
|
133 |
+
exclude_keys: []
|
134 |
+
runtime:
|
135 |
+
version: 1.1.2
|
136 |
+
cwd: /usr/src/app
|
137 |
+
config_sources:
|
138 |
+
- path: hydra.conf
|
139 |
+
schema: pkg
|
140 |
+
provider: hydra
|
141 |
+
- path: /usr/src/app/configs
|
142 |
+
schema: file
|
143 |
+
provider: main
|
144 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
145 |
+
schema: pkg
|
146 |
+
provider: hydra-colorlog
|
147 |
+
- path: ''
|
148 |
+
schema: structured
|
149 |
+
provider: schema
|
150 |
+
choices:
|
151 |
+
local: default.yaml
|
152 |
+
hparams_search: null
|
153 |
+
debug: null
|
154 |
+
experiment: focusConvReLU_MSE_150.yaml
|
155 |
+
log_dir: default.yaml
|
156 |
+
trainer: default.yaml
|
157 |
+
logger: many_loggers
|
158 |
+
callbacks: default.yaml
|
159 |
+
model: focusConvReLU_150.yaml
|
160 |
+
datamodule: focus150.yaml
|
161 |
+
hydra/env: default
|
162 |
+
hydra/callbacks: null
|
163 |
+
hydra/job_logging: colorlog
|
164 |
+
hydra/hydra_logging: colorlog
|
165 |
+
hydra/hydra_help: default
|
166 |
+
hydra/help: default
|
167 |
+
hydra/sweeper: basic
|
168 |
+
hydra/launcher: basic
|
169 |
+
hydra/output: default
|
170 |
+
verbose: false
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-34-32/.hydra/overrides.yaml
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
- experiment=focusConvReLU_MSE_150.yaml
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/config.yaml
ADDED
@@ -0,0 +1,81 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
original_work_dir: ${hydra:runtime.cwd}
|
2 |
+
data_dir: ${original_work_dir}/data
|
3 |
+
print_config: true
|
4 |
+
ignore_warnings: true
|
5 |
+
train: true
|
6 |
+
test: true
|
7 |
+
seed: 12345
|
8 |
+
name: focusConvReLU_MSE_150
|
9 |
+
datamodule:
|
10 |
+
_target_: src.datamodules.focus_datamodule.FocusDataModule
|
11 |
+
data_dir: ${data_dir}/focus150
|
12 |
+
csv_train_file: ${data_dir}/focus150/train_metadata.csv
|
13 |
+
csv_val_file: ${data_dir}/focus150/validation_metadata.csv
|
14 |
+
csv_test_file: ${data_dir}/focus150/test_metadata.csv
|
15 |
+
batch_size: 64
|
16 |
+
num_workers: 0
|
17 |
+
pin_memory: false
|
18 |
+
augmentation: true
|
19 |
+
model:
|
20 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
21 |
+
image_size: 150
|
22 |
+
pool_size: 2
|
23 |
+
conv1_size: 3
|
24 |
+
conv1_channels: 9
|
25 |
+
conv2_size: 7
|
26 |
+
conv2_channels: 6
|
27 |
+
lin1_size: 32
|
28 |
+
lin2_size: 72
|
29 |
+
output_size: 1
|
30 |
+
lr: 0.001
|
31 |
+
weight_decay: 0.0005
|
32 |
+
callbacks:
|
33 |
+
model_checkpoint:
|
34 |
+
_target_: pytorch_lightning.callbacks.ModelCheckpoint
|
35 |
+
monitor: val/mae
|
36 |
+
mode: min
|
37 |
+
save_top_k: 1
|
38 |
+
save_last: true
|
39 |
+
verbose: false
|
40 |
+
dirpath: checkpoints/
|
41 |
+
filename: epoch_{epoch:03d}
|
42 |
+
auto_insert_metric_name: false
|
43 |
+
early_stopping:
|
44 |
+
_target_: pytorch_lightning.callbacks.EarlyStopping
|
45 |
+
monitor: val/mae
|
46 |
+
mode: min
|
47 |
+
patience: 100
|
48 |
+
min_delta: 0
|
49 |
+
model_summary:
|
50 |
+
_target_: pytorch_lightning.callbacks.RichModelSummary
|
51 |
+
max_depth: -1
|
52 |
+
rich_progress_bar:
|
53 |
+
_target_: pytorch_lightning.callbacks.RichProgressBar
|
54 |
+
logger:
|
55 |
+
csv:
|
56 |
+
_target_: pytorch_lightning.loggers.csv_logs.CSVLogger
|
57 |
+
save_dir: .
|
58 |
+
name: csv/
|
59 |
+
prefix: ''
|
60 |
+
mlflow:
|
61 |
+
_target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
|
62 |
+
experiment_name: ${name}
|
63 |
+
tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
|
64 |
+
tags: null
|
65 |
+
save_dir: ./mlruns
|
66 |
+
prefix: ''
|
67 |
+
artifact_location: null
|
68 |
+
tensorboard:
|
69 |
+
_target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
|
70 |
+
save_dir: tensorboard/
|
71 |
+
name: null
|
72 |
+
version: ${name}
|
73 |
+
log_graph: false
|
74 |
+
default_hp_metric: true
|
75 |
+
prefix: ''
|
76 |
+
trainer:
|
77 |
+
_target_: pytorch_lightning.Trainer
|
78 |
+
gpus: 1
|
79 |
+
min_epochs: 1
|
80 |
+
max_epochs: 100
|
81 |
+
resume_from_checkpoint: null
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/hydra.yaml
ADDED
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
hydra:
|
2 |
+
run:
|
3 |
+
dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
4 |
+
sweep:
|
5 |
+
dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
|
6 |
+
subdir: ${hydra.job.num}
|
7 |
+
launcher:
|
8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
9 |
+
sweeper:
|
10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
11 |
+
max_batch_size: null
|
12 |
+
help:
|
13 |
+
app_name: ${hydra.job.name}
|
14 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
15 |
+
|
16 |
+
'
|
17 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
18 |
+
|
19 |
+
Use --hydra-help to view Hydra specific help
|
20 |
+
|
21 |
+
'
|
22 |
+
template: '${hydra.help.header}
|
23 |
+
|
24 |
+
== Configuration groups ==
|
25 |
+
|
26 |
+
Compose your configuration from those groups (group=option)
|
27 |
+
|
28 |
+
|
29 |
+
$APP_CONFIG_GROUPS
|
30 |
+
|
31 |
+
|
32 |
+
== Config ==
|
33 |
+
|
34 |
+
Override anything in the config (foo.bar=value)
|
35 |
+
|
36 |
+
|
37 |
+
$CONFIG
|
38 |
+
|
39 |
+
|
40 |
+
${hydra.help.footer}
|
41 |
+
|
42 |
+
'
|
43 |
+
hydra_help:
|
44 |
+
template: 'Hydra (${hydra.runtime.version})
|
45 |
+
|
46 |
+
See https://hydra.cc for more info.
|
47 |
+
|
48 |
+
|
49 |
+
== Flags ==
|
50 |
+
|
51 |
+
$FLAGS_HELP
|
52 |
+
|
53 |
+
|
54 |
+
== Configuration groups ==
|
55 |
+
|
56 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
57 |
+
to command line)
|
58 |
+
|
59 |
+
|
60 |
+
$HYDRA_CONFIG_GROUPS
|
61 |
+
|
62 |
+
|
63 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
64 |
+
|
65 |
+
'
|
66 |
+
hydra_help: ???
|
67 |
+
hydra_logging:
|
68 |
+
version: 1
|
69 |
+
formatters:
|
70 |
+
colorlog:
|
71 |
+
(): colorlog.ColoredFormatter
|
72 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
73 |
+
handlers:
|
74 |
+
console:
|
75 |
+
class: logging.StreamHandler
|
76 |
+
formatter: colorlog
|
77 |
+
stream: ext://sys.stdout
|
78 |
+
root:
|
79 |
+
level: INFO
|
80 |
+
handlers:
|
81 |
+
- console
|
82 |
+
disable_existing_loggers: false
|
83 |
+
job_logging:
|
84 |
+
version: 1
|
85 |
+
formatters:
|
86 |
+
simple:
|
87 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
88 |
+
colorlog:
|
89 |
+
(): colorlog.ColoredFormatter
|
90 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
91 |
+
- %(message)s'
|
92 |
+
log_colors:
|
93 |
+
DEBUG: purple
|
94 |
+
INFO: green
|
95 |
+
WARNING: yellow
|
96 |
+
ERROR: red
|
97 |
+
CRITICAL: red
|
98 |
+
handlers:
|
99 |
+
console:
|
100 |
+
class: logging.StreamHandler
|
101 |
+
formatter: colorlog
|
102 |
+
stream: ext://sys.stdout
|
103 |
+
file:
|
104 |
+
class: logging.FileHandler
|
105 |
+
formatter: simple
|
106 |
+
filename: ${hydra.job.name}.log
|
107 |
+
root:
|
108 |
+
level: INFO
|
109 |
+
handlers:
|
110 |
+
- console
|
111 |
+
- file
|
112 |
+
disable_existing_loggers: false
|
113 |
+
env: {}
|
114 |
+
searchpath: []
|
115 |
+
callbacks: {}
|
116 |
+
output_subdir: .hydra
|
117 |
+
overrides:
|
118 |
+
hydra: []
|
119 |
+
task:
|
120 |
+
- experiment=focusConvReLU_MSE_150.yaml
|
121 |
+
job:
|
122 |
+
name: train
|
123 |
+
override_dirname: experiment=focusConvReLU_MSE_150.yaml
|
124 |
+
id: ???
|
125 |
+
num: ???
|
126 |
+
config_name: train.yaml
|
127 |
+
env_set: {}
|
128 |
+
env_copy: []
|
129 |
+
config:
|
130 |
+
override_dirname:
|
131 |
+
kv_sep: '='
|
132 |
+
item_sep: ','
|
133 |
+
exclude_keys: []
|
134 |
+
runtime:
|
135 |
+
version: 1.1.2
|
136 |
+
cwd: /usr/src/app
|
137 |
+
config_sources:
|
138 |
+
- path: hydra.conf
|
139 |
+
schema: pkg
|
140 |
+
provider: hydra
|
141 |
+
- path: /usr/src/app/configs
|
142 |
+
schema: file
|
143 |
+
provider: main
|
144 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
145 |
+
schema: pkg
|
146 |
+
provider: hydra-colorlog
|
147 |
+
- path: ''
|
148 |
+
schema: structured
|
149 |
+
provider: schema
|
150 |
+
choices:
|
151 |
+
local: default.yaml
|
152 |
+
hparams_search: null
|
153 |
+
debug: null
|
154 |
+
experiment: focusConvReLU_MSE_150.yaml
|
155 |
+
log_dir: default.yaml
|
156 |
+
trainer: default.yaml
|
157 |
+
logger: many_loggers
|
158 |
+
callbacks: default.yaml
|
159 |
+
model: focusConvReLU_150.yaml
|
160 |
+
datamodule: focus150.yaml
|
161 |
+
hydra/env: default
|
162 |
+
hydra/callbacks: null
|
163 |
+
hydra/job_logging: colorlog
|
164 |
+
hydra/hydra_logging: colorlog
|
165 |
+
hydra/hydra_help: default
|
166 |
+
hydra/help: default
|
167 |
+
hydra/sweeper: basic
|
168 |
+
hydra/launcher: basic
|
169 |
+
hydra/output: default
|
170 |
+
verbose: false
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/.hydra/overrides.yaml
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
- experiment=focusConvReLU_MSE_150.yaml
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/epoch_096.ckpt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3e9159964f903b3063b0a2170eb2baa7d138b219689f46620567a0b821cbf98
|
3 |
+
size 2612879
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/checkpoints/last.ckpt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3d046d8e08e9ab7ff46290261d0f10a025a04c4c43ff48de76df4a91d77ee0f6
|
3 |
+
size 2612879
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/hparams.yaml
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
trainer:
|
2 |
+
_target_: pytorch_lightning.Trainer
|
3 |
+
gpus: 1
|
4 |
+
min_epochs: 1
|
5 |
+
max_epochs: 100
|
6 |
+
resume_from_checkpoint: null
|
7 |
+
model:
|
8 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
9 |
+
image_size: 150
|
10 |
+
pool_size: 2
|
11 |
+
conv1_size: 3
|
12 |
+
conv1_channels: 9
|
13 |
+
conv2_size: 7
|
14 |
+
conv2_channels: 6
|
15 |
+
lin1_size: 32
|
16 |
+
lin2_size: 72
|
17 |
+
output_size: 1
|
18 |
+
lr: 0.001
|
19 |
+
weight_decay: 0.0005
|
20 |
+
datamodule:
|
21 |
+
_target_: src.datamodules.focus_datamodule.FocusDataModule
|
22 |
+
data_dir: /usr/src/app/data/focus150
|
23 |
+
csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
|
24 |
+
csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
|
25 |
+
csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
|
26 |
+
batch_size: 64
|
27 |
+
num_workers: 0
|
28 |
+
pin_memory: false
|
29 |
+
augmentation: true
|
30 |
+
seed: 12345
|
31 |
+
callbacks:
|
32 |
+
model_checkpoint:
|
33 |
+
_target_: pytorch_lightning.callbacks.ModelCheckpoint
|
34 |
+
monitor: val/mae
|
35 |
+
mode: min
|
36 |
+
save_top_k: 1
|
37 |
+
save_last: true
|
38 |
+
verbose: false
|
39 |
+
dirpath: checkpoints/
|
40 |
+
filename: epoch_{epoch:03d}
|
41 |
+
auto_insert_metric_name: false
|
42 |
+
early_stopping:
|
43 |
+
_target_: pytorch_lightning.callbacks.EarlyStopping
|
44 |
+
monitor: val/mae
|
45 |
+
mode: min
|
46 |
+
patience: 100
|
47 |
+
min_delta: 0
|
48 |
+
model_summary:
|
49 |
+
_target_: pytorch_lightning.callbacks.RichModelSummary
|
50 |
+
max_depth: -1
|
51 |
+
rich_progress_bar:
|
52 |
+
_target_: pytorch_lightning.callbacks.RichProgressBar
|
53 |
+
model/params/total: 216231
|
54 |
+
model/params/trainable: 216231
|
55 |
+
model/params/non_trainable: 0
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/csv/version_0/metrics.csv
ADDED
@@ -0,0 +1,202 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
|
2 |
+
0.0010634043719619513,0.02607358619570732,0.02607358619570732,0,47,,,,
|
3 |
+
,,,1,47,0.015712032094597816,0.07011217623949051,,
|
4 |
+
0.0008477172814309597,0.023356162011623383,0.023356162011623383,1,95,,,,
|
5 |
+
,,,2,95,0.0006998287281021476,0.021280117332935333,,
|
6 |
+
0.0006586583331227303,0.02111729420721531,0.02111729420721531,2,143,,,,
|
7 |
+
,,,3,143,0.0006262350943870842,0.020297976210713387,,
|
8 |
+
0.0006095922435633838,0.020395344123244286,0.020395344123244286,3,191,,,,
|
9 |
+
,,,4,191,0.0006112757255323231,0.020039746537804604,,
|
10 |
+
0.000579743820708245,0.02008357085287571,0.02008357085287571,4,239,,,,
|
11 |
+
,,,5,239,0.0005892604240216315,0.019787773489952087,,
|
12 |
+
0.000539788044989109,0.019109351560473442,0.019109351560473442,5,287,,,,
|
13 |
+
,,,6,287,0.0004991526948288083,0.018337024375796318,,
|
14 |
+
0.0005701080663129687,0.020062454044818878,0.019109351560473442,6,335,,,,
|
15 |
+
,,,7,335,0.00046538031892850995,0.017815634608268738,,
|
16 |
+
0.0005549467750824988,0.018724771216511726,0.018724771216511726,7,383,,,,
|
17 |
+
,,,8,383,0.0004903865046799183,0.017987234517931938,,
|
18 |
+
0.0005555949755944312,0.01976374350488186,0.018724771216511726,8,431,,,,
|
19 |
+
,,,9,431,0.0004449250118341297,0.017149975523352623,,
|
20 |
+
0.0004789295489899814,0.018088672310113907,0.018088672310113907,9,479,,,,
|
21 |
+
,,,10,479,0.0004655240918509662,0.01764928176999092,,
|
22 |
+
0.0005822633975185454,0.02009824477136135,0.018088672310113907,10,527,,,,
|
23 |
+
,,,11,527,0.00044383766362443566,0.017262466251850128,,
|
24 |
+
0.00048447673907503486,0.017433589324355125,0.017433589324355125,11,575,,,,
|
25 |
+
,,,12,575,0.0004356526769697666,0.017195068299770355,,
|
26 |
+
0.00046181678771972656,0.018075214698910713,0.017433589324355125,12,623,,,,
|
27 |
+
,,,13,623,0.00041521640378050506,0.01676551066339016,,
|
28 |
+
0.00041472347220405936,0.016927972435951233,0.016927972435951233,13,671,,,,
|
29 |
+
,,,14,671,0.0003966042131651193,0.016418082639575005,,
|
30 |
+
0.0005150702781975269,0.01795283891260624,0.016927972435951233,14,719,,,,
|
31 |
+
,,,15,719,0.0004069829301442951,0.016572223976254463,,
|
32 |
+
0.00043978836038149893,0.017694905400276184,0.016927972435951233,15,767,,,,
|
33 |
+
,,,16,767,0.0004001975175924599,0.016431566327810287,,
|
34 |
+
0.0004524872056208551,0.017309539020061493,0.016927972435951233,16,815,,,,
|
35 |
+
,,,17,815,0.000411732355132699,0.016624923795461655,,
|
36 |
+
0.0004895858000963926,0.017853353172540665,0.016927972435951233,17,863,,,,
|
37 |
+
,,,18,863,0.0003962284536100924,0.016285978257656097,,
|
38 |
+
0.00048630539095029235,0.017320722341537476,0.016927972435951233,18,911,,,,
|
39 |
+
,,,19,911,0.0003755573998205364,0.015845887362957,,
|
40 |
+
0.00045636019785888493,0.017117636278271675,0.016927972435951233,19,959,,,,
|
41 |
+
,,,20,959,0.00040089385583996773,0.016440516337752342,,
|
42 |
+
0.00042457759263925254,0.01685478165745735,0.01685478165745735,20,1007,,,,
|
43 |
+
,,,21,1007,0.00037927881930954754,0.015909109264612198,,
|
44 |
+
0.0005504836444742978,0.01946953497827053,0.01685478165745735,21,1055,,,,
|
45 |
+
,,,22,1055,0.00037079042522236705,0.015598192811012268,,
|
46 |
+
0.0005828228895552456,0.019640492275357246,0.01685478165745735,22,1103,,,,
|
47 |
+
,,,23,1103,0.0003636741894297302,0.015478463843464851,,
|
48 |
+
0.000536997162271291,0.01877320185303688,0.01685478165745735,23,1151,,,,
|
49 |
+
,,,24,1151,0.0003473551769275218,0.014973259530961514,,
|
50 |
+
0.0004296917177271098,0.016905883327126503,0.01685478165745735,24,1199,,,,
|
51 |
+
,,,25,1199,0.0003502711479086429,0.014955243095755577,,
|
52 |
+
0.000369079876691103,0.015234655700623989,0.015234655700623989,25,1247,,,,
|
53 |
+
,,,26,1247,0.0003374406951479614,0.014734053052961826,,
|
54 |
+
0.0005687376251444221,0.019466156139969826,0.015234655700623989,26,1295,,,,
|
55 |
+
,,,27,1295,0.0003077377623412758,0.013931847177445889,,
|
56 |
+
0.0003877778071910143,0.01577661745250225,0.015234655700623989,27,1343,,,,
|
57 |
+
,,,28,1343,0.00030867080204188824,0.013890136033296585,,
|
58 |
+
0.0007391179096885026,0.021662525832653046,0.015234655700623989,28,1391,,,,
|
59 |
+
,,,29,1391,0.0003139588516205549,0.013945831917226315,,
|
60 |
+
0.0002891223703045398,0.013201462104916573,0.013201462104916573,29,1439,,,,
|
61 |
+
,,,30,1439,0.00029595510568469763,0.013475497253239155,,
|
62 |
+
0.00136762042529881,0.03221268951892853,0.013201462104916573,30,1487,,,,
|
63 |
+
,,,31,1487,0.0002615793200675398,0.012566673569381237,,
|
64 |
+
0.0005622635362669826,0.018861766904592514,0.013201462104916573,31,1535,,,,
|
65 |
+
,,,32,1535,0.0002453128108754754,0.012109467759728432,,
|
66 |
+
0.0002613224496599287,0.012555716559290886,0.012555716559290886,32,1583,,,,
|
67 |
+
,,,33,1583,0.0002475597430020571,0.01209949143230915,,
|
68 |
+
0.00026621491997502744,0.01243590097874403,0.01243590097874403,33,1631,,,,
|
69 |
+
,,,34,1631,0.00021561248286161572,0.011168932542204857,,
|
70 |
+
0.0003965417272411287,0.015400874428451061,0.01243590097874403,34,1679,,,,
|
71 |
+
,,,35,1679,0.0002476295630913228,0.012050420045852661,,
|
72 |
+
0.0006301697576418519,0.019847549498081207,0.01243590097874403,35,1727,,,,
|
73 |
+
,,,36,1727,0.00022967555560171604,0.011594988405704498,,
|
74 |
+
0.0004612059856299311,0.01633538119494915,0.01243590097874403,36,1775,,,,
|
75 |
+
,,,37,1775,0.0002333874726900831,0.011586652137339115,,
|
76 |
+
0.00025420173187740147,0.012520545162260532,0.01243590097874403,37,1823,,,,
|
77 |
+
,,,38,1823,0.00023667479399591684,0.01181836612522602,,
|
78 |
+
0.0003383368020877242,0.015009582974016666,0.01243590097874403,38,1871,,,,
|
79 |
+
,,,39,1871,0.0001987472496693954,0.010663148015737534,,
|
80 |
+
0.0003864422906190157,0.016114825382828712,0.01243590097874403,39,1919,,,,
|
81 |
+
,,,40,1919,0.00019275647355243564,0.010579877533018589,,
|
82 |
+
0.00022850626555737108,0.011395555920898914,0.011395555920898914,40,1967,,,,
|
83 |
+
,,,41,1967,0.00020515960932243615,0.010919580236077309,,
|
84 |
+
0.00036217307206243277,0.015877339988946915,0.011395555920898914,41,2015,,,,
|
85 |
+
,,,42,2015,0.00020770113042090088,0.010937352664768696,,
|
86 |
+
0.00026963415439240634,0.012874774634838104,0.011395555920898914,42,2063,,,,
|
87 |
+
,,,43,2063,0.00020995769591536373,0.010980397462844849,,
|
88 |
+
0.0003081822069361806,0.01374065037816763,0.011395555920898914,43,2111,,,,
|
89 |
+
,,,44,2111,0.0001873430737759918,0.010383875109255314,,
|
90 |
+
0.0008588534547016025,0.02455970272421837,0.011395555920898914,44,2159,,,,
|
91 |
+
,,,45,2159,0.0001887517428258434,0.010248820297420025,,
|
92 |
+
0.00025588576681911945,0.012964467518031597,0.011395555920898914,45,2207,,,,
|
93 |
+
,,,46,2207,0.00017797888722270727,0.009944379329681396,,
|
94 |
+
0.0012111393734812737,0.030586877837777138,0.011395555920898914,46,2255,,,,
|
95 |
+
,,,47,2255,0.00017462909454479814,0.009952726773917675,,
|
96 |
+
0.0005991487414576113,0.01970861665904522,0.011395555920898914,47,2303,,,,
|
97 |
+
,,,48,2303,0.00018134394485969096,0.010111919604241848,,
|
98 |
+
0.00027215175214223564,0.012582853436470032,0.011395555920898914,48,2351,,,,
|
99 |
+
,,,49,2351,0.0002078687393805012,0.010890776291489601,,
|
100 |
+
0.0002326530375285074,0.011895284987986088,0.011395555920898914,49,2399,,,,
|
101 |
+
,,,50,2399,0.00016190602036658674,0.009552615694701672,,
|
102 |
+
0.00034161240910179913,0.014539946801960468,0.011395555920898914,50,2447,,,,
|
103 |
+
,,,51,2447,0.0001574503257870674,0.009343606419861317,,
|
104 |
+
0.0002047311281785369,0.010513107292354107,0.010513107292354107,51,2495,,,,
|
105 |
+
,,,52,2495,0.00017259498417843133,0.00977871473878622,,
|
106 |
+
0.0002189291553804651,0.011585352011024952,0.010513107292354107,52,2543,,,,
|
107 |
+
,,,53,2543,0.00016215378127526492,0.009527333080768585,,
|
108 |
+
0.0008305492228828371,0.025230778381228447,0.010513107292354107,53,2591,,,,
|
109 |
+
,,,54,2591,0.00017192678933497518,0.009975402615964413,,
|
110 |
+
0.0003245118132326752,0.01448488887399435,0.010513107292354107,54,2639,,,,
|
111 |
+
,,,55,2639,0.00014025233394932002,0.008768484927713871,,
|
112 |
+
0.00043747638119384646,0.016850745305418968,0.010513107292354107,55,2687,,,,
|
113 |
+
,,,56,2687,0.00014388358977157623,0.008977909572422504,,
|
114 |
+
0.0003352710627950728,0.014916492626070976,0.010513107292354107,56,2735,,,,
|
115 |
+
,,,57,2735,0.00013614425552077591,0.008777343668043613,,
|
116 |
+
0.00017761132039595395,0.009772807359695435,0.009772807359695435,57,2783,,,,
|
117 |
+
,,,58,2783,0.00014166149776428938,0.008911573328077793,,
|
118 |
+
0.0002451064356137067,0.012413904070854187,0.009772807359695435,58,2831,,,,
|
119 |
+
,,,59,2831,0.00014307026867754757,0.008944649249315262,,
|
120 |
+
0.00045808948925696313,0.017842639237642288,0.009772807359695435,59,2879,,,,
|
121 |
+
,,,60,2879,0.00012790328764822334,0.008465489372611046,,
|
122 |
+
0.00016568403225392103,0.009426051750779152,0.009426051750779152,60,2927,,,,
|
123 |
+
,,,61,2927,0.00013250892516225576,0.008481234312057495,,
|
124 |
+
0.0002005849382840097,0.011247269809246063,0.009426051750779152,61,2975,,,,
|
125 |
+
,,,62,2975,0.00012225286627653986,0.008217735216021538,,
|
126 |
+
0.0004337916907388717,0.017680322751402855,0.009426051750779152,62,3023,,,,
|
127 |
+
,,,63,3023,0.00013184384442865849,0.008489103056490421,,
|
128 |
+
0.00043202925007790327,0.016938434913754463,0.009426051750779152,63,3071,,,,
|
129 |
+
,,,64,3071,0.0001139832820626907,0.00794004462659359,,
|
130 |
+
0.0006070298259146512,0.020868053659796715,0.009426051750779152,64,3119,,,,
|
131 |
+
,,,65,3119,0.00012551495456136763,0.00839218683540821,,
|
132 |
+
0.00125686835963279,0.03198228031396866,0.009426051750779152,65,3167,,,,
|
133 |
+
,,,66,3167,0.00013683686847798526,0.008686883375048637,,
|
134 |
+
0.0003458949795458466,0.01481217984110117,0.009426051750779152,66,3215,,,,
|
135 |
+
,,,67,3215,0.0001252585934707895,0.008334343321621418,,
|
136 |
+
0.00023582404537592083,0.01217477023601532,0.009426051750779152,67,3263,,,,
|
137 |
+
,,,68,3263,0.0001234338415088132,0.008174673654139042,,
|
138 |
+
0.00043027440551668406,0.0175019484013319,0.009426051750779152,68,3311,,,,
|
139 |
+
,,,69,3311,0.0001386790390824899,0.008791987784206867,,
|
140 |
+
0.00018619702314026654,0.01059352234005928,0.009426051750779152,69,3359,,,,
|
141 |
+
,,,70,3359,0.000133068097056821,0.008551678620278835,,
|
142 |
+
0.0004644407599698752,0.017586909234523773,0.009426051750779152,70,3407,,,,
|
143 |
+
,,,71,3407,0.0001232567592523992,0.008277542889118195,,
|
144 |
+
0.00025989525602199137,0.013158340007066727,0.009426051750779152,71,3455,,,,
|
145 |
+
,,,72,3455,0.00010765787737909704,0.007683832198381424,,
|
146 |
+
0.000271034223260358,0.013016403652727604,0.009426051750779152,72,3503,,,,
|
147 |
+
,,,73,3503,0.00010420197213534266,0.007514127530157566,,
|
148 |
+
0.00015595319564454257,0.00941098015755415,0.00941098015755415,73,3551,,,,
|
149 |
+
,,,74,3551,0.00011743942741304636,0.008030056953430176,,
|
150 |
+
0.00018362083937972784,0.010173797607421875,0.00941098015755415,74,3599,,,,
|
151 |
+
,,,75,3599,0.00013075751485303044,0.008549727499485016,,
|
152 |
+
0.0001723619207041338,0.010034985840320587,0.00941098015755415,75,3647,,,,
|
153 |
+
,,,76,3647,0.00012383297143969685,0.008334793150424957,,
|
154 |
+
0.00026339536998420954,0.01230852585285902,0.00941098015755415,76,3695,,,,
|
155 |
+
,,,77,3695,0.00012997408339288086,0.00849801953881979,,
|
156 |
+
0.0022478969767689705,0.043913356959819794,0.00941098015755415,77,3743,,,,
|
157 |
+
,,,78,3743,0.0001171917756437324,0.00794607400894165,,
|
158 |
+
0.00043688094592653215,0.01787818782031536,0.00941098015755415,78,3791,,,,
|
159 |
+
,,,79,3791,0.0001384491624776274,0.008685767650604248,,
|
160 |
+
0.000253666948992759,0.012860489077866077,0.00941098015755415,79,3839,,,,
|
161 |
+
,,,80,3839,0.00011071155313402414,0.007812082301825285,,
|
162 |
+
0.0007027353858575225,0.02276649698615074,0.00941098015755415,80,3887,,,,
|
163 |
+
,,,81,3887,0.00012242942466400564,0.008264751173555851,,
|
164 |
+
0.00036985089536756277,0.015199720859527588,0.00941098015755415,81,3935,,,,
|
165 |
+
,,,82,3935,0.00012077093560947105,0.008274727500975132,,
|
166 |
+
0.00042630801908671856,0.016393927857279778,0.00941098015755415,82,3983,,,,
|
167 |
+
,,,83,3983,0.00011708390229614452,0.008091365918517113,,
|
168 |
+
0.00018512520182412118,0.010693005286157131,0.00941098015755415,83,4031,,,,
|
169 |
+
,,,84,4031,0.00013511900033336133,0.008610659278929234,,
|
170 |
+
0.0006041005253791809,0.02147362008690834,0.00941098015755415,84,4079,,,,
|
171 |
+
,,,85,4079,0.00010951827425742522,0.007766522467136383,,
|
172 |
+
0.0005739088519476354,0.019695602357387543,0.00941098015755415,85,4127,,,,
|
173 |
+
,,,86,4127,0.00011793604062404484,0.008045573718845844,,
|
174 |
+
0.00020085058349650353,0.010959284380078316,0.00941098015755415,86,4175,,,,
|
175 |
+
,,,87,4175,0.00011460384848760441,0.007821462117135525,,
|
176 |
+
0.00035394675796851516,0.014944094233214855,0.00941098015755415,87,4223,,,,
|
177 |
+
,,,88,4223,0.00012705886911135167,0.008322164416313171,,
|
178 |
+
0.001467306399717927,0.03422271087765694,0.00941098015755415,88,4271,,,,
|
179 |
+
,,,89,4271,0.00011695512512233108,0.008126244880259037,,
|
180 |
+
0.0003908797516487539,0.015736214816570282,0.00941098015755415,89,4319,,,,
|
181 |
+
,,,90,4319,0.00011259406892349944,0.007857552729547024,,
|
182 |
+
0.0001805946958484128,0.009996251203119755,0.00941098015755415,90,4367,,,,
|
183 |
+
,,,91,4367,0.00011745867959689349,0.008035344071686268,,
|
184 |
+
0.00035339672467671335,0.015099494718015194,0.00941098015755415,91,4415,,,,
|
185 |
+
,,,92,4415,0.00013093209418002516,0.008491233922541142,,
|
186 |
+
0.00040678223012946546,0.015847915783524513,0.00941098015755415,92,4463,,,,
|
187 |
+
,,,93,4463,0.00013953317829873413,0.008733570575714111,,
|
188 |
+
0.00018399060354568064,0.010453518480062485,0.00941098015755415,93,4511,,,,
|
189 |
+
,,,94,4511,0.00012743366823997349,0.008426659740507603,,
|
190 |
+
0.0005399276851676404,0.018864858895540237,0.00941098015755415,94,4559,,,,
|
191 |
+
,,,95,4559,0.00011808565614046529,0.008046725764870644,,
|
192 |
+
0.00029979468672536314,0.01469734963029623,0.00941098015755415,95,4607,,,,
|
193 |
+
,,,96,4607,0.00015584049106109887,0.009322610683739185,,
|
194 |
+
0.00013096585462335497,0.00869279820472002,0.00869279820472002,96,4655,,,,
|
195 |
+
,,,97,4655,0.00013836825382895768,0.008852057158946991,,
|
196 |
+
0.00021977233700454235,0.011555126868188381,0.00869279820472002,97,4703,,,,
|
197 |
+
,,,98,4703,0.00012112926197005436,0.008111957460641861,,
|
198 |
+
0.00024704227689653635,0.01307331956923008,0.00869279820472002,98,4751,,,,
|
199 |
+
,,,99,4751,0.00012230231368448585,0.008170234970748425,,
|
200 |
+
0.0003437100094743073,0.01433614082634449,0.00869279820472002,99,4799,,,,
|
201 |
+
,,,100,4799,0.00011741576599888504,0.008008270524442196,,
|
202 |
+
,,,96,4800,,,0.00014260406896937639,0.008947640657424927
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652279809.c49d1a46d86b.1.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c87544f737c7a20d4e44d813742d70c0b9acd226f353912b880bbdbf298495a
|
3 |
+
size 36870
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/events.out.tfevents.1652280195.c49d1a46d86b.1.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7bdc9d2c64d369ea43ddafc00d36574863f5c53b08e1a856563afc76f942d11c
|
3 |
+
size 179
|
logs/experiments/runs/focusConvReLU_MSE_150/2022-05-11_14-36-47/tensorboard/focusConvReLU_MSE_150/hparams.yaml
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
trainer:
|
2 |
+
_target_: pytorch_lightning.Trainer
|
3 |
+
gpus: 1
|
4 |
+
min_epochs: 1
|
5 |
+
max_epochs: 100
|
6 |
+
resume_from_checkpoint: null
|
7 |
+
model:
|
8 |
+
_target_: src.models.focus_conv_module.FocusConvReLULitModule
|
9 |
+
image_size: 150
|
10 |
+
pool_size: 2
|
11 |
+
conv1_size: 3
|
12 |
+
conv1_channels: 9
|
13 |
+
conv2_size: 7
|
14 |
+
conv2_channels: 6
|
15 |
+
lin1_size: 32
|
16 |
+
lin2_size: 72
|
17 |
+
output_size: 1
|
18 |
+
lr: 0.001
|
19 |
+
weight_decay: 0.0005
|
20 |
+
datamodule:
|
21 |
+
_target_: src.datamodules.focus_datamodule.FocusDataModule
|
22 |
+
data_dir: /usr/src/app/data/focus150
|
23 |
+
csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
|
24 |
+
csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
|
25 |
+
csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
|
26 |
+
batch_size: 64
|
27 |
+
num_workers: 0
|
28 |
+
pin_memory: false
|
29 |
+
augmentation: true
|
30 |
+
seed: 12345
|
31 |
+
callbacks:
|
32 |
+
model_checkpoint:
|
33 |
+
_target_: pytorch_lightning.callbacks.ModelCheckpoint
|
34 |
+
monitor: val/mae
|
35 |
+
mode: min
|
36 |
+
save_top_k: 1
|
37 |
+
save_last: true
|
38 |
+
verbose: false
|
39 |
+
dirpath: checkpoints/
|
40 |
+
filename: epoch_{epoch:03d}
|
41 |
+
auto_insert_metric_name: false
|
42 |
+
early_stopping:
|
43 |
+
_target_: pytorch_lightning.callbacks.EarlyStopping
|
44 |
+
monitor: val/mae
|
45 |
+
mode: min
|
46 |
+
patience: 100
|
47 |
+
min_delta: 0
|
48 |
+
model_summary:
|
49 |
+
_target_: pytorch_lightning.callbacks.RichModelSummary
|
50 |
+
max_depth: -1
|
51 |
+
rich_progress_bar:
|
52 |
+
_target_: pytorch_lightning.callbacks.RichProgressBar
|
53 |
+
model/params/total: 216231
|
54 |
+
model/params/trainable: 216231
|
55 |
+
model/params/non_trainable: 0
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/meta.yaml
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
artifact_uri: /usr/src/app/logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/artifacts
|
2 |
+
end_time: 1652280195396
|
3 |
+
entry_point_name: ''
|
4 |
+
experiment_id: '11'
|
5 |
+
lifecycle_stage: active
|
6 |
+
name: ''
|
7 |
+
run_id: fc9d4c9c610e4ae9a51aeec7e544684b
|
8 |
+
run_uuid: fc9d4c9c610e4ae9a51aeec7e544684b
|
9 |
+
source_name: ''
|
10 |
+
source_type: 4
|
11 |
+
source_version: ''
|
12 |
+
start_time: 1652279809740
|
13 |
+
status: 3
|
14 |
+
tags: []
|
15 |
+
user_id: unknown
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/epoch
ADDED
@@ -0,0 +1,201 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1652279820704 0 47
|
2 |
+
1652279820749 1 47
|
3 |
+
1652279824357 1 95
|
4 |
+
1652279824405 2 95
|
5 |
+
1652279828044 2 143
|
6 |
+
1652279828096 3 143
|
7 |
+
1652279831676 3 191
|
8 |
+
1652279831720 4 191
|
9 |
+
1652279835446 4 239
|
10 |
+
1652279835495 5 239
|
11 |
+
1652279839271 5 287
|
12 |
+
1652279839319 6 287
|
13 |
+
1652279843079 6 335
|
14 |
+
1652279843119 7 335
|
15 |
+
1652279846830 7 383
|
16 |
+
1652279846878 8 383
|
17 |
+
1652279850614 8 431
|
18 |
+
1652279850653 9 431
|
19 |
+
1652279854390 9 479
|
20 |
+
1652279854434 10 479
|
21 |
+
1652279858241 10 527
|
22 |
+
1652279858281 11 527
|
23 |
+
1652279862079 11 575
|
24 |
+
1652279862126 12 575
|
25 |
+
1652279865967 12 623
|
26 |
+
1652279866012 13 623
|
27 |
+
1652279869769 13 671
|
28 |
+
1652279869817 14 671
|
29 |
+
1652279873629 14 719
|
30 |
+
1652279873668 15 719
|
31 |
+
1652279877456 15 767
|
32 |
+
1652279877494 16 767
|
33 |
+
1652279881261 16 815
|
34 |
+
1652279881296 17 815
|
35 |
+
1652279885044 17 863
|
36 |
+
1652279885082 18 863
|
37 |
+
1652279888829 18 911
|
38 |
+
1652279888864 19 911
|
39 |
+
1652279892527 19 959
|
40 |
+
1652279892566 20 959
|
41 |
+
1652279896237 20 1007
|
42 |
+
1652279896290 21 1007
|
43 |
+
1652279900002 21 1055
|
44 |
+
1652279900037 22 1055
|
45 |
+
1652279903680 22 1103
|
46 |
+
1652279903717 23 1103
|
47 |
+
1652279907516 23 1151
|
48 |
+
1652279907559 24 1151
|
49 |
+
1652279911251 24 1199
|
50 |
+
1652279911367 25 1199
|
51 |
+
1652279915084 25 1247
|
52 |
+
1652279915129 26 1247
|
53 |
+
1652279918947 26 1295
|
54 |
+
1652279918986 27 1295
|
55 |
+
1652279922767 27 1343
|
56 |
+
1652279922806 28 1343
|
57 |
+
1652279926483 28 1391
|
58 |
+
1652279926523 29 1391
|
59 |
+
1652279930329 29 1439
|
60 |
+
1652279930377 30 1439
|
61 |
+
1652279934121 30 1487
|
62 |
+
1652279934160 31 1487
|
63 |
+
1652279937833 31 1535
|
64 |
+
1652279937872 32 1535
|
65 |
+
1652279941623 32 1583
|
66 |
+
1652279941677 33 1583
|
67 |
+
1652279945268 33 1631
|
68 |
+
1652279945317 34 1631
|
69 |
+
1652279949005 34 1679
|
70 |
+
1652279949040 35 1679
|
71 |
+
1652279952772 35 1727
|
72 |
+
1652279952812 36 1727
|
73 |
+
1652279956496 36 1775
|
74 |
+
1652279956535 37 1775
|
75 |
+
1652279960298 37 1823
|
76 |
+
1652279960344 38 1823
|
77 |
+
1652279964073 38 1871
|
78 |
+
1652279964110 39 1871
|
79 |
+
1652279967844 39 1919
|
80 |
+
1652279967883 40 1919
|
81 |
+
1652279971446 40 1967
|
82 |
+
1652279971491 41 1967
|
83 |
+
1652279975126 41 2015
|
84 |
+
1652279975165 42 2015
|
85 |
+
1652279978811 42 2063
|
86 |
+
1652279978850 43 2063
|
87 |
+
1652279982539 43 2111
|
88 |
+
1652279982578 44 2111
|
89 |
+
1652279986321 44 2159
|
90 |
+
1652279986365 45 2159
|
91 |
+
1652279990134 45 2207
|
92 |
+
1652279990170 46 2207
|
93 |
+
1652279993983 46 2255
|
94 |
+
1652279994023 47 2255
|
95 |
+
1652279997788 47 2303
|
96 |
+
1652279997827 48 2303
|
97 |
+
1652280001609 48 2351
|
98 |
+
1652280001649 49 2351
|
99 |
+
1652280005406 49 2399
|
100 |
+
1652280005492 50 2399
|
101 |
+
1652280009144 50 2447
|
102 |
+
1652280009187 51 2447
|
103 |
+
1652280012811 51 2495
|
104 |
+
1652280012862 52 2495
|
105 |
+
1652280016584 52 2543
|
106 |
+
1652280016623 53 2543
|
107 |
+
1652280020312 53 2591
|
108 |
+
1652280020351 54 2591
|
109 |
+
1652280024138 54 2639
|
110 |
+
1652280024177 55 2639
|
111 |
+
1652280027939 55 2687
|
112 |
+
1652280027981 56 2687
|
113 |
+
1652280031799 56 2735
|
114 |
+
1652280031839 57 2735
|
115 |
+
1652280035376 57 2783
|
116 |
+
1652280035425 58 2783
|
117 |
+
1652280039147 58 2831
|
118 |
+
1652280039190 59 2831
|
119 |
+
1652280043008 59 2879
|
120 |
+
1652280043047 60 2879
|
121 |
+
1652280046735 60 2927
|
122 |
+
1652280046780 61 2927
|
123 |
+
1652280050532 61 2975
|
124 |
+
1652280050567 62 2975
|
125 |
+
1652280054457 62 3023
|
126 |
+
1652280054496 63 3023
|
127 |
+
1652280058279 63 3071
|
128 |
+
1652280058318 64 3071
|
129 |
+
1652280062047 64 3119
|
130 |
+
1652280062086 65 3119
|
131 |
+
1652280065810 65 3167
|
132 |
+
1652280065848 66 3167
|
133 |
+
1652280069610 66 3215
|
134 |
+
1652280069646 67 3215
|
135 |
+
1652280073394 67 3263
|
136 |
+
1652280073440 68 3263
|
137 |
+
1652280077179 68 3311
|
138 |
+
1652280077219 69 3311
|
139 |
+
1652280081010 69 3359
|
140 |
+
1652280081049 70 3359
|
141 |
+
1652280084783 70 3407
|
142 |
+
1652280084823 71 3407
|
143 |
+
1652280088575 71 3455
|
144 |
+
1652280088615 72 3455
|
145 |
+
1652280092368 72 3503
|
146 |
+
1652280092408 73 3503
|
147 |
+
1652280096159 73 3551
|
148 |
+
1652280096210 74 3551
|
149 |
+
1652280099894 74 3599
|
150 |
+
1652280099978 75 3599
|
151 |
+
1652280103580 75 3647
|
152 |
+
1652280103622 76 3647
|
153 |
+
1652280107354 76 3695
|
154 |
+
1652280107393 77 3695
|
155 |
+
1652280111162 77 3743
|
156 |
+
1652280111203 78 3743
|
157 |
+
1652280114987 78 3791
|
158 |
+
1652280115112 79 3791
|
159 |
+
1652280118870 79 3839
|
160 |
+
1652280118906 80 3839
|
161 |
+
1652280122629 80 3887
|
162 |
+
1652280122669 81 3887
|
163 |
+
1652280126475 81 3935
|
164 |
+
1652280126515 82 3935
|
165 |
+
1652280130198 82 3983
|
166 |
+
1652280130237 83 3983
|
167 |
+
1652280134070 83 4031
|
168 |
+
1652280134111 84 4031
|
169 |
+
1652280137810 84 4079
|
170 |
+
1652280137849 85 4079
|
171 |
+
1652280141642 85 4127
|
172 |
+
1652280141678 86 4127
|
173 |
+
1652280145396 86 4175
|
174 |
+
1652280145432 87 4175
|
175 |
+
1652280149183 87 4223
|
176 |
+
1652280149222 88 4223
|
177 |
+
1652280152927 88 4271
|
178 |
+
1652280152969 89 4271
|
179 |
+
1652280156721 89 4319
|
180 |
+
1652280156762 90 4319
|
181 |
+
1652280160468 90 4367
|
182 |
+
1652280160504 91 4367
|
183 |
+
1652280164338 91 4415
|
184 |
+
1652280164378 92 4415
|
185 |
+
1652280168035 92 4463
|
186 |
+
1652280168071 93 4463
|
187 |
+
1652280171787 93 4511
|
188 |
+
1652280171823 94 4511
|
189 |
+
1652280175551 94 4559
|
190 |
+
1652280175591 95 4559
|
191 |
+
1652280179385 95 4607
|
192 |
+
1652280179426 96 4607
|
193 |
+
1652280183124 96 4655
|
194 |
+
1652280183168 97 4655
|
195 |
+
1652280186940 97 4703
|
196 |
+
1652280186980 98 4703
|
197 |
+
1652280190746 98 4751
|
198 |
+
1652280190782 99 4751
|
199 |
+
1652280194552 99 4799
|
200 |
+
1652280194640 100 4799
|
201 |
+
1652280195352 96 4800
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/loss
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
1652280195352 0.00014260406896937639 4800
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/test/mae
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
1652280195352 0.008947640657424927 4800
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/loss
ADDED
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1652279820749 0.015712032094597816 47
|
2 |
+
1652279824405 0.0006998287281021476 95
|
3 |
+
1652279828096 0.0006262350943870842 143
|
4 |
+
1652279831720 0.0006112757255323231 191
|
5 |
+
1652279835495 0.0005892604240216315 239
|
6 |
+
1652279839319 0.0004991526948288083 287
|
7 |
+
1652279843119 0.00046538031892850995 335
|
8 |
+
1652279846878 0.0004903865046799183 383
|
9 |
+
1652279850653 0.0004449250118341297 431
|
10 |
+
1652279854434 0.0004655240918509662 479
|
11 |
+
1652279858281 0.00044383766362443566 527
|
12 |
+
1652279862126 0.0004356526769697666 575
|
13 |
+
1652279866012 0.00041521640378050506 623
|
14 |
+
1652279869817 0.0003966042131651193 671
|
15 |
+
1652279873668 0.0004069829301442951 719
|
16 |
+
1652279877494 0.0004001975175924599 767
|
17 |
+
1652279881296 0.000411732355132699 815
|
18 |
+
1652279885082 0.0003962284536100924 863
|
19 |
+
1652279888864 0.0003755573998205364 911
|
20 |
+
1652279892566 0.00040089385583996773 959
|
21 |
+
1652279896290 0.00037927881930954754 1007
|
22 |
+
1652279900037 0.00037079042522236705 1055
|
23 |
+
1652279903717 0.0003636741894297302 1103
|
24 |
+
1652279907559 0.0003473551769275218 1151
|
25 |
+
1652279911367 0.0003502711479086429 1199
|
26 |
+
1652279915129 0.0003374406951479614 1247
|
27 |
+
1652279918986 0.0003077377623412758 1295
|
28 |
+
1652279922806 0.00030867080204188824 1343
|
29 |
+
1652279926523 0.0003139588516205549 1391
|
30 |
+
1652279930377 0.00029595510568469763 1439
|
31 |
+
1652279934160 0.0002615793200675398 1487
|
32 |
+
1652279937872 0.0002453128108754754 1535
|
33 |
+
1652279941677 0.0002475597430020571 1583
|
34 |
+
1652279945317 0.00021561248286161572 1631
|
35 |
+
1652279949040 0.0002476295630913228 1679
|
36 |
+
1652279952812 0.00022967555560171604 1727
|
37 |
+
1652279956535 0.0002333874726900831 1775
|
38 |
+
1652279960344 0.00023667479399591684 1823
|
39 |
+
1652279964110 0.0001987472496693954 1871
|
40 |
+
1652279967883 0.00019275647355243564 1919
|
41 |
+
1652279971491 0.00020515960932243615 1967
|
42 |
+
1652279975165 0.00020770113042090088 2015
|
43 |
+
1652279978850 0.00020995769591536373 2063
|
44 |
+
1652279982578 0.0001873430737759918 2111
|
45 |
+
1652279986365 0.0001887517428258434 2159
|
46 |
+
1652279990170 0.00017797888722270727 2207
|
47 |
+
1652279994023 0.00017462909454479814 2255
|
48 |
+
1652279997827 0.00018134394485969096 2303
|
49 |
+
1652280001649 0.0002078687393805012 2351
|
50 |
+
1652280005492 0.00016190602036658674 2399
|
51 |
+
1652280009187 0.0001574503257870674 2447
|
52 |
+
1652280012862 0.00017259498417843133 2495
|
53 |
+
1652280016623 0.00016215378127526492 2543
|
54 |
+
1652280020351 0.00017192678933497518 2591
|
55 |
+
1652280024177 0.00014025233394932002 2639
|
56 |
+
1652280027981 0.00014388358977157623 2687
|
57 |
+
1652280031839 0.00013614425552077591 2735
|
58 |
+
1652280035425 0.00014166149776428938 2783
|
59 |
+
1652280039190 0.00014307026867754757 2831
|
60 |
+
1652280043047 0.00012790328764822334 2879
|
61 |
+
1652280046780 0.00013250892516225576 2927
|
62 |
+
1652280050567 0.00012225286627653986 2975
|
63 |
+
1652280054496 0.00013184384442865849 3023
|
64 |
+
1652280058318 0.0001139832820626907 3071
|
65 |
+
1652280062086 0.00012551495456136763 3119
|
66 |
+
1652280065848 0.00013683686847798526 3167
|
67 |
+
1652280069646 0.0001252585934707895 3215
|
68 |
+
1652280073440 0.0001234338415088132 3263
|
69 |
+
1652280077219 0.0001386790390824899 3311
|
70 |
+
1652280081049 0.000133068097056821 3359
|
71 |
+
1652280084823 0.0001232567592523992 3407
|
72 |
+
1652280088615 0.00010765787737909704 3455
|
73 |
+
1652280092408 0.00010420197213534266 3503
|
74 |
+
1652280096210 0.00011743942741304636 3551
|
75 |
+
1652280099978 0.00013075751485303044 3599
|
76 |
+
1652280103622 0.00012383297143969685 3647
|
77 |
+
1652280107393 0.00012997408339288086 3695
|
78 |
+
1652280111203 0.0001171917756437324 3743
|
79 |
+
1652280115112 0.0001384491624776274 3791
|
80 |
+
1652280118906 0.00011071155313402414 3839
|
81 |
+
1652280122669 0.00012242942466400564 3887
|
82 |
+
1652280126515 0.00012077093560947105 3935
|
83 |
+
1652280130237 0.00011708390229614452 3983
|
84 |
+
1652280134111 0.00013511900033336133 4031
|
85 |
+
1652280137849 0.00010951827425742522 4079
|
86 |
+
1652280141678 0.00011793604062404484 4127
|
87 |
+
1652280145432 0.00011460384848760441 4175
|
88 |
+
1652280149222 0.00012705886911135167 4223
|
89 |
+
1652280152969 0.00011695512512233108 4271
|
90 |
+
1652280156762 0.00011259406892349944 4319
|
91 |
+
1652280160504 0.00011745867959689349 4367
|
92 |
+
1652280164378 0.00013093209418002516 4415
|
93 |
+
1652280168071 0.00013953317829873413 4463
|
94 |
+
1652280171823 0.00012743366823997349 4511
|
95 |
+
1652280175591 0.00011808565614046529 4559
|
96 |
+
1652280179426 0.00015584049106109887 4607
|
97 |
+
1652280183168 0.00013836825382895768 4655
|
98 |
+
1652280186980 0.00012112926197005436 4703
|
99 |
+
1652280190782 0.00012230231368448585 4751
|
100 |
+
1652280194640 0.00011741576599888504 4799
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/train/mae
ADDED
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1652279820749 0.07011217623949051 47
|
2 |
+
1652279824405 0.021280117332935333 95
|
3 |
+
1652279828096 0.020297976210713387 143
|
4 |
+
1652279831720 0.020039746537804604 191
|
5 |
+
1652279835495 0.019787773489952087 239
|
6 |
+
1652279839319 0.018337024375796318 287
|
7 |
+
1652279843119 0.017815634608268738 335
|
8 |
+
1652279846878 0.017987234517931938 383
|
9 |
+
1652279850653 0.017149975523352623 431
|
10 |
+
1652279854434 0.01764928176999092 479
|
11 |
+
1652279858281 0.017262466251850128 527
|
12 |
+
1652279862126 0.017195068299770355 575
|
13 |
+
1652279866012 0.01676551066339016 623
|
14 |
+
1652279869817 0.016418082639575005 671
|
15 |
+
1652279873668 0.016572223976254463 719
|
16 |
+
1652279877494 0.016431566327810287 767
|
17 |
+
1652279881296 0.016624923795461655 815
|
18 |
+
1652279885082 0.016285978257656097 863
|
19 |
+
1652279888864 0.015845887362957 911
|
20 |
+
1652279892566 0.016440516337752342 959
|
21 |
+
1652279896290 0.015909109264612198 1007
|
22 |
+
1652279900037 0.015598192811012268 1055
|
23 |
+
1652279903717 0.015478463843464851 1103
|
24 |
+
1652279907559 0.014973259530961514 1151
|
25 |
+
1652279911367 0.014955243095755577 1199
|
26 |
+
1652279915129 0.014734053052961826 1247
|
27 |
+
1652279918986 0.013931847177445889 1295
|
28 |
+
1652279922806 0.013890136033296585 1343
|
29 |
+
1652279926523 0.013945831917226315 1391
|
30 |
+
1652279930377 0.013475497253239155 1439
|
31 |
+
1652279934160 0.012566673569381237 1487
|
32 |
+
1652279937872 0.012109467759728432 1535
|
33 |
+
1652279941677 0.01209949143230915 1583
|
34 |
+
1652279945317 0.011168932542204857 1631
|
35 |
+
1652279949040 0.012050420045852661 1679
|
36 |
+
1652279952812 0.011594988405704498 1727
|
37 |
+
1652279956535 0.011586652137339115 1775
|
38 |
+
1652279960344 0.01181836612522602 1823
|
39 |
+
1652279964110 0.010663148015737534 1871
|
40 |
+
1652279967883 0.010579877533018589 1919
|
41 |
+
1652279971491 0.010919580236077309 1967
|
42 |
+
1652279975165 0.010937352664768696 2015
|
43 |
+
1652279978850 0.010980397462844849 2063
|
44 |
+
1652279982578 0.010383875109255314 2111
|
45 |
+
1652279986365 0.010248820297420025 2159
|
46 |
+
1652279990170 0.009944379329681396 2207
|
47 |
+
1652279994023 0.009952726773917675 2255
|
48 |
+
1652279997827 0.010111919604241848 2303
|
49 |
+
1652280001649 0.010890776291489601 2351
|
50 |
+
1652280005492 0.009552615694701672 2399
|
51 |
+
1652280009187 0.009343606419861317 2447
|
52 |
+
1652280012862 0.00977871473878622 2495
|
53 |
+
1652280016623 0.009527333080768585 2543
|
54 |
+
1652280020351 0.009975402615964413 2591
|
55 |
+
1652280024177 0.008768484927713871 2639
|
56 |
+
1652280027981 0.008977909572422504 2687
|
57 |
+
1652280031839 0.008777343668043613 2735
|
58 |
+
1652280035425 0.008911573328077793 2783
|
59 |
+
1652280039190 0.008944649249315262 2831
|
60 |
+
1652280043047 0.008465489372611046 2879
|
61 |
+
1652280046780 0.008481234312057495 2927
|
62 |
+
1652280050567 0.008217735216021538 2975
|
63 |
+
1652280054496 0.008489103056490421 3023
|
64 |
+
1652280058318 0.00794004462659359 3071
|
65 |
+
1652280062086 0.00839218683540821 3119
|
66 |
+
1652280065848 0.008686883375048637 3167
|
67 |
+
1652280069646 0.008334343321621418 3215
|
68 |
+
1652280073440 0.008174673654139042 3263
|
69 |
+
1652280077219 0.008791987784206867 3311
|
70 |
+
1652280081049 0.008551678620278835 3359
|
71 |
+
1652280084823 0.008277542889118195 3407
|
72 |
+
1652280088615 0.007683832198381424 3455
|
73 |
+
1652280092408 0.007514127530157566 3503
|
74 |
+
1652280096210 0.008030056953430176 3551
|
75 |
+
1652280099978 0.008549727499485016 3599
|
76 |
+
1652280103622 0.008334793150424957 3647
|
77 |
+
1652280107393 0.00849801953881979 3695
|
78 |
+
1652280111203 0.00794607400894165 3743
|
79 |
+
1652280115112 0.008685767650604248 3791
|
80 |
+
1652280118906 0.007812082301825285 3839
|
81 |
+
1652280122669 0.008264751173555851 3887
|
82 |
+
1652280126515 0.008274727500975132 3935
|
83 |
+
1652280130237 0.008091365918517113 3983
|
84 |
+
1652280134111 0.008610659278929234 4031
|
85 |
+
1652280137849 0.007766522467136383 4079
|
86 |
+
1652280141678 0.008045573718845844 4127
|
87 |
+
1652280145432 0.007821462117135525 4175
|
88 |
+
1652280149222 0.008322164416313171 4223
|
89 |
+
1652280152969 0.008126244880259037 4271
|
90 |
+
1652280156762 0.007857552729547024 4319
|
91 |
+
1652280160504 0.008035344071686268 4367
|
92 |
+
1652280164378 0.008491233922541142 4415
|
93 |
+
1652280168071 0.008733570575714111 4463
|
94 |
+
1652280171823 0.008426659740507603 4511
|
95 |
+
1652280175591 0.008046725764870644 4559
|
96 |
+
1652280179426 0.009322610683739185 4607
|
97 |
+
1652280183168 0.008852057158946991 4655
|
98 |
+
1652280186980 0.008111957460641861 4703
|
99 |
+
1652280190782 0.008170234970748425 4751
|
100 |
+
1652280194640 0.008008270524442196 4799
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/loss
ADDED
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1652279820704 0.0010634043719619513 47
|
2 |
+
1652279824357 0.0008477172814309597 95
|
3 |
+
1652279828044 0.0006586583331227303 143
|
4 |
+
1652279831676 0.0006095922435633838 191
|
5 |
+
1652279835446 0.000579743820708245 239
|
6 |
+
1652279839271 0.000539788044989109 287
|
7 |
+
1652279843079 0.0005701080663129687 335
|
8 |
+
1652279846830 0.0005549467750824988 383
|
9 |
+
1652279850614 0.0005555949755944312 431
|
10 |
+
1652279854390 0.0004789295489899814 479
|
11 |
+
1652279858241 0.0005822633975185454 527
|
12 |
+
1652279862079 0.00048447673907503486 575
|
13 |
+
1652279865967 0.00046181678771972656 623
|
14 |
+
1652279869769 0.00041472347220405936 671
|
15 |
+
1652279873629 0.0005150702781975269 719
|
16 |
+
1652279877456 0.00043978836038149893 767
|
17 |
+
1652279881261 0.0004524872056208551 815
|
18 |
+
1652279885044 0.0004895858000963926 863
|
19 |
+
1652279888829 0.00048630539095029235 911
|
20 |
+
1652279892527 0.00045636019785888493 959
|
21 |
+
1652279896237 0.00042457759263925254 1007
|
22 |
+
1652279900002 0.0005504836444742978 1055
|
23 |
+
1652279903680 0.0005828228895552456 1103
|
24 |
+
1652279907516 0.000536997162271291 1151
|
25 |
+
1652279911251 0.0004296917177271098 1199
|
26 |
+
1652279915084 0.000369079876691103 1247
|
27 |
+
1652279918947 0.0005687376251444221 1295
|
28 |
+
1652279922767 0.0003877778071910143 1343
|
29 |
+
1652279926483 0.0007391179096885026 1391
|
30 |
+
1652279930329 0.0002891223703045398 1439
|
31 |
+
1652279934121 0.00136762042529881 1487
|
32 |
+
1652279937833 0.0005622635362669826 1535
|
33 |
+
1652279941623 0.0002613224496599287 1583
|
34 |
+
1652279945268 0.00026621491997502744 1631
|
35 |
+
1652279949005 0.0003965417272411287 1679
|
36 |
+
1652279952772 0.0006301697576418519 1727
|
37 |
+
1652279956496 0.0004612059856299311 1775
|
38 |
+
1652279960298 0.00025420173187740147 1823
|
39 |
+
1652279964073 0.0003383368020877242 1871
|
40 |
+
1652279967844 0.0003864422906190157 1919
|
41 |
+
1652279971446 0.00022850626555737108 1967
|
42 |
+
1652279975126 0.00036217307206243277 2015
|
43 |
+
1652279978811 0.00026963415439240634 2063
|
44 |
+
1652279982539 0.0003081822069361806 2111
|
45 |
+
1652279986321 0.0008588534547016025 2159
|
46 |
+
1652279990134 0.00025588576681911945 2207
|
47 |
+
1652279993983 0.0012111393734812737 2255
|
48 |
+
1652279997788 0.0005991487414576113 2303
|
49 |
+
1652280001609 0.00027215175214223564 2351
|
50 |
+
1652280005406 0.0002326530375285074 2399
|
51 |
+
1652280009144 0.00034161240910179913 2447
|
52 |
+
1652280012811 0.0002047311281785369 2495
|
53 |
+
1652280016584 0.0002189291553804651 2543
|
54 |
+
1652280020312 0.0008305492228828371 2591
|
55 |
+
1652280024138 0.0003245118132326752 2639
|
56 |
+
1652280027939 0.00043747638119384646 2687
|
57 |
+
1652280031799 0.0003352710627950728 2735
|
58 |
+
1652280035376 0.00017761132039595395 2783
|
59 |
+
1652280039147 0.0002451064356137067 2831
|
60 |
+
1652280043008 0.00045808948925696313 2879
|
61 |
+
1652280046735 0.00016568403225392103 2927
|
62 |
+
1652280050532 0.0002005849382840097 2975
|
63 |
+
1652280054457 0.0004337916907388717 3023
|
64 |
+
1652280058279 0.00043202925007790327 3071
|
65 |
+
1652280062047 0.0006070298259146512 3119
|
66 |
+
1652280065810 0.00125686835963279 3167
|
67 |
+
1652280069610 0.0003458949795458466 3215
|
68 |
+
1652280073394 0.00023582404537592083 3263
|
69 |
+
1652280077179 0.00043027440551668406 3311
|
70 |
+
1652280081010 0.00018619702314026654 3359
|
71 |
+
1652280084783 0.0004644407599698752 3407
|
72 |
+
1652280088575 0.00025989525602199137 3455
|
73 |
+
1652280092368 0.000271034223260358 3503
|
74 |
+
1652280096159 0.00015595319564454257 3551
|
75 |
+
1652280099894 0.00018362083937972784 3599
|
76 |
+
1652280103580 0.0001723619207041338 3647
|
77 |
+
1652280107354 0.00026339536998420954 3695
|
78 |
+
1652280111162 0.0022478969767689705 3743
|
79 |
+
1652280114987 0.00043688094592653215 3791
|
80 |
+
1652280118870 0.000253666948992759 3839
|
81 |
+
1652280122629 0.0007027353858575225 3887
|
82 |
+
1652280126475 0.00036985089536756277 3935
|
83 |
+
1652280130198 0.00042630801908671856 3983
|
84 |
+
1652280134070 0.00018512520182412118 4031
|
85 |
+
1652280137810 0.0006041005253791809 4079
|
86 |
+
1652280141642 0.0005739088519476354 4127
|
87 |
+
1652280145396 0.00020085058349650353 4175
|
88 |
+
1652280149183 0.00035394675796851516 4223
|
89 |
+
1652280152927 0.001467306399717927 4271
|
90 |
+
1652280156721 0.0003908797516487539 4319
|
91 |
+
1652280160468 0.0001805946958484128 4367
|
92 |
+
1652280164338 0.00035339672467671335 4415
|
93 |
+
1652280168035 0.00040678223012946546 4463
|
94 |
+
1652280171787 0.00018399060354568064 4511
|
95 |
+
1652280175551 0.0005399276851676404 4559
|
96 |
+
1652280179385 0.00029979468672536314 4607
|
97 |
+
1652280183124 0.00013096585462335497 4655
|
98 |
+
1652280186940 0.00021977233700454235 4703
|
99 |
+
1652280190746 0.00024704227689653635 4751
|
100 |
+
1652280194552 0.0003437100094743073 4799
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae
ADDED
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1652279820704 0.02607358619570732 47
|
2 |
+
1652279824357 0.023356162011623383 95
|
3 |
+
1652279828044 0.02111729420721531 143
|
4 |
+
1652279831676 0.020395344123244286 191
|
5 |
+
1652279835446 0.02008357085287571 239
|
6 |
+
1652279839271 0.019109351560473442 287
|
7 |
+
1652279843079 0.020062454044818878 335
|
8 |
+
1652279846830 0.018724771216511726 383
|
9 |
+
1652279850614 0.01976374350488186 431
|
10 |
+
1652279854390 0.018088672310113907 479
|
11 |
+
1652279858241 0.02009824477136135 527
|
12 |
+
1652279862079 0.017433589324355125 575
|
13 |
+
1652279865967 0.018075214698910713 623
|
14 |
+
1652279869769 0.016927972435951233 671
|
15 |
+
1652279873629 0.01795283891260624 719
|
16 |
+
1652279877456 0.017694905400276184 767
|
17 |
+
1652279881261 0.017309539020061493 815
|
18 |
+
1652279885044 0.017853353172540665 863
|
19 |
+
1652279888829 0.017320722341537476 911
|
20 |
+
1652279892527 0.017117636278271675 959
|
21 |
+
1652279896237 0.01685478165745735 1007
|
22 |
+
1652279900002 0.01946953497827053 1055
|
23 |
+
1652279903680 0.019640492275357246 1103
|
24 |
+
1652279907516 0.01877320185303688 1151
|
25 |
+
1652279911251 0.016905883327126503 1199
|
26 |
+
1652279915084 0.015234655700623989 1247
|
27 |
+
1652279918947 0.019466156139969826 1295
|
28 |
+
1652279922767 0.01577661745250225 1343
|
29 |
+
1652279926483 0.021662525832653046 1391
|
30 |
+
1652279930329 0.013201462104916573 1439
|
31 |
+
1652279934121 0.03221268951892853 1487
|
32 |
+
1652279937833 0.018861766904592514 1535
|
33 |
+
1652279941623 0.012555716559290886 1583
|
34 |
+
1652279945268 0.01243590097874403 1631
|
35 |
+
1652279949005 0.015400874428451061 1679
|
36 |
+
1652279952772 0.019847549498081207 1727
|
37 |
+
1652279956496 0.01633538119494915 1775
|
38 |
+
1652279960298 0.012520545162260532 1823
|
39 |
+
1652279964073 0.015009582974016666 1871
|
40 |
+
1652279967844 0.016114825382828712 1919
|
41 |
+
1652279971446 0.011395555920898914 1967
|
42 |
+
1652279975126 0.015877339988946915 2015
|
43 |
+
1652279978811 0.012874774634838104 2063
|
44 |
+
1652279982539 0.01374065037816763 2111
|
45 |
+
1652279986321 0.02455970272421837 2159
|
46 |
+
1652279990134 0.012964467518031597 2207
|
47 |
+
1652279993983 0.030586877837777138 2255
|
48 |
+
1652279997788 0.01970861665904522 2303
|
49 |
+
1652280001609 0.012582853436470032 2351
|
50 |
+
1652280005406 0.011895284987986088 2399
|
51 |
+
1652280009144 0.014539946801960468 2447
|
52 |
+
1652280012811 0.010513107292354107 2495
|
53 |
+
1652280016584 0.011585352011024952 2543
|
54 |
+
1652280020312 0.025230778381228447 2591
|
55 |
+
1652280024138 0.01448488887399435 2639
|
56 |
+
1652280027939 0.016850745305418968 2687
|
57 |
+
1652280031799 0.014916492626070976 2735
|
58 |
+
1652280035376 0.009772807359695435 2783
|
59 |
+
1652280039147 0.012413904070854187 2831
|
60 |
+
1652280043008 0.017842639237642288 2879
|
61 |
+
1652280046735 0.009426051750779152 2927
|
62 |
+
1652280050532 0.011247269809246063 2975
|
63 |
+
1652280054457 0.017680322751402855 3023
|
64 |
+
1652280058279 0.016938434913754463 3071
|
65 |
+
1652280062047 0.020868053659796715 3119
|
66 |
+
1652280065810 0.03198228031396866 3167
|
67 |
+
1652280069610 0.01481217984110117 3215
|
68 |
+
1652280073394 0.01217477023601532 3263
|
69 |
+
1652280077179 0.0175019484013319 3311
|
70 |
+
1652280081010 0.01059352234005928 3359
|
71 |
+
1652280084783 0.017586909234523773 3407
|
72 |
+
1652280088575 0.013158340007066727 3455
|
73 |
+
1652280092368 0.013016403652727604 3503
|
74 |
+
1652280096159 0.00941098015755415 3551
|
75 |
+
1652280099894 0.010173797607421875 3599
|
76 |
+
1652280103580 0.010034985840320587 3647
|
77 |
+
1652280107354 0.01230852585285902 3695
|
78 |
+
1652280111162 0.043913356959819794 3743
|
79 |
+
1652280114987 0.01787818782031536 3791
|
80 |
+
1652280118870 0.012860489077866077 3839
|
81 |
+
1652280122629 0.02276649698615074 3887
|
82 |
+
1652280126475 0.015199720859527588 3935
|
83 |
+
1652280130198 0.016393927857279778 3983
|
84 |
+
1652280134070 0.010693005286157131 4031
|
85 |
+
1652280137810 0.02147362008690834 4079
|
86 |
+
1652280141642 0.019695602357387543 4127
|
87 |
+
1652280145396 0.010959284380078316 4175
|
88 |
+
1652280149183 0.014944094233214855 4223
|
89 |
+
1652280152927 0.03422271087765694 4271
|
90 |
+
1652280156721 0.015736214816570282 4319
|
91 |
+
1652280160468 0.009996251203119755 4367
|
92 |
+
1652280164338 0.015099494718015194 4415
|
93 |
+
1652280168035 0.015847915783524513 4463
|
94 |
+
1652280171787 0.010453518480062485 4511
|
95 |
+
1652280175551 0.018864858895540237 4559
|
96 |
+
1652280179385 0.01469734963029623 4607
|
97 |
+
1652280183124 0.00869279820472002 4655
|
98 |
+
1652280186940 0.011555126868188381 4703
|
99 |
+
1652280190746 0.01307331956923008 4751
|
100 |
+
1652280194552 0.01433614082634449 4799
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/metrics/val/mae_best
ADDED
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
1652279820704 0.02607358619570732 47
|
2 |
+
1652279824357 0.023356162011623383 95
|
3 |
+
1652279828044 0.02111729420721531 143
|
4 |
+
1652279831676 0.020395344123244286 191
|
5 |
+
1652279835446 0.02008357085287571 239
|
6 |
+
1652279839271 0.019109351560473442 287
|
7 |
+
1652279843079 0.019109351560473442 335
|
8 |
+
1652279846830 0.018724771216511726 383
|
9 |
+
1652279850614 0.018724771216511726 431
|
10 |
+
1652279854390 0.018088672310113907 479
|
11 |
+
1652279858241 0.018088672310113907 527
|
12 |
+
1652279862079 0.017433589324355125 575
|
13 |
+
1652279865967 0.017433589324355125 623
|
14 |
+
1652279869769 0.016927972435951233 671
|
15 |
+
1652279873629 0.016927972435951233 719
|
16 |
+
1652279877456 0.016927972435951233 767
|
17 |
+
1652279881261 0.016927972435951233 815
|
18 |
+
1652279885044 0.016927972435951233 863
|
19 |
+
1652279888829 0.016927972435951233 911
|
20 |
+
1652279892527 0.016927972435951233 959
|
21 |
+
1652279896237 0.01685478165745735 1007
|
22 |
+
1652279900002 0.01685478165745735 1055
|
23 |
+
1652279903680 0.01685478165745735 1103
|
24 |
+
1652279907516 0.01685478165745735 1151
|
25 |
+
1652279911251 0.01685478165745735 1199
|
26 |
+
1652279915084 0.015234655700623989 1247
|
27 |
+
1652279918947 0.015234655700623989 1295
|
28 |
+
1652279922767 0.015234655700623989 1343
|
29 |
+
1652279926483 0.015234655700623989 1391
|
30 |
+
1652279930329 0.013201462104916573 1439
|
31 |
+
1652279934121 0.013201462104916573 1487
|
32 |
+
1652279937833 0.013201462104916573 1535
|
33 |
+
1652279941623 0.012555716559290886 1583
|
34 |
+
1652279945268 0.01243590097874403 1631
|
35 |
+
1652279949005 0.01243590097874403 1679
|
36 |
+
1652279952772 0.01243590097874403 1727
|
37 |
+
1652279956496 0.01243590097874403 1775
|
38 |
+
1652279960298 0.01243590097874403 1823
|
39 |
+
1652279964073 0.01243590097874403 1871
|
40 |
+
1652279967844 0.01243590097874403 1919
|
41 |
+
1652279971446 0.011395555920898914 1967
|
42 |
+
1652279975126 0.011395555920898914 2015
|
43 |
+
1652279978811 0.011395555920898914 2063
|
44 |
+
1652279982539 0.011395555920898914 2111
|
45 |
+
1652279986321 0.011395555920898914 2159
|
46 |
+
1652279990134 0.011395555920898914 2207
|
47 |
+
1652279993983 0.011395555920898914 2255
|
48 |
+
1652279997788 0.011395555920898914 2303
|
49 |
+
1652280001609 0.011395555920898914 2351
|
50 |
+
1652280005406 0.011395555920898914 2399
|
51 |
+
1652280009144 0.011395555920898914 2447
|
52 |
+
1652280012811 0.010513107292354107 2495
|
53 |
+
1652280016584 0.010513107292354107 2543
|
54 |
+
1652280020312 0.010513107292354107 2591
|
55 |
+
1652280024138 0.010513107292354107 2639
|
56 |
+
1652280027939 0.010513107292354107 2687
|
57 |
+
1652280031799 0.010513107292354107 2735
|
58 |
+
1652280035376 0.009772807359695435 2783
|
59 |
+
1652280039147 0.009772807359695435 2831
|
60 |
+
1652280043008 0.009772807359695435 2879
|
61 |
+
1652280046735 0.009426051750779152 2927
|
62 |
+
1652280050532 0.009426051750779152 2975
|
63 |
+
1652280054457 0.009426051750779152 3023
|
64 |
+
1652280058279 0.009426051750779152 3071
|
65 |
+
1652280062047 0.009426051750779152 3119
|
66 |
+
1652280065810 0.009426051750779152 3167
|
67 |
+
1652280069610 0.009426051750779152 3215
|
68 |
+
1652280073394 0.009426051750779152 3263
|
69 |
+
1652280077179 0.009426051750779152 3311
|
70 |
+
1652280081010 0.009426051750779152 3359
|
71 |
+
1652280084783 0.009426051750779152 3407
|
72 |
+
1652280088575 0.009426051750779152 3455
|
73 |
+
1652280092368 0.009426051750779152 3503
|
74 |
+
1652280096159 0.00941098015755415 3551
|
75 |
+
1652280099894 0.00941098015755415 3599
|
76 |
+
1652280103580 0.00941098015755415 3647
|
77 |
+
1652280107354 0.00941098015755415 3695
|
78 |
+
1652280111162 0.00941098015755415 3743
|
79 |
+
1652280114987 0.00941098015755415 3791
|
80 |
+
1652280118870 0.00941098015755415 3839
|
81 |
+
1652280122629 0.00941098015755415 3887
|
82 |
+
1652280126475 0.00941098015755415 3935
|
83 |
+
1652280130198 0.00941098015755415 3983
|
84 |
+
1652280134070 0.00941098015755415 4031
|
85 |
+
1652280137810 0.00941098015755415 4079
|
86 |
+
1652280141642 0.00941098015755415 4127
|
87 |
+
1652280145396 0.00941098015755415 4175
|
88 |
+
1652280149183 0.00941098015755415 4223
|
89 |
+
1652280152927 0.00941098015755415 4271
|
90 |
+
1652280156721 0.00941098015755415 4319
|
91 |
+
1652280160468 0.00941098015755415 4367
|
92 |
+
1652280164338 0.00941098015755415 4415
|
93 |
+
1652280168035 0.00941098015755415 4463
|
94 |
+
1652280171787 0.00941098015755415 4511
|
95 |
+
1652280175551 0.00941098015755415 4559
|
96 |
+
1652280179385 0.00941098015755415 4607
|
97 |
+
1652280183124 0.00869279820472002 4655
|
98 |
+
1652280186940 0.00869279820472002 4703
|
99 |
+
1652280190746 0.00869279820472002 4751
|
100 |
+
1652280194552 0.00869279820472002 4799
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/_target_
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
pytorch_lightning.callbacks.EarlyStopping
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/min_delta
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
0
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/mode
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
min
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/monitor
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
val/mae
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/early_stopping/patience
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
100
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/_target_
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
pytorch_lightning.callbacks.ModelCheckpoint
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/auto_insert_metric_name
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
False
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/dirpath
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
checkpoints/
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/filename
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
epoch_{epoch:03d}
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/mode
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
min
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/monitor
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
val/mae
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_last
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
True
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/save_top_k
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
1
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_checkpoint/verbose
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
False
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/_target_
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
pytorch_lightning.callbacks.RichModelSummary
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/model_summary/max_depth
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
-1
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/callbacks/rich_progress_bar/_target_
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
pytorch_lightning.callbacks.RichProgressBar
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/_target_
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
src.datamodules.focus_datamodule.FocusDataModule
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/augmentation
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
True
|
logs/mlflow/mlruns/11/fc9d4c9c610e4ae9a51aeec7e544684b/params/datamodule/batch_size
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
64
|