Hannes Kuchelmeister commited on
Commit
72d30c1
·
1 Parent(s): 3faec92

add hyperparameter optimisation run

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/.hydra/config.yaml +71 -0
  2. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/.hydra/hydra.yaml +226 -0
  3. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/.hydra/overrides.yaml +7 -0
  4. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/csv/version_0/hparams.yaml +52 -0
  5. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/csv/version_0/metrics.csv +202 -0
  6. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646931592.37a92d36c649.1.0 +3 -0
  7. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646933528.37a92d36c649.1.1 +3 -0
  8. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml +52 -0
  9. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/.hydra/config.yaml +71 -0
  10. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/.hydra/hydra.yaml +226 -0
  11. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/.hydra/overrides.yaml +7 -0
  12. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/csv/version_0/hparams.yaml +52 -0
  13. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/csv/version_0/metrics.csv +202 -0
  14. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646933529.37a92d36c649.1.2 +3 -0
  15. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646937064.37a92d36c649.1.3 +3 -0
  16. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml +52 -0
  17. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/.hydra/config.yaml +71 -0
  18. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/.hydra/hydra.yaml +226 -0
  19. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/.hydra/overrides.yaml +7 -0
  20. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/csv/version_0/hparams.yaml +52 -0
  21. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/csv/version_0/metrics.csv +202 -0
  22. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646950714.37a92d36c649.1.20 +3 -0
  23. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646951667.37a92d36c649.1.21 +3 -0
  24. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml +52 -0
  25. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/.hydra/config.yaml +71 -0
  26. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/.hydra/hydra.yaml +226 -0
  27. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/.hydra/overrides.yaml +7 -0
  28. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/csv/version_0/hparams.yaml +52 -0
  29. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/csv/version_0/metrics.csv +202 -0
  30. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646951668.37a92d36c649.1.22 +3 -0
  31. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646952623.37a92d36c649.1.23 +3 -0
  32. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml +52 -0
  33. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/.hydra/config.yaml +71 -0
  34. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/.hydra/hydra.yaml +226 -0
  35. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/.hydra/overrides.yaml +7 -0
  36. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/csv/version_0/hparams.yaml +52 -0
  37. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/csv/version_0/metrics.csv +202 -0
  38. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646952624.37a92d36c649.1.24 +3 -0
  39. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646953611.37a92d36c649.1.25 +3 -0
  40. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml +52 -0
  41. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/.hydra/config.yaml +71 -0
  42. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/.hydra/hydra.yaml +226 -0
  43. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/.hydra/overrides.yaml +7 -0
  44. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/csv/version_0/hparams.yaml +52 -0
  45. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/csv/version_0/metrics.csv +202 -0
  46. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646953611.37a92d36c649.1.26 +3 -0
  47. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646954564.37a92d36c649.1.27 +3 -0
  48. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml +52 -0
  49. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/14/.hydra/config.yaml +71 -0
  50. logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/14/.hydra/hydra.yaml +226 -0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 512
24
+ lin2_size: 512
25
+ lin3_size: 1024
26
+ output_size: 1
27
+ lr: 0.03686537045425119
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.03686537045425119
171
+ - model.lin1_size=512
172
+ - model.lin2_size=512
173
+ - model.lin3_size=1024
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=512,model.lin2_size=512,model.lin3_size=1024,model.lr=0.03686537045425119,trainer.gpus=0
179
+ id: '0'
180
+ num: 0
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - datamodule.batch_size=64
2
+ - model.lr=0.03686537045425119
3
+ - model.lin1_size=512
4
+ - model.lin2_size=512
5
+ - model.lin3_size=1024
6
+ - hparams_search=focusMAE_150_optuna
7
+ - trainer.gpus=0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 512
11
+ lin2_size: 512
12
+ lin3_size: 1024
13
+ output_size: 1
14
+ lr: 0.03686537045425119
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 35353601
51
+ model/params/trainable: 35353601
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 1.5817662477493286,1.5817662477493286,1.5817662477493286,0,26,,,,
3
+ ,,,0,26,2.9489288330078125,2.9489288330078125,,
4
+ 1.519407033920288,1.519407033920288,1.519407033920288,1,53,,,,
5
+ ,,,1,53,1.596929907798767,1.596929907798767,,
6
+ 1.7517845630645752,1.7517845630645752,1.519407033920288,2,80,,,,
7
+ ,,,2,80,1.6025917530059814,1.6025917530059814,,
8
+ 1.533663034439087,1.533663034439087,1.519407033920288,3,107,,,,
9
+ ,,,3,107,1.5848336219787598,1.5848336219787598,,
10
+ 1.5713140964508057,1.5713140964508057,1.519407033920288,4,134,,,,
11
+ ,,,4,134,1.5639420747756958,1.5639420747756958,,
12
+ 2.1142706871032715,2.1142706871032715,1.519407033920288,5,161,,,,
13
+ ,,,5,161,1.565294861793518,1.565294861793518,,
14
+ 1.6672890186309814,1.6672890186309814,1.519407033920288,6,188,,,,
15
+ ,,,6,188,1.5487409830093384,1.5487409830093384,,
16
+ 2.123723268508911,2.123723268508911,1.519407033920288,7,215,,,,
17
+ ,,,7,215,1.5563510656356812,1.5563510656356812,,
18
+ 1.5217065811157227,1.5217065811157227,1.519407033920288,8,242,,,,
19
+ ,,,8,242,1.5466948747634888,1.5466948747634888,,
20
+ 1.628381371498108,1.628381371498108,1.519407033920288,9,269,,,,
21
+ ,,,9,269,1.5386103391647339,1.5386103391647339,,
22
+ 1.5665384531021118,1.5665384531021118,1.519407033920288,10,296,,,,
23
+ ,,,10,296,1.537092685699463,1.537092685699463,,
24
+ 1.647467851638794,1.647467851638794,1.519407033920288,11,323,,,,
25
+ ,,,11,323,1.5201250314712524,1.5201250314712524,,
26
+ 1.7197288274765015,1.7197288274765015,1.519407033920288,12,350,,,,
27
+ ,,,12,350,1.5263408422470093,1.5263408422470093,,
28
+ 1.87315034866333,1.87315034866333,1.519407033920288,13,377,,,,
29
+ ,,,13,377,1.5487321615219116,1.5487321615219116,,
30
+ 1.6700716018676758,1.6700716018676758,1.519407033920288,14,404,,,,
31
+ ,,,14,404,1.5312869548797607,1.5312869548797607,,
32
+ 1.5027695894241333,1.5027695894241333,1.5027695894241333,15,431,,,,
33
+ ,,,15,431,1.5504881143569946,1.5504881143569946,,
34
+ 1.5581985712051392,1.5581985712051392,1.5027695894241333,16,458,,,,
35
+ ,,,16,458,1.518345594406128,1.518345594406128,,
36
+ 1.5221569538116455,1.5221569538116455,1.5027695894241333,17,485,,,,
37
+ ,,,17,485,1.546921968460083,1.546921968460083,,
38
+ 1.529903531074524,1.529903531074524,1.5027695894241333,18,512,,,,
39
+ ,,,18,512,1.5361909866333008,1.5361909866333008,,
40
+ 2.162675619125366,2.162675619125366,1.5027695894241333,19,539,,,,
41
+ ,,,19,539,1.541818618774414,1.541818618774414,,
42
+ 2.095296859741211,2.095296859741211,1.5027695894241333,20,566,,,,
43
+ ,,,20,566,1.542490005493164,1.542490005493164,,
44
+ 2.2468199729919434,2.2468199729919434,1.5027695894241333,21,593,,,,
45
+ ,,,21,593,1.5121276378631592,1.5121276378631592,,
46
+ 1.619676947593689,1.619676947593689,1.5027695894241333,22,620,,,,
47
+ ,,,22,620,1.545835256576538,1.545835256576538,,
48
+ 1.6065987348556519,1.6065987348556519,1.5027695894241333,23,647,,,,
49
+ ,,,23,647,1.5323158502578735,1.5323158502578735,,
50
+ 1.4961801767349243,1.4961801767349243,1.4961801767349243,24,674,,,,
51
+ ,,,24,674,1.5414543151855469,1.5414543151855469,,
52
+ 1.51967453956604,1.51967453956604,1.4961801767349243,25,701,,,,
53
+ ,,,25,701,1.5501147508621216,1.5501147508621216,,
54
+ 1.535367727279663,1.535367727279663,1.4961801767349243,26,728,,,,
55
+ ,,,26,728,1.5353726148605347,1.5353726148605347,,
56
+ 1.4789810180664062,1.4789810180664062,1.4789810180664062,27,755,,,,
57
+ ,,,27,755,1.5555198192596436,1.5555198192596436,,
58
+ 1.4862595796585083,1.4862595796585083,1.4789810180664062,28,782,,,,
59
+ ,,,28,782,1.5383554697036743,1.5383554697036743,,
60
+ 1.5236232280731201,1.5236232280731201,1.4789810180664062,29,809,,,,
61
+ ,,,29,809,1.5491728782653809,1.5491728782653809,,
62
+ 1.6988039016723633,1.6988039016723633,1.4789810180664062,30,836,,,,
63
+ ,,,30,836,1.5513640642166138,1.5513640642166138,,
64
+ 1.6100221872329712,1.6100221872329712,1.4789810180664062,31,863,,,,
65
+ ,,,31,863,1.5313589572906494,1.5313589572906494,,
66
+ 2.0812690258026123,2.0812690258026123,1.4789810180664062,32,890,,,,
67
+ ,,,32,890,1.5378966331481934,1.5378966331481934,,
68
+ 1.5452286005020142,1.5452286005020142,1.4789810180664062,33,917,,,,
69
+ ,,,33,917,1.531347393989563,1.531347393989563,,
70
+ 2.168597936630249,2.168597936630249,1.4789810180664062,34,944,,,,
71
+ ,,,34,944,1.5500789880752563,1.5500789880752563,,
72
+ 1.6888200044631958,1.6888200044631958,1.4789810180664062,35,971,,,,
73
+ ,,,35,971,1.5606012344360352,1.5606012344360352,,
74
+ 1.644452452659607,1.644452452659607,1.4789810180664062,36,998,,,,
75
+ ,,,36,998,1.5430793762207031,1.5430793762207031,,
76
+ 1.93430495262146,1.93430495262146,1.4789810180664062,37,1025,,,,
77
+ ,,,37,1025,1.5269699096679688,1.5269699096679688,,
78
+ 1.6087878942489624,1.6087878942489624,1.4789810180664062,38,1052,,,,
79
+ ,,,38,1052,1.544972538948059,1.544972538948059,,
80
+ 1.5186485052108765,1.5186485052108765,1.4789810180664062,39,1079,,,,
81
+ ,,,39,1079,1.539893388748169,1.539893388748169,,
82
+ 1.52126145362854,1.52126145362854,1.4789810180664062,40,1106,,,,
83
+ ,,,40,1106,1.53793203830719,1.53793203830719,,
84
+ 1.70045804977417,1.70045804977417,1.4789810180664062,41,1133,,,,
85
+ ,,,41,1133,1.545230746269226,1.545230746269226,,
86
+ 1.5828505754470825,1.5828505754470825,1.4789810180664062,42,1160,,,,
87
+ ,,,42,1160,1.5217416286468506,1.5217416286468506,,
88
+ 1.5459929704666138,1.5459929704666138,1.4789810180664062,43,1187,,,,
89
+ ,,,43,1187,1.550054669380188,1.550054669380188,,
90
+ 1.6890517473220825,1.6890517473220825,1.4789810180664062,44,1214,,,,
91
+ ,,,44,1214,1.5414174795150757,1.5414174795150757,,
92
+ 1.5879722833633423,1.5879722833633423,1.4789810180664062,45,1241,,,,
93
+ ,,,45,1241,1.5494052171707153,1.5494052171707153,,
94
+ 1.9080801010131836,1.9080801010131836,1.4789810180664062,46,1268,,,,
95
+ ,,,46,1268,1.5503209829330444,1.5503209829330444,,
96
+ 1.7443077564239502,1.7443077564239502,1.4789810180664062,47,1295,,,,
97
+ ,,,47,1295,1.5583010911941528,1.5583010911941528,,
98
+ 1.8529819250106812,1.8529819250106812,1.4789810180664062,48,1322,,,,
99
+ ,,,48,1322,1.5445235967636108,1.5445235967636108,,
100
+ 2.1755263805389404,2.1755263805389404,1.4789810180664062,49,1349,,,,
101
+ ,,,49,1349,1.5581728219985962,1.5581728219985962,,
102
+ 1.5154876708984375,1.5154876708984375,1.4789810180664062,50,1376,,,,
103
+ ,,,50,1376,1.5434831380844116,1.5434831380844116,,
104
+ 1.5233850479125977,1.5233850479125977,1.4789810180664062,51,1403,,,,
105
+ ,,,51,1403,1.5547490119934082,1.5547490119934082,,
106
+ 3.795761823654175,3.795761823654175,1.4789810180664062,52,1430,,,,
107
+ ,,,52,1430,1.511307716369629,1.511307716369629,,
108
+ 1.524182915687561,1.524182915687561,1.4789810180664062,53,1457,,,,
109
+ ,,,53,1457,1.5449738502502441,1.5449738502502441,,
110
+ 2.197523593902588,2.197523593902588,1.4789810180664062,54,1484,,,,
111
+ ,,,54,1484,1.554245948791504,1.554245948791504,,
112
+ 1.4860424995422363,1.4860424995422363,1.4789810180664062,55,1511,,,,
113
+ ,,,55,1511,1.5498542785644531,1.5498542785644531,,
114
+ 1.494900107383728,1.494900107383728,1.4789810180664062,56,1538,,,,
115
+ ,,,56,1538,1.545539379119873,1.545539379119873,,
116
+ 1.630722999572754,1.630722999572754,1.4789810180664062,57,1565,,,,
117
+ ,,,57,1565,1.5449497699737549,1.5449497699737549,,
118
+ 1.5924216508865356,1.5924216508865356,1.4789810180664062,58,1592,,,,
119
+ ,,,58,1592,1.5530821084976196,1.5530821084976196,,
120
+ 1.675205111503601,1.675205111503601,1.4789810180664062,59,1619,,,,
121
+ ,,,59,1619,1.5506713390350342,1.5506713390350342,,
122
+ 1.7345980405807495,1.7345980405807495,1.4789810180664062,60,1646,,,,
123
+ ,,,60,1646,1.5539727210998535,1.5539727210998535,,
124
+ 2.1135194301605225,2.1135194301605225,1.4789810180664062,61,1673,,,,
125
+ ,,,61,1673,1.5415444374084473,1.5415444374084473,,
126
+ 1.6276977062225342,1.6276977062225342,1.4789810180664062,62,1700,,,,
127
+ ,,,62,1700,1.5567636489868164,1.5567636489868164,,
128
+ 1.5710558891296387,1.5710558891296387,1.4789810180664062,63,1727,,,,
129
+ ,,,63,1727,1.5625300407409668,1.5625300407409668,,
130
+ 1.6335477828979492,1.6335477828979492,1.4789810180664062,64,1754,,,,
131
+ ,,,64,1754,1.5535508394241333,1.5535508394241333,,
132
+ 1.942752480506897,1.942752480506897,1.4789810180664062,65,1781,,,,
133
+ ,,,65,1781,1.5470272302627563,1.5470272302627563,,
134
+ 1.6966638565063477,1.6966638565063477,1.4789810180664062,66,1808,,,,
135
+ ,,,66,1808,1.5541692972183228,1.5541692972183228,,
136
+ 1.9469621181488037,1.9469621181488037,1.4789810180664062,67,1835,,,,
137
+ ,,,67,1835,1.5689467191696167,1.5689467191696167,,
138
+ 1.537792682647705,1.537792682647705,1.4789810180664062,68,1862,,,,
139
+ ,,,68,1862,1.5576432943344116,1.5576432943344116,,
140
+ 2.756701946258545,2.756701946258545,1.4789810180664062,69,1889,,,,
141
+ ,,,69,1889,1.554655909538269,1.554655909538269,,
142
+ 1.8372505903244019,1.8372505903244019,1.4789810180664062,70,1916,,,,
143
+ ,,,70,1916,1.5448646545410156,1.5448646545410156,,
144
+ 3.46052885055542,3.46052885055542,1.4789810180664062,71,1943,,,,
145
+ ,,,71,1943,1.5456660985946655,1.5456660985946655,,
146
+ 1.6675915718078613,1.6675915718078613,1.4789810180664062,72,1970,,,,
147
+ ,,,72,1970,1.5650439262390137,1.5650439262390137,,
148
+ 1.7358143329620361,1.7358143329620361,1.4789810180664062,73,1997,,,,
149
+ ,,,73,1997,1.5670278072357178,1.5670278072357178,,
150
+ 5.314837455749512,5.314837455749512,1.4789810180664062,74,2024,,,,
151
+ ,,,74,2024,1.5667744874954224,1.5667744874954224,,
152
+ 2.545644760131836,2.545644760131836,1.4789810180664062,75,2051,,,,
153
+ ,,,75,2051,1.5524017810821533,1.5524017810821533,,
154
+ 1.531238079071045,1.531238079071045,1.4789810180664062,76,2078,,,,
155
+ ,,,76,2078,1.5600334405899048,1.5600334405899048,,
156
+ 2.5285863876342773,2.5285863876342773,1.4789810180664062,77,2105,,,,
157
+ ,,,77,2105,1.5680941343307495,1.5680941343307495,,
158
+ 1.5376269817352295,1.5376269817352295,1.4789810180664062,78,2132,,,,
159
+ ,,,78,2132,1.5522637367248535,1.5522637367248535,,
160
+ 1.8609051704406738,1.8609051704406738,1.4789810180664062,79,2159,,,,
161
+ ,,,79,2159,1.5448931455612183,1.5448931455612183,,
162
+ 1.527772307395935,1.527772307395935,1.4789810180664062,80,2186,,,,
163
+ ,,,80,2186,1.5477863550186157,1.5477863550186157,,
164
+ 1.6045533418655396,1.6045533418655396,1.4789810180664062,81,2213,,,,
165
+ ,,,81,2213,1.5636101961135864,1.5636101961135864,,
166
+ 1.742592692375183,1.742592692375183,1.4789810180664062,82,2240,,,,
167
+ ,,,82,2240,1.5515958070755005,1.5515958070755005,,
168
+ 2.126866579055786,2.126866579055786,1.4789810180664062,83,2267,,,,
169
+ ,,,83,2267,1.574913501739502,1.574913501739502,,
170
+ 1.5660400390625,1.5660400390625,1.4789810180664062,84,2294,,,,
171
+ ,,,84,2294,1.5619218349456787,1.5619218349456787,,
172
+ 4.364716053009033,4.364716053009033,1.4789810180664062,85,2321,,,,
173
+ ,,,85,2321,1.5909316539764404,1.5909316539764404,,
174
+ 1.5759717226028442,1.5759717226028442,1.4789810180664062,86,2348,,,,
175
+ ,,,86,2348,1.5548707246780396,1.5548707246780396,,
176
+ 2.038386821746826,2.038386821746826,1.4789810180664062,87,2375,,,,
177
+ ,,,87,2375,1.5722819566726685,1.5722819566726685,,
178
+ 1.49917733669281,1.49917733669281,1.4789810180664062,88,2402,,,,
179
+ ,,,88,2402,1.5589559078216553,1.5589559078216553,,
180
+ 1.7344645261764526,1.7344645261764526,1.4789810180664062,89,2429,,,,
181
+ ,,,89,2429,1.5571664571762085,1.5571664571762085,,
182
+ 4.255862712860107,4.255862712860107,1.4789810180664062,90,2456,,,,
183
+ ,,,90,2456,1.541688084602356,1.541688084602356,,
184
+ 16.791975021362305,16.791975021362305,1.4789810180664062,91,2483,,,,
185
+ ,,,91,2483,1.5512487888336182,1.5512487888336182,,
186
+ 1.7317806482315063,1.7317806482315063,1.4789810180664062,92,2510,,,,
187
+ ,,,92,2510,1.575599193572998,1.575599193572998,,
188
+ 1.5712475776672363,1.5712475776672363,1.4789810180664062,93,2537,,,,
189
+ ,,,93,2537,1.5583535432815552,1.5583535432815552,,
190
+ 2.531160593032837,2.531160593032837,1.4789810180664062,94,2564,,,,
191
+ ,,,94,2564,1.5624415874481201,1.5624415874481201,,
192
+ 1.6169700622558594,1.6169700622558594,1.4789810180664062,95,2591,,,,
193
+ ,,,95,2591,1.5781086683273315,1.5781086683273315,,
194
+ 4.9033613204956055,4.9033613204956055,1.4789810180664062,96,2618,,,,
195
+ ,,,96,2618,1.573669672012329,1.573669672012329,,
196
+ 1.681107759475708,1.681107759475708,1.4789810180664062,97,2645,,,,
197
+ ,,,97,2645,1.5686882734298706,1.5686882734298706,,
198
+ 1.5722702741622925,1.5722702741622925,1.4789810180664062,98,2672,,,,
199
+ ,,,98,2672,1.5601458549499512,1.5601458549499512,,
200
+ 1.940112590789795,1.940112590789795,1.4789810180664062,99,2699,,,,
201
+ ,,,99,2699,1.5403550863265991,1.5403550863265991,,
202
+ ,,,28,756,,,1.6055947542190552,1.6055947542190552
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646931592.37a92d36c649.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a8044884ff2266e669ba3c5c1d3840fb4e4250dc3e9def0e0423c2cf616ecd1
3
+ size 36426
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646933528.37a92d36c649.1.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bf59fbf508fb5f007752e2b2eb88d6e07b7ed718437c52dbddaa8b8b5bdfd24
3
+ size 179
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/0/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 512
11
+ lin2_size: 512
12
+ lin3_size: 1024
13
+ output_size: 1
14
+ lr: 0.03686537045425119
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 35353601
51
+ model/params/trainable: 35353601
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 1024
24
+ lin2_size: 64
25
+ lin3_size: 128
26
+ output_size: 1
27
+ lr: 0.12853081804637867
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.12853081804637867
171
+ - model.lin1_size=1024
172
+ - model.lin2_size=64
173
+ - model.lin3_size=128
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=1024,model.lin2_size=64,model.lin3_size=128,model.lr=0.12853081804637867,trainer.gpus=0
179
+ id: '1'
180
+ num: 1
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - datamodule.batch_size=64
2
+ - model.lr=0.12853081804637867
3
+ - model.lin1_size=1024
4
+ - model.lin2_size=64
5
+ - model.lin3_size=128
6
+ - hparams_search=focusMAE_150_optuna
7
+ - trainer.gpus=0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 1024
11
+ lin2_size: 64
12
+ lin3_size: 128
13
+ output_size: 1
14
+ lr: 0.12853081804637867
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 69197505
51
+ model/params/trainable: 69197505
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 1.5457894802093506,1.5457894802093506,1.5457894802093506,0,26,,,,
3
+ ,,,0,26,1.93147873878479,1.93147873878479,,
4
+ 1.518662452697754,1.518662452697754,1.518662452697754,1,53,,,,
5
+ ,,,1,53,1.5900448560714722,1.5900448560714722,,
6
+ 1.5237871408462524,1.5237871408462524,1.518662452697754,2,80,,,,
7
+ ,,,2,80,1.5704622268676758,1.5704622268676758,,
8
+ 1.5101613998413086,1.5101613998413086,1.5101613998413086,3,107,,,,
9
+ ,,,3,107,1.5570534467697144,1.5570534467697144,,
10
+ 1.5968084335327148,1.5968084335327148,1.5101613998413086,4,134,,,,
11
+ ,,,4,134,1.576980710029602,1.576980710029602,,
12
+ 1.4964513778686523,1.4964513778686523,1.4964513778686523,5,161,,,,
13
+ ,,,5,161,1.5659037828445435,1.5659037828445435,,
14
+ 1.5253665447235107,1.5253665447235107,1.4964513778686523,6,188,,,,
15
+ ,,,6,188,1.5655879974365234,1.5655879974365234,,
16
+ 1.6784342527389526,1.6784342527389526,1.4964513778686523,7,215,,,,
17
+ ,,,7,215,1.562648892402649,1.562648892402649,,
18
+ 1.513913869857788,1.513913869857788,1.4964513778686523,8,242,,,,
19
+ ,,,8,242,1.5687941312789917,1.5687941312789917,,
20
+ 4.911733627319336,4.911733627319336,1.4964513778686523,9,269,,,,
21
+ ,,,9,269,1.5633389949798584,1.5633389949798584,,
22
+ 3.7279114723205566,3.7279114723205566,1.4964513778686523,10,296,,,,
23
+ ,,,10,296,1.569581389427185,1.569581389427185,,
24
+ 1.623799204826355,1.623799204826355,1.4964513778686523,11,323,,,,
25
+ ,,,11,323,1.5841323137283325,1.5841323137283325,,
26
+ 3.1766417026519775,3.1766417026519775,1.4964513778686523,12,350,,,,
27
+ ,,,12,350,1.5681496858596802,1.5681496858596802,,
28
+ 1.5450650453567505,1.5450650453567505,1.4964513778686523,13,377,,,,
29
+ ,,,13,377,1.599021553993225,1.599021553993225,,
30
+ 1.7169760465621948,1.7169760465621948,1.4964513778686523,14,404,,,,
31
+ ,,,14,404,1.566733956336975,1.566733956336975,,
32
+ 1.7649874687194824,1.7649874687194824,1.4964513778686523,15,431,,,,
33
+ ,,,15,431,1.60713529586792,1.60713529586792,,
34
+ 1.5147674083709717,1.5147674083709717,1.4964513778686523,16,458,,,,
35
+ ,,,16,458,1.5965725183486938,1.5965725183486938,,
36
+ 1.5292353630065918,1.5292353630065918,1.4964513778686523,17,485,,,,
37
+ ,,,17,485,1.5831682682037354,1.5831682682037354,,
38
+ 1.4994009733200073,1.4994009733200073,1.4964513778686523,18,512,,,,
39
+ ,,,18,512,1.5718650817871094,1.5718650817871094,,
40
+ 1.5470983982086182,1.5470983982086182,1.4964513778686523,19,539,,,,
41
+ ,,,19,539,1.586369514465332,1.586369514465332,,
42
+ 1.5612688064575195,1.5612688064575195,1.4964513778686523,20,566,,,,
43
+ ,,,20,566,1.5651100873947144,1.5651100873947144,,
44
+ 10.558342933654785,10.558342933654785,1.4964513778686523,21,593,,,,
45
+ ,,,21,593,1.566519021987915,1.566519021987915,,
46
+ 1.752294898033142,1.752294898033142,1.4964513778686523,22,620,,,,
47
+ ,,,22,620,1.5655397176742554,1.5655397176742554,,
48
+ 4.056135654449463,4.056135654449463,1.4964513778686523,23,647,,,,
49
+ ,,,23,647,1.5815690755844116,1.5815690755844116,,
50
+ 1.5049797296524048,1.5049797296524048,1.4964513778686523,24,674,,,,
51
+ ,,,24,674,1.58075749874115,1.58075749874115,,
52
+ 1.5338116884231567,1.5338116884231567,1.4964513778686523,25,701,,,,
53
+ ,,,25,701,1.5529197454452515,1.5529197454452515,,
54
+ 1.552352786064148,1.552352786064148,1.4964513778686523,26,728,,,,
55
+ ,,,26,728,1.5741243362426758,1.5741243362426758,,
56
+ 1.524463176727295,1.524463176727295,1.4964513778686523,27,755,,,,
57
+ ,,,27,755,1.6238592863082886,1.6238592863082886,,
58
+ 1.8281292915344238,1.8281292915344238,1.4964513778686523,28,782,,,,
59
+ ,,,28,782,1.5760538578033447,1.5760538578033447,,
60
+ 1.513440489768982,1.513440489768982,1.4964513778686523,29,809,,,,
61
+ ,,,29,809,1.5889009237289429,1.5889009237289429,,
62
+ 28.168088912963867,28.168088912963867,1.4964513778686523,30,836,,,,
63
+ ,,,30,836,1.6001774072647095,1.6001774072647095,,
64
+ 1.5468473434448242,1.5468473434448242,1.4964513778686523,31,863,,,,
65
+ ,,,31,863,1.599015474319458,1.599015474319458,,
66
+ 4.666551113128662,4.666551113128662,1.4964513778686523,32,890,,,,
67
+ ,,,32,890,1.5989962816238403,1.5989962816238403,,
68
+ 1.578824758529663,1.578824758529663,1.4964513778686523,33,917,,,,
69
+ ,,,33,917,1.6038649082183838,1.6038649082183838,,
70
+ 1.69940984249115,1.69940984249115,1.4964513778686523,34,944,,,,
71
+ ,,,34,944,1.5932281017303467,1.5932281017303467,,
72
+ 1.6619060039520264,1.6619060039520264,1.4964513778686523,35,971,,,,
73
+ ,,,35,971,1.6159247159957886,1.6159247159957886,,
74
+ 5.854098796844482,5.854098796844482,1.4964513778686523,36,998,,,,
75
+ ,,,36,998,1.5982983112335205,1.5982983112335205,,
76
+ 1.5281752347946167,1.5281752347946167,1.4964513778686523,37,1025,,,,
77
+ ,,,37,1025,1.5835182666778564,1.5835182666778564,,
78
+ 2.8466644287109375,2.8466644287109375,1.4964513778686523,38,1052,,,,
79
+ ,,,38,1052,1.6168947219848633,1.6168947219848633,,
80
+ 1.520732045173645,1.520732045173645,1.4964513778686523,39,1079,,,,
81
+ ,,,39,1079,1.600001335144043,1.600001335144043,,
82
+ 6.837627410888672,6.837627410888672,1.4964513778686523,40,1106,,,,
83
+ ,,,40,1106,1.586019515991211,1.586019515991211,,
84
+ 1.5251741409301758,1.5251741409301758,1.4964513778686523,41,1133,,,,
85
+ ,,,41,1133,1.596055507659912,1.596055507659912,,
86
+ 1.5217574834823608,1.5217574834823608,1.4964513778686523,42,1160,,,,
87
+ ,,,42,1160,1.5796046257019043,1.5796046257019043,,
88
+ 1.8967450857162476,1.8967450857162476,1.4964513778686523,43,1187,,,,
89
+ ,,,43,1187,1.6189672946929932,1.6189672946929932,,
90
+ 3.550907850265503,3.550907850265503,1.4964513778686523,44,1214,,,,
91
+ ,,,44,1214,1.6081902980804443,1.6081902980804443,,
92
+ 1.53545343875885,1.53545343875885,1.4964513778686523,45,1241,,,,
93
+ ,,,45,1241,1.6205562353134155,1.6205562353134155,,
94
+ 1.5376633405685425,1.5376633405685425,1.4964513778686523,46,1268,,,,
95
+ ,,,46,1268,1.6020132303237915,1.6020132303237915,,
96
+ 2.720132827758789,2.720132827758789,1.4964513778686523,47,1295,,,,
97
+ ,,,47,1295,1.59677255153656,1.59677255153656,,
98
+ 1.5386885404586792,1.5386885404586792,1.4964513778686523,48,1322,,,,
99
+ ,,,48,1322,1.6013177633285522,1.6013177633285522,,
100
+ 1.9883198738098145,1.9883198738098145,1.4964513778686523,49,1349,,,,
101
+ ,,,49,1349,1.5789361000061035,1.5789361000061035,,
102
+ 1.593371033668518,1.593371033668518,1.4964513778686523,50,1376,,,,
103
+ ,,,50,1376,1.6008371114730835,1.6008371114730835,,
104
+ 1.5711454153060913,1.5711454153060913,1.4964513778686523,51,1403,,,,
105
+ ,,,51,1403,1.5797780752182007,1.5797780752182007,,
106
+ 2.4740052223205566,2.4740052223205566,1.4964513778686523,52,1430,,,,
107
+ ,,,52,1430,1.565558910369873,1.565558910369873,,
108
+ 1.5136743783950806,1.5136743783950806,1.4964513778686523,53,1457,,,,
109
+ ,,,53,1457,1.5826765298843384,1.5826765298843384,,
110
+ 1.5344640016555786,1.5344640016555786,1.4964513778686523,54,1484,,,,
111
+ ,,,54,1484,1.5864404439926147,1.5864404439926147,,
112
+ 13.738432884216309,13.738432884216309,1.4964513778686523,55,1511,,,,
113
+ ,,,55,1511,1.587600588798523,1.587600588798523,,
114
+ 42.461551666259766,42.461551666259766,1.4964513778686523,56,1538,,,,
115
+ ,,,56,1538,1.5916862487792969,1.5916862487792969,,
116
+ 1.5237947702407837,1.5237947702407837,1.4964513778686523,57,1565,,,,
117
+ ,,,57,1565,1.6166635751724243,1.6166635751724243,,
118
+ 5.10219144821167,5.10219144821167,1.4964513778686523,58,1592,,,,
119
+ ,,,58,1592,1.6150850057601929,1.6150850057601929,,
120
+ 1.550328016281128,1.550328016281128,1.4964513778686523,59,1619,,,,
121
+ ,,,59,1619,1.6161441802978516,1.6161441802978516,,
122
+ 4.7871222496032715,4.7871222496032715,1.4964513778686523,60,1646,,,,
123
+ ,,,60,1646,1.6073036193847656,1.6073036193847656,,
124
+ 1.5274826288223267,1.5274826288223267,1.4964513778686523,61,1673,,,,
125
+ ,,,61,1673,1.6093207597732544,1.6093207597732544,,
126
+ 1.8601120710372925,1.8601120710372925,1.4964513778686523,62,1700,,,,
127
+ ,,,62,1700,1.6090507507324219,1.6090507507324219,,
128
+ 1.539394736289978,1.539394736289978,1.4964513778686523,63,1727,,,,
129
+ ,,,63,1727,1.6039584875106812,1.6039584875106812,,
130
+ 1.5111101865768433,1.5111101865768433,1.4964513778686523,64,1754,,,,
131
+ ,,,64,1754,1.601883888244629,1.601883888244629,,
132
+ 1.5048625469207764,1.5048625469207764,1.4964513778686523,65,1781,,,,
133
+ ,,,65,1781,1.5932526588439941,1.5932526588439941,,
134
+ 1.5920627117156982,1.5920627117156982,1.4964513778686523,66,1808,,,,
135
+ ,,,66,1808,1.5937992334365845,1.5937992334365845,,
136
+ 1.5307120084762573,1.5307120084762573,1.4964513778686523,67,1835,,,,
137
+ ,,,67,1835,1.606019377708435,1.606019377708435,,
138
+ 1.526431679725647,1.526431679725647,1.4964513778686523,68,1862,,,,
139
+ ,,,68,1862,1.6092278957366943,1.6092278957366943,,
140
+ 1.527294635772705,1.527294635772705,1.4964513778686523,69,1889,,,,
141
+ ,,,69,1889,1.6061241626739502,1.6061241626739502,,
142
+ 1.535790205001831,1.535790205001831,1.4964513778686523,70,1916,,,,
143
+ ,,,70,1916,1.6027199029922485,1.6027199029922485,,
144
+ 7.108047962188721,7.108047962188721,1.4964513778686523,71,1943,,,,
145
+ ,,,71,1943,1.60779869556427,1.60779869556427,,
146
+ 1.5393939018249512,1.5393939018249512,1.4964513778686523,72,1970,,,,
147
+ ,,,72,1970,1.624327540397644,1.624327540397644,,
148
+ 1.526472806930542,1.526472806930542,1.4964513778686523,73,1997,,,,
149
+ ,,,73,1997,1.5907820463180542,1.5907820463180542,,
150
+ 1.5240964889526367,1.5240964889526367,1.4964513778686523,74,2024,,,,
151
+ ,,,74,2024,1.6047561168670654,1.6047561168670654,,
152
+ 9.961384773254395,9.961384773254395,1.4964513778686523,75,2051,,,,
153
+ ,,,75,2051,1.5908408164978027,1.5908408164978027,,
154
+ 1.5271530151367188,1.5271530151367188,1.4964513778686523,76,2078,,,,
155
+ ,,,76,2078,1.6088954210281372,1.6088954210281372,,
156
+ 1.5703331232070923,1.5703331232070923,1.4964513778686523,77,2105,,,,
157
+ ,,,77,2105,1.6130024194717407,1.6130024194717407,,
158
+ 1.6119712591171265,1.6119712591171265,1.4964513778686523,78,2132,,,,
159
+ ,,,78,2132,1.6129977703094482,1.6129977703094482,,
160
+ 1.5369189977645874,1.5369189977645874,1.4964513778686523,79,2159,,,,
161
+ ,,,79,2159,1.5994534492492676,1.5994534492492676,,
162
+ 6.256889820098877,6.256889820098877,1.4964513778686523,80,2186,,,,
163
+ ,,,80,2186,1.5921787023544312,1.5921787023544312,,
164
+ 1.5237727165222168,1.5237727165222168,1.4964513778686523,81,2213,,,,
165
+ ,,,81,2213,1.5913481712341309,1.5913481712341309,,
166
+ 1.5433704853057861,1.5433704853057861,1.4964513778686523,82,2240,,,,
167
+ ,,,82,2240,1.6063846349716187,1.6063846349716187,,
168
+ 1.5143684148788452,1.5143684148788452,1.4964513778686523,83,2267,,,,
169
+ ,,,83,2267,1.6061906814575195,1.6061906814575195,,
170
+ 1.5512630939483643,1.5512630939483643,1.4964513778686523,84,2294,,,,
171
+ ,,,84,2294,1.61200749874115,1.61200749874115,,
172
+ 1.748395323753357,1.748395323753357,1.4964513778686523,85,2321,,,,
173
+ ,,,85,2321,1.6335575580596924,1.6335575580596924,,
174
+ 1.560850739479065,1.560850739479065,1.4964513778686523,86,2348,,,,
175
+ ,,,86,2348,1.6014795303344727,1.6014795303344727,,
176
+ 1.5232243537902832,1.5232243537902832,1.4964513778686523,87,2375,,,,
177
+ ,,,87,2375,1.6086633205413818,1.6086633205413818,,
178
+ 1.5236783027648926,1.5236783027648926,1.4964513778686523,88,2402,,,,
179
+ ,,,88,2402,1.601511836051941,1.601511836051941,,
180
+ 1.672797679901123,1.672797679901123,1.4964513778686523,89,2429,,,,
181
+ ,,,89,2429,1.6028125286102295,1.6028125286102295,,
182
+ 1.5533705949783325,1.5533705949783325,1.4964513778686523,90,2456,,,,
183
+ ,,,90,2456,1.5999672412872314,1.5999672412872314,,
184
+ 1.5226218700408936,1.5226218700408936,1.4964513778686523,91,2483,,,,
185
+ ,,,91,2483,1.6082600355148315,1.6082600355148315,,
186
+ 1.5232043266296387,1.5232043266296387,1.4964513778686523,92,2510,,,,
187
+ ,,,92,2510,1.6300674676895142,1.6300674676895142,,
188
+ 1.5347774028778076,1.5347774028778076,1.4964513778686523,93,2537,,,,
189
+ ,,,93,2537,1.6055806875228882,1.6055806875228882,,
190
+ 42.29332733154297,42.29332733154297,1.4964513778686523,94,2564,,,,
191
+ ,,,94,2564,1.6171374320983887,1.6171374320983887,,
192
+ 7.51504373550415,7.51504373550415,1.4964513778686523,95,2591,,,,
193
+ ,,,95,2591,1.6072918176651,1.6072918176651,,
194
+ 1.5361765623092651,1.5361765623092651,1.4964513778686523,96,2618,,,,
195
+ ,,,96,2618,1.6097540855407715,1.6097540855407715,,
196
+ 10.45922565460205,10.45922565460205,1.4964513778686523,97,2645,,,,
197
+ ,,,97,2645,1.6088382005691528,1.6088382005691528,,
198
+ 1.8529367446899414,1.8529367446899414,1.4964513778686523,98,2672,,,,
199
+ ,,,98,2672,1.6240711212158203,1.6240711212158203,,
200
+ 1.5809720754623413,1.5809720754623413,1.4964513778686523,99,2699,,,,
201
+ ,,,99,2699,1.61771821975708,1.61771821975708,,
202
+ ,,,6,162,,,1.6138743162155151,1.6138743162155151
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646933529.37a92d36c649.1.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e5b01d3df96e71bf8fd7431c9e7910a218d168fee07c1fe3fda2d03dbd6a8d0
3
+ size 36426
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646937064.37a92d36c649.1.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2855e1b2dcfbc5fd0ced549aa91a3c096e1bb04ca44be6075e440128d517bf6
3
+ size 179
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/1/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 1024
11
+ lin2_size: 64
12
+ lin3_size: 128
13
+ output_size: 1
14
+ lr: 0.12853081804637867
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 69197505
51
+ model/params/trainable: 69197505
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 256
24
+ lin2_size: 128
25
+ lin3_size: 256
26
+ output_size: 1
27
+ lr: 0.02282695974132741
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.02282695974132741
171
+ - model.lin1_size=256
172
+ - model.lin2_size=128
173
+ - model.lin3_size=256
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=256,model.lr=0.02282695974132741,trainer.gpus=0
179
+ id: '10'
180
+ num: 10
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - datamodule.batch_size=64
2
+ - model.lr=0.02282695974132741
3
+ - model.lin1_size=256
4
+ - model.lin2_size=128
5
+ - model.lin3_size=256
6
+ - hparams_search=focusMAE_150_optuna
7
+ - trainer.gpus=0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.02282695974132741
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 1.54915452003479,1.54915452003479,1.54915452003479,0,26,,,,
3
+ ,,,0,26,1.7947152853012085,1.7947152853012085,,
4
+ 1.4813045263290405,1.4813045263290405,1.4813045263290405,1,53,,,,
5
+ ,,,1,53,1.5661964416503906,1.5661964416503906,,
6
+ 1.4941850900650024,1.4941850900650024,1.4813045263290405,2,80,,,,
7
+ ,,,2,80,1.5476306676864624,1.5476306676864624,,
8
+ 1.559279441833496,1.559279441833496,1.4813045263290405,3,107,,,,
9
+ ,,,3,107,1.5460481643676758,1.5460481643676758,,
10
+ 1.5001925230026245,1.5001925230026245,1.4813045263290405,4,134,,,,
11
+ ,,,4,134,1.5567209720611572,1.5567209720611572,,
12
+ 1.5714366436004639,1.5714366436004639,1.4813045263290405,5,161,,,,
13
+ ,,,5,161,1.5221930742263794,1.5221930742263794,,
14
+ 1.5059216022491455,1.5059216022491455,1.4813045263290405,6,188,,,,
15
+ ,,,6,188,1.5445741415023804,1.5445741415023804,,
16
+ 1.5441499948501587,1.5441499948501587,1.4813045263290405,7,215,,,,
17
+ ,,,7,215,1.5316859483718872,1.5316859483718872,,
18
+ 1.5908784866333008,1.5908784866333008,1.4813045263290405,8,242,,,,
19
+ ,,,8,242,1.556628704071045,1.556628704071045,,
20
+ 1.4949480295181274,1.4949480295181274,1.4813045263290405,9,269,,,,
21
+ ,,,9,269,1.5524144172668457,1.5524144172668457,,
22
+ 1.4900106191635132,1.4900106191635132,1.4813045263290405,10,296,,,,
23
+ ,,,10,296,1.5266354084014893,1.5266354084014893,,
24
+ 1.4854364395141602,1.4854364395141602,1.4813045263290405,11,323,,,,
25
+ ,,,11,323,1.5311683416366577,1.5311683416366577,,
26
+ 1.5203735828399658,1.5203735828399658,1.4813045263290405,12,350,,,,
27
+ ,,,12,350,1.535923957824707,1.535923957824707,,
28
+ 1.498347520828247,1.498347520828247,1.4813045263290405,13,377,,,,
29
+ ,,,13,377,1.5193228721618652,1.5193228721618652,,
30
+ 1.5191665887832642,1.5191665887832642,1.4813045263290405,14,404,,,,
31
+ ,,,14,404,1.5163968801498413,1.5163968801498413,,
32
+ 1.8561465740203857,1.8561465740203857,1.4813045263290405,15,431,,,,
33
+ ,,,15,431,1.5274927616119385,1.5274927616119385,,
34
+ 1.5576380491256714,1.5576380491256714,1.4813045263290405,16,458,,,,
35
+ ,,,16,458,1.5263643264770508,1.5263643264770508,,
36
+ 1.5795142650604248,1.5795142650604248,1.4813045263290405,17,485,,,,
37
+ ,,,17,485,1.5001400709152222,1.5001400709152222,,
38
+ 1.4902113676071167,1.4902113676071167,1.4813045263290405,18,512,,,,
39
+ ,,,18,512,1.516677737236023,1.516677737236023,,
40
+ 5.073453903198242,5.073453903198242,1.4813045263290405,19,539,,,,
41
+ ,,,19,539,1.5088609457015991,1.5088609457015991,,
42
+ 1.933295726776123,1.933295726776123,1.4813045263290405,20,566,,,,
43
+ ,,,20,566,1.5014426708221436,1.5014426708221436,,
44
+ 1.5227258205413818,1.5227258205413818,1.4813045263290405,21,593,,,,
45
+ ,,,21,593,1.5077985525131226,1.5077985525131226,,
46
+ 1.5633625984191895,1.5633625984191895,1.4813045263290405,22,620,,,,
47
+ ,,,22,620,1.479738712310791,1.479738712310791,,
48
+ 1.7504345178604126,1.7504345178604126,1.4813045263290405,23,647,,,,
49
+ ,,,23,647,1.4974632263183594,1.4974632263183594,,
50
+ 1.7892953157424927,1.7892953157424927,1.4813045263290405,24,674,,,,
51
+ ,,,24,674,1.5153311491012573,1.5153311491012573,,
52
+ 1.5206643342971802,1.5206643342971802,1.4813045263290405,25,701,,,,
53
+ ,,,25,701,1.5057305097579956,1.5057305097579956,,
54
+ 1.8115839958190918,1.8115839958190918,1.4813045263290405,26,728,,,,
55
+ ,,,26,728,1.5112546682357788,1.5112546682357788,,
56
+ 2.030661106109619,2.030661106109619,1.4813045263290405,27,755,,,,
57
+ ,,,27,755,1.5197839736938477,1.5197839736938477,,
58
+ 1.86220383644104,1.86220383644104,1.4813045263290405,28,782,,,,
59
+ ,,,28,782,1.492752194404602,1.492752194404602,,
60
+ 1.561750054359436,1.561750054359436,1.4813045263290405,29,809,,,,
61
+ ,,,29,809,1.5225181579589844,1.5225181579589844,,
62
+ 2.5055038928985596,2.5055038928985596,1.4813045263290405,30,836,,,,
63
+ ,,,30,836,1.488628625869751,1.488628625869751,,
64
+ 1.4965015649795532,1.4965015649795532,1.4813045263290405,31,863,,,,
65
+ ,,,31,863,1.519391417503357,1.519391417503357,,
66
+ 1.5533944368362427,1.5533944368362427,1.4813045263290405,32,890,,,,
67
+ ,,,32,890,1.4976305961608887,1.4976305961608887,,
68
+ 2.142780065536499,2.142780065536499,1.4813045263290405,33,917,,,,
69
+ ,,,33,917,1.4807366132736206,1.4807366132736206,,
70
+ 1.472182273864746,1.472182273864746,1.472182273864746,34,944,,,,
71
+ ,,,34,944,1.5067917108535767,1.5067917108535767,,
72
+ 1.8617546558380127,1.8617546558380127,1.472182273864746,35,971,,,,
73
+ ,,,35,971,1.4976863861083984,1.4976863861083984,,
74
+ 2.134965419769287,2.134965419769287,1.472182273864746,36,998,,,,
75
+ ,,,36,998,1.4933010339736938,1.4933010339736938,,
76
+ 1.52902352809906,1.52902352809906,1.472182273864746,37,1025,,,,
77
+ ,,,37,1025,1.4979575872421265,1.4979575872421265,,
78
+ 1.608336091041565,1.608336091041565,1.472182273864746,38,1052,,,,
79
+ ,,,38,1052,1.4944689273834229,1.4944689273834229,,
80
+ 2.0778658390045166,2.0778658390045166,1.472182273864746,39,1079,,,,
81
+ ,,,39,1079,1.4775594472885132,1.4775594472885132,,
82
+ 1.5630720853805542,1.5630720853805542,1.472182273864746,40,1106,,,,
83
+ ,,,40,1106,1.5007765293121338,1.5007765293121338,,
84
+ 1.4986320734024048,1.4986320734024048,1.472182273864746,41,1133,,,,
85
+ ,,,41,1133,1.499476432800293,1.499476432800293,,
86
+ 1.5532052516937256,1.5532052516937256,1.472182273864746,42,1160,,,,
87
+ ,,,42,1160,1.5045232772827148,1.5045232772827148,,
88
+ 1.5133055448532104,1.5133055448532104,1.472182273864746,43,1187,,,,
89
+ ,,,43,1187,1.51310133934021,1.51310133934021,,
90
+ 1.640570878982544,1.640570878982544,1.472182273864746,44,1214,,,,
91
+ ,,,44,1214,1.5313445329666138,1.5313445329666138,,
92
+ 2.2327382564544678,2.2327382564544678,1.472182273864746,45,1241,,,,
93
+ ,,,45,1241,1.5123286247253418,1.5123286247253418,,
94
+ 1.5145050287246704,1.5145050287246704,1.472182273864746,46,1268,,,,
95
+ ,,,46,1268,1.5127230882644653,1.5127230882644653,,
96
+ 1.5009477138519287,1.5009477138519287,1.472182273864746,47,1295,,,,
97
+ ,,,47,1295,1.5293209552764893,1.5293209552764893,,
98
+ 1.5876779556274414,1.5876779556274414,1.472182273864746,48,1322,,,,
99
+ ,,,48,1322,1.503609538078308,1.503609538078308,,
100
+ 1.7818883657455444,1.7818883657455444,1.472182273864746,49,1349,,,,
101
+ ,,,49,1349,1.5058997869491577,1.5058997869491577,,
102
+ 1.4677724838256836,1.4677724838256836,1.4677724838256836,50,1376,,,,
103
+ ,,,50,1376,1.5166460275650024,1.5166460275650024,,
104
+ 1.531506061553955,1.531506061553955,1.4677724838256836,51,1403,,,,
105
+ ,,,51,1403,1.5017387866973877,1.5017387866973877,,
106
+ 1.881100058555603,1.881100058555603,1.4677724838256836,52,1430,,,,
107
+ ,,,52,1430,1.5125443935394287,1.5125443935394287,,
108
+ 1.4942419528961182,1.4942419528961182,1.4677724838256836,53,1457,,,,
109
+ ,,,53,1457,1.5166507959365845,1.5166507959365845,,
110
+ 1.692086100578308,1.692086100578308,1.4677724838256836,54,1484,,,,
111
+ ,,,54,1484,1.5313019752502441,1.5313019752502441,,
112
+ 1.6460152864456177,1.6460152864456177,1.4677724838256836,55,1511,,,,
113
+ ,,,55,1511,1.4998795986175537,1.4998795986175537,,
114
+ 1.836544394493103,1.836544394493103,1.4677724838256836,56,1538,,,,
115
+ ,,,56,1538,1.506105899810791,1.506105899810791,,
116
+ 1.5427119731903076,1.5427119731903076,1.4677724838256836,57,1565,,,,
117
+ ,,,57,1565,1.4999603033065796,1.4999603033065796,,
118
+ 1.5046226978302002,1.5046226978302002,1.4677724838256836,58,1592,,,,
119
+ ,,,58,1592,1.5169248580932617,1.5169248580932617,,
120
+ 1.5429843664169312,1.5429843664169312,1.4677724838256836,59,1619,,,,
121
+ ,,,59,1619,1.5131078958511353,1.5131078958511353,,
122
+ 1.6014117002487183,1.6014117002487183,1.4677724838256836,60,1646,,,,
123
+ ,,,60,1646,1.5238933563232422,1.5238933563232422,,
124
+ 1.5440922975540161,1.5440922975540161,1.4677724838256836,61,1673,,,,
125
+ ,,,61,1673,1.5364078283309937,1.5364078283309937,,
126
+ 1.6102505922317505,1.6102505922317505,1.4677724838256836,62,1700,,,,
127
+ ,,,62,1700,1.5143425464630127,1.5143425464630127,,
128
+ 1.5728535652160645,1.5728535652160645,1.4677724838256836,63,1727,,,,
129
+ ,,,63,1727,1.5125833749771118,1.5125833749771118,,
130
+ 1.593582272529602,1.593582272529602,1.4677724838256836,64,1754,,,,
131
+ ,,,64,1754,1.527673363685608,1.527673363685608,,
132
+ 1.521716594696045,1.521716594696045,1.4677724838256836,65,1781,,,,
133
+ ,,,65,1781,1.532250165939331,1.532250165939331,,
134
+ 1.67029869556427,1.67029869556427,1.4677724838256836,66,1808,,,,
135
+ ,,,66,1808,1.509299635887146,1.509299635887146,,
136
+ 2.1808524131774902,2.1808524131774902,1.4677724838256836,67,1835,,,,
137
+ ,,,67,1835,1.5237321853637695,1.5237321853637695,,
138
+ 1.7052583694458008,1.7052583694458008,1.4677724838256836,68,1862,,,,
139
+ ,,,68,1862,1.5353375673294067,1.5353375673294067,,
140
+ 1.671800971031189,1.671800971031189,1.4677724838256836,69,1889,,,,
141
+ ,,,69,1889,1.5197367668151855,1.5197367668151855,,
142
+ 1.5118688344955444,1.5118688344955444,1.4677724838256836,70,1916,,,,
143
+ ,,,70,1916,1.5191370248794556,1.5191370248794556,,
144
+ 1.5275362730026245,1.5275362730026245,1.4677724838256836,71,1943,,,,
145
+ ,,,71,1943,1.5351717472076416,1.5351717472076416,,
146
+ 1.5210256576538086,1.5210256576538086,1.4677724838256836,72,1970,,,,
147
+ ,,,72,1970,1.5208426713943481,1.5208426713943481,,
148
+ 1.654943585395813,1.654943585395813,1.4677724838256836,73,1997,,,,
149
+ ,,,73,1997,1.530107021331787,1.530107021331787,,
150
+ 1.5063841342926025,1.5063841342926025,1.4677724838256836,74,2024,,,,
151
+ ,,,74,2024,1.5324897766113281,1.5324897766113281,,
152
+ 1.9869394302368164,1.9869394302368164,1.4677724838256836,75,2051,,,,
153
+ ,,,75,2051,1.5222572088241577,1.5222572088241577,,
154
+ 1.620804786682129,1.620804786682129,1.4677724838256836,76,2078,,,,
155
+ ,,,76,2078,1.5140538215637207,1.5140538215637207,,
156
+ 1.527064561843872,1.527064561843872,1.4677724838256836,77,2105,,,,
157
+ ,,,77,2105,1.5159393548965454,1.5159393548965454,,
158
+ 1.7091588973999023,1.7091588973999023,1.4677724838256836,78,2132,,,,
159
+ ,,,78,2132,1.5180113315582275,1.5180113315582275,,
160
+ 1.9504739046096802,1.9504739046096802,1.4677724838256836,79,2159,,,,
161
+ ,,,79,2159,1.5056763887405396,1.5056763887405396,,
162
+ 1.5281766653060913,1.5281766653060913,1.4677724838256836,80,2186,,,,
163
+ ,,,80,2186,1.516516089439392,1.516516089439392,,
164
+ 2.3494374752044678,2.3494374752044678,1.4677724838256836,81,2213,,,,
165
+ ,,,81,2213,1.513085126876831,1.513085126876831,,
166
+ 1.5083938837051392,1.5083938837051392,1.4677724838256836,82,2240,,,,
167
+ ,,,82,2240,1.5092164278030396,1.5092164278030396,,
168
+ 1.9437925815582275,1.9437925815582275,1.4677724838256836,83,2267,,,,
169
+ ,,,83,2267,1.5361279249191284,1.5361279249191284,,
170
+ 1.4560555219650269,1.4560555219650269,1.4560555219650269,84,2294,,,,
171
+ ,,,84,2294,1.5336685180664062,1.5336685180664062,,
172
+ 1.499437689781189,1.499437689781189,1.4560555219650269,85,2321,,,,
173
+ ,,,85,2321,1.5202487707138062,1.5202487707138062,,
174
+ 1.5475561618804932,1.5475561618804932,1.4560555219650269,86,2348,,,,
175
+ ,,,86,2348,1.5198792219161987,1.5198792219161987,,
176
+ 1.6395263671875,1.6395263671875,1.4560555219650269,87,2375,,,,
177
+ ,,,87,2375,1.5371838808059692,1.5371838808059692,,
178
+ 1.561975359916687,1.561975359916687,1.4560555219650269,88,2402,,,,
179
+ ,,,88,2402,1.5324909687042236,1.5324909687042236,,
180
+ 1.5383996963500977,1.5383996963500977,1.4560555219650269,89,2429,,,,
181
+ ,,,89,2429,1.5167973041534424,1.5167973041534424,,
182
+ 1.5120161771774292,1.5120161771774292,1.4560555219650269,90,2456,,,,
183
+ ,,,90,2456,1.5161545276641846,1.5161545276641846,,
184
+ 1.5026012659072876,1.5026012659072876,1.4560555219650269,91,2483,,,,
185
+ ,,,91,2483,1.5210195779800415,1.5210195779800415,,
186
+ 1.5586644411087036,1.5586644411087036,1.4560555219650269,92,2510,,,,
187
+ ,,,92,2510,1.5279392004013062,1.5279392004013062,,
188
+ 1.4975413084030151,1.4975413084030151,1.4560555219650269,93,2537,,,,
189
+ ,,,93,2537,1.5191806554794312,1.5191806554794312,,
190
+ 1.6976767778396606,1.6976767778396606,1.4560555219650269,94,2564,,,,
191
+ ,,,94,2564,1.5128707885742188,1.5128707885742188,,
192
+ 2.470996856689453,2.470996856689453,1.4560555219650269,95,2591,,,,
193
+ ,,,95,2591,1.5361442565917969,1.5361442565917969,,
194
+ 1.5277329683303833,1.5277329683303833,1.4560555219650269,96,2618,,,,
195
+ ,,,96,2618,1.5311845541000366,1.5311845541000366,,
196
+ 1.5454241037368774,1.5454241037368774,1.4560555219650269,97,2645,,,,
197
+ ,,,97,2645,1.5216116905212402,1.5216116905212402,,
198
+ 1.930446743965149,1.930446743965149,1.4560555219650269,98,2672,,,,
199
+ ,,,98,2672,1.5277382135391235,1.5277382135391235,,
200
+ 2.723839044570923,2.723839044570923,1.4560555219650269,99,2699,,,,
201
+ ,,,99,2699,1.5109692811965942,1.5109692811965942,,
202
+ ,,,85,2295,,,1.6256794929504395,1.6256794929504395
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646950714.37a92d36c649.1.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7b9683865febbf1b51407a8fbfcb1d3c1632d7c08ec1dd533338c91148f1210
3
+ size 36426
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646951667.37a92d36c649.1.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d56341fd9fda699ba019fcd5dd2ee48c2d1955e371489950a745656c17e0c39
3
+ size 179
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/10/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.02282695974132741
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 256
24
+ lin2_size: 128
25
+ lin3_size: 256
26
+ output_size: 1
27
+ lr: 0.0051297750484612
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.0051297750484612
171
+ - model.lin1_size=256
172
+ - model.lin2_size=128
173
+ - model.lin3_size=256
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=256,model.lr=0.0051297750484612,trainer.gpus=0
179
+ id: '11'
180
+ num: 11
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - datamodule.batch_size=64
2
+ - model.lr=0.0051297750484612
3
+ - model.lin1_size=256
4
+ - model.lin2_size=128
5
+ - model.lin3_size=256
6
+ - hparams_search=focusMAE_150_optuna
7
+ - trainer.gpus=0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.0051297750484612
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 1.5529874563217163,1.5529874563217163,1.5529874563217163,0,26,,,,
3
+ ,,,0,26,1.7005412578582764,1.7005412578582764,,
4
+ 1.4758678674697876,1.4758678674697876,1.4758678674697876,1,53,,,,
5
+ ,,,1,53,1.552437424659729,1.552437424659729,,
6
+ 1.5317124128341675,1.5317124128341675,1.4758678674697876,2,80,,,,
7
+ ,,,2,80,1.5342624187469482,1.5342624187469482,,
8
+ 1.5215035676956177,1.5215035676956177,1.4758678674697876,3,107,,,,
9
+ ,,,3,107,1.5154995918273926,1.5154995918273926,,
10
+ 1.5599119663238525,1.5599119663238525,1.4758678674697876,4,134,,,,
11
+ ,,,4,134,1.4910739660263062,1.4910739660263062,,
12
+ 1.5603760480880737,1.5603760480880737,1.4758678674697876,5,161,,,,
13
+ ,,,5,161,1.4519827365875244,1.4519827365875244,,
14
+ 1.7191616296768188,1.7191616296768188,1.4758678674697876,6,188,,,,
15
+ ,,,6,188,1.445921540260315,1.445921540260315,,
16
+ 1.5689905881881714,1.5689905881881714,1.4758678674697876,7,215,,,,
17
+ ,,,7,215,1.4619702100753784,1.4619702100753784,,
18
+ 1.5409839153289795,1.5409839153289795,1.4758678674697876,8,242,,,,
19
+ ,,,8,242,1.4217779636383057,1.4217779636383057,,
20
+ 1.5249755382537842,1.5249755382537842,1.4758678674697876,9,269,,,,
21
+ ,,,9,269,1.432146668434143,1.432146668434143,,
22
+ 1.591946005821228,1.591946005821228,1.4758678674697876,10,296,,,,
23
+ ,,,10,296,1.4111212491989136,1.4111212491989136,,
24
+ 1.5828744173049927,1.5828744173049927,1.4758678674697876,11,323,,,,
25
+ ,,,11,323,1.4136533737182617,1.4136533737182617,,
26
+ 1.5061739683151245,1.5061739683151245,1.4758678674697876,12,350,,,,
27
+ ,,,12,350,1.410377025604248,1.410377025604248,,
28
+ 1.599090576171875,1.599090576171875,1.4758678674697876,13,377,,,,
29
+ ,,,13,377,1.4391021728515625,1.4391021728515625,,
30
+ 1.6281280517578125,1.6281280517578125,1.4758678674697876,14,404,,,,
31
+ ,,,14,404,1.4367587566375732,1.4367587566375732,,
32
+ 1.5029304027557373,1.5029304027557373,1.4758678674697876,15,431,,,,
33
+ ,,,15,431,1.4189836978912354,1.4189836978912354,,
34
+ 1.5762006044387817,1.5762006044387817,1.4758678674697876,16,458,,,,
35
+ ,,,16,458,1.4034911394119263,1.4034911394119263,,
36
+ 1.6103154420852661,1.6103154420852661,1.4758678674697876,17,485,,,,
37
+ ,,,17,485,1.431480050086975,1.431480050086975,,
38
+ 1.5456775426864624,1.5456775426864624,1.4758678674697876,18,512,,,,
39
+ ,,,18,512,1.4067394733428955,1.4067394733428955,,
40
+ 1.4778574705123901,1.4778574705123901,1.4758678674697876,19,539,,,,
41
+ ,,,19,539,1.4214212894439697,1.4214212894439697,,
42
+ 1.7477492094039917,1.7477492094039917,1.4758678674697876,20,566,,,,
43
+ ,,,20,566,1.411471962928772,1.411471962928772,,
44
+ 1.6120436191558838,1.6120436191558838,1.4758678674697876,21,593,,,,
45
+ ,,,21,593,1.4197815656661987,1.4197815656661987,,
46
+ 1.5599509477615356,1.5599509477615356,1.4758678674697876,22,620,,,,
47
+ ,,,22,620,1.4050796031951904,1.4050796031951904,,
48
+ 1.5538502931594849,1.5538502931594849,1.4758678674697876,23,647,,,,
49
+ ,,,23,647,1.404669165611267,1.404669165611267,,
50
+ 2.643871784210205,2.643871784210205,1.4758678674697876,24,674,,,,
51
+ ,,,24,674,1.3815333843231201,1.3815333843231201,,
52
+ 2.229886531829834,2.229886531829834,1.4758678674697876,25,701,,,,
53
+ ,,,25,701,1.3973186016082764,1.3973186016082764,,
54
+ 1.6227734088897705,1.6227734088897705,1.4758678674697876,26,728,,,,
55
+ ,,,26,728,1.4046413898468018,1.4046413898468018,,
56
+ 1.9245754480361938,1.9245754480361938,1.4758678674697876,27,755,,,,
57
+ ,,,27,755,1.372873067855835,1.372873067855835,,
58
+ 1.6182373762130737,1.6182373762130737,1.4758678674697876,28,782,,,,
59
+ ,,,28,782,1.4318004846572876,1.4318004846572876,,
60
+ 1.695565104484558,1.695565104484558,1.4758678674697876,29,809,,,,
61
+ ,,,29,809,1.4029852151870728,1.4029852151870728,,
62
+ 1.4891357421875,1.4891357421875,1.4758678674697876,30,836,,,,
63
+ ,,,30,836,1.3658686876296997,1.3658686876296997,,
64
+ 1.4806747436523438,1.4806747436523438,1.4758678674697876,31,863,,,,
65
+ ,,,31,863,1.4287527799606323,1.4287527799606323,,
66
+ 1.5332704782485962,1.5332704782485962,1.4758678674697876,32,890,,,,
67
+ ,,,32,890,1.4089083671569824,1.4089083671569824,,
68
+ 1.6101855039596558,1.6101855039596558,1.4758678674697876,33,917,,,,
69
+ ,,,33,917,1.4272220134735107,1.4272220134735107,,
70
+ 2.158543586730957,2.158543586730957,1.4758678674697876,34,944,,,,
71
+ ,,,34,944,1.3930636644363403,1.3930636644363403,,
72
+ 1.7488014698028564,1.7488014698028564,1.4758678674697876,35,971,,,,
73
+ ,,,35,971,1.3615875244140625,1.3615875244140625,,
74
+ 1.5668308734893799,1.5668308734893799,1.4758678674697876,36,998,,,,
75
+ ,,,36,998,1.3733593225479126,1.3733593225479126,,
76
+ 1.4741945266723633,1.4741945266723633,1.4741945266723633,37,1025,,,,
77
+ ,,,37,1025,1.3845897912979126,1.3845897912979126,,
78
+ 1.489251971244812,1.489251971244812,1.4741945266723633,38,1052,,,,
79
+ ,,,38,1052,1.3629571199417114,1.3629571199417114,,
80
+ 1.543493390083313,1.543493390083313,1.4741945266723633,39,1079,,,,
81
+ ,,,39,1079,1.3832662105560303,1.3832662105560303,,
82
+ 2.1800224781036377,2.1800224781036377,1.4741945266723633,40,1106,,,,
83
+ ,,,40,1106,1.348641037940979,1.348641037940979,,
84
+ 1.4829375743865967,1.4829375743865967,1.4741945266723633,41,1133,,,,
85
+ ,,,41,1133,1.400350570678711,1.400350570678711,,
86
+ 1.643622636795044,1.643622636795044,1.4741945266723633,42,1160,,,,
87
+ ,,,42,1160,1.3576005697250366,1.3576005697250366,,
88
+ 2.2521026134490967,2.2521026134490967,1.4741945266723633,43,1187,,,,
89
+ ,,,43,1187,1.3368573188781738,1.3368573188781738,,
90
+ 1.4667986631393433,1.4667986631393433,1.4667986631393433,44,1214,,,,
91
+ ,,,44,1214,1.3384252786636353,1.3384252786636353,,
92
+ 2.0049800872802734,2.0049800872802734,1.4667986631393433,45,1241,,,,
93
+ ,,,45,1241,1.3784995079040527,1.3784995079040527,,
94
+ 2.3140408992767334,2.3140408992767334,1.4667986631393433,46,1268,,,,
95
+ ,,,46,1268,1.3434473276138306,1.3434473276138306,,
96
+ 1.4815415143966675,1.4815415143966675,1.4667986631393433,47,1295,,,,
97
+ ,,,47,1295,1.355631947517395,1.355631947517395,,
98
+ 3.095125675201416,3.095125675201416,1.4667986631393433,48,1322,,,,
99
+ ,,,48,1322,1.3348404169082642,1.3348404169082642,,
100
+ 1.6569138765335083,1.6569138765335083,1.4667986631393433,49,1349,,,,
101
+ ,,,49,1349,1.3125057220458984,1.3125057220458984,,
102
+ 2.4747583866119385,2.4747583866119385,1.4667986631393433,50,1376,,,,
103
+ ,,,50,1376,1.2978061437606812,1.2978061437606812,,
104
+ 1.7729928493499756,1.7729928493499756,1.4667986631393433,51,1403,,,,
105
+ ,,,51,1403,1.2893775701522827,1.2893775701522827,,
106
+ 1.4477797746658325,1.4477797746658325,1.4477797746658325,52,1430,,,,
107
+ ,,,52,1430,1.3354593515396118,1.3354593515396118,,
108
+ 1.565670132637024,1.565670132637024,1.4477797746658325,53,1457,,,,
109
+ ,,,53,1457,1.2869794368743896,1.2869794368743896,,
110
+ 1.6527693271636963,1.6527693271636963,1.4477797746658325,54,1484,,,,
111
+ ,,,54,1484,1.3248947858810425,1.3248947858810425,,
112
+ 1.609217882156372,1.609217882156372,1.4477797746658325,55,1511,,,,
113
+ ,,,55,1511,1.276477336883545,1.276477336883545,,
114
+ 2.0620369911193848,2.0620369911193848,1.4477797746658325,56,1538,,,,
115
+ ,,,56,1538,1.3122347593307495,1.3122347593307495,,
116
+ 1.6234723329544067,1.6234723329544067,1.4477797746658325,57,1565,,,,
117
+ ,,,57,1565,1.2983700037002563,1.2983700037002563,,
118
+ 1.3670010566711426,1.3670010566711426,1.3670010566711426,58,1592,,,,
119
+ ,,,58,1592,1.3156099319458008,1.3156099319458008,,
120
+ 1.5603337287902832,1.5603337287902832,1.3670010566711426,59,1619,,,,
121
+ ,,,59,1619,1.30414617061615,1.30414617061615,,
122
+ 4.9891228675842285,4.9891228675842285,1.3670010566711426,60,1646,,,,
123
+ ,,,60,1646,1.2855805158615112,1.2855805158615112,,
124
+ 1.5714200735092163,1.5714200735092163,1.3670010566711426,61,1673,,,,
125
+ ,,,61,1673,1.2686656713485718,1.2686656713485718,,
126
+ 2.2389798164367676,2.2389798164367676,1.3670010566711426,62,1700,,,,
127
+ ,,,62,1700,1.2702429294586182,1.2702429294586182,,
128
+ 1.5364372730255127,1.5364372730255127,1.3670010566711426,63,1727,,,,
129
+ ,,,63,1727,1.283432126045227,1.283432126045227,,
130
+ 1.7579401731491089,1.7579401731491089,1.3670010566711426,64,1754,,,,
131
+ ,,,64,1754,1.2375255823135376,1.2375255823135376,,
132
+ 1.6451730728149414,1.6451730728149414,1.3670010566711426,65,1781,,,,
133
+ ,,,65,1781,1.2785224914550781,1.2785224914550781,,
134
+ 1.5431175231933594,1.5431175231933594,1.3670010566711426,66,1808,,,,
135
+ ,,,66,1808,1.2635434865951538,1.2635434865951538,,
136
+ 1.6014606952667236,1.6014606952667236,1.3670010566711426,67,1835,,,,
137
+ ,,,67,1835,1.2820208072662354,1.2820208072662354,,
138
+ 1.5533331632614136,1.5533331632614136,1.3670010566711426,68,1862,,,,
139
+ ,,,68,1862,1.2117564678192139,1.2117564678192139,,
140
+ 2.6706387996673584,2.6706387996673584,1.3670010566711426,69,1889,,,,
141
+ ,,,69,1889,1.2495293617248535,1.2495293617248535,,
142
+ 1.6036927700042725,1.6036927700042725,1.3670010566711426,70,1916,,,,
143
+ ,,,70,1916,1.2474517822265625,1.2474517822265625,,
144
+ 1.666502594947815,1.666502594947815,1.3670010566711426,71,1943,,,,
145
+ ,,,71,1943,1.2659395933151245,1.2659395933151245,,
146
+ 1.55945885181427,1.55945885181427,1.3670010566711426,72,1970,,,,
147
+ ,,,72,1970,1.2158597707748413,1.2158597707748413,,
148
+ 3.0420000553131104,3.0420000553131104,1.3670010566711426,73,1997,,,,
149
+ ,,,73,1997,1.2569940090179443,1.2569940090179443,,
150
+ 1.8147947788238525,1.8147947788238525,1.3670010566711426,74,2024,,,,
151
+ ,,,74,2024,1.2612547874450684,1.2612547874450684,,
152
+ 4.089903831481934,4.089903831481934,1.3670010566711426,75,2051,,,,
153
+ ,,,75,2051,1.2399418354034424,1.2399418354034424,,
154
+ 1.5728484392166138,1.5728484392166138,1.3670010566711426,76,2078,,,,
155
+ ,,,76,2078,1.253847599029541,1.253847599029541,,
156
+ 1.9235353469848633,1.9235353469848633,1.3670010566711426,77,2105,,,,
157
+ ,,,77,2105,1.2418291568756104,1.2418291568756104,,
158
+ 2.069246530532837,2.069246530532837,1.3670010566711426,78,2132,,,,
159
+ ,,,78,2132,1.199945092201233,1.199945092201233,,
160
+ 1.6820194721221924,1.6820194721221924,1.3670010566711426,79,2159,,,,
161
+ ,,,79,2159,1.2384698390960693,1.2384698390960693,,
162
+ 2.007930040359497,2.007930040359497,1.3670010566711426,80,2186,,,,
163
+ ,,,80,2186,1.2008082866668701,1.2008082866668701,,
164
+ 1.764360785484314,1.764360785484314,1.3670010566711426,81,2213,,,,
165
+ ,,,81,2213,1.195989966392517,1.195989966392517,,
166
+ 4.695965766906738,4.695965766906738,1.3670010566711426,82,2240,,,,
167
+ ,,,82,2240,1.2075717449188232,1.2075717449188232,,
168
+ 1.7273682355880737,1.7273682355880737,1.3670010566711426,83,2267,,,,
169
+ ,,,83,2267,1.1842665672302246,1.1842665672302246,,
170
+ 1.524625539779663,1.524625539779663,1.3670010566711426,84,2294,,,,
171
+ ,,,84,2294,1.2273812294006348,1.2273812294006348,,
172
+ 3.537388563156128,3.537388563156128,1.3670010566711426,85,2321,,,,
173
+ ,,,85,2321,1.275421142578125,1.275421142578125,,
174
+ 1.5653022527694702,1.5653022527694702,1.3670010566711426,86,2348,,,,
175
+ ,,,86,2348,1.2298959493637085,1.2298959493637085,,
176
+ 1.463333010673523,1.463333010673523,1.3670010566711426,87,2375,,,,
177
+ ,,,87,2375,1.20870840549469,1.20870840549469,,
178
+ 1.6395522356033325,1.6395522356033325,1.3670010566711426,88,2402,,,,
179
+ ,,,88,2402,1.2088440656661987,1.2088440656661987,,
180
+ 1.3735792636871338,1.3735792636871338,1.3670010566711426,89,2429,,,,
181
+ ,,,89,2429,1.1960471868515015,1.1960471868515015,,
182
+ 1.5243204832077026,1.5243204832077026,1.3670010566711426,90,2456,,,,
183
+ ,,,90,2456,1.1506792306900024,1.1506792306900024,,
184
+ 1.8173081874847412,1.8173081874847412,1.3670010566711426,91,2483,,,,
185
+ ,,,91,2483,1.1671724319458008,1.1671724319458008,,
186
+ 1.45876145362854,1.45876145362854,1.3670010566711426,92,2510,,,,
187
+ ,,,92,2510,1.219564437866211,1.219564437866211,,
188
+ 1.6679826974868774,1.6679826974868774,1.3670010566711426,93,2537,,,,
189
+ ,,,93,2537,1.1834959983825684,1.1834959983825684,,
190
+ 1.59215247631073,1.59215247631073,1.3670010566711426,94,2564,,,,
191
+ ,,,94,2564,1.1848361492156982,1.1848361492156982,,
192
+ 1.50753653049469,1.50753653049469,1.3670010566711426,95,2591,,,,
193
+ ,,,95,2591,1.2056703567504883,1.2056703567504883,,
194
+ 1.7928040027618408,1.7928040027618408,1.3670010566711426,96,2618,,,,
195
+ ,,,96,2618,1.1338263750076294,1.1338263750076294,,
196
+ 1.3231185674667358,1.3231185674667358,1.3231185674667358,97,2645,,,,
197
+ ,,,97,2645,1.1223158836364746,1.1223158836364746,,
198
+ 6.021496295928955,6.021496295928955,1.3231185674667358,98,2672,,,,
199
+ ,,,98,2672,1.1094789505004883,1.1094789505004883,,
200
+ 1.7138872146606445,1.7138872146606445,1.3231185674667358,99,2699,,,,
201
+ ,,,99,2699,1.1203669309616089,1.1203669309616089,,
202
+ ,,,98,2646,,,1.4364614486694336,1.4364614486694336
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646951668.37a92d36c649.1.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28dcf8ba13934c17dbf521819555f79210af186b661af0a27ecd398bdece1197
3
+ size 36426
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646952623.37a92d36c649.1.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d50b436c453bcaf94fa14cdf643d30d70b5a058e25d077d2e05f8f513c631a3
3
+ size 179
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/11/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.0051297750484612
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 256
24
+ lin2_size: 128
25
+ lin3_size: 256
26
+ output_size: 1
27
+ lr: 0.001406127964835635
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.001406127964835635
171
+ - model.lin1_size=256
172
+ - model.lin2_size=128
173
+ - model.lin3_size=256
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=256,model.lr=0.001406127964835635,trainer.gpus=0
179
+ id: '12'
180
+ num: 12
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - datamodule.batch_size=64
2
+ - model.lr=0.001406127964835635
3
+ - model.lin1_size=256
4
+ - model.lin2_size=128
5
+ - model.lin3_size=256
6
+ - hparams_search=focusMAE_150_optuna
7
+ - trainer.gpus=0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.001406127964835635
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 1.870489239692688,1.870489239692688,1.870489239692688,0,26,,,,
3
+ ,,,0,26,1.733809232711792,1.733809232711792,,
4
+ 1.4888343811035156,1.4888343811035156,1.4888343811035156,1,53,,,,
5
+ ,,,1,53,1.4929286241531372,1.4929286241531372,,
6
+ 1.5004178285598755,1.5004178285598755,1.4888343811035156,2,80,,,,
7
+ ,,,2,80,1.4616612195968628,1.4616612195968628,,
8
+ 1.4358630180358887,1.4358630180358887,1.4358630180358887,3,107,,,,
9
+ ,,,3,107,1.4206945896148682,1.4206945896148682,,
10
+ 1.4641728401184082,1.4641728401184082,1.4358630180358887,4,134,,,,
11
+ ,,,4,134,1.3737001419067383,1.3737001419067383,,
12
+ 1.5375480651855469,1.5375480651855469,1.4358630180358887,5,161,,,,
13
+ ,,,5,161,1.3810197114944458,1.3810197114944458,,
14
+ 1.4193321466445923,1.4193321466445923,1.4193321466445923,6,188,,,,
15
+ ,,,6,188,1.3204692602157593,1.3204692602157593,,
16
+ 1.5184860229492188,1.5184860229492188,1.4193321466445923,7,215,,,,
17
+ ,,,7,215,1.281538963317871,1.281538963317871,,
18
+ 1.4773902893066406,1.4773902893066406,1.4193321466445923,8,242,,,,
19
+ ,,,8,242,1.3167067766189575,1.3167067766189575,,
20
+ 1.600895881652832,1.600895881652832,1.4193321466445923,9,269,,,,
21
+ ,,,9,269,1.267653465270996,1.267653465270996,,
22
+ 1.620168924331665,1.620168924331665,1.4193321466445923,10,296,,,,
23
+ ,,,10,296,1.2584946155548096,1.2584946155548096,,
24
+ 2.071122407913208,2.071122407913208,1.4193321466445923,11,323,,,,
25
+ ,,,11,323,1.2184756994247437,1.2184756994247437,,
26
+ 1.4684910774230957,1.4684910774230957,1.4193321466445923,12,350,,,,
27
+ ,,,12,350,1.2365461587905884,1.2365461587905884,,
28
+ 1.4508426189422607,1.4508426189422607,1.4193321466445923,13,377,,,,
29
+ ,,,13,377,1.2474862337112427,1.2474862337112427,,
30
+ 1.3424183130264282,1.3424183130264282,1.3424183130264282,14,404,,,,
31
+ ,,,14,404,1.1824089288711548,1.1824089288711548,,
32
+ 1.4531885385513306,1.4531885385513306,1.3424183130264282,15,431,,,,
33
+ ,,,15,431,1.1809911727905273,1.1809911727905273,,
34
+ 2.6511635780334473,2.6511635780334473,1.3424183130264282,16,458,,,,
35
+ ,,,16,458,1.1664263010025024,1.1664263010025024,,
36
+ 1.5276511907577515,1.5276511907577515,1.3424183130264282,17,485,,,,
37
+ ,,,17,485,1.186537504196167,1.186537504196167,,
38
+ 1.3154600858688354,1.3154600858688354,1.3154600858688354,18,512,,,,
39
+ ,,,18,512,1.143416404724121,1.143416404724121,,
40
+ 2.155818462371826,2.155818462371826,1.3154600858688354,19,539,,,,
41
+ ,,,19,539,1.1829451322555542,1.1829451322555542,,
42
+ 1.493385672569275,1.493385672569275,1.3154600858688354,20,566,,,,
43
+ ,,,20,566,1.1590404510498047,1.1590404510498047,,
44
+ 1.5980056524276733,1.5980056524276733,1.3154600858688354,21,593,,,,
45
+ ,,,21,593,1.106691598892212,1.106691598892212,,
46
+ 1.4680603742599487,1.4680603742599487,1.3154600858688354,22,620,,,,
47
+ ,,,22,620,1.1744014024734497,1.1744014024734497,,
48
+ 1.3440299034118652,1.3440299034118652,1.3154600858688354,23,647,,,,
49
+ ,,,23,647,1.1086127758026123,1.1086127758026123,,
50
+ 1.5388057231903076,1.5388057231903076,1.3154600858688354,24,674,,,,
51
+ ,,,24,674,1.1037986278533936,1.1037986278533936,,
52
+ 2.016345262527466,2.016345262527466,1.3154600858688354,25,701,,,,
53
+ ,,,25,701,1.1079531908035278,1.1079531908035278,,
54
+ 1.6513011455535889,1.6513011455535889,1.3154600858688354,26,728,,,,
55
+ ,,,26,728,1.1088963747024536,1.1088963747024536,,
56
+ 1.8563127517700195,1.8563127517700195,1.3154600858688354,27,755,,,,
57
+ ,,,27,755,1.1106961965560913,1.1106961965560913,,
58
+ 1.9838964939117432,1.9838964939117432,1.3154600858688354,28,782,,,,
59
+ ,,,28,782,1.0945570468902588,1.0945570468902588,,
60
+ 1.299761414527893,1.299761414527893,1.299761414527893,29,809,,,,
61
+ ,,,29,809,1.116113543510437,1.116113543510437,,
62
+ 1.4513648748397827,1.4513648748397827,1.299761414527893,30,836,,,,
63
+ ,,,30,836,1.0945316553115845,1.0945316553115845,,
64
+ 1.6059437990188599,1.6059437990188599,1.299761414527893,31,863,,,,
65
+ ,,,31,863,1.1260499954223633,1.1260499954223633,,
66
+ 1.3629494905471802,1.3629494905471802,1.299761414527893,32,890,,,,
67
+ ,,,32,890,1.1021392345428467,1.1021392345428467,,
68
+ 2.858036756515503,2.858036756515503,1.299761414527893,33,917,,,,
69
+ ,,,33,917,1.0759387016296387,1.0759387016296387,,
70
+ 1.434820532798767,1.434820532798767,1.299761414527893,34,944,,,,
71
+ ,,,34,944,1.0628936290740967,1.0628936290740967,,
72
+ 1.5864146947860718,1.5864146947860718,1.299761414527893,35,971,,,,
73
+ ,,,35,971,1.0650806427001953,1.0650806427001953,,
74
+ 1.5234899520874023,1.5234899520874023,1.299761414527893,36,998,,,,
75
+ ,,,36,998,1.0299108028411865,1.0299108028411865,,
76
+ 1.4551551342010498,1.4551551342010498,1.299761414527893,37,1025,,,,
77
+ ,,,37,1025,1.0748956203460693,1.0748956203460693,,
78
+ 1.8059290647506714,1.8059290647506714,1.299761414527893,38,1052,,,,
79
+ ,,,38,1052,1.0359222888946533,1.0359222888946533,,
80
+ 1.9063102006912231,1.9063102006912231,1.299761414527893,39,1079,,,,
81
+ ,,,39,1079,1.055390477180481,1.055390477180481,,
82
+ 1.587982177734375,1.587982177734375,1.299761414527893,40,1106,,,,
83
+ ,,,40,1106,1.0321245193481445,1.0321245193481445,,
84
+ 1.357559323310852,1.357559323310852,1.299761414527893,41,1133,,,,
85
+ ,,,41,1133,1.0249015092849731,1.0249015092849731,,
86
+ 1.960500955581665,1.960500955581665,1.299761414527893,42,1160,,,,
87
+ ,,,42,1160,1.049727439880371,1.049727439880371,,
88
+ 1.5254287719726562,1.5254287719726562,1.299761414527893,43,1187,,,,
89
+ ,,,43,1187,1.0426660776138306,1.0426660776138306,,
90
+ 2.032141923904419,2.032141923904419,1.299761414527893,44,1214,,,,
91
+ ,,,44,1214,1.048103928565979,1.048103928565979,,
92
+ 1.6677577495574951,1.6677577495574951,1.299761414527893,45,1241,,,,
93
+ ,,,45,1241,1.0212867259979248,1.0212867259979248,,
94
+ 1.5873496532440186,1.5873496532440186,1.299761414527893,46,1268,,,,
95
+ ,,,46,1268,0.9795964360237122,0.9795964360237122,,
96
+ 1.9190750122070312,1.9190750122070312,1.299761414527893,47,1295,,,,
97
+ ,,,47,1295,1.0137090682983398,1.0137090682983398,,
98
+ 1.375336766242981,1.375336766242981,1.299761414527893,48,1322,,,,
99
+ ,,,48,1322,1.0123286247253418,1.0123286247253418,,
100
+ 1.4369909763336182,1.4369909763336182,1.299761414527893,49,1349,,,,
101
+ ,,,49,1349,0.9647429585456848,0.9647429585456848,,
102
+ 2.0165135860443115,2.0165135860443115,1.299761414527893,50,1376,,,,
103
+ ,,,50,1376,0.9378822445869446,0.9378822445869446,,
104
+ 1.7321683168411255,1.7321683168411255,1.299761414527893,51,1403,,,,
105
+ ,,,51,1403,0.9276694059371948,0.9276694059371948,,
106
+ 2.779517889022827,2.779517889022827,1.299761414527893,52,1430,,,,
107
+ ,,,52,1430,0.9949976801872253,0.9949976801872253,,
108
+ 1.5267013311386108,1.5267013311386108,1.299761414527893,53,1457,,,,
109
+ ,,,53,1457,0.9365115761756897,0.9365115761756897,,
110
+ 2.3344242572784424,2.3344242572784424,1.299761414527893,54,1484,,,,
111
+ ,,,54,1484,0.9113747477531433,0.9113747477531433,,
112
+ 1.6980617046356201,1.6980617046356201,1.299761414527893,55,1511,,,,
113
+ ,,,55,1511,0.9396400451660156,0.9396400451660156,,
114
+ 1.7346097230911255,1.7346097230911255,1.299761414527893,56,1538,,,,
115
+ ,,,56,1538,0.9336511492729187,0.9336511492729187,,
116
+ 1.7183431386947632,1.7183431386947632,1.299761414527893,57,1565,,,,
117
+ ,,,57,1565,0.9190303683280945,0.9190303683280945,,
118
+ 1.290636420249939,1.290636420249939,1.290636420249939,58,1592,,,,
119
+ ,,,58,1592,0.9411647319793701,0.9411647319793701,,
120
+ 2.4577672481536865,2.4577672481536865,1.290636420249939,59,1619,,,,
121
+ ,,,59,1619,0.9275367259979248,0.9275367259979248,,
122
+ 2.2040932178497314,2.2040932178497314,1.290636420249939,60,1646,,,,
123
+ ,,,60,1646,0.9203812479972839,0.9203812479972839,,
124
+ 2.1028313636779785,2.1028313636779785,1.290636420249939,61,1673,,,,
125
+ ,,,61,1673,0.9135093688964844,0.9135093688964844,,
126
+ 1.6171327829360962,1.6171327829360962,1.290636420249939,62,1700,,,,
127
+ ,,,62,1700,0.9327650666236877,0.9327650666236877,,
128
+ 1.249523639678955,1.249523639678955,1.249523639678955,63,1727,,,,
129
+ ,,,63,1727,0.889938473701477,0.889938473701477,,
130
+ 1.5783754587173462,1.5783754587173462,1.249523639678955,64,1754,,,,
131
+ ,,,64,1754,0.854543149471283,0.854543149471283,,
132
+ 2.932495355606079,2.932495355606079,1.249523639678955,65,1781,,,,
133
+ ,,,65,1781,0.8996431827545166,0.8996431827545166,,
134
+ 1.6773712635040283,1.6773712635040283,1.249523639678955,66,1808,,,,
135
+ ,,,66,1808,0.8558576703071594,0.8558576703071594,,
136
+ 2.966348171234131,2.966348171234131,1.249523639678955,67,1835,,,,
137
+ ,,,67,1835,0.8499673008918762,0.8499673008918762,,
138
+ 1.3223453760147095,1.3223453760147095,1.249523639678955,68,1862,,,,
139
+ ,,,68,1862,0.8500609397888184,0.8500609397888184,,
140
+ 1.333121418952942,1.333121418952942,1.249523639678955,69,1889,,,,
141
+ ,,,69,1889,0.8425454497337341,0.8425454497337341,,
142
+ 1.377079963684082,1.377079963684082,1.249523639678955,70,1916,,,,
143
+ ,,,70,1916,0.8307526707649231,0.8307526707649231,,
144
+ 2.1237523555755615,2.1237523555755615,1.249523639678955,71,1943,,,,
145
+ ,,,71,1943,0.8052217364311218,0.8052217364311218,,
146
+ 2.1112334728240967,2.1112334728240967,1.249523639678955,72,1970,,,,
147
+ ,,,72,1970,0.821009635925293,0.821009635925293,,
148
+ 1.544689655303955,1.544689655303955,1.249523639678955,73,1997,,,,
149
+ ,,,73,1997,0.819486677646637,0.819486677646637,,
150
+ 1.5640920400619507,1.5640920400619507,1.249523639678955,74,2024,,,,
151
+ ,,,74,2024,0.7873777151107788,0.7873777151107788,,
152
+ 1.2791508436203003,1.2791508436203003,1.249523639678955,75,2051,,,,
153
+ ,,,75,2051,0.8059744834899902,0.8059744834899902,,
154
+ 1.7408134937286377,1.7408134937286377,1.249523639678955,76,2078,,,,
155
+ ,,,76,2078,0.8544288873672485,0.8544288873672485,,
156
+ 1.4483155012130737,1.4483155012130737,1.249523639678955,77,2105,,,,
157
+ ,,,77,2105,0.870209276676178,0.870209276676178,,
158
+ 2.0424013137817383,2.0424013137817383,1.249523639678955,78,2132,,,,
159
+ ,,,78,2132,0.7596953511238098,0.7596953511238098,,
160
+ 1.4385079145431519,1.4385079145431519,1.249523639678955,79,2159,,,,
161
+ ,,,79,2159,0.8185403347015381,0.8185403347015381,,
162
+ 1.6566975116729736,1.6566975116729736,1.249523639678955,80,2186,,,,
163
+ ,,,80,2186,0.824723482131958,0.824723482131958,,
164
+ 1.4256575107574463,1.4256575107574463,1.249523639678955,81,2213,,,,
165
+ ,,,81,2213,0.7928110957145691,0.7928110957145691,,
166
+ 2.618906259536743,2.618906259536743,1.249523639678955,82,2240,,,,
167
+ ,,,82,2240,0.8119828104972839,0.8119828104972839,,
168
+ 1.3975027799606323,1.3975027799606323,1.249523639678955,83,2267,,,,
169
+ ,,,83,2267,0.8413053750991821,0.8413053750991821,,
170
+ 3.4462430477142334,3.4462430477142334,1.249523639678955,84,2294,,,,
171
+ ,,,84,2294,0.7618115544319153,0.7618115544319153,,
172
+ 1.5943691730499268,1.5943691730499268,1.249523639678955,85,2321,,,,
173
+ ,,,85,2321,0.7695584893226624,0.7695584893226624,,
174
+ 1.452353596687317,1.452353596687317,1.249523639678955,86,2348,,,,
175
+ ,,,86,2348,0.8068476915359497,0.8068476915359497,,
176
+ 1.9820992946624756,1.9820992946624756,1.249523639678955,87,2375,,,,
177
+ ,,,87,2375,0.7745007276535034,0.7745007276535034,,
178
+ 1.4738150835037231,1.4738150835037231,1.249523639678955,88,2402,,,,
179
+ ,,,88,2402,0.7288912534713745,0.7288912534713745,,
180
+ 1.71302330493927,1.71302330493927,1.249523639678955,89,2429,,,,
181
+ ,,,89,2429,0.699465811252594,0.699465811252594,,
182
+ 1.3945368528366089,1.3945368528366089,1.249523639678955,90,2456,,,,
183
+ ,,,90,2456,0.7267603278160095,0.7267603278160095,,
184
+ 1.4475047588348389,1.4475047588348389,1.249523639678955,91,2483,,,,
185
+ ,,,91,2483,0.7543859481811523,0.7543859481811523,,
186
+ 1.542484998703003,1.542484998703003,1.249523639678955,92,2510,,,,
187
+ ,,,92,2510,0.7718124389648438,0.7718124389648438,,
188
+ 1.1037075519561768,1.1037075519561768,1.1037075519561768,93,2537,,,,
189
+ ,,,93,2537,0.774516224861145,0.774516224861145,,
190
+ 1.1909815073013306,1.1909815073013306,1.1037075519561768,94,2564,,,,
191
+ ,,,94,2564,0.7957005500793457,0.7957005500793457,,
192
+ 1.0443216562271118,1.0443216562271118,1.0443216562271118,95,2591,,,,
193
+ ,,,95,2591,0.7273293733596802,0.7273293733596802,,
194
+ 1.5082931518554688,1.5082931518554688,1.0443216562271118,96,2618,,,,
195
+ ,,,96,2618,0.6924417614936829,0.6924417614936829,,
196
+ 1.3130191564559937,1.3130191564559937,1.0443216562271118,97,2645,,,,
197
+ ,,,97,2645,0.6858658790588379,0.6858658790588379,,
198
+ 1.3467752933502197,1.3467752933502197,1.0443216562271118,98,2672,,,,
199
+ ,,,98,2672,0.7169437408447266,0.7169437408447266,,
200
+ 1.3395512104034424,1.3395512104034424,1.0443216562271118,99,2699,,,,
201
+ ,,,99,2699,0.7298117280006409,0.7298117280006409,,
202
+ ,,,96,2592,,,1.0170925855636597,1.0170925855636597
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646952624.37a92d36c649.1.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe56a66fe47358a8b11b458e1e1b747204ff36a5f37d5edcc3789cb2f9de5357
3
+ size 36426
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646953611.37a92d36c649.1.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:911e8952f8d1acdf412cd35a39b00482711573d87196f2e62e0ac3e54a4ad748
3
+ size 179
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/12/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.001406127964835635
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 256
24
+ lin2_size: 128
25
+ lin3_size: 256
26
+ output_size: 1
27
+ lr: 0.003667217846048454
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.003667217846048454
171
+ - model.lin1_size=256
172
+ - model.lin2_size=128
173
+ - model.lin3_size=256
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=256,model.lr=0.003667217846048454,trainer.gpus=0
179
+ id: '13'
180
+ num: 13
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - datamodule.batch_size=64
2
+ - model.lr=0.003667217846048454
3
+ - model.lin1_size=256
4
+ - model.lin2_size=128
5
+ - model.lin3_size=256
6
+ - hparams_search=focusMAE_150_optuna
7
+ - trainer.gpus=0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.003667217846048454
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 1.4606432914733887,1.4606432914733887,1.4606432914733887,0,26,,,,
3
+ ,,,0,26,1.6699919700622559,1.6699919700622559,,
4
+ 1.4746863842010498,1.4746863842010498,1.4606432914733887,1,53,,,,
5
+ ,,,1,53,1.5210121870040894,1.5210121870040894,,
6
+ 1.4380803108215332,1.4380803108215332,1.4380803108215332,2,80,,,,
7
+ ,,,2,80,1.4927997589111328,1.4927997589111328,,
8
+ 1.4777369499206543,1.4777369499206543,1.4380803108215332,3,107,,,,
9
+ ,,,3,107,1.465999960899353,1.465999960899353,,
10
+ 1.451737403869629,1.451737403869629,1.4380803108215332,4,134,,,,
11
+ ,,,4,134,1.4530463218688965,1.4530463218688965,,
12
+ 1.5097095966339111,1.5097095966339111,1.4380803108215332,5,161,,,,
13
+ ,,,5,161,1.4216750860214233,1.4216750860214233,,
14
+ 1.5579642057418823,1.5579642057418823,1.4380803108215332,6,188,,,,
15
+ ,,,6,188,1.401484727859497,1.401484727859497,,
16
+ 1.4673517942428589,1.4673517942428589,1.4380803108215332,7,215,,,,
17
+ ,,,7,215,1.4372293949127197,1.4372293949127197,,
18
+ 1.4666259288787842,1.4666259288787842,1.4380803108215332,8,242,,,,
19
+ ,,,8,242,1.4040248394012451,1.4040248394012451,,
20
+ 1.8650015592575073,1.8650015592575073,1.4380803108215332,9,269,,,,
21
+ ,,,9,269,1.4133046865463257,1.4133046865463257,,
22
+ 1.4947688579559326,1.4947688579559326,1.4380803108215332,10,296,,,,
23
+ ,,,10,296,1.4175314903259277,1.4175314903259277,,
24
+ 1.5075221061706543,1.5075221061706543,1.4380803108215332,11,323,,,,
25
+ ,,,11,323,1.3705891370773315,1.3705891370773315,,
26
+ 1.6019188165664673,1.6019188165664673,1.4380803108215332,12,350,,,,
27
+ ,,,12,350,1.3806273937225342,1.3806273937225342,,
28
+ 1.462353229522705,1.462353229522705,1.4380803108215332,13,377,,,,
29
+ ,,,13,377,1.3705203533172607,1.3705203533172607,,
30
+ 1.4787653684616089,1.4787653684616089,1.4380803108215332,14,404,,,,
31
+ ,,,14,404,1.3517358303070068,1.3517358303070068,,
32
+ 1.6885863542556763,1.6885863542556763,1.4380803108215332,15,431,,,,
33
+ ,,,15,431,1.357539176940918,1.357539176940918,,
34
+ 1.5303391218185425,1.5303391218185425,1.4380803108215332,16,458,,,,
35
+ ,,,16,458,1.3722567558288574,1.3722567558288574,,
36
+ 2.5266568660736084,2.5266568660736084,1.4380803108215332,17,485,,,,
37
+ ,,,17,485,1.3515821695327759,1.3515821695327759,,
38
+ 1.5134350061416626,1.5134350061416626,1.4380803108215332,18,512,,,,
39
+ ,,,18,512,1.3410217761993408,1.3410217761993408,,
40
+ 2.35429310798645,2.35429310798645,1.4380803108215332,19,539,,,,
41
+ ,,,19,539,1.3631525039672852,1.3631525039672852,,
42
+ 1.5652703046798706,1.5652703046798706,1.4380803108215332,20,566,,,,
43
+ ,,,20,566,1.3404223918914795,1.3404223918914795,,
44
+ 1.884360909461975,1.884360909461975,1.4380803108215332,21,593,,,,
45
+ ,,,21,593,1.3451640605926514,1.3451640605926514,,
46
+ 1.4733203649520874,1.4733203649520874,1.4380803108215332,22,620,,,,
47
+ ,,,22,620,1.291732907295227,1.291732907295227,,
48
+ 1.4617335796356201,1.4617335796356201,1.4380803108215332,23,647,,,,
49
+ ,,,23,647,1.3343158960342407,1.3343158960342407,,
50
+ 1.5419450998306274,1.5419450998306274,1.4380803108215332,24,674,,,,
51
+ ,,,24,674,1.3400871753692627,1.3400871753692627,,
52
+ 1.601158857345581,1.601158857345581,1.4380803108215332,25,701,,,,
53
+ ,,,25,701,1.355955719947815,1.355955719947815,,
54
+ 1.516626238822937,1.516626238822937,1.4380803108215332,26,728,,,,
55
+ ,,,26,728,1.2812522649765015,1.2812522649765015,,
56
+ 1.7504658699035645,1.7504658699035645,1.4380803108215332,27,755,,,,
57
+ ,,,27,755,1.311252474784851,1.311252474784851,,
58
+ 1.4654655456542969,1.4654655456542969,1.4380803108215332,28,782,,,,
59
+ ,,,28,782,1.2835952043533325,1.2835952043533325,,
60
+ 1.5266945362091064,1.5266945362091064,1.4380803108215332,29,809,,,,
61
+ ,,,29,809,1.2741574048995972,1.2741574048995972,,
62
+ 1.4983974695205688,1.4983974695205688,1.4380803108215332,30,836,,,,
63
+ ,,,30,836,1.2600045204162598,1.2600045204162598,,
64
+ 2.99694561958313,2.99694561958313,1.4380803108215332,31,863,,,,
65
+ ,,,31,863,1.2586629390716553,1.2586629390716553,,
66
+ 1.6219595670700073,1.6219595670700073,1.4380803108215332,32,890,,,,
67
+ ,,,32,890,1.2287592887878418,1.2287592887878418,,
68
+ 1.527970314025879,1.527970314025879,1.4380803108215332,33,917,,,,
69
+ ,,,33,917,1.175965666770935,1.175965666770935,,
70
+ 1.5289629697799683,1.5289629697799683,1.4380803108215332,34,944,,,,
71
+ ,,,34,944,1.2669771909713745,1.2669771909713745,,
72
+ 3.5157012939453125,3.5157012939453125,1.4380803108215332,35,971,,,,
73
+ ,,,35,971,1.201438069343567,1.201438069343567,,
74
+ 1.8071776628494263,1.8071776628494263,1.4380803108215332,36,998,,,,
75
+ ,,,36,998,1.2234296798706055,1.2234296798706055,,
76
+ 2.9550063610076904,2.9550063610076904,1.4380803108215332,37,1025,,,,
77
+ ,,,37,1025,1.1995052099227905,1.1995052099227905,,
78
+ 1.5931316614151,1.5931316614151,1.4380803108215332,38,1052,,,,
79
+ ,,,38,1052,1.1905255317687988,1.1905255317687988,,
80
+ 2.0785012245178223,2.0785012245178223,1.4380803108215332,39,1079,,,,
81
+ ,,,39,1079,1.1170796155929565,1.1170796155929565,,
82
+ 1.459867238998413,1.459867238998413,1.4380803108215332,40,1106,,,,
83
+ ,,,40,1106,1.1982049942016602,1.1982049942016602,,
84
+ 1.3608770370483398,1.3608770370483398,1.3608770370483398,41,1133,,,,
85
+ ,,,41,1133,1.0974856615066528,1.0974856615066528,,
86
+ 1.521744728088379,1.521744728088379,1.3608770370483398,42,1160,,,,
87
+ ,,,42,1160,1.1459871530532837,1.1459871530532837,,
88
+ 1.6354188919067383,1.6354188919067383,1.3608770370483398,43,1187,,,,
89
+ ,,,43,1187,1.2127324342727661,1.2127324342727661,,
90
+ 1.721358060836792,1.721358060836792,1.3608770370483398,44,1214,,,,
91
+ ,,,44,1214,1.1219037771224976,1.1219037771224976,,
92
+ 2.7165470123291016,2.7165470123291016,1.3608770370483398,45,1241,,,,
93
+ ,,,45,1241,1.1469695568084717,1.1469695568084717,,
94
+ 2.7102606296539307,2.7102606296539307,1.3608770370483398,46,1268,,,,
95
+ ,,,46,1268,1.0804805755615234,1.0804805755615234,,
96
+ 3.901205539703369,3.901205539703369,1.3608770370483398,47,1295,,,,
97
+ ,,,47,1295,1.0920984745025635,1.0920984745025635,,
98
+ 1.9869391918182373,1.9869391918182373,1.3608770370483398,48,1322,,,,
99
+ ,,,48,1322,1.0500808954238892,1.0500808954238892,,
100
+ 1.5710898637771606,1.5710898637771606,1.3608770370483398,49,1349,,,,
101
+ ,,,49,1349,1.1200019121170044,1.1200019121170044,,
102
+ 1.48117995262146,1.48117995262146,1.3608770370483398,50,1376,,,,
103
+ ,,,50,1376,1.124404788017273,1.124404788017273,,
104
+ 1.4984129667282104,1.4984129667282104,1.3608770370483398,51,1403,,,,
105
+ ,,,51,1403,1.037623643875122,1.037623643875122,,
106
+ 1.4938468933105469,1.4938468933105469,1.3608770370483398,52,1430,,,,
107
+ ,,,52,1430,1.0634758472442627,1.0634758472442627,,
108
+ 2.06067156791687,2.06067156791687,1.3608770370483398,53,1457,,,,
109
+ ,,,53,1457,1.0552468299865723,1.0552468299865723,,
110
+ 1.643050193786621,1.643050193786621,1.3608770370483398,54,1484,,,,
111
+ ,,,54,1484,1.1060487031936646,1.1060487031936646,,
112
+ 2.774390935897827,2.774390935897827,1.3608770370483398,55,1511,,,,
113
+ ,,,55,1511,1.056064248085022,1.056064248085022,,
114
+ 1.3282986879348755,1.3282986879348755,1.3282986879348755,56,1538,,,,
115
+ ,,,56,1538,1.0134131908416748,1.0134131908416748,,
116
+ 2.899129867553711,2.899129867553711,1.3282986879348755,57,1565,,,,
117
+ ,,,57,1565,0.9875587821006775,0.9875587821006775,,
118
+ 3.5410494804382324,3.5410494804382324,1.3282986879348755,58,1592,,,,
119
+ ,,,58,1592,1.027536392211914,1.027536392211914,,
120
+ 1.8087658882141113,1.8087658882141113,1.3282986879348755,59,1619,,,,
121
+ ,,,59,1619,0.9983460903167725,0.9983460903167725,,
122
+ 2.5392849445343018,2.5392849445343018,1.3282986879348755,60,1646,,,,
123
+ ,,,60,1646,0.9814966917037964,0.9814966917037964,,
124
+ 1.889186978340149,1.889186978340149,1.3282986879348755,61,1673,,,,
125
+ ,,,61,1673,1.04585862159729,1.04585862159729,,
126
+ 1.4085228443145752,1.4085228443145752,1.3282986879348755,62,1700,,,,
127
+ ,,,62,1700,1.019834280014038,1.019834280014038,,
128
+ 2.2238454818725586,2.2238454818725586,1.3282986879348755,63,1727,,,,
129
+ ,,,63,1727,0.9147136807441711,0.9147136807441711,,
130
+ 2.507706880569458,2.507706880569458,1.3282986879348755,64,1754,,,,
131
+ ,,,64,1754,0.950082540512085,0.950082540512085,,
132
+ 1.1386076211929321,1.1386076211929321,1.1386076211929321,65,1781,,,,
133
+ ,,,65,1781,0.9198989272117615,0.9198989272117615,,
134
+ 1.7515085935592651,1.7515085935592651,1.1386076211929321,66,1808,,,,
135
+ ,,,66,1808,0.9104466438293457,0.9104466438293457,,
136
+ 1.4851727485656738,1.4851727485656738,1.1386076211929321,67,1835,,,,
137
+ ,,,67,1835,0.9577696323394775,0.9577696323394775,,
138
+ 1.8228271007537842,1.8228271007537842,1.1386076211929321,68,1862,,,,
139
+ ,,,68,1862,0.9101367592811584,0.9101367592811584,,
140
+ 1.7904624938964844,1.7904624938964844,1.1386076211929321,69,1889,,,,
141
+ ,,,69,1889,0.9035717248916626,0.9035717248916626,,
142
+ 1.9228156805038452,1.9228156805038452,1.1386076211929321,70,1916,,,,
143
+ ,,,70,1916,0.8861273527145386,0.8861273527145386,,
144
+ 2.085308074951172,2.085308074951172,1.1386076211929321,71,1943,,,,
145
+ ,,,71,1943,0.9694860577583313,0.9694860577583313,,
146
+ 1.9828592538833618,1.9828592538833618,1.1386076211929321,72,1970,,,,
147
+ ,,,72,1970,0.9413881897926331,0.9413881897926331,,
148
+ 1.218909740447998,1.218909740447998,1.1386076211929321,73,1997,,,,
149
+ ,,,73,1997,0.9394506812095642,0.9394506812095642,,
150
+ 1.2931345701217651,1.2931345701217651,1.1386076211929321,74,2024,,,,
151
+ ,,,74,2024,0.8854947686195374,0.8854947686195374,,
152
+ 1.9957361221313477,1.9957361221313477,1.1386076211929321,75,2051,,,,
153
+ ,,,75,2051,0.8731582760810852,0.8731582760810852,,
154
+ 1.156192421913147,1.156192421913147,1.1386076211929321,76,2078,,,,
155
+ ,,,76,2078,0.8668397665023804,0.8668397665023804,,
156
+ 4.492703914642334,4.492703914642334,1.1386076211929321,77,2105,,,,
157
+ ,,,77,2105,0.8878253102302551,0.8878253102302551,,
158
+ 1.2137655019760132,1.2137655019760132,1.1386076211929321,78,2132,,,,
159
+ ,,,78,2132,0.8640716075897217,0.8640716075897217,,
160
+ 1.9379135370254517,1.9379135370254517,1.1386076211929321,79,2159,,,,
161
+ ,,,79,2159,0.8310316205024719,0.8310316205024719,,
162
+ 2.1287708282470703,2.1287708282470703,1.1386076211929321,80,2186,,,,
163
+ ,,,80,2186,0.8973637223243713,0.8973637223243713,,
164
+ 1.910030484199524,1.910030484199524,1.1386076211929321,81,2213,,,,
165
+ ,,,81,2213,0.9557178020477295,0.9557178020477295,,
166
+ 1.5638020038604736,1.5638020038604736,1.1386076211929321,82,2240,,,,
167
+ ,,,82,2240,0.902524471282959,0.902524471282959,,
168
+ 1.8373081684112549,1.8373081684112549,1.1386076211929321,83,2267,,,,
169
+ ,,,83,2267,0.8470232486724854,0.8470232486724854,,
170
+ 4.7020158767700195,4.7020158767700195,1.1386076211929321,84,2294,,,,
171
+ ,,,84,2294,0.8718170523643494,0.8718170523643494,,
172
+ 2.8360331058502197,2.8360331058502197,1.1386076211929321,85,2321,,,,
173
+ ,,,85,2321,0.9025630950927734,0.9025630950927734,,
174
+ 1.8024927377700806,1.8024927377700806,1.1386076211929321,86,2348,,,,
175
+ ,,,86,2348,0.8570308685302734,0.8570308685302734,,
176
+ 2.514446496963501,2.514446496963501,1.1386076211929321,87,2375,,,,
177
+ ,,,87,2375,0.8571733832359314,0.8571733832359314,,
178
+ 3.4728214740753174,3.4728214740753174,1.1386076211929321,88,2402,,,,
179
+ ,,,88,2402,0.9757309556007385,0.9757309556007385,,
180
+ 2.3552134037017822,2.3552134037017822,1.1386076211929321,89,2429,,,,
181
+ ,,,89,2429,0.8453519940376282,0.8453519940376282,,
182
+ 1.7544959783554077,1.7544959783554077,1.1386076211929321,90,2456,,,,
183
+ ,,,90,2456,0.8101593852043152,0.8101593852043152,,
184
+ 1.0175409317016602,1.0175409317016602,1.0175409317016602,91,2483,,,,
185
+ ,,,91,2483,0.8310639262199402,0.8310639262199402,,
186
+ 1.3298760652542114,1.3298760652542114,1.0175409317016602,92,2510,,,,
187
+ ,,,92,2510,0.9122585654258728,0.9122585654258728,,
188
+ 4.411593914031982,4.411593914031982,1.0175409317016602,93,2537,,,,
189
+ ,,,93,2537,0.7885783314704895,0.7885783314704895,,
190
+ 4.896975040435791,4.896975040435791,1.0175409317016602,94,2564,,,,
191
+ ,,,94,2564,0.7917547225952148,0.7917547225952148,,
192
+ 2.0306270122528076,2.0306270122528076,1.0175409317016602,95,2591,,,,
193
+ ,,,95,2591,0.8102917075157166,0.8102917075157166,,
194
+ 1.7351102828979492,1.7351102828979492,1.0175409317016602,96,2618,,,,
195
+ ,,,96,2618,0.8058015704154968,0.8058015704154968,,
196
+ 2.072267532348633,2.072267532348633,1.0175409317016602,97,2645,,,,
197
+ ,,,97,2645,0.7851160764694214,0.7851160764694214,,
198
+ 1.3256508111953735,1.3256508111953735,1.0175409317016602,98,2672,,,,
199
+ ,,,98,2672,0.8001875281333923,0.8001875281333923,,
200
+ 1.3688437938690186,1.3688437938690186,1.0175409317016602,99,2699,,,,
201
+ ,,,99,2699,0.8080283403396606,0.8080283403396606,,
202
+ ,,,92,2484,,,1.108647346496582,1.108647346496582
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646953611.37a92d36c649.1.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95143eeacee12ca761e053bce474364f45801be48f431279a017c9bb8ec0fa27
3
+ size 36426
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/tensorboard/focusMAE_150_hyperparameter_search/events.out.tfevents.1646954564.37a92d36c649.1.27 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac2b5a52ad73729c0a6e0296127a72bb3c420cdd6d8d92e3c659824fb84b5e88
3
+ size 179
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/13/tensorboard/focusMAE_150_hyperparameter_search/hparams.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 0
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_module.FocusLitModule
9
+ input_size: 67500
10
+ lin1_size: 256
11
+ lin2_size: 128
12
+ lin3_size: 256
13
+ output_size: 1
14
+ lr: 0.003667217846048454
15
+ weight_decay: 0.0005
16
+ datamodule:
17
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
18
+ data_dir: /usr/src/app/data/focus150
19
+ csv_file: /usr/src/app/data/focus150/metadata.csv
20
+ batch_size: 64
21
+ train_val_test_split_percentage:
22
+ - 0.7
23
+ - 0.15
24
+ - 0.15
25
+ num_workers: 0
26
+ pin_memory: false
27
+ seed: null
28
+ callbacks:
29
+ model_checkpoint:
30
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
31
+ monitor: val/mae
32
+ mode: min
33
+ save_top_k: 1
34
+ save_last: true
35
+ verbose: false
36
+ dirpath: checkpoints/
37
+ filename: epoch_{epoch:03d}
38
+ auto_insert_metric_name: false
39
+ early_stopping:
40
+ _target_: pytorch_lightning.callbacks.EarlyStopping
41
+ monitor: val/mae
42
+ mode: min
43
+ patience: 100
44
+ min_delta: 0
45
+ model_summary:
46
+ _target_: pytorch_lightning.callbacks.RichModelSummary
47
+ max_depth: -1
48
+ rich_progress_bar:
49
+ _target_: pytorch_lightning.callbacks.RichProgressBar
50
+ model/params/total: 17347713
51
+ model/params/trainable: 17347713
52
+ model/params/non_trainable: 0
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/14/.hydra/config.yaml ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: null
8
+ name: focusMAE_150_hyperparameter_search
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_file: ${data_dir}/focus150/metadata.csv
13
+ batch_size: 64
14
+ train_val_test_split_percentage:
15
+ - 0.7
16
+ - 0.15
17
+ - 0.15
18
+ num_workers: 0
19
+ pin_memory: false
20
+ model:
21
+ _target_: src.models.focus_module.FocusLitModule
22
+ input_size: 67500
23
+ lin1_size: 256
24
+ lin2_size: 128
25
+ lin3_size: 256
26
+ output_size: 1
27
+ lr: 0.059650958218651365
28
+ weight_decay: 0.0005
29
+ callbacks:
30
+ model_checkpoint:
31
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
32
+ monitor: val/mae
33
+ mode: min
34
+ save_top_k: 1
35
+ save_last: true
36
+ verbose: false
37
+ dirpath: checkpoints/
38
+ filename: epoch_{epoch:03d}
39
+ auto_insert_metric_name: false
40
+ early_stopping:
41
+ _target_: pytorch_lightning.callbacks.EarlyStopping
42
+ monitor: val/mae
43
+ mode: min
44
+ patience: 100
45
+ min_delta: 0
46
+ model_summary:
47
+ _target_: pytorch_lightning.callbacks.RichModelSummary
48
+ max_depth: -1
49
+ rich_progress_bar:
50
+ _target_: pytorch_lightning.callbacks.RichProgressBar
51
+ logger:
52
+ csv:
53
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
54
+ save_dir: .
55
+ name: csv/
56
+ prefix: ''
57
+ tensorboard:
58
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
59
+ save_dir: tensorboard/
60
+ name: null
61
+ version: ${name}
62
+ log_graph: false
63
+ default_hp_metric: true
64
+ prefix: ''
65
+ trainer:
66
+ _target_: pytorch_lightning.Trainer
67
+ gpus: 0
68
+ min_epochs: 1
69
+ max_epochs: 100
70
+ resume_from_checkpoint: null
71
+ optimized_metric: val/mae_best
logs/experiments/multiruns/focusMAE_150_hyperparameter_search/2022-03-10_16-59-48/14/.hydra/hydra.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ sampler:
11
+ _target_: optuna.samplers.TPESampler
12
+ seed: 12345
13
+ consider_prior: true
14
+ prior_weight: 1.0
15
+ consider_magic_clip: true
16
+ consider_endpoints: false
17
+ n_startup_trials: 10
18
+ n_ei_candidates: 24
19
+ multivariate: false
20
+ warn_independent_sampling: true
21
+ _target_: hydra_plugins.hydra_optuna_sweeper.optuna_sweeper.OptunaSweeper
22
+ direction: minimize
23
+ storage: null
24
+ study_name: focusMAE_150_hyperparameter_search
25
+ n_trials: 25
26
+ n_jobs: 1
27
+ search_space:
28
+ datamodule.batch_size:
29
+ type: categorical
30
+ choices:
31
+ - 64
32
+ - 128
33
+ model.lr:
34
+ type: float
35
+ low: 0.0001
36
+ high: 0.2
37
+ model.lin1_size:
38
+ type: categorical
39
+ choices:
40
+ - 64
41
+ - 128
42
+ - 256
43
+ - 512
44
+ - 1024
45
+ model.lin2_size:
46
+ type: categorical
47
+ choices:
48
+ - 64
49
+ - 128
50
+ - 256
51
+ - 512
52
+ - 1024
53
+ model.lin3_size:
54
+ type: categorical
55
+ choices:
56
+ - 64
57
+ - 128
58
+ - 256
59
+ - 512
60
+ - 1024
61
+ help:
62
+ app_name: ${hydra.job.name}
63
+ header: '${hydra.help.app_name} is powered by Hydra.
64
+
65
+ '
66
+ footer: 'Powered by Hydra (https://hydra.cc)
67
+
68
+ Use --hydra-help to view Hydra specific help
69
+
70
+ '
71
+ template: '${hydra.help.header}
72
+
73
+ == Configuration groups ==
74
+
75
+ Compose your configuration from those groups (group=option)
76
+
77
+
78
+ $APP_CONFIG_GROUPS
79
+
80
+
81
+ == Config ==
82
+
83
+ Override anything in the config (foo.bar=value)
84
+
85
+
86
+ $CONFIG
87
+
88
+
89
+ ${hydra.help.footer}
90
+
91
+ '
92
+ hydra_help:
93
+ template: 'Hydra (${hydra.runtime.version})
94
+
95
+ See https://hydra.cc for more info.
96
+
97
+
98
+ == Flags ==
99
+
100
+ $FLAGS_HELP
101
+
102
+
103
+ == Configuration groups ==
104
+
105
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
106
+ to command line)
107
+
108
+
109
+ $HYDRA_CONFIG_GROUPS
110
+
111
+
112
+ Use ''--cfg hydra'' to Show the Hydra config.
113
+
114
+ '
115
+ hydra_help: ???
116
+ hydra_logging:
117
+ version: 1
118
+ formatters:
119
+ colorlog:
120
+ (): colorlog.ColoredFormatter
121
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
122
+ handlers:
123
+ console:
124
+ class: logging.StreamHandler
125
+ formatter: colorlog
126
+ stream: ext://sys.stdout
127
+ root:
128
+ level: INFO
129
+ handlers:
130
+ - console
131
+ disable_existing_loggers: false
132
+ job_logging:
133
+ version: 1
134
+ formatters:
135
+ simple:
136
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
137
+ colorlog:
138
+ (): colorlog.ColoredFormatter
139
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
140
+ - %(message)s'
141
+ log_colors:
142
+ DEBUG: purple
143
+ INFO: green
144
+ WARNING: yellow
145
+ ERROR: red
146
+ CRITICAL: red
147
+ handlers:
148
+ console:
149
+ class: logging.StreamHandler
150
+ formatter: colorlog
151
+ stream: ext://sys.stdout
152
+ file:
153
+ class: logging.FileHandler
154
+ formatter: simple
155
+ filename: ${hydra.job.name}.log
156
+ root:
157
+ level: INFO
158
+ handlers:
159
+ - console
160
+ - file
161
+ disable_existing_loggers: false
162
+ env: {}
163
+ searchpath: []
164
+ callbacks: {}
165
+ output_subdir: .hydra
166
+ overrides:
167
+ hydra: []
168
+ task:
169
+ - datamodule.batch_size=64
170
+ - model.lr=0.059650958218651365
171
+ - model.lin1_size=256
172
+ - model.lin2_size=128
173
+ - model.lin3_size=256
174
+ - hparams_search=focusMAE_150_optuna
175
+ - trainer.gpus=0
176
+ job:
177
+ name: train
178
+ override_dirname: datamodule.batch_size=64,hparams_search=focusMAE_150_optuna,model.lin1_size=256,model.lin2_size=128,model.lin3_size=256,model.lr=0.059650958218651365,trainer.gpus=0
179
+ id: '14'
180
+ num: 14
181
+ config_name: train.yaml
182
+ env_set: {}
183
+ env_copy: []
184
+ config:
185
+ override_dirname:
186
+ kv_sep: '='
187
+ item_sep: ','
188
+ exclude_keys: []
189
+ runtime:
190
+ version: 1.1.1
191
+ cwd: /usr/src/app
192
+ config_sources:
193
+ - path: hydra.conf
194
+ schema: pkg
195
+ provider: hydra
196
+ - path: /usr/src/app/configs
197
+ schema: file
198
+ provider: main
199
+ - path: hydra_plugins.hydra_colorlog.conf
200
+ schema: pkg
201
+ provider: hydra-colorlog
202
+ - path: ''
203
+ schema: structured
204
+ provider: schema
205
+ choices:
206
+ local: default.yaml
207
+ hparams_search: focusMAE_150_optuna
208
+ debug: null
209
+ experiment: null
210
+ log_dir: default.yaml
211
+ trainer: long.yaml
212
+ logger: many_loggers
213
+ callbacks: default.yaml
214
+ model: focus150.yaml
215
+ datamodule: focus150.yaml
216
+ hydra/env: default
217
+ hydra/callbacks: null
218
+ hydra/job_logging: colorlog
219
+ hydra/hydra_logging: colorlog
220
+ hydra/hydra_help: default
221
+ hydra/help: default
222
+ hydra/sweeper: optuna
223
+ hydra/sweeper/sampler: tpe
224
+ hydra/launcher: basic
225
+ hydra/output: default
226
+ verbose: false