Hannes Kuchelmeister commited on
Commit
2833379
·
1 Parent(s): a317510

rerun and add logs for focusConvMSE_besthyp_150

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/.hydra/config.yaml +81 -0
  2. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/.hydra/hydra.yaml +170 -0
  3. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/.hydra/overrides.yaml +1 -0
  4. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/checkpoints/epoch_053.ckpt +3 -0
  5. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/checkpoints/last.ckpt +3 -0
  6. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/csv/version_0/hparams.yaml +55 -0
  7. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/csv/version_0/metrics.csv +202 -0
  8. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/tensorboard/focusConvMSE_150/events.out.tfevents.1647960193.0c30fcf2cd8d.1.0 +3 -0
  9. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/tensorboard/focusConvMSE_150/events.out.tfevents.1647960610.0c30fcf2cd8d.1.1 +3 -0
  10. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/tensorboard/focusConvMSE_150/hparams.yaml +55 -0
  11. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/meta.yaml +15 -0
  12. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/epoch +201 -0
  13. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/test/loss +1 -0
  14. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/test/mae +1 -0
  15. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/train/loss +100 -0
  16. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/train/mae +100 -0
  17. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/val/loss +100 -0
  18. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/val/mae +100 -0
  19. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/val/mae_best +100 -0
  20. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/_target_ +1 -0
  21. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/min_delta +1 -0
  22. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/mode +1 -0
  23. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/monitor +1 -0
  24. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/patience +1 -0
  25. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/_target_ +1 -0
  26. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/auto_insert_metric_name +1 -0
  27. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/dirpath +1 -0
  28. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/filename +1 -0
  29. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/mode +1 -0
  30. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/monitor +1 -0
  31. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/save_last +1 -0
  32. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/save_top_k +1 -0
  33. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/verbose +1 -0
  34. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_summary/_target_ +1 -0
  35. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_summary/max_depth +1 -0
  36. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/rich_progress_bar/_target_ +1 -0
  37. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/_target_ +1 -0
  38. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/augmentation +1 -0
  39. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/batch_size +1 -0
  40. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/csv_test_file +1 -0
  41. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/csv_train_file +1 -0
  42. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/csv_val_file +1 -0
  43. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/data_dir +1 -0
  44. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/num_workers +1 -0
  45. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/pin_memory +1 -0
  46. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/_target_ +1 -0
  47. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv1_channels +1 -0
  48. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv1_size +1 -0
  49. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv2_channels +1 -0
  50. logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv2_size +1 -0
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/.hydra/config.yaml ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: 12345
8
+ name: focusConvMSE_150
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_train_file: ${data_dir}/focus150/train_metadata.csv
13
+ csv_val_file: ${data_dir}/focus150/validation_metadata.csv
14
+ csv_test_file: ${data_dir}/focus150/test_metadata.csv
15
+ batch_size: 64
16
+ num_workers: 0
17
+ pin_memory: false
18
+ augmentation: true
19
+ model:
20
+ _target_: src.models.focus_conv_module.FocusConvLitModule
21
+ image_size: 150
22
+ pool_size: 2
23
+ conv1_size: 3
24
+ conv1_channels: 9
25
+ conv2_size: 7
26
+ conv2_channels: 6
27
+ lin1_size: 32
28
+ lin2_size: 72
29
+ output_size: 1
30
+ lr: 0.001
31
+ weight_decay: 0.0005
32
+ callbacks:
33
+ model_checkpoint:
34
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
35
+ monitor: val/mae
36
+ mode: min
37
+ save_top_k: 1
38
+ save_last: true
39
+ verbose: false
40
+ dirpath: checkpoints/
41
+ filename: epoch_{epoch:03d}
42
+ auto_insert_metric_name: false
43
+ early_stopping:
44
+ _target_: pytorch_lightning.callbacks.EarlyStopping
45
+ monitor: val/mae
46
+ mode: min
47
+ patience: 100
48
+ min_delta: 0
49
+ model_summary:
50
+ _target_: pytorch_lightning.callbacks.RichModelSummary
51
+ max_depth: -1
52
+ rich_progress_bar:
53
+ _target_: pytorch_lightning.callbacks.RichProgressBar
54
+ logger:
55
+ csv:
56
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
57
+ save_dir: .
58
+ name: csv/
59
+ prefix: ''
60
+ mlflow:
61
+ _target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
62
+ experiment_name: ${name}
63
+ tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
64
+ tags: null
65
+ save_dir: ./mlruns
66
+ prefix: ''
67
+ artifact_location: null
68
+ tensorboard:
69
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
70
+ save_dir: tensorboard/
71
+ name: null
72
+ version: ${name}
73
+ log_graph: false
74
+ default_hp_metric: true
75
+ prefix: ''
76
+ trainer:
77
+ _target_: pytorch_lightning.Trainer
78
+ gpus: 1
79
+ min_epochs: 1
80
+ max_epochs: 100
81
+ resume_from_checkpoint: null
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - experiment=focusConvMSE_besthyp_150
121
+ job:
122
+ name: train
123
+ override_dirname: experiment=focusConvMSE_besthyp_150
124
+ id: ???
125
+ num: ???
126
+ config_name: train.yaml
127
+ env_set: {}
128
+ env_copy: []
129
+ config:
130
+ override_dirname:
131
+ kv_sep: '='
132
+ item_sep: ','
133
+ exclude_keys: []
134
+ runtime:
135
+ version: 1.1.1
136
+ cwd: /usr/src/app
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /usr/src/app/configs
142
+ schema: file
143
+ provider: main
144
+ - path: hydra_plugins.hydra_colorlog.conf
145
+ schema: pkg
146
+ provider: hydra-colorlog
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ choices:
151
+ local: default.yaml
152
+ hparams_search: null
153
+ debug: null
154
+ experiment: focusConvMSE_besthyp_150
155
+ log_dir: default.yaml
156
+ trainer: default.yaml
157
+ logger: many_loggers
158
+ callbacks: default.yaml
159
+ model: focusConv_150.yaml
160
+ datamodule: focus150.yaml
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: colorlog
164
+ hydra/hydra_logging: colorlog
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=focusConvMSE_besthyp_150
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/checkpoints/epoch_053.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b924ea40c8e52cd3e9ae4b701cb3af3217fd2b37644b374263cac05466f86000
3
+ size 2605061
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/checkpoints/last.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dfc4ec0ce67172773d9441f9e4727324a10361baf1cea98ba55d703ae3909b5
3
+ size 2605061
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 1
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_conv_module.FocusConvLitModule
9
+ image_size: 150
10
+ pool_size: 2
11
+ conv1_size: 3
12
+ conv1_channels: 9
13
+ conv2_size: 7
14
+ conv2_channels: 6
15
+ lin1_size: 32
16
+ lin2_size: 72
17
+ output_size: 1
18
+ lr: 0.001
19
+ weight_decay: 0.0005
20
+ datamodule:
21
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
22
+ data_dir: /usr/src/app/data/focus150
23
+ csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
24
+ csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
25
+ csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
26
+ batch_size: 64
27
+ num_workers: 0
28
+ pin_memory: false
29
+ augmentation: true
30
+ seed: 12345
31
+ callbacks:
32
+ model_checkpoint:
33
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
34
+ monitor: val/mae
35
+ mode: min
36
+ save_top_k: 1
37
+ save_last: true
38
+ verbose: false
39
+ dirpath: checkpoints/
40
+ filename: epoch_{epoch:03d}
41
+ auto_insert_metric_name: false
42
+ early_stopping:
43
+ _target_: pytorch_lightning.callbacks.EarlyStopping
44
+ monitor: val/mae
45
+ mode: min
46
+ patience: 100
47
+ min_delta: 0
48
+ model_summary:
49
+ _target_: pytorch_lightning.callbacks.RichModelSummary
50
+ max_depth: -1
51
+ rich_progress_bar:
52
+ _target_: pytorch_lightning.callbacks.RichProgressBar
53
+ model/params/total: 216201
54
+ model/params/trainable: 216201
55
+ model/params/non_trainable: 0
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 0.0010520756477490067,0.027000371366739273,0.024726606905460358,0,47,,,,
3
+ ,,,0,47,0.07324540615081787,0.15608204901218414,,
4
+ 0.0004612061893567443,0.017776092514395714,0.017776092514395714,1,95,,,,
5
+ ,,,1,95,0.0005119902198202908,0.01848370023071766,,
6
+ 0.0004499765345826745,0.017643338069319725,0.017643338069319725,2,143,,,,
7
+ ,,,2,143,0.0004279178101569414,0.017223240807652473,,
8
+ 0.00044631771743297577,0.017654919996857643,0.017643338069319725,3,191,,,,
9
+ ,,,3,191,0.0004204828874208033,0.017105428501963615,,
10
+ 0.0004415122966747731,0.017546169459819794,0.017546169459819794,4,239,,,,
11
+ ,,,4,239,0.0004181377007625997,0.01710123009979725,,
12
+ 0.000438133894931525,0.017503635957837105,0.017503635957837105,5,287,,,,
13
+ ,,,5,287,0.0004158291267231107,0.017015699297189713,,
14
+ 0.00043563765939325094,0.017501212656497955,0.017501212656497955,6,335,,,,
15
+ ,,,6,335,0.0004123631224501878,0.016972672194242477,,
16
+ 0.0004368386580608785,0.017341317608952522,0.017341317608952522,7,383,,,,
17
+ ,,,7,383,0.00040746244485490024,0.016849098727107048,,
18
+ 0.0004356758145149797,0.017313677817583084,0.017313677817583084,8,431,,,,
19
+ ,,,8,431,0.00041121308458968997,0.016946695744991302,,
20
+ 0.00043189621646888554,0.0172542966902256,0.0172542966902256,9,479,,,,
21
+ ,,,9,479,0.00040692713810130954,0.0168300773948431,,
22
+ 0.000431660155300051,0.017524439841508865,0.0172542966902256,10,527,,,,
23
+ ,,,10,527,0.00040299276588484645,0.016762785613536835,,
24
+ 0.0004278685664758086,0.017209237441420555,0.017209237441420555,11,575,,,,
25
+ ,,,11,575,0.00040395042742602527,0.01680273935198784,,
26
+ 0.00042351684533059597,0.017291804775595665,0.017209237441420555,12,623,,,,
27
+ ,,,12,623,0.0004016980528831482,0.01675964891910553,,
28
+ 0.00042290761484764516,0.01727038435637951,0.017209237441420555,13,671,,,,
29
+ ,,,13,671,0.00040133227594196796,0.01673046499490738,,
30
+ 0.00042417016811668873,0.017126740887761116,0.017126740887761116,14,719,,,,
31
+ ,,,14,719,0.00040109638939611614,0.016756772994995117,,
32
+ 0.0004259877896402031,0.017122970893979073,0.017122970893979073,15,767,,,,
33
+ ,,,15,767,0.0004060292267240584,0.016798974946141243,,
34
+ 0.00042129578650929034,0.017265643924474716,0.017122970893979073,16,815,,,,
35
+ ,,,16,815,0.00040416873525828123,0.016754010692238808,,
36
+ 0.00042111988295800984,0.017123492434620857,0.017122970893979073,17,863,,,,
37
+ ,,,17,863,0.0004033875884488225,0.016784122213721275,,
38
+ 0.0004269638448022306,0.017454244196414948,0.017122970893979073,18,911,,,,
39
+ ,,,18,911,0.0003990893601439893,0.01666918769478798,,
40
+ 0.00042243217467330396,0.017306191846728325,0.017122970893979073,19,959,,,,
41
+ ,,,19,959,0.000404653896111995,0.016769438982009888,,
42
+ 0.0004220800183247775,0.017325717955827713,0.017122970893979073,20,1007,,,,
43
+ ,,,20,1007,0.00040122432983480394,0.016709182411432266,,
44
+ 0.0004301489389035851,0.017113929614424706,0.017113929614424706,21,1055,,,,
45
+ ,,,21,1055,0.0004034280136693269,0.016742847859859467,,
46
+ 0.00042216639849357307,0.01728595420718193,0.017113929614424706,22,1103,,,,
47
+ ,,,22,1103,0.0003990979748778045,0.016647830605506897,,
48
+ 0.0004318137071095407,0.017589587718248367,0.017113929614424706,23,1151,,,,
49
+ ,,,23,1151,0.00040245126001536846,0.016745105385780334,,
50
+ 0.0004194175999145955,0.017076712101697922,0.017076712101697922,24,1199,,,,
51
+ ,,,24,1199,0.00040643103420734406,0.016855906695127487,,
52
+ 0.00042441036202944815,0.01708088628947735,0.017076712101697922,25,1247,,,,
53
+ ,,,25,1247,0.000400546588934958,0.016742048785090446,,
54
+ 0.00042018183739855886,0.01709458790719509,0.017076712101697922,26,1295,,,,
55
+ ,,,26,1295,0.0003996045561507344,0.016682812944054604,,
56
+ 0.0004190116305835545,0.017154676839709282,0.017076712101697922,27,1343,,,,
57
+ ,,,27,1343,0.00039723876398056746,0.016645090654492378,,
58
+ 0.0004195934161543846,0.017233334481716156,0.017076712101697922,28,1391,,,,
59
+ ,,,28,1391,0.0003959255700465292,0.016611488536000252,,
60
+ 0.00041890854481607676,0.017193002626299858,0.017076712101697922,29,1439,,,,
61
+ ,,,29,1439,0.000394946604501456,0.01659465581178665,,
62
+ 0.0004188215534668416,0.017117813229560852,0.017076712101697922,30,1487,,,,
63
+ ,,,30,1487,0.0003974179271608591,0.016647476702928543,,
64
+ 0.0004203834687359631,0.01704377681016922,0.01704377681016922,31,1535,,,,
65
+ ,,,31,1535,0.00039746827678754926,0.01665402390062809,,
66
+ 0.0004258453845977783,0.017442181706428528,0.01704377681016922,32,1583,,,,
67
+ ,,,32,1583,0.0003963193448726088,0.016615083441138268,,
68
+ 0.0004177540831733495,0.017155330628156662,0.01704377681016922,33,1631,,,,
69
+ ,,,33,1631,0.0003986273950431496,0.016656169667840004,,
70
+ 0.00042232274427078664,0.017022596672177315,0.017022596672177315,34,1679,,,,
71
+ ,,,34,1679,0.000402714591473341,0.01677883230149746,,
72
+ 0.0004189989995211363,0.017094330862164497,0.017022596672177315,35,1727,,,,
73
+ ,,,35,1727,0.00039801615639589727,0.016634121537208557,,
74
+ 0.00042335118632763624,0.01704075001180172,0.017022596672177315,36,1775,,,,
75
+ ,,,36,1775,0.0004030045529361814,0.016751982271671295,,
76
+ 0.00041969664744101465,0.017027610912919044,0.017022596672177315,37,1823,,,,
77
+ ,,,37,1823,0.0004063526284880936,0.016803236678242683,,
78
+ 0.00041658285772427917,0.01709923706948757,0.017022596672177315,38,1871,,,,
79
+ ,,,38,1871,0.0003981548943556845,0.016650604084134102,,
80
+ 0.000418454670580104,0.01721111498773098,0.017022596672177315,39,1919,,,,
81
+ ,,,39,1919,0.0003961462643928826,0.016624847427010536,,
82
+ 0.00041677828994579613,0.01710767298936844,0.017022596672177315,40,1967,,,,
83
+ ,,,40,1967,0.0004012832068838179,0.01667129620909691,,
84
+ 0.00042211331310682,0.017018204554915428,0.017018204554915428,41,2015,,,,
85
+ ,,,41,2015,0.0003976260486524552,0.016667962074279785,,
86
+ 0.0004176416841801256,0.017177345231175423,0.017018204554915428,42,2063,,,,
87
+ ,,,42,2063,0.0003990065888501704,0.016656573861837387,,
88
+ 0.0004183136916253716,0.01718120276927948,0.017018204554915428,43,2111,,,,
89
+ ,,,43,2111,0.00039825798012316227,0.016642922535538673,,
90
+ 0.0004184833960607648,0.017036592587828636,0.017018204554915428,44,2159,,,,
91
+ ,,,44,2159,0.00039614003617316484,0.016659162938594818,,
92
+ 0.00042137890704907477,0.017030268907546997,0.017018204554915428,45,2207,,,,
93
+ ,,,45,2207,0.00039644099888391793,0.016649728640913963,,
94
+ 0.00041784756467677653,0.017211318016052246,0.017018204554915428,46,2255,,,,
95
+ ,,,46,2255,0.00039509835187345743,0.016588879749178886,,
96
+ 0.00041607263847254217,0.017052004113793373,0.017018204554915428,47,2303,,,,
97
+ ,,,47,2303,0.00039680046029388905,0.01660764217376709,,
98
+ 0.00041994749335572124,0.01700209081172943,0.01700209081172943,48,2351,,,,
99
+ ,,,48,2351,0.00039851333713158965,0.016675254330039024,,
100
+ 0.00041497754864394665,0.01702883280813694,0.01700209081172943,49,2399,,,,
101
+ ,,,49,2399,0.00039479872793890536,0.016594236716628075,,
102
+ 0.00041593355126678944,0.01700270175933838,0.01700209081172943,50,2447,,,,
103
+ ,,,50,2447,0.0003952273400500417,0.01660444773733616,,
104
+ 0.0004291844379622489,0.017490429803729057,0.01700209081172943,51,2495,,,,
105
+ ,,,51,2495,0.0003985257644671947,0.01664111576974392,,
106
+ 0.00042258668690919876,0.017022263258695602,0.01700209081172943,52,2543,,,,
107
+ ,,,52,2543,0.00039632743573747575,0.0166194848716259,,
108
+ 0.00041858889744617045,0.017000695690512657,0.017000695690512657,53,2591,,,,
109
+ ,,,53,2591,0.0003982774505857378,0.016623923555016518,,
110
+ 0.0004160885000601411,0.01703418791294098,0.017000695690512657,54,2639,,,,
111
+ ,,,54,2639,0.00039480795385316014,0.016582660377025604,,
112
+ 0.000415789894759655,0.017159059643745422,0.017000695690512657,55,2687,,,,
113
+ ,,,55,2687,0.00040211877785623074,0.016737496480345726,,
114
+ 0.00041934612090699375,0.01727544330060482,0.017000695690512657,56,2735,,,,
115
+ ,,,56,2735,0.0003965380019508302,0.016589321196079254,,
116
+ 0.0004232667852193117,0.017047345638275146,0.017000695690512657,57,2783,,,,
117
+ ,,,57,2783,0.00039496805402450264,0.016631722450256348,,
118
+ 0.00041306737693957984,0.017058594152331352,0.017000695690512657,58,2831,,,,
119
+ ,,,58,2831,0.00039568173815496266,0.016634881496429443,,
120
+ 0.00041552481707185507,0.017004605382680893,0.017000695690512657,59,2879,,,,
121
+ ,,,59,2879,0.00039406149880960584,0.01660664565861225,,
122
+ 0.00042583755566738546,0.01705414429306984,0.017000695690512657,60,2927,,,,
123
+ ,,,60,2927,0.00039627321530133486,0.016636522486805916,,
124
+ 0.00041783798951655626,0.017244363203644753,0.017000695690512657,61,2975,,,,
125
+ ,,,61,2975,0.0003948824596591294,0.016557961702346802,,
126
+ 0.0004157483053859323,0.017151813954114914,0.017000695690512657,62,3023,,,,
127
+ ,,,62,3023,0.0004056752077303827,0.016779622063040733,,
128
+ 0.00041274423711001873,0.017053205519914627,0.017000695690512657,63,3071,,,,
129
+ ,,,63,3071,0.00039979253779165447,0.01668638363480568,,
130
+ 0.0004246786411385983,0.01736396737396717,0.017000695690512657,64,3119,,,,
131
+ ,,,64,3119,0.0003951334801968187,0.016622306779026985,,
132
+ 0.00042355924961157143,0.01705130562186241,0.017000695690512657,65,3167,,,,
133
+ ,,,65,3167,0.00039837457006797194,0.016673840582370758,,
134
+ 0.0004138355434406549,0.017032425850629807,0.017000695690512657,66,3215,,,,
135
+ ,,,66,3215,0.00040074967546388507,0.01668577454984188,,
136
+ 0.00042258683242835104,0.017326809465885162,0.017000695690512657,67,3263,,,,
137
+ ,,,67,3263,0.00039572324021719396,0.016614973545074463,,
138
+ 0.00041341199539601803,0.017015334218740463,0.017000695690512657,68,3311,,,,
139
+ ,,,68,3311,0.00039784895488992333,0.016692474484443665,,
140
+ 0.0004134845221415162,0.017086541280150414,0.017000695690512657,69,3359,,,,
141
+ ,,,69,3359,0.0003974255232606083,0.016638388857245445,,
142
+ 0.0004133299516979605,0.017100200057029724,0.017000695690512657,70,3407,,,,
143
+ ,,,70,3407,0.00039733582525514066,0.01667010597884655,,
144
+ 0.0004191308398731053,0.017036352306604385,0.017000695690512657,71,3455,,,,
145
+ ,,,71,3455,0.0003985513758379966,0.01668703928589821,,
146
+ 0.00041330925887450576,0.017060358077287674,0.017000695690512657,72,3503,,,,
147
+ ,,,72,3503,0.00040199141949415207,0.016739370301365852,,
148
+ 0.0004210953484289348,0.017061863094568253,0.017000695690512657,73,3551,,,,
149
+ ,,,73,3551,0.0003950465761590749,0.016600387170910835,,
150
+ 0.00042797168134711683,0.01714470051229,0.017000695690512657,74,3599,,,,
151
+ ,,,74,3599,0.00039893726352602243,0.01670648530125618,,
152
+ 0.0004146322899032384,0.017042651772499084,0.017000695690512657,75,3647,,,,
153
+ ,,,75,3647,0.00039998453576117754,0.016690267249941826,,
154
+ 0.0004134388000238687,0.01703176274895668,0.017000695690512657,76,3695,,,,
155
+ ,,,76,3695,0.00039601398748345673,0.016635997220873833,,
156
+ 0.00041642290307208896,0.017208458855748177,0.017000695690512657,77,3743,,,,
157
+ ,,,77,3743,0.00039686643867753446,0.016633793711662292,,
158
+ 0.00042115477845072746,0.017059164121747017,0.017000695690512657,78,3791,,,,
159
+ ,,,78,3791,0.00039846799336373806,0.016688905656337738,,
160
+ 0.000415122602134943,0.01717168651521206,0.017000695690512657,79,3839,,,,
161
+ ,,,79,3839,0.00039516520337201655,0.01661309413611889,,
162
+ 0.0004131546011194587,0.017082393169403076,0.017000695690512657,80,3887,,,,
163
+ ,,,80,3887,0.00039921721327118576,0.01671804115176201,,
164
+ 0.0004128122527617961,0.017054999247193336,0.017000695690512657,81,3935,,,,
165
+ ,,,81,3935,0.00039697522879578173,0.016647279262542725,,
166
+ 0.0004128477012272924,0.017052575945854187,0.017000695690512657,82,3983,,,,
167
+ ,,,82,3983,0.0003956857544835657,0.016641169786453247,,
168
+ 0.000417727860622108,0.017034592106938362,0.017000695690512657,83,4031,,,,
169
+ ,,,83,4031,0.0003957526932936162,0.01663590781390667,,
170
+ 0.0004379573801998049,0.017268991097807884,0.017000695690512657,84,4079,,,,
171
+ ,,,84,4079,0.0003935681888833642,0.016566380858421326,,
172
+ 0.0004142831894569099,0.01714366301894188,0.017000695690512657,85,4127,,,,
173
+ ,,,85,4127,0.0003956035361625254,0.01663059927523136,,
174
+ 0.00041444782982580364,0.017031202092766762,0.017000695690512657,86,4175,,,,
175
+ ,,,86,4175,0.0003978180466219783,0.016641462221741676,,
176
+ 0.0004190559557173401,0.017260581254959106,0.017000695690512657,87,4223,,,,
177
+ ,,,87,4223,0.0003951934340875596,0.01661202684044838,,
178
+ 0.00041441599023528397,0.017036981880664825,0.017000695690512657,88,4271,,,,
179
+ ,,,88,4271,0.0003961081965826452,0.016663121059536934,,
180
+ 0.00041358114685863256,0.017102306708693504,0.017000695690512657,89,4319,,,,
181
+ ,,,89,4319,0.0004015740123577416,0.016726816073060036,,
182
+ 0.0004130680172238499,0.017052575945854187,0.017000695690512657,90,4367,,,,
183
+ ,,,90,4367,0.0003977353044319898,0.016681544482707977,,
184
+ 0.0004177606606390327,0.017053550109267235,0.017000695690512657,91,4415,,,,
185
+ ,,,91,4415,0.00039689752156846225,0.01666596531867981,,
186
+ 0.0004127973807044327,0.01709304377436638,0.017000695690512657,92,4463,,,,
187
+ ,,,92,4463,0.00039606288191862404,0.016615284606814384,,
188
+ 0.0004147265281062573,0.017038505524396896,0.017000695690512657,93,4511,,,,
189
+ ,,,93,4511,0.0003999987675342709,0.016702203080058098,,
190
+ 0.0004209577164147049,0.017128383740782738,0.017000695690512657,94,4559,,,,
191
+ ,,,94,4559,0.00040548722608946264,0.01684720255434513,,
192
+ 0.0004143503319937736,0.01708775945007801,0.017000695690512657,95,4607,,,,
193
+ ,,,95,4607,0.0004048148693982512,0.016792071983218193,,
194
+ 0.00041501803207211196,0.01715242862701416,0.017000695690512657,96,4655,,,,
195
+ ,,,96,4655,0.0003974545397795737,0.016677362844347954,,
196
+ 0.00041411182610318065,0.01708676666021347,0.017000695690512657,97,4703,,,,
197
+ ,,,97,4703,0.0004028358671348542,0.016793260350823402,,
198
+ 0.0004176621150691062,0.01721743308007717,0.017000695690512657,98,4751,,,,
199
+ ,,,98,4751,0.0004061829822603613,0.01681651920080185,,
200
+ 0.0004173672932665795,0.017215680330991745,0.017000695690512657,99,4799,,,,
201
+ ,,,99,4799,0.0004006385861430317,0.016772150993347168,,
202
+ ,,,54,2592,,,0.00040279902168549597,0.016626913100481033
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/tensorboard/focusConvMSE_150/events.out.tfevents.1647960193.0c30fcf2cd8d.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49eb2c644c8762c51da4b305c064b66211cadbfa1a8b19aa784b4f3a291c5fdb
3
+ size 36866
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/tensorboard/focusConvMSE_150/events.out.tfevents.1647960610.0c30fcf2cd8d.1.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d68ce1026be4c14b8a4d9a2db6edfd221b801a67024344d4afccc71f13311c5
3
+ size 179
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-43-11/tensorboard/focusConvMSE_150/hparams.yaml ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 1
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_conv_module.FocusConvLitModule
9
+ image_size: 150
10
+ pool_size: 2
11
+ conv1_size: 3
12
+ conv1_channels: 9
13
+ conv2_size: 7
14
+ conv2_channels: 6
15
+ lin1_size: 32
16
+ lin2_size: 72
17
+ output_size: 1
18
+ lr: 0.001
19
+ weight_decay: 0.0005
20
+ datamodule:
21
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
22
+ data_dir: /usr/src/app/data/focus150
23
+ csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
24
+ csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
25
+ csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
26
+ batch_size: 64
27
+ num_workers: 0
28
+ pin_memory: false
29
+ augmentation: true
30
+ seed: 12345
31
+ callbacks:
32
+ model_checkpoint:
33
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
34
+ monitor: val/mae
35
+ mode: min
36
+ save_top_k: 1
37
+ save_last: true
38
+ verbose: false
39
+ dirpath: checkpoints/
40
+ filename: epoch_{epoch:03d}
41
+ auto_insert_metric_name: false
42
+ early_stopping:
43
+ _target_: pytorch_lightning.callbacks.EarlyStopping
44
+ monitor: val/mae
45
+ mode: min
46
+ patience: 100
47
+ min_delta: 0
48
+ model_summary:
49
+ _target_: pytorch_lightning.callbacks.RichModelSummary
50
+ max_depth: -1
51
+ rich_progress_bar:
52
+ _target_: pytorch_lightning.callbacks.RichProgressBar
53
+ model/params/total: 216201
54
+ model/params/trainable: 216201
55
+ model/params/non_trainable: 0
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/meta.yaml ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ artifact_uri: /usr/src/app/logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/artifacts
2
+ end_time: 1647960610283
3
+ entry_point_name: ''
4
+ experiment_id: '1'
5
+ lifecycle_stage: active
6
+ name: ''
7
+ run_id: 40b80b1e6d4e48918cf7254779a3cb9a
8
+ run_uuid: 40b80b1e6d4e48918cf7254779a3cb9a
9
+ source_name: ''
10
+ source_type: 4
11
+ source_version: ''
12
+ start_time: 1647960193779
13
+ status: 3
14
+ tags: []
15
+ user_id: unknown
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/epoch ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647960203460 0 47
2
+ 1647960203502 0 47
3
+ 1647960207355 1 95
4
+ 1647960207402 1 95
5
+ 1647960211265 2 143
6
+ 1647960211334 2 143
7
+ 1647960215069 3 191
8
+ 1647960215108 3 191
9
+ 1647960218802 4 239
10
+ 1647960218844 4 239
11
+ 1647960222628 5 287
12
+ 1647960222676 5 287
13
+ 1647960226535 6 335
14
+ 1647960226580 6 335
15
+ 1647960230437 7 383
16
+ 1647960230480 7 383
17
+ 1647960234336 8 431
18
+ 1647960234376 8 431
19
+ 1647960238394 9 479
20
+ 1647960238433 9 479
21
+ 1647960242373 10 527
22
+ 1647960242406 10 527
23
+ 1647960246290 11 575
24
+ 1647960246329 11 575
25
+ 1647960250343 12 623
26
+ 1647960250375 12 623
27
+ 1647960254256 13 671
28
+ 1647960254288 13 671
29
+ 1647960258205 14 719
30
+ 1647960258251 14 719
31
+ 1647960262182 15 767
32
+ 1647960262224 15 767
33
+ 1647960266171 16 815
34
+ 1647960266207 16 815
35
+ 1647960270080 17 863
36
+ 1647960270116 17 863
37
+ 1647960273989 18 911
38
+ 1647960274023 18 911
39
+ 1647960277985 19 959
40
+ 1647960278022 19 959
41
+ 1647960281884 20 1007
42
+ 1647960281927 20 1007
43
+ 1647960285897 21 1055
44
+ 1647960285946 21 1055
45
+ 1647960289891 22 1103
46
+ 1647960289931 22 1103
47
+ 1647960293812 23 1151
48
+ 1647960293852 23 1151
49
+ 1647960297773 24 1199
50
+ 1647960297855 24 1199
51
+ 1647960301818 25 1247
52
+ 1647960301863 25 1247
53
+ 1647960305717 26 1295
54
+ 1647960305757 26 1295
55
+ 1647960309717 27 1343
56
+ 1647960309764 27 1343
57
+ 1647960313710 28 1391
58
+ 1647960313763 28 1391
59
+ 1647960317652 29 1439
60
+ 1647960317691 29 1439
61
+ 1647960321790 30 1487
62
+ 1647960321823 30 1487
63
+ 1647960325805 31 1535
64
+ 1647960325851 31 1535
65
+ 1647960329769 32 1583
66
+ 1647960329803 32 1583
67
+ 1647960333766 33 1631
68
+ 1647960333801 33 1631
69
+ 1647960337730 34 1679
70
+ 1647960337773 34 1679
71
+ 1647960341774 35 1727
72
+ 1647960341810 35 1727
73
+ 1647960345738 36 1775
74
+ 1647960345778 36 1775
75
+ 1647960349713 37 1823
76
+ 1647960349764 37 1823
77
+ 1647960353519 38 1871
78
+ 1647960353552 38 1871
79
+ 1647960357560 39 1919
80
+ 1647960357601 39 1919
81
+ 1647960361641 40 1967
82
+ 1647960361674 40 1967
83
+ 1647960365536 41 2015
84
+ 1647960365576 41 2015
85
+ 1647960369371 42 2063
86
+ 1647960369404 42 2063
87
+ 1647960373310 43 2111
88
+ 1647960373344 43 2111
89
+ 1647960377280 44 2159
90
+ 1647960377320 44 2159
91
+ 1647960381231 45 2207
92
+ 1647960381273 45 2207
93
+ 1647960385241 46 2255
94
+ 1647960385282 46 2255
95
+ 1647960389218 47 2303
96
+ 1647960389255 47 2303
97
+ 1647960393083 48 2351
98
+ 1647960393129 48 2351
99
+ 1647960397103 49 2399
100
+ 1647960397175 49 2399
101
+ 1647960401102 50 2447
102
+ 1647960401145 50 2447
103
+ 1647960405089 51 2495
104
+ 1647960405121 51 2495
105
+ 1647960409069 52 2543
106
+ 1647960409105 52 2543
107
+ 1647960413019 53 2591
108
+ 1647960413064 53 2591
109
+ 1647960416969 54 2639
110
+ 1647960417010 54 2639
111
+ 1647960420975 55 2687
112
+ 1647960421014 55 2687
113
+ 1647960424779 56 2735
114
+ 1647960424816 56 2735
115
+ 1647960428682 57 2783
116
+ 1647960428727 57 2783
117
+ 1647960432629 58 2831
118
+ 1647960432665 58 2831
119
+ 1647960436520 59 2879
120
+ 1647960436558 59 2879
121
+ 1647960440395 60 2927
122
+ 1647960440428 60 2927
123
+ 1647960444342 61 2975
124
+ 1647960444376 61 2975
125
+ 1647960448284 62 3023
126
+ 1647960448320 62 3023
127
+ 1647960452162 63 3071
128
+ 1647960452199 63 3071
129
+ 1647960456158 64 3119
130
+ 1647960456194 64 3119
131
+ 1647960460171 65 3167
132
+ 1647960460205 65 3167
133
+ 1647960464064 66 3215
134
+ 1647960464103 66 3215
135
+ 1647960467957 67 3263
136
+ 1647960467997 67 3263
137
+ 1647960473299 68 3311
138
+ 1647960473343 68 3311
139
+ 1647960477374 69 3359
140
+ 1647960477408 69 3359
141
+ 1647960481415 70 3407
142
+ 1647960481449 70 3407
143
+ 1647960485799 71 3455
144
+ 1647960485844 71 3455
145
+ 1647960489849 72 3503
146
+ 1647960489902 72 3503
147
+ 1647960494420 73 3551
148
+ 1647960494469 73 3551
149
+ 1647960499349 74 3599
150
+ 1647960499849 74 3599
151
+ 1647960504459 75 3647
152
+ 1647960504564 75 3647
153
+ 1647960508707 76 3695
154
+ 1647960508914 76 3695
155
+ 1647960513159 77 3743
156
+ 1647960513367 77 3743
157
+ 1647960517546 78 3791
158
+ 1647960517697 78 3791
159
+ 1647960522033 79 3839
160
+ 1647960522220 79 3839
161
+ 1647960526250 80 3887
162
+ 1647960526296 80 3887
163
+ 1647960530977 81 3935
164
+ 1647960531017 81 3935
165
+ 1647960535476 82 3983
166
+ 1647960535708 82 3983
167
+ 1647960540098 83 4031
168
+ 1647960540290 83 4031
169
+ 1647960544532 84 4079
170
+ 1647960544773 84 4079
171
+ 1647960548960 85 4127
172
+ 1647960549111 85 4127
173
+ 1647960553203 86 4175
174
+ 1647960553366 86 4175
175
+ 1647960557539 87 4223
176
+ 1647960557627 87 4223
177
+ 1647960562267 88 4271
178
+ 1647960562440 88 4271
179
+ 1647960566815 89 4319
180
+ 1647960567054 89 4319
181
+ 1647960571435 90 4367
182
+ 1647960571841 90 4367
183
+ 1647960575958 91 4415
184
+ 1647960576331 91 4415
185
+ 1647960580400 92 4463
186
+ 1647960580568 92 4463
187
+ 1647960584677 93 4511
188
+ 1647960584854 93 4511
189
+ 1647960588839 94 4559
190
+ 1647960588927 94 4559
191
+ 1647960593001 95 4607
192
+ 1647960593098 95 4607
193
+ 1647960597173 96 4655
194
+ 1647960597237 96 4655
195
+ 1647960601382 97 4703
196
+ 1647960601471 97 4703
197
+ 1647960605473 98 4751
198
+ 1647960605555 98 4751
199
+ 1647960609428 99 4799
200
+ 1647960609584 99 4799
201
+ 1647960610145 54 2592
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/test/loss ADDED
@@ -0,0 +1 @@
 
 
1
+ 1647960610145 0.00040279902168549597 2592
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/test/mae ADDED
@@ -0,0 +1 @@
 
 
1
+ 1647960610145 0.016626913100481033 2592
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/train/loss ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647960203502 0.07324540615081787 47
2
+ 1647960207402 0.0005119902198202908 95
3
+ 1647960211334 0.0004279178101569414 143
4
+ 1647960215108 0.0004204828874208033 191
5
+ 1647960218844 0.0004181377007625997 239
6
+ 1647960222676 0.0004158291267231107 287
7
+ 1647960226580 0.0004123631224501878 335
8
+ 1647960230480 0.00040746244485490024 383
9
+ 1647960234376 0.00041121308458968997 431
10
+ 1647960238433 0.00040692713810130954 479
11
+ 1647960242406 0.00040299276588484645 527
12
+ 1647960246329 0.00040395042742602527 575
13
+ 1647960250375 0.0004016980528831482 623
14
+ 1647960254288 0.00040133227594196796 671
15
+ 1647960258251 0.00040109638939611614 719
16
+ 1647960262224 0.0004060292267240584 767
17
+ 1647960266207 0.00040416873525828123 815
18
+ 1647960270116 0.0004033875884488225 863
19
+ 1647960274023 0.0003990893601439893 911
20
+ 1647960278022 0.000404653896111995 959
21
+ 1647960281927 0.00040122432983480394 1007
22
+ 1647960285946 0.0004034280136693269 1055
23
+ 1647960289931 0.0003990979748778045 1103
24
+ 1647960293852 0.00040245126001536846 1151
25
+ 1647960297855 0.00040643103420734406 1199
26
+ 1647960301863 0.000400546588934958 1247
27
+ 1647960305757 0.0003996045561507344 1295
28
+ 1647960309764 0.00039723876398056746 1343
29
+ 1647960313763 0.0003959255700465292 1391
30
+ 1647960317691 0.000394946604501456 1439
31
+ 1647960321823 0.0003974179271608591 1487
32
+ 1647960325851 0.00039746827678754926 1535
33
+ 1647960329803 0.0003963193448726088 1583
34
+ 1647960333801 0.0003986273950431496 1631
35
+ 1647960337773 0.000402714591473341 1679
36
+ 1647960341810 0.00039801615639589727 1727
37
+ 1647960345778 0.0004030045529361814 1775
38
+ 1647960349764 0.0004063526284880936 1823
39
+ 1647960353552 0.0003981548943556845 1871
40
+ 1647960357601 0.0003961462643928826 1919
41
+ 1647960361674 0.0004012832068838179 1967
42
+ 1647960365576 0.0003976260486524552 2015
43
+ 1647960369404 0.0003990065888501704 2063
44
+ 1647960373344 0.00039825798012316227 2111
45
+ 1647960377320 0.00039614003617316484 2159
46
+ 1647960381273 0.00039644099888391793 2207
47
+ 1647960385282 0.00039509835187345743 2255
48
+ 1647960389255 0.00039680046029388905 2303
49
+ 1647960393129 0.00039851333713158965 2351
50
+ 1647960397175 0.00039479872793890536 2399
51
+ 1647960401145 0.0003952273400500417 2447
52
+ 1647960405121 0.0003985257644671947 2495
53
+ 1647960409105 0.00039632743573747575 2543
54
+ 1647960413064 0.0003982774505857378 2591
55
+ 1647960417010 0.00039480795385316014 2639
56
+ 1647960421014 0.00040211877785623074 2687
57
+ 1647960424816 0.0003965380019508302 2735
58
+ 1647960428727 0.00039496805402450264 2783
59
+ 1647960432665 0.00039568173815496266 2831
60
+ 1647960436558 0.00039406149880960584 2879
61
+ 1647960440428 0.00039627321530133486 2927
62
+ 1647960444376 0.0003948824596591294 2975
63
+ 1647960448320 0.0004056752077303827 3023
64
+ 1647960452199 0.00039979253779165447 3071
65
+ 1647960456194 0.0003951334801968187 3119
66
+ 1647960460205 0.00039837457006797194 3167
67
+ 1647960464103 0.00040074967546388507 3215
68
+ 1647960467997 0.00039572324021719396 3263
69
+ 1647960473343 0.00039784895488992333 3311
70
+ 1647960477408 0.0003974255232606083 3359
71
+ 1647960481449 0.00039733582525514066 3407
72
+ 1647960485844 0.0003985513758379966 3455
73
+ 1647960489902 0.00040199141949415207 3503
74
+ 1647960494469 0.0003950465761590749 3551
75
+ 1647960499849 0.00039893726352602243 3599
76
+ 1647960504564 0.00039998453576117754 3647
77
+ 1647960508914 0.00039601398748345673 3695
78
+ 1647960513367 0.00039686643867753446 3743
79
+ 1647960517697 0.00039846799336373806 3791
80
+ 1647960522220 0.00039516520337201655 3839
81
+ 1647960526296 0.00039921721327118576 3887
82
+ 1647960531017 0.00039697522879578173 3935
83
+ 1647960535708 0.0003956857544835657 3983
84
+ 1647960540290 0.0003957526932936162 4031
85
+ 1647960544773 0.0003935681888833642 4079
86
+ 1647960549111 0.0003956035361625254 4127
87
+ 1647960553366 0.0003978180466219783 4175
88
+ 1647960557627 0.0003951934340875596 4223
89
+ 1647960562440 0.0003961081965826452 4271
90
+ 1647960567054 0.0004015740123577416 4319
91
+ 1647960571841 0.0003977353044319898 4367
92
+ 1647960576331 0.00039689752156846225 4415
93
+ 1647960580568 0.00039606288191862404 4463
94
+ 1647960584854 0.0003999987675342709 4511
95
+ 1647960588927 0.00040548722608946264 4559
96
+ 1647960593098 0.0004048148693982512 4607
97
+ 1647960597237 0.0003974545397795737 4655
98
+ 1647960601471 0.0004028358671348542 4703
99
+ 1647960605555 0.0004061829822603613 4751
100
+ 1647960609584 0.0004006385861430317 4799
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/train/mae ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647960203502 0.15608204901218414 47
2
+ 1647960207402 0.01848370023071766 95
3
+ 1647960211334 0.017223240807652473 143
4
+ 1647960215108 0.017105428501963615 191
5
+ 1647960218844 0.01710123009979725 239
6
+ 1647960222676 0.017015699297189713 287
7
+ 1647960226580 0.016972672194242477 335
8
+ 1647960230480 0.016849098727107048 383
9
+ 1647960234376 0.016946695744991302 431
10
+ 1647960238433 0.0168300773948431 479
11
+ 1647960242406 0.016762785613536835 527
12
+ 1647960246329 0.01680273935198784 575
13
+ 1647960250375 0.01675964891910553 623
14
+ 1647960254288 0.01673046499490738 671
15
+ 1647960258251 0.016756772994995117 719
16
+ 1647960262224 0.016798974946141243 767
17
+ 1647960266207 0.016754010692238808 815
18
+ 1647960270116 0.016784122213721275 863
19
+ 1647960274023 0.01666918769478798 911
20
+ 1647960278022 0.016769438982009888 959
21
+ 1647960281927 0.016709182411432266 1007
22
+ 1647960285946 0.016742847859859467 1055
23
+ 1647960289931 0.016647830605506897 1103
24
+ 1647960293852 0.016745105385780334 1151
25
+ 1647960297855 0.016855906695127487 1199
26
+ 1647960301863 0.016742048785090446 1247
27
+ 1647960305757 0.016682812944054604 1295
28
+ 1647960309764 0.016645090654492378 1343
29
+ 1647960313763 0.016611488536000252 1391
30
+ 1647960317691 0.01659465581178665 1439
31
+ 1647960321823 0.016647476702928543 1487
32
+ 1647960325851 0.01665402390062809 1535
33
+ 1647960329803 0.016615083441138268 1583
34
+ 1647960333801 0.016656169667840004 1631
35
+ 1647960337773 0.01677883230149746 1679
36
+ 1647960341810 0.016634121537208557 1727
37
+ 1647960345778 0.016751982271671295 1775
38
+ 1647960349764 0.016803236678242683 1823
39
+ 1647960353552 0.016650604084134102 1871
40
+ 1647960357601 0.016624847427010536 1919
41
+ 1647960361674 0.01667129620909691 1967
42
+ 1647960365576 0.016667962074279785 2015
43
+ 1647960369404 0.016656573861837387 2063
44
+ 1647960373344 0.016642922535538673 2111
45
+ 1647960377320 0.016659162938594818 2159
46
+ 1647960381273 0.016649728640913963 2207
47
+ 1647960385282 0.016588879749178886 2255
48
+ 1647960389255 0.01660764217376709 2303
49
+ 1647960393129 0.016675254330039024 2351
50
+ 1647960397175 0.016594236716628075 2399
51
+ 1647960401145 0.01660444773733616 2447
52
+ 1647960405121 0.01664111576974392 2495
53
+ 1647960409105 0.0166194848716259 2543
54
+ 1647960413064 0.016623923555016518 2591
55
+ 1647960417010 0.016582660377025604 2639
56
+ 1647960421014 0.016737496480345726 2687
57
+ 1647960424816 0.016589321196079254 2735
58
+ 1647960428727 0.016631722450256348 2783
59
+ 1647960432665 0.016634881496429443 2831
60
+ 1647960436558 0.01660664565861225 2879
61
+ 1647960440428 0.016636522486805916 2927
62
+ 1647960444376 0.016557961702346802 2975
63
+ 1647960448320 0.016779622063040733 3023
64
+ 1647960452199 0.01668638363480568 3071
65
+ 1647960456194 0.016622306779026985 3119
66
+ 1647960460205 0.016673840582370758 3167
67
+ 1647960464103 0.01668577454984188 3215
68
+ 1647960467997 0.016614973545074463 3263
69
+ 1647960473343 0.016692474484443665 3311
70
+ 1647960477408 0.016638388857245445 3359
71
+ 1647960481449 0.01667010597884655 3407
72
+ 1647960485844 0.01668703928589821 3455
73
+ 1647960489902 0.016739370301365852 3503
74
+ 1647960494469 0.016600387170910835 3551
75
+ 1647960499849 0.01670648530125618 3599
76
+ 1647960504564 0.016690267249941826 3647
77
+ 1647960508914 0.016635997220873833 3695
78
+ 1647960513367 0.016633793711662292 3743
79
+ 1647960517697 0.016688905656337738 3791
80
+ 1647960522220 0.01661309413611889 3839
81
+ 1647960526296 0.01671804115176201 3887
82
+ 1647960531017 0.016647279262542725 3935
83
+ 1647960535708 0.016641169786453247 3983
84
+ 1647960540290 0.01663590781390667 4031
85
+ 1647960544773 0.016566380858421326 4079
86
+ 1647960549111 0.01663059927523136 4127
87
+ 1647960553366 0.016641462221741676 4175
88
+ 1647960557627 0.01661202684044838 4223
89
+ 1647960562440 0.016663121059536934 4271
90
+ 1647960567054 0.016726816073060036 4319
91
+ 1647960571841 0.016681544482707977 4367
92
+ 1647960576331 0.01666596531867981 4415
93
+ 1647960580568 0.016615284606814384 4463
94
+ 1647960584854 0.016702203080058098 4511
95
+ 1647960588927 0.01684720255434513 4559
96
+ 1647960593098 0.016792071983218193 4607
97
+ 1647960597237 0.016677362844347954 4655
98
+ 1647960601471 0.016793260350823402 4703
99
+ 1647960605555 0.01681651920080185 4751
100
+ 1647960609584 0.016772150993347168 4799
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/val/loss ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647960203460 0.0010520756477490067 47
2
+ 1647960207355 0.0004612061893567443 95
3
+ 1647960211265 0.0004499765345826745 143
4
+ 1647960215069 0.00044631771743297577 191
5
+ 1647960218802 0.0004415122966747731 239
6
+ 1647960222628 0.000438133894931525 287
7
+ 1647960226535 0.00043563765939325094 335
8
+ 1647960230437 0.0004368386580608785 383
9
+ 1647960234336 0.0004356758145149797 431
10
+ 1647960238394 0.00043189621646888554 479
11
+ 1647960242373 0.000431660155300051 527
12
+ 1647960246290 0.0004278685664758086 575
13
+ 1647960250343 0.00042351684533059597 623
14
+ 1647960254256 0.00042290761484764516 671
15
+ 1647960258205 0.00042417016811668873 719
16
+ 1647960262182 0.0004259877896402031 767
17
+ 1647960266171 0.00042129578650929034 815
18
+ 1647960270080 0.00042111988295800984 863
19
+ 1647960273989 0.0004269638448022306 911
20
+ 1647960277985 0.00042243217467330396 959
21
+ 1647960281884 0.0004220800183247775 1007
22
+ 1647960285897 0.0004301489389035851 1055
23
+ 1647960289891 0.00042216639849357307 1103
24
+ 1647960293812 0.0004318137071095407 1151
25
+ 1647960297773 0.0004194175999145955 1199
26
+ 1647960301818 0.00042441036202944815 1247
27
+ 1647960305717 0.00042018183739855886 1295
28
+ 1647960309717 0.0004190116305835545 1343
29
+ 1647960313710 0.0004195934161543846 1391
30
+ 1647960317652 0.00041890854481607676 1439
31
+ 1647960321790 0.0004188215534668416 1487
32
+ 1647960325805 0.0004203834687359631 1535
33
+ 1647960329769 0.0004258453845977783 1583
34
+ 1647960333766 0.0004177540831733495 1631
35
+ 1647960337730 0.00042232274427078664 1679
36
+ 1647960341774 0.0004189989995211363 1727
37
+ 1647960345738 0.00042335118632763624 1775
38
+ 1647960349713 0.00041969664744101465 1823
39
+ 1647960353519 0.00041658285772427917 1871
40
+ 1647960357560 0.000418454670580104 1919
41
+ 1647960361641 0.00041677828994579613 1967
42
+ 1647960365536 0.00042211331310682 2015
43
+ 1647960369371 0.0004176416841801256 2063
44
+ 1647960373310 0.0004183136916253716 2111
45
+ 1647960377280 0.0004184833960607648 2159
46
+ 1647960381231 0.00042137890704907477 2207
47
+ 1647960385241 0.00041784756467677653 2255
48
+ 1647960389218 0.00041607263847254217 2303
49
+ 1647960393083 0.00041994749335572124 2351
50
+ 1647960397103 0.00041497754864394665 2399
51
+ 1647960401102 0.00041593355126678944 2447
52
+ 1647960405089 0.0004291844379622489 2495
53
+ 1647960409069 0.00042258668690919876 2543
54
+ 1647960413019 0.00041858889744617045 2591
55
+ 1647960416969 0.0004160885000601411 2639
56
+ 1647960420975 0.000415789894759655 2687
57
+ 1647960424779 0.00041934612090699375 2735
58
+ 1647960428682 0.0004232667852193117 2783
59
+ 1647960432629 0.00041306737693957984 2831
60
+ 1647960436520 0.00041552481707185507 2879
61
+ 1647960440395 0.00042583755566738546 2927
62
+ 1647960444342 0.00041783798951655626 2975
63
+ 1647960448284 0.0004157483053859323 3023
64
+ 1647960452162 0.00041274423711001873 3071
65
+ 1647960456158 0.0004246786411385983 3119
66
+ 1647960460171 0.00042355924961157143 3167
67
+ 1647960464064 0.0004138355434406549 3215
68
+ 1647960467957 0.00042258683242835104 3263
69
+ 1647960473299 0.00041341199539601803 3311
70
+ 1647960477374 0.0004134845221415162 3359
71
+ 1647960481415 0.0004133299516979605 3407
72
+ 1647960485799 0.0004191308398731053 3455
73
+ 1647960489849 0.00041330925887450576 3503
74
+ 1647960494420 0.0004210953484289348 3551
75
+ 1647960499349 0.00042797168134711683 3599
76
+ 1647960504459 0.0004146322899032384 3647
77
+ 1647960508707 0.0004134388000238687 3695
78
+ 1647960513159 0.00041642290307208896 3743
79
+ 1647960517546 0.00042115477845072746 3791
80
+ 1647960522033 0.000415122602134943 3839
81
+ 1647960526250 0.0004131546011194587 3887
82
+ 1647960530977 0.0004128122527617961 3935
83
+ 1647960535476 0.0004128477012272924 3983
84
+ 1647960540098 0.000417727860622108 4031
85
+ 1647960544532 0.0004379573801998049 4079
86
+ 1647960548960 0.0004142831894569099 4127
87
+ 1647960553203 0.00041444782982580364 4175
88
+ 1647960557539 0.0004190559557173401 4223
89
+ 1647960562267 0.00041441599023528397 4271
90
+ 1647960566815 0.00041358114685863256 4319
91
+ 1647960571435 0.0004130680172238499 4367
92
+ 1647960575958 0.0004177606606390327 4415
93
+ 1647960580400 0.0004127973807044327 4463
94
+ 1647960584677 0.0004147265281062573 4511
95
+ 1647960588839 0.0004209577164147049 4559
96
+ 1647960593001 0.0004143503319937736 4607
97
+ 1647960597173 0.00041501803207211196 4655
98
+ 1647960601382 0.00041411182610318065 4703
99
+ 1647960605473 0.0004176621150691062 4751
100
+ 1647960609428 0.0004173672932665795 4799
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/val/mae ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647960203460 0.027000371366739273 47
2
+ 1647960207355 0.017776092514395714 95
3
+ 1647960211265 0.017643338069319725 143
4
+ 1647960215069 0.017654919996857643 191
5
+ 1647960218802 0.017546169459819794 239
6
+ 1647960222628 0.017503635957837105 287
7
+ 1647960226535 0.017501212656497955 335
8
+ 1647960230437 0.017341317608952522 383
9
+ 1647960234336 0.017313677817583084 431
10
+ 1647960238394 0.0172542966902256 479
11
+ 1647960242373 0.017524439841508865 527
12
+ 1647960246290 0.017209237441420555 575
13
+ 1647960250343 0.017291804775595665 623
14
+ 1647960254256 0.01727038435637951 671
15
+ 1647960258205 0.017126740887761116 719
16
+ 1647960262182 0.017122970893979073 767
17
+ 1647960266171 0.017265643924474716 815
18
+ 1647960270080 0.017123492434620857 863
19
+ 1647960273989 0.017454244196414948 911
20
+ 1647960277985 0.017306191846728325 959
21
+ 1647960281884 0.017325717955827713 1007
22
+ 1647960285897 0.017113929614424706 1055
23
+ 1647960289891 0.01728595420718193 1103
24
+ 1647960293812 0.017589587718248367 1151
25
+ 1647960297773 0.017076712101697922 1199
26
+ 1647960301818 0.01708088628947735 1247
27
+ 1647960305717 0.01709458790719509 1295
28
+ 1647960309717 0.017154676839709282 1343
29
+ 1647960313710 0.017233334481716156 1391
30
+ 1647960317652 0.017193002626299858 1439
31
+ 1647960321790 0.017117813229560852 1487
32
+ 1647960325805 0.01704377681016922 1535
33
+ 1647960329769 0.017442181706428528 1583
34
+ 1647960333766 0.017155330628156662 1631
35
+ 1647960337730 0.017022596672177315 1679
36
+ 1647960341774 0.017094330862164497 1727
37
+ 1647960345738 0.01704075001180172 1775
38
+ 1647960349713 0.017027610912919044 1823
39
+ 1647960353519 0.01709923706948757 1871
40
+ 1647960357560 0.01721111498773098 1919
41
+ 1647960361641 0.01710767298936844 1967
42
+ 1647960365536 0.017018204554915428 2015
43
+ 1647960369371 0.017177345231175423 2063
44
+ 1647960373310 0.01718120276927948 2111
45
+ 1647960377280 0.017036592587828636 2159
46
+ 1647960381231 0.017030268907546997 2207
47
+ 1647960385241 0.017211318016052246 2255
48
+ 1647960389218 0.017052004113793373 2303
49
+ 1647960393083 0.01700209081172943 2351
50
+ 1647960397103 0.01702883280813694 2399
51
+ 1647960401102 0.01700270175933838 2447
52
+ 1647960405089 0.017490429803729057 2495
53
+ 1647960409069 0.017022263258695602 2543
54
+ 1647960413019 0.017000695690512657 2591
55
+ 1647960416969 0.01703418791294098 2639
56
+ 1647960420975 0.017159059643745422 2687
57
+ 1647960424779 0.01727544330060482 2735
58
+ 1647960428682 0.017047345638275146 2783
59
+ 1647960432629 0.017058594152331352 2831
60
+ 1647960436520 0.017004605382680893 2879
61
+ 1647960440395 0.01705414429306984 2927
62
+ 1647960444342 0.017244363203644753 2975
63
+ 1647960448284 0.017151813954114914 3023
64
+ 1647960452162 0.017053205519914627 3071
65
+ 1647960456158 0.01736396737396717 3119
66
+ 1647960460171 0.01705130562186241 3167
67
+ 1647960464064 0.017032425850629807 3215
68
+ 1647960467957 0.017326809465885162 3263
69
+ 1647960473299 0.017015334218740463 3311
70
+ 1647960477374 0.017086541280150414 3359
71
+ 1647960481415 0.017100200057029724 3407
72
+ 1647960485799 0.017036352306604385 3455
73
+ 1647960489849 0.017060358077287674 3503
74
+ 1647960494420 0.017061863094568253 3551
75
+ 1647960499349 0.01714470051229 3599
76
+ 1647960504459 0.017042651772499084 3647
77
+ 1647960508707 0.01703176274895668 3695
78
+ 1647960513159 0.017208458855748177 3743
79
+ 1647960517546 0.017059164121747017 3791
80
+ 1647960522033 0.01717168651521206 3839
81
+ 1647960526250 0.017082393169403076 3887
82
+ 1647960530977 0.017054999247193336 3935
83
+ 1647960535476 0.017052575945854187 3983
84
+ 1647960540098 0.017034592106938362 4031
85
+ 1647960544532 0.017268991097807884 4079
86
+ 1647960548960 0.01714366301894188 4127
87
+ 1647960553203 0.017031202092766762 4175
88
+ 1647960557539 0.017260581254959106 4223
89
+ 1647960562267 0.017036981880664825 4271
90
+ 1647960566815 0.017102306708693504 4319
91
+ 1647960571435 0.017052575945854187 4367
92
+ 1647960575958 0.017053550109267235 4415
93
+ 1647960580400 0.01709304377436638 4463
94
+ 1647960584677 0.017038505524396896 4511
95
+ 1647960588839 0.017128383740782738 4559
96
+ 1647960593001 0.01708775945007801 4607
97
+ 1647960597173 0.01715242862701416 4655
98
+ 1647960601382 0.01708676666021347 4703
99
+ 1647960605473 0.01721743308007717 4751
100
+ 1647960609428 0.017215680330991745 4799
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/metrics/val/mae_best ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647960203460 0.024726606905460358 47
2
+ 1647960207355 0.017776092514395714 95
3
+ 1647960211265 0.017643338069319725 143
4
+ 1647960215069 0.017643338069319725 191
5
+ 1647960218802 0.017546169459819794 239
6
+ 1647960222628 0.017503635957837105 287
7
+ 1647960226535 0.017501212656497955 335
8
+ 1647960230437 0.017341317608952522 383
9
+ 1647960234336 0.017313677817583084 431
10
+ 1647960238394 0.0172542966902256 479
11
+ 1647960242373 0.0172542966902256 527
12
+ 1647960246290 0.017209237441420555 575
13
+ 1647960250343 0.017209237441420555 623
14
+ 1647960254256 0.017209237441420555 671
15
+ 1647960258205 0.017126740887761116 719
16
+ 1647960262182 0.017122970893979073 767
17
+ 1647960266171 0.017122970893979073 815
18
+ 1647960270080 0.017122970893979073 863
19
+ 1647960273989 0.017122970893979073 911
20
+ 1647960277985 0.017122970893979073 959
21
+ 1647960281884 0.017122970893979073 1007
22
+ 1647960285897 0.017113929614424706 1055
23
+ 1647960289891 0.017113929614424706 1103
24
+ 1647960293812 0.017113929614424706 1151
25
+ 1647960297773 0.017076712101697922 1199
26
+ 1647960301818 0.017076712101697922 1247
27
+ 1647960305717 0.017076712101697922 1295
28
+ 1647960309717 0.017076712101697922 1343
29
+ 1647960313710 0.017076712101697922 1391
30
+ 1647960317652 0.017076712101697922 1439
31
+ 1647960321790 0.017076712101697922 1487
32
+ 1647960325805 0.01704377681016922 1535
33
+ 1647960329769 0.01704377681016922 1583
34
+ 1647960333766 0.01704377681016922 1631
35
+ 1647960337730 0.017022596672177315 1679
36
+ 1647960341774 0.017022596672177315 1727
37
+ 1647960345738 0.017022596672177315 1775
38
+ 1647960349713 0.017022596672177315 1823
39
+ 1647960353519 0.017022596672177315 1871
40
+ 1647960357560 0.017022596672177315 1919
41
+ 1647960361641 0.017022596672177315 1967
42
+ 1647960365536 0.017018204554915428 2015
43
+ 1647960369371 0.017018204554915428 2063
44
+ 1647960373310 0.017018204554915428 2111
45
+ 1647960377280 0.017018204554915428 2159
46
+ 1647960381231 0.017018204554915428 2207
47
+ 1647960385241 0.017018204554915428 2255
48
+ 1647960389218 0.017018204554915428 2303
49
+ 1647960393083 0.01700209081172943 2351
50
+ 1647960397103 0.01700209081172943 2399
51
+ 1647960401102 0.01700209081172943 2447
52
+ 1647960405089 0.01700209081172943 2495
53
+ 1647960409069 0.01700209081172943 2543
54
+ 1647960413019 0.017000695690512657 2591
55
+ 1647960416969 0.017000695690512657 2639
56
+ 1647960420975 0.017000695690512657 2687
57
+ 1647960424779 0.017000695690512657 2735
58
+ 1647960428682 0.017000695690512657 2783
59
+ 1647960432629 0.017000695690512657 2831
60
+ 1647960436520 0.017000695690512657 2879
61
+ 1647960440395 0.017000695690512657 2927
62
+ 1647960444342 0.017000695690512657 2975
63
+ 1647960448284 0.017000695690512657 3023
64
+ 1647960452162 0.017000695690512657 3071
65
+ 1647960456158 0.017000695690512657 3119
66
+ 1647960460171 0.017000695690512657 3167
67
+ 1647960464064 0.017000695690512657 3215
68
+ 1647960467957 0.017000695690512657 3263
69
+ 1647960473299 0.017000695690512657 3311
70
+ 1647960477374 0.017000695690512657 3359
71
+ 1647960481415 0.017000695690512657 3407
72
+ 1647960485799 0.017000695690512657 3455
73
+ 1647960489849 0.017000695690512657 3503
74
+ 1647960494420 0.017000695690512657 3551
75
+ 1647960499349 0.017000695690512657 3599
76
+ 1647960504459 0.017000695690512657 3647
77
+ 1647960508707 0.017000695690512657 3695
78
+ 1647960513159 0.017000695690512657 3743
79
+ 1647960517546 0.017000695690512657 3791
80
+ 1647960522033 0.017000695690512657 3839
81
+ 1647960526250 0.017000695690512657 3887
82
+ 1647960530977 0.017000695690512657 3935
83
+ 1647960535476 0.017000695690512657 3983
84
+ 1647960540098 0.017000695690512657 4031
85
+ 1647960544532 0.017000695690512657 4079
86
+ 1647960548960 0.017000695690512657 4127
87
+ 1647960553203 0.017000695690512657 4175
88
+ 1647960557539 0.017000695690512657 4223
89
+ 1647960562267 0.017000695690512657 4271
90
+ 1647960566815 0.017000695690512657 4319
91
+ 1647960571435 0.017000695690512657 4367
92
+ 1647960575958 0.017000695690512657 4415
93
+ 1647960580400 0.017000695690512657 4463
94
+ 1647960584677 0.017000695690512657 4511
95
+ 1647960588839 0.017000695690512657 4559
96
+ 1647960593001 0.017000695690512657 4607
97
+ 1647960597173 0.017000695690512657 4655
98
+ 1647960601382 0.017000695690512657 4703
99
+ 1647960605473 0.017000695690512657 4751
100
+ 1647960609428 0.017000695690512657 4799
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.EarlyStopping
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/min_delta ADDED
@@ -0,0 +1 @@
 
 
1
+ 0
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/mode ADDED
@@ -0,0 +1 @@
 
 
1
+ min
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/monitor ADDED
@@ -0,0 +1 @@
 
 
1
+ val/mae
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/early_stopping/patience ADDED
@@ -0,0 +1 @@
 
 
1
+ 100
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.ModelCheckpoint
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/auto_insert_metric_name ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/dirpath ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoints/
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/filename ADDED
@@ -0,0 +1 @@
 
 
1
+ epoch_{epoch:03d}
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/mode ADDED
@@ -0,0 +1 @@
 
 
1
+ min
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/monitor ADDED
@@ -0,0 +1 @@
 
 
1
+ val/mae
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/save_last ADDED
@@ -0,0 +1 @@
 
 
1
+ True
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/save_top_k ADDED
@@ -0,0 +1 @@
 
 
1
+ 1
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_checkpoint/verbose ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_summary/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.RichModelSummary
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/model_summary/max_depth ADDED
@@ -0,0 +1 @@
 
 
1
+ -1
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/callbacks/rich_progress_bar/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.RichProgressBar
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ src.datamodules.focus_datamodule.FocusDataModule
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/augmentation ADDED
@@ -0,0 +1 @@
 
 
1
+ True
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/batch_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 64
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/csv_test_file ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150/test_metadata.csv
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/csv_train_file ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150/train_metadata.csv
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/csv_val_file ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150/validation_metadata.csv
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/data_dir ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/num_workers ADDED
@@ -0,0 +1 @@
 
 
1
+ 0
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/datamodule/pin_memory ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ src.models.focus_conv_module.FocusConvLitModule
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv1_channels ADDED
@@ -0,0 +1 @@
 
 
1
+ 9
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv1_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 3
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv2_channels ADDED
@@ -0,0 +1 @@
 
 
1
+ 6
logs/mlflow/mlruns/1/40b80b1e6d4e48918cf7254779a3cb9a/params/model/conv2_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 7