Hannes Kuchelmeister commited on
Commit
a317510
·
1 Parent(s): eb23ded

rerun and add logs for focusConvMSE_150

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/.hydra/config.yaml +80 -0
  2. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/.hydra/hydra.yaml +170 -0
  3. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/.hydra/overrides.yaml +1 -0
  4. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/checkpoints/epoch_086.ckpt +3 -0
  5. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/checkpoints/last.ckpt +3 -0
  6. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/csv/version_0/hparams.yaml +54 -0
  7. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/csv/version_0/metrics.csv +202 -0
  8. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/tensorboard/focusConvMSE_150/events.out.tfevents.1647959877.0649b520f1a6.1.0 +3 -0
  9. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/tensorboard/focusConvMSE_150/events.out.tfevents.1647960130.0649b520f1a6.1.1 +3 -0
  10. logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/tensorboard/focusConvMSE_150/hparams.yaml +54 -0
  11. logs/mlflow/mlruns/0/meta.yaml +4 -0
  12. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/meta.yaml +15 -0
  13. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/epoch +201 -0
  14. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/test/loss +1 -0
  15. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/test/mae +1 -0
  16. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/train/loss +100 -0
  17. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/train/mae +100 -0
  18. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/val/loss +100 -0
  19. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/val/mae +100 -0
  20. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/val/mae_best +100 -0
  21. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/_target_ +1 -0
  22. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/min_delta +1 -0
  23. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/mode +1 -0
  24. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/monitor +1 -0
  25. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/patience +1 -0
  26. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/_target_ +1 -0
  27. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/auto_insert_metric_name +1 -0
  28. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/dirpath +1 -0
  29. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/filename +1 -0
  30. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/mode +1 -0
  31. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/monitor +1 -0
  32. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/save_last +1 -0
  33. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/save_top_k +1 -0
  34. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/verbose +1 -0
  35. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_summary/_target_ +1 -0
  36. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_summary/max_depth +1 -0
  37. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/rich_progress_bar/_target_ +1 -0
  38. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/_target_ +1 -0
  39. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/batch_size +1 -0
  40. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/csv_test_file +1 -0
  41. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/csv_train_file +1 -0
  42. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/csv_val_file +1 -0
  43. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/data_dir +1 -0
  44. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/num_workers +1 -0
  45. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/pin_memory +1 -0
  46. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/_target_ +1 -0
  47. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv1_channels +1 -0
  48. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv1_size +1 -0
  49. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv2_channels +1 -0
  50. logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv2_size +1 -0
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/.hydra/config.yaml ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ original_work_dir: ${hydra:runtime.cwd}
2
+ data_dir: ${original_work_dir}/data
3
+ print_config: true
4
+ ignore_warnings: true
5
+ train: true
6
+ test: true
7
+ seed: 12345
8
+ name: focusConvMSE_150
9
+ datamodule:
10
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
11
+ data_dir: ${data_dir}/focus150
12
+ csv_train_file: ${data_dir}/focus150/train_metadata.csv
13
+ csv_val_file: ${data_dir}/focus150/validation_metadata.csv
14
+ csv_test_file: ${data_dir}/focus150/test_metadata.csv
15
+ batch_size: 128
16
+ num_workers: 0
17
+ pin_memory: false
18
+ model:
19
+ _target_: src.models.focus_conv_module.FocusConvLitModule
20
+ image_size: 150
21
+ pool_size: 2
22
+ conv1_size: 5
23
+ conv1_channels: 6
24
+ conv2_size: 5
25
+ conv2_channels: 16
26
+ lin1_size: 100
27
+ lin2_size: 80
28
+ output_size: 1
29
+ lr: 0.001
30
+ weight_decay: 0.0005
31
+ callbacks:
32
+ model_checkpoint:
33
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
34
+ monitor: val/mae
35
+ mode: min
36
+ save_top_k: 1
37
+ save_last: true
38
+ verbose: false
39
+ dirpath: checkpoints/
40
+ filename: epoch_{epoch:03d}
41
+ auto_insert_metric_name: false
42
+ early_stopping:
43
+ _target_: pytorch_lightning.callbacks.EarlyStopping
44
+ monitor: val/mae
45
+ mode: min
46
+ patience: 100
47
+ min_delta: 0
48
+ model_summary:
49
+ _target_: pytorch_lightning.callbacks.RichModelSummary
50
+ max_depth: -1
51
+ rich_progress_bar:
52
+ _target_: pytorch_lightning.callbacks.RichProgressBar
53
+ logger:
54
+ csv:
55
+ _target_: pytorch_lightning.loggers.csv_logs.CSVLogger
56
+ save_dir: .
57
+ name: csv/
58
+ prefix: ''
59
+ mlflow:
60
+ _target_: pytorch_lightning.loggers.mlflow.MLFlowLogger
61
+ experiment_name: ${name}
62
+ tracking_uri: ${original_work_dir}/logs/mlflow/mlruns
63
+ tags: null
64
+ save_dir: ./mlruns
65
+ prefix: ''
66
+ artifact_location: null
67
+ tensorboard:
68
+ _target_: pytorch_lightning.loggers.tensorboard.TensorBoardLogger
69
+ save_dir: tensorboard/
70
+ name: null
71
+ version: ${name}
72
+ log_graph: false
73
+ default_hp_metric: true
74
+ prefix: ''
75
+ trainer:
76
+ _target_: pytorch_lightning.Trainer
77
+ gpus: 1
78
+ min_epochs: 1
79
+ max_epochs: 100
80
+ resume_from_checkpoint: null
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: logs/experiments/runs/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: logs/experiments/multiruns/${name}/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - experiment=focusConvMSE_150
121
+ job:
122
+ name: train
123
+ override_dirname: experiment=focusConvMSE_150
124
+ id: ???
125
+ num: ???
126
+ config_name: train.yaml
127
+ env_set: {}
128
+ env_copy: []
129
+ config:
130
+ override_dirname:
131
+ kv_sep: '='
132
+ item_sep: ','
133
+ exclude_keys: []
134
+ runtime:
135
+ version: 1.1.1
136
+ cwd: /usr/src/app
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /usr/src/app/configs
142
+ schema: file
143
+ provider: main
144
+ - path: hydra_plugins.hydra_colorlog.conf
145
+ schema: pkg
146
+ provider: hydra-colorlog
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ choices:
151
+ local: default.yaml
152
+ hparams_search: null
153
+ debug: null
154
+ experiment: focusConvMSE_150
155
+ log_dir: default.yaml
156
+ trainer: default.yaml
157
+ logger: many_loggers
158
+ callbacks: default.yaml
159
+ model: focusConv_150.yaml
160
+ datamodule: focus150.yaml
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: colorlog
164
+ hydra/hydra_logging: colorlog
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=focusConvMSE_150
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/checkpoints/epoch_086.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66a9be868aac9bdf754525634e9e81c5b36206ad9609bea027b346536ad12bd2
3
+ size 22352261
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/checkpoints/last.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35c98c898f4733e7f5cb8154d7806523e7bf972a5ab454578a78e70cb5d6a75b
3
+ size 22352261
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/csv/version_0/hparams.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 1
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_conv_module.FocusConvLitModule
9
+ image_size: 150
10
+ pool_size: 2
11
+ conv1_size: 5
12
+ conv1_channels: 6
13
+ conv2_size: 5
14
+ conv2_channels: 16
15
+ lin1_size: 100
16
+ lin2_size: 80
17
+ output_size: 1
18
+ lr: 0.001
19
+ weight_decay: 0.0005
20
+ datamodule:
21
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
22
+ data_dir: /usr/src/app/data/focus150
23
+ csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
24
+ csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
25
+ csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
26
+ batch_size: 128
27
+ num_workers: 0
28
+ pin_memory: false
29
+ seed: 12345
30
+ callbacks:
31
+ model_checkpoint:
32
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
33
+ monitor: val/mae
34
+ mode: min
35
+ save_top_k: 1
36
+ save_last: true
37
+ verbose: false
38
+ dirpath: checkpoints/
39
+ filename: epoch_{epoch:03d}
40
+ auto_insert_metric_name: false
41
+ early_stopping:
42
+ _target_: pytorch_lightning.callbacks.EarlyStopping
43
+ monitor: val/mae
44
+ mode: min
45
+ patience: 100
46
+ min_delta: 0
47
+ model_summary:
48
+ _target_: pytorch_lightning.callbacks.RichModelSummary
49
+ max_depth: -1
50
+ rich_progress_bar:
51
+ _target_: pytorch_lightning.callbacks.RichProgressBar
52
+ model/params/total: 1861789
53
+ model/params/trainable: 1861789
54
+ model/params/non_trainable: 0
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/csv/version_0/metrics.csv ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ val/loss,val/mae,val/mae_best,epoch,step,train/loss,train/mae,test/loss,test/mae
2
+ 0.014177942648530006,0.09954570978879929,0.038817837834358215,0,23,,,,
3
+ ,,,0,23,4.342347621917725,1.3158199787139893,,
4
+ 0.024941448122262955,0.1495475023984909,0.038817837834358215,1,47,,,,
5
+ ,,,1,47,0.05730731412768364,0.19751663506031036,,
6
+ 0.0022524979431182146,0.04012170061469078,0.038817837834358215,2,71,,,,
7
+ ,,,2,71,0.006286133546382189,0.06423742324113846,,
8
+ 0.000798797351308167,0.023267723619937897,0.023267723619937897,3,95,,,,
9
+ ,,,3,95,0.0013532182201743126,0.02940954454243183,,
10
+ 0.000731383275706321,0.021397700533270836,0.021397700533270836,4,119,,,,
11
+ ,,,4,119,0.0008239876478910446,0.02320658229291439,,
12
+ 0.0006743474514223635,0.02102094329893589,0.02102094329893589,5,143,,,,
13
+ ,,,5,143,0.0007202242850326002,0.021659735590219498,,
14
+ 0.0006443106685765088,0.02113427221775055,0.02102094329893589,6,167,,,,
15
+ ,,,6,167,0.0006690375157631934,0.021050428971648216,,
16
+ 0.0006375169032253325,0.02006976678967476,0.02006976678967476,7,191,,,,
17
+ ,,,7,191,0.0006276399944908917,0.02056770585477352,,
18
+ 0.0005847099237143993,0.020206814631819725,0.02006976678967476,8,215,,,,
19
+ ,,,8,215,0.0006050210795365274,0.020237302407622337,,
20
+ 0.0005691925762221217,0.020232098177075386,0.02006976678967476,9,239,,,,
21
+ ,,,9,239,0.0005693044513463974,0.019613152369856834,,
22
+ 0.0005465119611471891,0.019510304555296898,0.019510304555296898,10,263,,,,
23
+ ,,,10,263,0.0005463812849484384,0.019444193691015244,,
24
+ 0.0005308444960974157,0.01932346075773239,0.01932346075773239,11,287,,,,
25
+ ,,,11,287,0.0005229813395999372,0.019030658528208733,,
26
+ 0.000525242299772799,0.018989086151123047,0.018989086151123047,12,311,,,,
27
+ ,,,12,311,0.0005098703550174832,0.01886916160583496,,
28
+ 0.0005212906980887055,0.018810341134667397,0.018810341134667397,13,335,,,,
29
+ ,,,13,335,0.000495909305755049,0.018634887412190437,,
30
+ 0.0005010279710404575,0.018762608990073204,0.018762608990073204,14,359,,,,
31
+ ,,,14,359,0.00048210768727585673,0.018486270681023598,,
32
+ 0.0004964997060596943,0.018582377582788467,0.018582377582788467,15,383,,,,
33
+ ,,,15,383,0.000473074265755713,0.018334897235035896,,
34
+ 0.00048635710845701396,0.018572870641946793,0.018572870641946793,16,407,,,,
35
+ ,,,16,407,0.0004635636869352311,0.018168363720178604,,
36
+ 0.0004799317684955895,0.01848633959889412,0.01848633959889412,17,431,,,,
37
+ ,,,17,431,0.00045512395445257425,0.018060654401779175,,
38
+ 0.0004763782490044832,0.018363453447818756,0.018363453447818756,18,455,,,,
39
+ ,,,18,455,0.00045438535744324327,0.0180768221616745,,
40
+ 0.0004732394590973854,0.018254607915878296,0.018254607915878296,19,479,,,,
41
+ ,,,19,479,0.00044914596946910024,0.017933538183569908,,
42
+ 0.00046858476707711816,0.018206853419542313,0.018206853419542313,20,503,,,,
43
+ ,,,20,503,0.00044510987936519086,0.017830582335591316,,
44
+ 0.0004627775924745947,0.018315646797418594,0.018206853419542313,21,527,,,,
45
+ ,,,21,527,0.00043568635010160506,0.01768254116177559,,
46
+ 0.00048223865451291203,0.018926186487078667,0.018206853419542313,22,551,,,,
47
+ ,,,22,551,0.00043914539855904877,0.01769491843879223,,
48
+ 0.00045650775427930057,0.01812773011624813,0.01812773011624813,23,575,,,,
49
+ ,,,23,575,0.00043160346103832126,0.017608648166060448,,
50
+ 0.00045403436524793506,0.01813136413693428,0.01812773011624813,24,599,,,,
51
+ ,,,24,599,0.00042612897232174873,0.017463266849517822,,
52
+ 0.0004518564383033663,0.017970262095332146,0.017970262095332146,25,623,,,,
53
+ ,,,25,623,0.0004240336420480162,0.017441386356949806,,
54
+ 0.000449806684628129,0.01790132373571396,0.01790132373571396,26,647,,,,
55
+ ,,,26,647,0.00042002907139249146,0.017370155081152916,,
56
+ 0.0004594481724780053,0.018357347697019577,0.01790132373571396,27,671,,,,
57
+ ,,,27,671,0.00041983669507317245,0.017345596104860306,,
58
+ 0.0004483773373067379,0.01778845489025116,0.01778845489025116,28,695,,,,
59
+ ,,,28,695,0.0004216474189888686,0.017313584685325623,,
60
+ 0.0004582132969517261,0.017787707969546318,0.017787707969546318,29,719,,,,
61
+ ,,,29,719,0.0004165133577771485,0.017224453389644623,,
62
+ 0.000451582862297073,0.01773429661989212,0.01773429661989212,30,743,,,,
63
+ ,,,30,743,0.0004142482648603618,0.01721722073853016,,
64
+ 0.0004431355628184974,0.017908763140439987,0.01773429661989212,31,767,,,,
65
+ ,,,31,767,0.0004127994761802256,0.017110856249928474,,
66
+ 0.000467679143184796,0.01842939481139183,0.01773429661989212,32,791,,,,
67
+ ,,,32,791,0.00042158510768786073,0.01726309210062027,,
68
+ 0.0004435107402969152,0.017640704289078712,0.017640704289078712,33,815,,,,
69
+ ,,,33,815,0.00041695087566040456,0.017146525904536247,,
70
+ 0.00044180749682709575,0.0178662296384573,0.017640704289078712,34,839,,,,
71
+ ,,,34,839,0.0004059165657963604,0.016986556351184845,,
72
+ 0.00043626612750813365,0.01761534810066223,0.01761534810066223,35,863,,,,
73
+ ,,,35,863,0.0004062257066834718,0.0169779434800148,,
74
+ 0.0004603431443683803,0.01775723695755005,0.01761534810066223,36,887,,,,
75
+ ,,,36,887,0.000408758845878765,0.017033370211720467,,
76
+ 0.00044433787115849555,0.017887664958834648,0.01761534810066223,37,911,,,,
77
+ ,,,37,911,0.00041210511699318886,0.017063522711396217,,
78
+ 0.0004384913481771946,0.017551790922880173,0.017551790922880173,38,935,,,,
79
+ ,,,38,935,0.0004000916378572583,0.016827763989567757,,
80
+ 0.00043312823981978,0.017565051093697548,0.017551790922880173,39,959,,,,
81
+ ,,,39,959,0.00039786219713278115,0.016777368262410164,,
82
+ 0.0004400752077344805,0.017544588074088097,0.017544588074088097,40,983,,,,
83
+ ,,,40,983,0.000396965304389596,0.016790589317679405,,
84
+ 0.0004321136511862278,0.017519894987344742,0.017519894987344742,41,1007,,,,
85
+ ,,,41,1007,0.00040004777838476,0.01677679643034935,,
86
+ 0.0004558261134661734,0.01801685057580471,0.017519894987344742,42,1031,,,,
87
+ ,,,42,1031,0.0004005195223726332,0.01680240035057068,,
88
+ 0.00044323524343781173,0.01776825450360775,0.017519894987344742,43,1055,,,,
89
+ ,,,43,1055,0.00039835445932112634,0.01677827537059784,,
90
+ 0.00043952741543762386,0.01769193448126316,0.017519894987344742,44,1079,,,,
91
+ ,,,44,1079,0.00039412121986970305,0.016687458381056786,,
92
+ 0.00043559601181186736,0.017593219876289368,0.017519894987344742,45,1103,,,,
93
+ ,,,45,1103,0.0003929083759430796,0.01665475219488144,,
94
+ 0.00044441656791605055,0.01759103313088417,0.017519894987344742,46,1127,,,,
95
+ ,,,46,1127,0.00040109854307956994,0.016871973872184753,,
96
+ 0.00043277026270516217,0.01745220273733139,0.01745220273733139,47,1151,,,,
97
+ ,,,47,1151,0.0003939596936106682,0.016618330031633377,,
98
+ 0.0004414572031237185,0.01768527552485466,0.01745220273733139,48,1175,,,,
99
+ ,,,48,1175,0.0003906012570951134,0.016592541709542274,,
100
+ 0.0004329547518864274,0.01749493181705475,0.01745220273733139,49,1199,,,,
101
+ ,,,49,1199,0.0003921556053683162,0.01663227006793022,,
102
+ 0.0004438141768332571,0.0175788477063179,0.01745220273733139,50,1223,,,,
103
+ ,,,50,1223,0.0003916115965694189,0.01655171997845173,,
104
+ 0.0004302943416405469,0.0174209326505661,0.0174209326505661,51,1247,,,,
105
+ ,,,51,1247,0.00039053973159752786,0.016555769369006157,,
106
+ 0.0004304391623008996,0.017427116632461548,0.0174209326505661,52,1271,,,,
107
+ ,,,52,1271,0.00038710495573468506,0.016491563990712166,,
108
+ 0.0004304780450183898,0.017413398250937462,0.017413398250937462,53,1295,,,,
109
+ ,,,53,1295,0.00038700783625245094,0.016510063782334328,,
110
+ 0.0004549633013084531,0.01783713698387146,0.017413398250937462,54,1319,,,,
111
+ ,,,54,1319,0.00039118644781410694,0.016560301184654236,,
112
+ 0.0004564271657727659,0.017778683453798294,0.017413398250937462,55,1343,,,,
113
+ ,,,55,1343,0.00039927964098751545,0.016744039952754974,,
114
+ 0.0004305720212869346,0.017405666410923004,0.017405666410923004,56,1367,,,,
115
+ ,,,56,1367,0.0003894182445947081,0.01651095040142536,,
116
+ 0.00043586301035247743,0.017491891980171204,0.017405666410923004,57,1391,,,,
117
+ ,,,57,1391,0.0003865647013299167,0.016462186351418495,,
118
+ 0.00044148581218905747,0.01758340559899807,0.017405666410923004,58,1415,,,,
119
+ ,,,58,1415,0.0003892740933224559,0.01648084446787834,,
120
+ 0.0004328140348661691,0.017423313111066818,0.017405666410923004,59,1439,,,,
121
+ ,,,59,1439,0.00038562898407690227,0.016439443454146385,,
122
+ 0.00044432288268581033,0.017613204196095467,0.017405666410923004,60,1463,,,,
123
+ ,,,60,1463,0.00039640345494262874,0.016590487211942673,,
124
+ 0.00043844786705449224,0.01751849800348282,0.017405666410923004,61,1487,,,,
125
+ ,,,61,1487,0.0003929836966563016,0.016544204205274582,,
126
+ 0.000432810396887362,0.017426760867238045,0.017405666410923004,62,1511,,,,
127
+ ,,,62,1511,0.0003826528263743967,0.01636682078242302,,
128
+ 0.0004337234713602811,0.017433997243642807,0.017405666410923004,63,1535,,,,
129
+ ,,,63,1535,0.00039758102502673864,0.016689548268914223,,
130
+ 0.0004506470577325672,0.017688874155282974,0.017405666410923004,64,1559,,,,
131
+ ,,,64,1559,0.00038526495336554945,0.016391392797231674,,
132
+ 0.0005121462745591998,0.018620284274220467,0.017405666410923004,65,1583,,,,
133
+ ,,,65,1583,0.0003987601085100323,0.016639119014143944,,
134
+ 0.00043341098353266716,0.017417754977941513,0.017405666410923004,66,1607,,,,
135
+ ,,,66,1607,0.0003918004804290831,0.01652211882174015,,
136
+ 0.0004314535763114691,0.01739826239645481,0.01739826239645481,67,1631,,,,
137
+ ,,,67,1631,0.00038548445445485413,0.016405444592237473,,
138
+ 0.00046141925849951804,0.017827652394771576,0.01739826239645481,68,1655,,,,
139
+ ,,,68,1655,0.00038453712477348745,0.016384242102503777,,
140
+ 0.0004310157091822475,0.017386266961693764,0.017386266961693764,69,1679,,,,
141
+ ,,,69,1679,0.00038578943349421024,0.016403140500187874,,
142
+ 0.00043788779294118285,0.017491240054368973,0.017386266961693764,70,1703,,,,
143
+ ,,,70,1703,0.0003904291370417923,0.01646820455789566,,
144
+ 0.00043278036173433065,0.01741820201277733,0.017386266961693764,71,1727,,,,
145
+ ,,,71,1727,0.00038028054405003786,0.016317954286932945,,
146
+ 0.000435496709542349,0.017456887289881706,0.017386266961693764,72,1751,,,,
147
+ ,,,72,1751,0.0003783942374866456,0.016258543357253075,,
148
+ 0.0004316089325584471,0.017395686358213425,0.017386266961693764,73,1775,,,,
149
+ ,,,73,1775,0.00037868719664402306,0.01626788266003132,,
150
+ 0.00043277398799546063,0.017391294240951538,0.017386266961693764,74,1799,,,,
151
+ ,,,74,1799,0.0003772037453018129,0.016254499554634094,,
152
+ 0.0004360957827884704,0.017426805570721626,0.017386266961693764,75,1823,,,,
153
+ ,,,75,1823,0.0003772091295104474,0.01622932218015194,,
154
+ 0.00043179106432944536,0.01736919768154621,0.01736919768154621,76,1847,,,,
155
+ ,,,76,1847,0.00037675860221497715,0.016197772696614265,,
156
+ 0.00043282663682475686,0.017420921474695206,0.01736919768154621,77,1871,,,,
157
+ ,,,77,1871,0.00037759728729724884,0.016238529235124588,,
158
+ 0.00043416841072030365,0.01739104837179184,0.01736919768154621,78,1895,,,,
159
+ ,,,78,1895,0.00037590478314086795,0.016197707504034042,,
160
+ 0.000432082248153165,0.017376182600855827,0.01736919768154621,79,1919,,,,
161
+ ,,,79,1919,0.00037703668931499124,0.016217660158872604,,
162
+ 0.00043247482972219586,0.017374588176608086,0.01736919768154621,80,1943,,,,
163
+ ,,,80,1943,0.0003931742685381323,0.016501346603035927,,
164
+ 0.0004352308751549572,0.01739008165895939,0.01736919768154621,81,1967,,,,
165
+ ,,,81,1967,0.00037598126800730824,0.016180992126464844,,
166
+ 0.00043458439176902175,0.017396951094269753,0.01736919768154621,82,1991,,,,
167
+ ,,,82,1991,0.0003758900857064873,0.016161207109689713,,
168
+ 0.0004796235589310527,0.01820252649486065,0.01736919768154621,83,2015,,,,
169
+ ,,,83,2015,0.0003811830247286707,0.016300171613693237,,
170
+ 0.0004338948638178408,0.017369873821735382,0.01736919768154621,84,2039,,,,
171
+ ,,,84,2039,0.00038812385173514485,0.01642199419438839,,
172
+ 0.00044887603144161403,0.017721308395266533,0.01736919768154621,85,2063,,,,
173
+ ,,,85,2063,0.00037520512705668807,0.01615377515554428,,
174
+ 0.00043250841554254293,0.01735648140311241,0.01735648140311241,86,2087,,,,
175
+ ,,,86,2087,0.00038031762233003974,0.016325831413269043,,
176
+ 0.00044269993668422103,0.017486795783042908,0.01735648140311241,87,2111,,,,
177
+ ,,,87,2111,0.00037949124816805124,0.01625642739236355,,
178
+ 0.00043156533502042294,0.01737273670732975,0.01735648140311241,88,2135,,,,
179
+ ,,,88,2135,0.0003777534584514797,0.016191186383366585,,
180
+ 0.00043633393943309784,0.01747000217437744,0.01735648140311241,89,2159,,,,
181
+ ,,,89,2159,0.0003773653006646782,0.016204534098505974,,
182
+ 0.000434521323768422,0.017369193956255913,0.01735648140311241,90,2183,,,,
183
+ ,,,90,2183,0.00037604026147164404,0.016208775341510773,,
184
+ 0.00043653271859511733,0.01748810149729252,0.01735648140311241,91,2207,,,,
185
+ ,,,91,2207,0.00037361629074439406,0.016117053106427193,,
186
+ 0.00044135679490864277,0.017475992441177368,0.01735648140311241,92,2231,,,,
187
+ ,,,92,2231,0.00038605634472332895,0.0163609329611063,,
188
+ 0.00043531309347599745,0.01737702079117298,0.01735648140311241,93,2255,,,,
189
+ ,,,93,2255,0.0003846309264190495,0.016337504610419273,,
190
+ 0.000436639558756724,0.01747899316251278,0.01735648140311241,94,2279,,,,
191
+ ,,,94,2279,0.0003783575084526092,0.01624019630253315,,
192
+ 0.0004387025546748191,0.01744064688682556,0.01735648140311241,95,2303,,,,
193
+ ,,,95,2303,0.0003748334711417556,0.01612277328968048,,
194
+ 0.0004409730900079012,0.01744801551103592,0.01735648140311241,96,2327,,,,
195
+ ,,,96,2327,0.00038411852438002825,0.0163507629185915,,
196
+ 0.00045858731027692556,0.01768660731613636,0.01735648140311241,97,2351,,,,
197
+ ,,,97,2351,0.00037512960261665285,0.016116727143526077,,
198
+ 0.00044121488463133574,0.017449431121349335,0.01735648140311241,98,2375,,,,
199
+ ,,,98,2375,0.00039327985723502934,0.016511688008904457,,
200
+ 0.0005107350298203528,0.018421221524477005,0.01735648140311241,99,2399,,,,
201
+ ,,,99,2399,0.00037736291415058076,0.016158610582351685,,
202
+ ,,,87,2088,,,0.000402881734771654,0.016752799972891808
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/tensorboard/focusConvMSE_150/events.out.tfevents.1647959877.0649b520f1a6.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38fdc0288de684ce041c53683daff58459a2cf862931904dc06158f58af8c5b2
3
+ size 36778
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/tensorboard/focusConvMSE_150/events.out.tfevents.1647960130.0649b520f1a6.1.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18a875526c9033ce626bd53efa68bb47ece58730afe4712d52591efedc1e5006
3
+ size 179
logs/experiments/runs/focusConvMSE_150/2022-03-22_14-37-55/tensorboard/focusConvMSE_150/hparams.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainer:
2
+ _target_: pytorch_lightning.Trainer
3
+ gpus: 1
4
+ min_epochs: 1
5
+ max_epochs: 100
6
+ resume_from_checkpoint: null
7
+ model:
8
+ _target_: src.models.focus_conv_module.FocusConvLitModule
9
+ image_size: 150
10
+ pool_size: 2
11
+ conv1_size: 5
12
+ conv1_channels: 6
13
+ conv2_size: 5
14
+ conv2_channels: 16
15
+ lin1_size: 100
16
+ lin2_size: 80
17
+ output_size: 1
18
+ lr: 0.001
19
+ weight_decay: 0.0005
20
+ datamodule:
21
+ _target_: src.datamodules.focus_datamodule.FocusDataModule
22
+ data_dir: /usr/src/app/data/focus150
23
+ csv_train_file: /usr/src/app/data/focus150/train_metadata.csv
24
+ csv_val_file: /usr/src/app/data/focus150/validation_metadata.csv
25
+ csv_test_file: /usr/src/app/data/focus150/test_metadata.csv
26
+ batch_size: 128
27
+ num_workers: 0
28
+ pin_memory: false
29
+ seed: 12345
30
+ callbacks:
31
+ model_checkpoint:
32
+ _target_: pytorch_lightning.callbacks.ModelCheckpoint
33
+ monitor: val/mae
34
+ mode: min
35
+ save_top_k: 1
36
+ save_last: true
37
+ verbose: false
38
+ dirpath: checkpoints/
39
+ filename: epoch_{epoch:03d}
40
+ auto_insert_metric_name: false
41
+ early_stopping:
42
+ _target_: pytorch_lightning.callbacks.EarlyStopping
43
+ monitor: val/mae
44
+ mode: min
45
+ patience: 100
46
+ min_delta: 0
47
+ model_summary:
48
+ _target_: pytorch_lightning.callbacks.RichModelSummary
49
+ max_depth: -1
50
+ rich_progress_bar:
51
+ _target_: pytorch_lightning.callbacks.RichProgressBar
52
+ model/params/total: 1861789
53
+ model/params/trainable: 1861789
54
+ model/params/non_trainable: 0
logs/mlflow/mlruns/0/meta.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ artifact_location: /usr/src/app/logs/mlflow/mlruns/0
2
+ experiment_id: '0'
3
+ lifecycle_stage: active
4
+ name: Default
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/meta.yaml ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ artifact_uri: /usr/src/app/logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/artifacts
2
+ end_time: 1647960130595
3
+ entry_point_name: ''
4
+ experiment_id: '1'
5
+ lifecycle_stage: active
6
+ name: ''
7
+ run_id: 62c4a535287f4b44bd537524cf678fff
8
+ run_uuid: 62c4a535287f4b44bd537524cf678fff
9
+ source_name: ''
10
+ source_type: 4
11
+ source_version: ''
12
+ start_time: 1647959877721
13
+ status: 3
14
+ tags: []
15
+ user_id: unknown
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/epoch ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647959886393 0 23
2
+ 1647959886488 0 23
3
+ 1647959888772 1 47
4
+ 1647959888894 1 47
5
+ 1647959891185 2 71
6
+ 1647959891311 2 71
7
+ 1647959893603 3 95
8
+ 1647959893784 3 95
9
+ 1647959896075 4 119
10
+ 1647959896230 4 119
11
+ 1647959898546 5 143
12
+ 1647959898735 5 143
13
+ 1647959901038 6 167
14
+ 1647959901168 6 167
15
+ 1647959903443 7 191
16
+ 1647959903624 7 191
17
+ 1647959905912 8 215
18
+ 1647959906037 8 215
19
+ 1647959908338 9 239
20
+ 1647959908487 9 239
21
+ 1647959910764 10 263
22
+ 1647959910953 10 263
23
+ 1647959913246 11 287
24
+ 1647959913446 11 287
25
+ 1647959915770 12 311
26
+ 1647959915962 12 311
27
+ 1647959918244 13 335
28
+ 1647959918436 13 335
29
+ 1647959920693 14 359
30
+ 1647959920870 14 359
31
+ 1647959923162 15 383
32
+ 1647959923340 15 383
33
+ 1647959925629 16 407
34
+ 1647959925773 16 407
35
+ 1647959928086 17 431
36
+ 1647959928259 17 431
37
+ 1647959930519 18 455
38
+ 1647959930708 18 455
39
+ 1647959932958 19 479
40
+ 1647959933134 19 479
41
+ 1647959935471 20 503
42
+ 1647959935640 20 503
43
+ 1647959937907 21 527
44
+ 1647959938070 21 527
45
+ 1647959940384 22 551
46
+ 1647959940513 22 551
47
+ 1647959942822 23 575
48
+ 1647959943042 23 575
49
+ 1647959945335 24 599
50
+ 1647959945545 24 599
51
+ 1647959947814 25 623
52
+ 1647959948033 25 623
53
+ 1647959950303 26 647
54
+ 1647959950491 26 647
55
+ 1647959952770 27 671
56
+ 1647959952968 27 671
57
+ 1647959955260 28 695
58
+ 1647959955455 28 695
59
+ 1647959957774 29 719
60
+ 1647959957965 29 719
61
+ 1647959960220 30 743
62
+ 1647959960410 30 743
63
+ 1647959962648 31 767
64
+ 1647959962775 31 767
65
+ 1647959965011 32 791
66
+ 1647959965199 32 791
67
+ 1647959967529 33 815
68
+ 1647959967750 33 815
69
+ 1647959969999 34 839
70
+ 1647959970196 34 839
71
+ 1647959972477 35 863
72
+ 1647959972661 35 863
73
+ 1647959974978 36 887
74
+ 1647959975107 36 887
75
+ 1647959977437 37 911
76
+ 1647959977562 37 911
77
+ 1647959979782 38 935
78
+ 1647959979962 38 935
79
+ 1647959982227 39 959
80
+ 1647959982366 39 959
81
+ 1647959984625 40 983
82
+ 1647959984855 40 983
83
+ 1647959987165 41 1007
84
+ 1647959987351 41 1007
85
+ 1647959989661 42 1031
86
+ 1647959989788 42 1031
87
+ 1647959992143 43 1055
88
+ 1647959992297 43 1055
89
+ 1647959994568 44 1079
90
+ 1647959994698 44 1079
91
+ 1647959997016 45 1103
92
+ 1647959997166 45 1103
93
+ 1647959999457 46 1127
94
+ 1647959999651 46 1127
95
+ 1647960001942 47 1151
96
+ 1647960002132 47 1151
97
+ 1647960004419 48 1175
98
+ 1647960004609 48 1175
99
+ 1647960006913 49 1199
100
+ 1647960007107 49 1199
101
+ 1647960009432 50 1223
102
+ 1647960009559 50 1223
103
+ 1647960011843 51 1247
104
+ 1647960012033 51 1247
105
+ 1647960014340 52 1271
106
+ 1647960014466 52 1271
107
+ 1647960016727 53 1295
108
+ 1647960016920 53 1295
109
+ 1647960019263 54 1319
110
+ 1647960019395 54 1319
111
+ 1647960021699 55 1343
112
+ 1647960021828 55 1343
113
+ 1647960024113 56 1367
114
+ 1647960024302 56 1367
115
+ 1647960026595 57 1391
116
+ 1647960026728 57 1391
117
+ 1647960029045 58 1415
118
+ 1647960029198 58 1415
119
+ 1647960031574 59 1439
120
+ 1647960031776 59 1439
121
+ 1647960034043 60 1463
122
+ 1647960034168 60 1463
123
+ 1647960036419 61 1487
124
+ 1647960036549 61 1487
125
+ 1647960038897 62 1511
126
+ 1647960039049 62 1511
127
+ 1647960041334 63 1535
128
+ 1647960041464 63 1535
129
+ 1647960043726 64 1559
130
+ 1647960043874 64 1559
131
+ 1647960046213 65 1583
132
+ 1647960046406 65 1583
133
+ 1647960048755 66 1607
134
+ 1647960048881 66 1607
135
+ 1647960051173 67 1631
136
+ 1647960051372 67 1631
137
+ 1647960053693 68 1655
138
+ 1647960053881 68 1655
139
+ 1647960056189 69 1679
140
+ 1647960056386 69 1679
141
+ 1647960058707 70 1703
142
+ 1647960058849 70 1703
143
+ 1647960061152 71 1727
144
+ 1647960061292 71 1727
145
+ 1647960063577 72 1751
146
+ 1647960063774 72 1751
147
+ 1647960066081 73 1775
148
+ 1647960066210 73 1775
149
+ 1647960068438 74 1799
150
+ 1647960068645 74 1799
151
+ 1647960070938 75 1823
152
+ 1647960071131 75 1823
153
+ 1647960073367 76 1847
154
+ 1647960073563 76 1847
155
+ 1647960075858 77 1871
156
+ 1647960075986 77 1871
157
+ 1647960078256 78 1895
158
+ 1647960078451 78 1895
159
+ 1647960080766 79 1919
160
+ 1647960080892 79 1919
161
+ 1647960083180 80 1943
162
+ 1647960083380 80 1943
163
+ 1647960085683 81 1967
164
+ 1647960085811 81 1967
165
+ 1647960088103 82 1991
166
+ 1647960088232 82 1991
167
+ 1647960090580 83 2015
168
+ 1647960090774 83 2015
169
+ 1647960093077 84 2039
170
+ 1647960093204 84 2039
171
+ 1647960095491 85 2063
172
+ 1647960095613 85 2063
173
+ 1647960097919 86 2087
174
+ 1647960098130 86 2087
175
+ 1647960100420 87 2111
176
+ 1647960100547 87 2111
177
+ 1647960102813 88 2135
178
+ 1647960102939 88 2135
179
+ 1647960105272 89 2159
180
+ 1647960105405 89 2159
181
+ 1647960107693 90 2183
182
+ 1647960107820 90 2183
183
+ 1647960110156 91 2207
184
+ 1647960110301 91 2207
185
+ 1647960112584 92 2231
186
+ 1647960112713 92 2231
187
+ 1647960115022 93 2255
188
+ 1647960115176 93 2255
189
+ 1647960117460 94 2279
190
+ 1647960117593 94 2279
191
+ 1647960119888 95 2303
192
+ 1647960120090 95 2303
193
+ 1647960122371 96 2327
194
+ 1647960122511 96 2327
195
+ 1647960124793 97 2351
196
+ 1647960124921 97 2351
197
+ 1647960127251 98 2375
198
+ 1647960127449 98 2375
199
+ 1647960129726 99 2399
200
+ 1647960129926 99 2399
201
+ 1647960130550 87 2088
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/test/loss ADDED
@@ -0,0 +1 @@
 
 
1
+ 1647960130550 0.000402881734771654 2088
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/test/mae ADDED
@@ -0,0 +1 @@
 
 
1
+ 1647960130550 0.016752799972891808 2088
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/train/loss ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647959886488 4.342347621917725 23
2
+ 1647959888894 0.05730731412768364 47
3
+ 1647959891311 0.006286133546382189 71
4
+ 1647959893784 0.0013532182201743126 95
5
+ 1647959896230 0.0008239876478910446 119
6
+ 1647959898735 0.0007202242850326002 143
7
+ 1647959901168 0.0006690375157631934 167
8
+ 1647959903624 0.0006276399944908917 191
9
+ 1647959906037 0.0006050210795365274 215
10
+ 1647959908487 0.0005693044513463974 239
11
+ 1647959910953 0.0005463812849484384 263
12
+ 1647959913446 0.0005229813395999372 287
13
+ 1647959915962 0.0005098703550174832 311
14
+ 1647959918436 0.000495909305755049 335
15
+ 1647959920870 0.00048210768727585673 359
16
+ 1647959923340 0.000473074265755713 383
17
+ 1647959925773 0.0004635636869352311 407
18
+ 1647959928259 0.00045512395445257425 431
19
+ 1647959930708 0.00045438535744324327 455
20
+ 1647959933134 0.00044914596946910024 479
21
+ 1647959935640 0.00044510987936519086 503
22
+ 1647959938070 0.00043568635010160506 527
23
+ 1647959940513 0.00043914539855904877 551
24
+ 1647959943042 0.00043160346103832126 575
25
+ 1647959945545 0.00042612897232174873 599
26
+ 1647959948033 0.0004240336420480162 623
27
+ 1647959950491 0.00042002907139249146 647
28
+ 1647959952968 0.00041983669507317245 671
29
+ 1647959955455 0.0004216474189888686 695
30
+ 1647959957965 0.0004165133577771485 719
31
+ 1647959960410 0.0004142482648603618 743
32
+ 1647959962775 0.0004127994761802256 767
33
+ 1647959965199 0.00042158510768786073 791
34
+ 1647959967750 0.00041695087566040456 815
35
+ 1647959970196 0.0004059165657963604 839
36
+ 1647959972661 0.0004062257066834718 863
37
+ 1647959975107 0.000408758845878765 887
38
+ 1647959977562 0.00041210511699318886 911
39
+ 1647959979962 0.0004000916378572583 935
40
+ 1647959982366 0.00039786219713278115 959
41
+ 1647959984855 0.000396965304389596 983
42
+ 1647959987351 0.00040004777838476 1007
43
+ 1647959989788 0.0004005195223726332 1031
44
+ 1647959992297 0.00039835445932112634 1055
45
+ 1647959994698 0.00039412121986970305 1079
46
+ 1647959997166 0.0003929083759430796 1103
47
+ 1647959999651 0.00040109854307956994 1127
48
+ 1647960002132 0.0003939596936106682 1151
49
+ 1647960004609 0.0003906012570951134 1175
50
+ 1647960007107 0.0003921556053683162 1199
51
+ 1647960009559 0.0003916115965694189 1223
52
+ 1647960012033 0.00039053973159752786 1247
53
+ 1647960014466 0.00038710495573468506 1271
54
+ 1647960016920 0.00038700783625245094 1295
55
+ 1647960019395 0.00039118644781410694 1319
56
+ 1647960021828 0.00039927964098751545 1343
57
+ 1647960024302 0.0003894182445947081 1367
58
+ 1647960026728 0.0003865647013299167 1391
59
+ 1647960029198 0.0003892740933224559 1415
60
+ 1647960031776 0.00038562898407690227 1439
61
+ 1647960034168 0.00039640345494262874 1463
62
+ 1647960036549 0.0003929836966563016 1487
63
+ 1647960039049 0.0003826528263743967 1511
64
+ 1647960041464 0.00039758102502673864 1535
65
+ 1647960043874 0.00038526495336554945 1559
66
+ 1647960046406 0.0003987601085100323 1583
67
+ 1647960048881 0.0003918004804290831 1607
68
+ 1647960051372 0.00038548445445485413 1631
69
+ 1647960053881 0.00038453712477348745 1655
70
+ 1647960056386 0.00038578943349421024 1679
71
+ 1647960058849 0.0003904291370417923 1703
72
+ 1647960061292 0.00038028054405003786 1727
73
+ 1647960063774 0.0003783942374866456 1751
74
+ 1647960066210 0.00037868719664402306 1775
75
+ 1647960068645 0.0003772037453018129 1799
76
+ 1647960071131 0.0003772091295104474 1823
77
+ 1647960073563 0.00037675860221497715 1847
78
+ 1647960075986 0.00037759728729724884 1871
79
+ 1647960078451 0.00037590478314086795 1895
80
+ 1647960080892 0.00037703668931499124 1919
81
+ 1647960083380 0.0003931742685381323 1943
82
+ 1647960085811 0.00037598126800730824 1967
83
+ 1647960088232 0.0003758900857064873 1991
84
+ 1647960090774 0.0003811830247286707 2015
85
+ 1647960093204 0.00038812385173514485 2039
86
+ 1647960095613 0.00037520512705668807 2063
87
+ 1647960098130 0.00038031762233003974 2087
88
+ 1647960100547 0.00037949124816805124 2111
89
+ 1647960102939 0.0003777534584514797 2135
90
+ 1647960105405 0.0003773653006646782 2159
91
+ 1647960107820 0.00037604026147164404 2183
92
+ 1647960110301 0.00037361629074439406 2207
93
+ 1647960112713 0.00038605634472332895 2231
94
+ 1647960115176 0.0003846309264190495 2255
95
+ 1647960117593 0.0003783575084526092 2279
96
+ 1647960120090 0.0003748334711417556 2303
97
+ 1647960122511 0.00038411852438002825 2327
98
+ 1647960124921 0.00037512960261665285 2351
99
+ 1647960127449 0.00039327985723502934 2375
100
+ 1647960129926 0.00037736291415058076 2399
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/train/mae ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647959886488 1.3158199787139893 23
2
+ 1647959888894 0.19751663506031036 47
3
+ 1647959891311 0.06423742324113846 71
4
+ 1647959893784 0.02940954454243183 95
5
+ 1647959896230 0.02320658229291439 119
6
+ 1647959898735 0.021659735590219498 143
7
+ 1647959901168 0.021050428971648216 167
8
+ 1647959903624 0.02056770585477352 191
9
+ 1647959906037 0.020237302407622337 215
10
+ 1647959908487 0.019613152369856834 239
11
+ 1647959910953 0.019444193691015244 263
12
+ 1647959913446 0.019030658528208733 287
13
+ 1647959915962 0.01886916160583496 311
14
+ 1647959918436 0.018634887412190437 335
15
+ 1647959920870 0.018486270681023598 359
16
+ 1647959923340 0.018334897235035896 383
17
+ 1647959925773 0.018168363720178604 407
18
+ 1647959928259 0.018060654401779175 431
19
+ 1647959930708 0.0180768221616745 455
20
+ 1647959933134 0.017933538183569908 479
21
+ 1647959935640 0.017830582335591316 503
22
+ 1647959938070 0.01768254116177559 527
23
+ 1647959940513 0.01769491843879223 551
24
+ 1647959943042 0.017608648166060448 575
25
+ 1647959945545 0.017463266849517822 599
26
+ 1647959948033 0.017441386356949806 623
27
+ 1647959950491 0.017370155081152916 647
28
+ 1647959952968 0.017345596104860306 671
29
+ 1647959955455 0.017313584685325623 695
30
+ 1647959957965 0.017224453389644623 719
31
+ 1647959960410 0.01721722073853016 743
32
+ 1647959962775 0.017110856249928474 767
33
+ 1647959965199 0.01726309210062027 791
34
+ 1647959967750 0.017146525904536247 815
35
+ 1647959970196 0.016986556351184845 839
36
+ 1647959972661 0.0169779434800148 863
37
+ 1647959975107 0.017033370211720467 887
38
+ 1647959977562 0.017063522711396217 911
39
+ 1647959979962 0.016827763989567757 935
40
+ 1647959982366 0.016777368262410164 959
41
+ 1647959984855 0.016790589317679405 983
42
+ 1647959987351 0.01677679643034935 1007
43
+ 1647959989788 0.01680240035057068 1031
44
+ 1647959992297 0.01677827537059784 1055
45
+ 1647959994698 0.016687458381056786 1079
46
+ 1647959997166 0.01665475219488144 1103
47
+ 1647959999651 0.016871973872184753 1127
48
+ 1647960002132 0.016618330031633377 1151
49
+ 1647960004609 0.016592541709542274 1175
50
+ 1647960007107 0.01663227006793022 1199
51
+ 1647960009559 0.01655171997845173 1223
52
+ 1647960012033 0.016555769369006157 1247
53
+ 1647960014466 0.016491563990712166 1271
54
+ 1647960016920 0.016510063782334328 1295
55
+ 1647960019395 0.016560301184654236 1319
56
+ 1647960021828 0.016744039952754974 1343
57
+ 1647960024302 0.01651095040142536 1367
58
+ 1647960026728 0.016462186351418495 1391
59
+ 1647960029198 0.01648084446787834 1415
60
+ 1647960031776 0.016439443454146385 1439
61
+ 1647960034168 0.016590487211942673 1463
62
+ 1647960036549 0.016544204205274582 1487
63
+ 1647960039049 0.01636682078242302 1511
64
+ 1647960041464 0.016689548268914223 1535
65
+ 1647960043874 0.016391392797231674 1559
66
+ 1647960046406 0.016639119014143944 1583
67
+ 1647960048881 0.01652211882174015 1607
68
+ 1647960051372 0.016405444592237473 1631
69
+ 1647960053881 0.016384242102503777 1655
70
+ 1647960056386 0.016403140500187874 1679
71
+ 1647960058849 0.01646820455789566 1703
72
+ 1647960061292 0.016317954286932945 1727
73
+ 1647960063774 0.016258543357253075 1751
74
+ 1647960066210 0.01626788266003132 1775
75
+ 1647960068645 0.016254499554634094 1799
76
+ 1647960071131 0.01622932218015194 1823
77
+ 1647960073563 0.016197772696614265 1847
78
+ 1647960075986 0.016238529235124588 1871
79
+ 1647960078451 0.016197707504034042 1895
80
+ 1647960080892 0.016217660158872604 1919
81
+ 1647960083380 0.016501346603035927 1943
82
+ 1647960085811 0.016180992126464844 1967
83
+ 1647960088232 0.016161207109689713 1991
84
+ 1647960090774 0.016300171613693237 2015
85
+ 1647960093204 0.01642199419438839 2039
86
+ 1647960095613 0.01615377515554428 2063
87
+ 1647960098130 0.016325831413269043 2087
88
+ 1647960100547 0.01625642739236355 2111
89
+ 1647960102939 0.016191186383366585 2135
90
+ 1647960105405 0.016204534098505974 2159
91
+ 1647960107820 0.016208775341510773 2183
92
+ 1647960110301 0.016117053106427193 2207
93
+ 1647960112713 0.0163609329611063 2231
94
+ 1647960115176 0.016337504610419273 2255
95
+ 1647960117593 0.01624019630253315 2279
96
+ 1647960120090 0.01612277328968048 2303
97
+ 1647960122511 0.0163507629185915 2327
98
+ 1647960124921 0.016116727143526077 2351
99
+ 1647960127449 0.016511688008904457 2375
100
+ 1647960129926 0.016158610582351685 2399
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/val/loss ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647959886393 0.014177942648530006 23
2
+ 1647959888772 0.024941448122262955 47
3
+ 1647959891185 0.0022524979431182146 71
4
+ 1647959893603 0.000798797351308167 95
5
+ 1647959896075 0.000731383275706321 119
6
+ 1647959898546 0.0006743474514223635 143
7
+ 1647959901038 0.0006443106685765088 167
8
+ 1647959903443 0.0006375169032253325 191
9
+ 1647959905912 0.0005847099237143993 215
10
+ 1647959908338 0.0005691925762221217 239
11
+ 1647959910764 0.0005465119611471891 263
12
+ 1647959913246 0.0005308444960974157 287
13
+ 1647959915770 0.000525242299772799 311
14
+ 1647959918244 0.0005212906980887055 335
15
+ 1647959920693 0.0005010279710404575 359
16
+ 1647959923162 0.0004964997060596943 383
17
+ 1647959925629 0.00048635710845701396 407
18
+ 1647959928086 0.0004799317684955895 431
19
+ 1647959930519 0.0004763782490044832 455
20
+ 1647959932958 0.0004732394590973854 479
21
+ 1647959935471 0.00046858476707711816 503
22
+ 1647959937907 0.0004627775924745947 527
23
+ 1647959940384 0.00048223865451291203 551
24
+ 1647959942822 0.00045650775427930057 575
25
+ 1647959945335 0.00045403436524793506 599
26
+ 1647959947814 0.0004518564383033663 623
27
+ 1647959950303 0.000449806684628129 647
28
+ 1647959952770 0.0004594481724780053 671
29
+ 1647959955260 0.0004483773373067379 695
30
+ 1647959957774 0.0004582132969517261 719
31
+ 1647959960220 0.000451582862297073 743
32
+ 1647959962648 0.0004431355628184974 767
33
+ 1647959965011 0.000467679143184796 791
34
+ 1647959967529 0.0004435107402969152 815
35
+ 1647959969999 0.00044180749682709575 839
36
+ 1647959972477 0.00043626612750813365 863
37
+ 1647959974978 0.0004603431443683803 887
38
+ 1647959977437 0.00044433787115849555 911
39
+ 1647959979782 0.0004384913481771946 935
40
+ 1647959982227 0.00043312823981978 959
41
+ 1647959984625 0.0004400752077344805 983
42
+ 1647959987165 0.0004321136511862278 1007
43
+ 1647959989661 0.0004558261134661734 1031
44
+ 1647959992143 0.00044323524343781173 1055
45
+ 1647959994568 0.00043952741543762386 1079
46
+ 1647959997016 0.00043559601181186736 1103
47
+ 1647959999457 0.00044441656791605055 1127
48
+ 1647960001942 0.00043277026270516217 1151
49
+ 1647960004419 0.0004414572031237185 1175
50
+ 1647960006913 0.0004329547518864274 1199
51
+ 1647960009432 0.0004438141768332571 1223
52
+ 1647960011843 0.0004302943416405469 1247
53
+ 1647960014340 0.0004304391623008996 1271
54
+ 1647960016727 0.0004304780450183898 1295
55
+ 1647960019263 0.0004549633013084531 1319
56
+ 1647960021699 0.0004564271657727659 1343
57
+ 1647960024113 0.0004305720212869346 1367
58
+ 1647960026595 0.00043586301035247743 1391
59
+ 1647960029045 0.00044148581218905747 1415
60
+ 1647960031574 0.0004328140348661691 1439
61
+ 1647960034043 0.00044432288268581033 1463
62
+ 1647960036419 0.00043844786705449224 1487
63
+ 1647960038897 0.000432810396887362 1511
64
+ 1647960041334 0.0004337234713602811 1535
65
+ 1647960043726 0.0004506470577325672 1559
66
+ 1647960046213 0.0005121462745591998 1583
67
+ 1647960048755 0.00043341098353266716 1607
68
+ 1647960051173 0.0004314535763114691 1631
69
+ 1647960053693 0.00046141925849951804 1655
70
+ 1647960056189 0.0004310157091822475 1679
71
+ 1647960058707 0.00043788779294118285 1703
72
+ 1647960061152 0.00043278036173433065 1727
73
+ 1647960063577 0.000435496709542349 1751
74
+ 1647960066081 0.0004316089325584471 1775
75
+ 1647960068438 0.00043277398799546063 1799
76
+ 1647960070938 0.0004360957827884704 1823
77
+ 1647960073367 0.00043179106432944536 1847
78
+ 1647960075858 0.00043282663682475686 1871
79
+ 1647960078256 0.00043416841072030365 1895
80
+ 1647960080766 0.000432082248153165 1919
81
+ 1647960083180 0.00043247482972219586 1943
82
+ 1647960085683 0.0004352308751549572 1967
83
+ 1647960088103 0.00043458439176902175 1991
84
+ 1647960090580 0.0004796235589310527 2015
85
+ 1647960093077 0.0004338948638178408 2039
86
+ 1647960095491 0.00044887603144161403 2063
87
+ 1647960097919 0.00043250841554254293 2087
88
+ 1647960100420 0.00044269993668422103 2111
89
+ 1647960102813 0.00043156533502042294 2135
90
+ 1647960105272 0.00043633393943309784 2159
91
+ 1647960107693 0.000434521323768422 2183
92
+ 1647960110156 0.00043653271859511733 2207
93
+ 1647960112584 0.00044135679490864277 2231
94
+ 1647960115022 0.00043531309347599745 2255
95
+ 1647960117460 0.000436639558756724 2279
96
+ 1647960119888 0.0004387025546748191 2303
97
+ 1647960122371 0.0004409730900079012 2327
98
+ 1647960124793 0.00045858731027692556 2351
99
+ 1647960127251 0.00044121488463133574 2375
100
+ 1647960129726 0.0005107350298203528 2399
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/val/mae ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647959886393 0.09954570978879929 23
2
+ 1647959888772 0.1495475023984909 47
3
+ 1647959891185 0.04012170061469078 71
4
+ 1647959893603 0.023267723619937897 95
5
+ 1647959896075 0.021397700533270836 119
6
+ 1647959898546 0.02102094329893589 143
7
+ 1647959901038 0.02113427221775055 167
8
+ 1647959903443 0.02006976678967476 191
9
+ 1647959905912 0.020206814631819725 215
10
+ 1647959908338 0.020232098177075386 239
11
+ 1647959910764 0.019510304555296898 263
12
+ 1647959913246 0.01932346075773239 287
13
+ 1647959915770 0.018989086151123047 311
14
+ 1647959918244 0.018810341134667397 335
15
+ 1647959920693 0.018762608990073204 359
16
+ 1647959923162 0.018582377582788467 383
17
+ 1647959925629 0.018572870641946793 407
18
+ 1647959928086 0.01848633959889412 431
19
+ 1647959930519 0.018363453447818756 455
20
+ 1647959932958 0.018254607915878296 479
21
+ 1647959935471 0.018206853419542313 503
22
+ 1647959937907 0.018315646797418594 527
23
+ 1647959940384 0.018926186487078667 551
24
+ 1647959942822 0.01812773011624813 575
25
+ 1647959945335 0.01813136413693428 599
26
+ 1647959947814 0.017970262095332146 623
27
+ 1647959950303 0.01790132373571396 647
28
+ 1647959952770 0.018357347697019577 671
29
+ 1647959955260 0.01778845489025116 695
30
+ 1647959957774 0.017787707969546318 719
31
+ 1647959960220 0.01773429661989212 743
32
+ 1647959962648 0.017908763140439987 767
33
+ 1647959965011 0.01842939481139183 791
34
+ 1647959967529 0.017640704289078712 815
35
+ 1647959969999 0.0178662296384573 839
36
+ 1647959972477 0.01761534810066223 863
37
+ 1647959974978 0.01775723695755005 887
38
+ 1647959977437 0.017887664958834648 911
39
+ 1647959979782 0.017551790922880173 935
40
+ 1647959982227 0.017565051093697548 959
41
+ 1647959984625 0.017544588074088097 983
42
+ 1647959987165 0.017519894987344742 1007
43
+ 1647959989661 0.01801685057580471 1031
44
+ 1647959992143 0.01776825450360775 1055
45
+ 1647959994568 0.01769193448126316 1079
46
+ 1647959997016 0.017593219876289368 1103
47
+ 1647959999457 0.01759103313088417 1127
48
+ 1647960001942 0.01745220273733139 1151
49
+ 1647960004419 0.01768527552485466 1175
50
+ 1647960006913 0.01749493181705475 1199
51
+ 1647960009432 0.0175788477063179 1223
52
+ 1647960011843 0.0174209326505661 1247
53
+ 1647960014340 0.017427116632461548 1271
54
+ 1647960016727 0.017413398250937462 1295
55
+ 1647960019263 0.01783713698387146 1319
56
+ 1647960021699 0.017778683453798294 1343
57
+ 1647960024113 0.017405666410923004 1367
58
+ 1647960026595 0.017491891980171204 1391
59
+ 1647960029045 0.01758340559899807 1415
60
+ 1647960031574 0.017423313111066818 1439
61
+ 1647960034043 0.017613204196095467 1463
62
+ 1647960036419 0.01751849800348282 1487
63
+ 1647960038897 0.017426760867238045 1511
64
+ 1647960041334 0.017433997243642807 1535
65
+ 1647960043726 0.017688874155282974 1559
66
+ 1647960046213 0.018620284274220467 1583
67
+ 1647960048755 0.017417754977941513 1607
68
+ 1647960051173 0.01739826239645481 1631
69
+ 1647960053693 0.017827652394771576 1655
70
+ 1647960056189 0.017386266961693764 1679
71
+ 1647960058707 0.017491240054368973 1703
72
+ 1647960061152 0.01741820201277733 1727
73
+ 1647960063577 0.017456887289881706 1751
74
+ 1647960066081 0.017395686358213425 1775
75
+ 1647960068438 0.017391294240951538 1799
76
+ 1647960070938 0.017426805570721626 1823
77
+ 1647960073367 0.01736919768154621 1847
78
+ 1647960075858 0.017420921474695206 1871
79
+ 1647960078256 0.01739104837179184 1895
80
+ 1647960080766 0.017376182600855827 1919
81
+ 1647960083180 0.017374588176608086 1943
82
+ 1647960085683 0.01739008165895939 1967
83
+ 1647960088103 0.017396951094269753 1991
84
+ 1647960090580 0.01820252649486065 2015
85
+ 1647960093077 0.017369873821735382 2039
86
+ 1647960095491 0.017721308395266533 2063
87
+ 1647960097919 0.01735648140311241 2087
88
+ 1647960100420 0.017486795783042908 2111
89
+ 1647960102813 0.01737273670732975 2135
90
+ 1647960105272 0.01747000217437744 2159
91
+ 1647960107693 0.017369193956255913 2183
92
+ 1647960110156 0.01748810149729252 2207
93
+ 1647960112584 0.017475992441177368 2231
94
+ 1647960115022 0.01737702079117298 2255
95
+ 1647960117460 0.01747899316251278 2279
96
+ 1647960119888 0.01744064688682556 2303
97
+ 1647960122371 0.01744801551103592 2327
98
+ 1647960124793 0.01768660731613636 2351
99
+ 1647960127251 0.017449431121349335 2375
100
+ 1647960129726 0.018421221524477005 2399
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/metrics/val/mae_best ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1647959886393 0.038817837834358215 23
2
+ 1647959888772 0.038817837834358215 47
3
+ 1647959891185 0.038817837834358215 71
4
+ 1647959893603 0.023267723619937897 95
5
+ 1647959896075 0.021397700533270836 119
6
+ 1647959898546 0.02102094329893589 143
7
+ 1647959901038 0.02102094329893589 167
8
+ 1647959903443 0.02006976678967476 191
9
+ 1647959905912 0.02006976678967476 215
10
+ 1647959908338 0.02006976678967476 239
11
+ 1647959910764 0.019510304555296898 263
12
+ 1647959913246 0.01932346075773239 287
13
+ 1647959915770 0.018989086151123047 311
14
+ 1647959918244 0.018810341134667397 335
15
+ 1647959920693 0.018762608990073204 359
16
+ 1647959923162 0.018582377582788467 383
17
+ 1647959925629 0.018572870641946793 407
18
+ 1647959928086 0.01848633959889412 431
19
+ 1647959930519 0.018363453447818756 455
20
+ 1647959932958 0.018254607915878296 479
21
+ 1647959935471 0.018206853419542313 503
22
+ 1647959937907 0.018206853419542313 527
23
+ 1647959940384 0.018206853419542313 551
24
+ 1647959942822 0.01812773011624813 575
25
+ 1647959945335 0.01812773011624813 599
26
+ 1647959947814 0.017970262095332146 623
27
+ 1647959950303 0.01790132373571396 647
28
+ 1647959952770 0.01790132373571396 671
29
+ 1647959955260 0.01778845489025116 695
30
+ 1647959957774 0.017787707969546318 719
31
+ 1647959960220 0.01773429661989212 743
32
+ 1647959962648 0.01773429661989212 767
33
+ 1647959965011 0.01773429661989212 791
34
+ 1647959967529 0.017640704289078712 815
35
+ 1647959969999 0.017640704289078712 839
36
+ 1647959972477 0.01761534810066223 863
37
+ 1647959974978 0.01761534810066223 887
38
+ 1647959977437 0.01761534810066223 911
39
+ 1647959979782 0.017551790922880173 935
40
+ 1647959982227 0.017551790922880173 959
41
+ 1647959984625 0.017544588074088097 983
42
+ 1647959987165 0.017519894987344742 1007
43
+ 1647959989661 0.017519894987344742 1031
44
+ 1647959992143 0.017519894987344742 1055
45
+ 1647959994568 0.017519894987344742 1079
46
+ 1647959997016 0.017519894987344742 1103
47
+ 1647959999457 0.017519894987344742 1127
48
+ 1647960001942 0.01745220273733139 1151
49
+ 1647960004419 0.01745220273733139 1175
50
+ 1647960006913 0.01745220273733139 1199
51
+ 1647960009432 0.01745220273733139 1223
52
+ 1647960011843 0.0174209326505661 1247
53
+ 1647960014340 0.0174209326505661 1271
54
+ 1647960016727 0.017413398250937462 1295
55
+ 1647960019263 0.017413398250937462 1319
56
+ 1647960021699 0.017413398250937462 1343
57
+ 1647960024113 0.017405666410923004 1367
58
+ 1647960026595 0.017405666410923004 1391
59
+ 1647960029045 0.017405666410923004 1415
60
+ 1647960031574 0.017405666410923004 1439
61
+ 1647960034043 0.017405666410923004 1463
62
+ 1647960036419 0.017405666410923004 1487
63
+ 1647960038897 0.017405666410923004 1511
64
+ 1647960041334 0.017405666410923004 1535
65
+ 1647960043726 0.017405666410923004 1559
66
+ 1647960046213 0.017405666410923004 1583
67
+ 1647960048755 0.017405666410923004 1607
68
+ 1647960051173 0.01739826239645481 1631
69
+ 1647960053693 0.01739826239645481 1655
70
+ 1647960056189 0.017386266961693764 1679
71
+ 1647960058707 0.017386266961693764 1703
72
+ 1647960061152 0.017386266961693764 1727
73
+ 1647960063577 0.017386266961693764 1751
74
+ 1647960066081 0.017386266961693764 1775
75
+ 1647960068438 0.017386266961693764 1799
76
+ 1647960070938 0.017386266961693764 1823
77
+ 1647960073367 0.01736919768154621 1847
78
+ 1647960075858 0.01736919768154621 1871
79
+ 1647960078256 0.01736919768154621 1895
80
+ 1647960080766 0.01736919768154621 1919
81
+ 1647960083180 0.01736919768154621 1943
82
+ 1647960085683 0.01736919768154621 1967
83
+ 1647960088103 0.01736919768154621 1991
84
+ 1647960090580 0.01736919768154621 2015
85
+ 1647960093077 0.01736919768154621 2039
86
+ 1647960095491 0.01736919768154621 2063
87
+ 1647960097919 0.01735648140311241 2087
88
+ 1647960100420 0.01735648140311241 2111
89
+ 1647960102813 0.01735648140311241 2135
90
+ 1647960105272 0.01735648140311241 2159
91
+ 1647960107693 0.01735648140311241 2183
92
+ 1647960110156 0.01735648140311241 2207
93
+ 1647960112584 0.01735648140311241 2231
94
+ 1647960115022 0.01735648140311241 2255
95
+ 1647960117460 0.01735648140311241 2279
96
+ 1647960119888 0.01735648140311241 2303
97
+ 1647960122371 0.01735648140311241 2327
98
+ 1647960124793 0.01735648140311241 2351
99
+ 1647960127251 0.01735648140311241 2375
100
+ 1647960129726 0.01735648140311241 2399
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.EarlyStopping
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/min_delta ADDED
@@ -0,0 +1 @@
 
 
1
+ 0
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/mode ADDED
@@ -0,0 +1 @@
 
 
1
+ min
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/monitor ADDED
@@ -0,0 +1 @@
 
 
1
+ val/mae
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/early_stopping/patience ADDED
@@ -0,0 +1 @@
 
 
1
+ 100
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.ModelCheckpoint
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/auto_insert_metric_name ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/dirpath ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoints/
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/filename ADDED
@@ -0,0 +1 @@
 
 
1
+ epoch_{epoch:03d}
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/mode ADDED
@@ -0,0 +1 @@
 
 
1
+ min
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/monitor ADDED
@@ -0,0 +1 @@
 
 
1
+ val/mae
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/save_last ADDED
@@ -0,0 +1 @@
 
 
1
+ True
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/save_top_k ADDED
@@ -0,0 +1 @@
 
 
1
+ 1
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_checkpoint/verbose ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_summary/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.RichModelSummary
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/model_summary/max_depth ADDED
@@ -0,0 +1 @@
 
 
1
+ -1
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/callbacks/rich_progress_bar/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ pytorch_lightning.callbacks.RichProgressBar
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ src.datamodules.focus_datamodule.FocusDataModule
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/batch_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 128
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/csv_test_file ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150/test_metadata.csv
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/csv_train_file ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150/train_metadata.csv
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/csv_val_file ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150/validation_metadata.csv
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/data_dir ADDED
@@ -0,0 +1 @@
 
 
1
+ /usr/src/app/data/focus150
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/num_workers ADDED
@@ -0,0 +1 @@
 
 
1
+ 0
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/datamodule/pin_memory ADDED
@@ -0,0 +1 @@
 
 
1
+ False
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/_target_ ADDED
@@ -0,0 +1 @@
 
 
1
+ src.models.focus_conv_module.FocusConvLitModule
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv1_channels ADDED
@@ -0,0 +1 @@
 
 
1
+ 6
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv1_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 5
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv2_channels ADDED
@@ -0,0 +1 @@
 
 
1
+ 16
logs/mlflow/mlruns/1/62c4a535287f4b44bd537524cf678fff/params/model/conv2_size ADDED
@@ -0,0 +1 @@
 
 
1
+ 5