tomas-gajarsky commited on
Commit
c12eda1
·
verified ·
1 Parent(s): cb4e60a

Update config.merged.gpu.yml

Browse files
Files changed (1) hide show
  1. config.merged.gpu.yml +74 -129
config.merged.gpu.yml CHANGED
@@ -17,9 +17,11 @@ analyzer:
17
  detector:
18
  _target_: facetorch.analyzer.detector.FaceDetector
19
  downloader:
20
- _target_: facetorch.downloader.DownloaderGDrive
21
  file_id: 154x2VjmTQVqmowB0yZw4Uck7uQs2vVBs
22
- path_local: /home/user/code/models/torchscript/detector/1/model.pt
 
 
23
  device:
24
  _target_: torch.device
25
  type: ${analyzer.device}
@@ -30,14 +32,8 @@ analyzer:
30
  _target_: torchvision.transforms.Compose
31
  transforms:
32
  - _target_: torchvision.transforms.Normalize
33
- mean:
34
- - 123.0
35
- - 117.0
36
- - 104.0
37
- std:
38
- - 1.0
39
- - 1.0
40
- - 1.0
41
  device:
42
  _target_: torch.device
43
  type: ${analyzer.device}
@@ -58,20 +54,12 @@ analyzer:
58
  prior_box:
59
  _target_: facetorch.analyzer.detector.post.PriorBox
60
  min_sizes:
61
- - - 16
62
- - 32
63
- - - 64
64
- - 128
65
- - - 256
66
- - 512
67
- steps:
68
- - 8
69
- - 16
70
- - 32
71
  clip: false
72
- variance:
73
- - 0.1
74
- - 0.2
75
  reverse_colors: ${analyzer.detector.reverse_colors}
76
  expand_box_ratio: 0.1
77
  unifier:
@@ -80,18 +68,10 @@ analyzer:
80
  _target_: torchvision.transforms.Compose
81
  transforms:
82
  - _target_: torchvision.transforms.Normalize
83
- mean:
84
- - -123.0
85
- - -117.0
86
- - -104.0
87
- std:
88
- - 255.0
89
- - 255.0
90
- - 255.0
91
  - _target_: torchvision.transforms.Resize
92
- size:
93
- - 380
94
- - 380
95
  device:
96
  _target_: torch.device
97
  type: ${analyzer.device}
@@ -100,9 +80,11 @@ analyzer:
100
  embed:
101
  _target_: facetorch.analyzer.predictor.FacePredictor
102
  downloader:
103
- _target_: facetorch.downloader.DownloaderGDrive
104
  file_id: 19h3kqar1wlELAmM5hDyj9tlrUh8yjrCl
105
- path_local: /home/user/code/models/torchscript/predictor/embed/1/model.pt
 
 
106
  device:
107
  _target_: torch.device
108
  type: ${analyzer.device}
@@ -112,18 +94,10 @@ analyzer:
112
  _target_: torchvision.transforms.Compose
113
  transforms:
114
  - _target_: torchvision.transforms.Resize
115
- size:
116
- - 244
117
- - 244
118
  - _target_: torchvision.transforms.Normalize
119
- mean:
120
- - 0.485
121
- - 0.456
122
- - 0.406
123
- std:
124
- - 0.228
125
- - 0.224
126
- - 0.225
127
  device:
128
  _target_: torch.device
129
  type: ${analyzer.predictor.fer.device.type}
@@ -136,14 +110,15 @@ analyzer:
136
  _target_: torch.device
137
  type: ${analyzer.predictor.fer.device.type}
138
  optimize_transform: ${analyzer.optimize_transforms}
139
- labels:
140
- - abstract
141
  verify:
142
  _target_: facetorch.analyzer.predictor.FacePredictor
143
  downloader:
144
- _target_: facetorch.downloader.DownloaderGDrive
145
  file_id: 1H-aPtFd9C5D7y1vzoWsObKAxeIBE9QPd
146
- path_local: /home/user/code/models/torchscript/predictor/verify/1/model.pt
 
 
147
  device:
148
  _target_: torch.device
149
  type: ${analyzer.device}
@@ -153,18 +128,10 @@ analyzer:
153
  _target_: torchvision.transforms.Compose
154
  transforms:
155
  - _target_: torchvision.transforms.Resize
156
- size:
157
- - 112
158
- - 112
159
  - _target_: torchvision.transforms.Normalize
160
- mean:
161
- - 0.485
162
- - 0.456
163
- - 0.406
164
- std:
165
- - 0.229
166
- - 0.224
167
- - 0.225
168
  device:
169
  _target_: torch.device
170
  type: ${analyzer.predictor.verify.device.type}
@@ -177,14 +144,15 @@ analyzer:
177
  _target_: torch.device
178
  type: ${analyzer.predictor.verify.device.type}
179
  optimize_transform: ${analyzer.optimize_transforms}
180
- labels:
181
- - abstract
182
  fer:
183
  _target_: facetorch.analyzer.predictor.FacePredictor
184
  downloader:
185
- _target_: facetorch.downloader.DownloaderGDrive
186
  file_id: 1xoB5VYOd0XLjb-rQqqHWCkQvma4NytEd
187
- path_local: /home/user/code/models/torchscript/predictor/fer/2/model.pt
 
 
188
  device:
189
  _target_: torch.device
190
  type: ${analyzer.device}
@@ -194,18 +162,10 @@ analyzer:
194
  _target_: torchvision.transforms.Compose
195
  transforms:
196
  - _target_: torchvision.transforms.Resize
197
- size:
198
- - 260
199
- - 260
200
  - _target_: torchvision.transforms.Normalize
201
- mean:
202
- - 0.485
203
- - 0.456
204
- - 0.406
205
- std:
206
- - 0.229
207
- - 0.224
208
- - 0.225
209
  device:
210
  _target_: torch.device
211
  type: ${analyzer.predictor.fer.device.type}
@@ -231,31 +191,33 @@ analyzer:
231
  au:
232
  _target_: facetorch.analyzer.predictor.FacePredictor
233
  downloader:
234
- _target_: facetorch.downloader.DownloaderGDrive
235
  file_id: 1uoVX9suSA5JVWTms3hEtJKzwO-CUR_jV
236
- path_local: /home/user/opt/facetorch/models/torchscript/predictor/au/1/model.pt # str
 
 
237
  device:
238
  _target_: torch.device
239
- type: ${analyzer.device} # str
240
  preprocessor:
241
  _target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor
242
  transform:
243
  _target_: torchvision.transforms.Compose
244
  transforms:
245
- - _target_: torchvision.transforms.Resize
246
- size: [224, 224] # List[int]
247
- - _target_: torchvision.transforms.Normalize
248
- mean: [0.485, 0.456, 0.406] # List[float]
249
- std: [0.229, 0.224, 0.225] # List[float]
250
- device:
251
  _target_: torch.device
252
  type: ${analyzer.predictor.au.device.type}
253
  optimize_transform: ${analyzer.optimize_transforms}
254
- reverse_colors: False # bool
255
  postprocessor:
256
  _target_: facetorch.analyzer.predictor.post.PostMultiLabel
257
  transform: None
258
- device:
259
  _target_: torch.device
260
  type: ${analyzer.predictor.au.device.type}
261
  optimize_transform: ${analyzer.optimize_transforms}
@@ -306,9 +268,11 @@ analyzer:
306
  va:
307
  _target_: facetorch.analyzer.predictor.FacePredictor
308
  downloader:
309
- _target_: facetorch.downloader.DownloaderGDrive
310
  file_id: 1Xl4ilNCU_DgKNhITrXb3UyQUUdm3VTKS
311
- path_local: /home/user/code/models/torchscript/predictor/va/1/model.pt
 
 
312
  device:
313
  _target_: torch.device
314
  type: ${analyzer.device}
@@ -318,19 +282,11 @@ analyzer:
318
  _target_: torchvision.transforms.Compose
319
  transforms:
320
  - _target_: torchvision.transforms.Resize
321
- size:
322
- - 224
323
- - 224
324
  antialias: true
325
  - _target_: torchvision.transforms.Normalize
326
- mean:
327
- - 0.485
328
- - 0.456
329
- - 0.406
330
- std:
331
- - 0.229
332
- - 0.224
333
- - 0.225
334
  device:
335
  _target_: torch.device
336
  type: ${analyzer.predictor.va.device.type}
@@ -343,15 +299,15 @@ analyzer:
343
  _target_: torch.device
344
  type: ${analyzer.predictor.va.device.type}
345
  optimize_transform: ${analyzer.optimize_transforms}
346
- labels:
347
- - valence
348
- - arousal
349
  deepfake:
350
  _target_: facetorch.analyzer.predictor.FacePredictor
351
  downloader:
352
- _target_: facetorch.downloader.DownloaderGDrive
353
  file_id: 1GjDTwQpvrkCjXOdiBy1oMkzm7nt-bXFg
354
- path_local: /home/user/code/models/torchscript/predictor/deepfake/1/model.pt
 
 
355
  device:
356
  _target_: torch.device
357
  type: ${analyzer.device}
@@ -361,18 +317,10 @@ analyzer:
361
  _target_: torchvision.transforms.Compose
362
  transforms:
363
  - _target_: torchvision.transforms.Resize
364
- size:
365
- - 380
366
- - 380
367
  - _target_: torchvision.transforms.Normalize
368
- mean:
369
- - 0.485
370
- - 0.456
371
- - 0.406
372
- std:
373
- - 0.229
374
- - 0.224
375
- - 0.225
376
  device:
377
  _target_: torch.device
378
  type: ${analyzer.device}
@@ -385,16 +333,16 @@ analyzer:
385
  _target_: torch.device
386
  type: ${analyzer.device}
387
  optimize_transform: ${analyzer.optimize_transforms}
388
- labels:
389
- - Real
390
- - Fake
391
  threshold: 0.7
392
  align:
393
  _target_: facetorch.analyzer.predictor.FacePredictor
394
  downloader:
395
- _target_: facetorch.downloader.DownloaderGDrive
396
  file_id: 16gNFQdEH2nWvW3zTbdIAniKIbPAp6qBA
397
- path_local: /home/user/code/models/torchscript/predictor/align/1/model.pt
 
 
398
  device:
399
  _target_: torch.device
400
  type: ${analyzer.device}
@@ -404,9 +352,7 @@ analyzer:
404
  _target_: torchvision.transforms.Compose
405
  transforms:
406
  - _target_: torchvision.transforms.Resize
407
- size:
408
- - 120
409
- - 120
410
  device:
411
  _target_: torch.device
412
  type: ${analyzer.predictor.align.device.type}
@@ -419,8 +365,7 @@ analyzer:
419
  _target_: torch.device
420
  type: ${analyzer.predictor.align.device.type}
421
  optimize_transform: ${analyzer.optimize_transforms}
422
- labels:
423
- - abstract
424
  utilizer:
425
  align:
426
  _target_: facetorch.analyzer.utilizer.align.Lmk3DMeshPose
@@ -430,9 +375,9 @@ analyzer:
430
  type: ${analyzer.device}
431
  optimize_transform: false
432
  downloader_meta:
433
- _target_: facetorch.downloader.DownloaderGDrive
434
  file_id: 11tdAcFuSXqCCf58g52WT1Rpa8KuQwe2o
435
- path_local: /home/user/code/data/3dmm/meta.pt
436
  image_size: 120
437
  draw_boxes:
438
  _target_: facetorch.analyzer.utilizer.draw.BoxDrawer
 
17
  detector:
18
  _target_: facetorch.analyzer.detector.FaceDetector
19
  downloader:
20
+ _target_: facetorch.downloader.DownloaderHuggingFace
21
  file_id: 154x2VjmTQVqmowB0yZw4Uck7uQs2vVBs
22
+ path_local: /opt/facetorch/models/torchscript/detector/1/model.pt
23
+ repo_id: tomas-gajarsky/facetorch-detector-retinaface
24
+ filename: model.pt
25
  device:
26
  _target_: torch.device
27
  type: ${analyzer.device}
 
32
  _target_: torchvision.transforms.Compose
33
  transforms:
34
  - _target_: torchvision.transforms.Normalize
35
+ mean: [123.0, 117.0, 104.0]
36
+ std: [1.0, 1.0, 1.0]
 
 
 
 
 
 
37
  device:
38
  _target_: torch.device
39
  type: ${analyzer.device}
 
54
  prior_box:
55
  _target_: facetorch.analyzer.detector.post.PriorBox
56
  min_sizes:
57
+ - [16, 32]
58
+ - [64, 128]
59
+ - [256, 512]
60
+ steps: [8, 16, 32]
 
 
 
 
 
 
61
  clip: false
62
+ variance: [0.1, 0.2]
 
 
63
  reverse_colors: ${analyzer.detector.reverse_colors}
64
  expand_box_ratio: 0.1
65
  unifier:
 
68
  _target_: torchvision.transforms.Compose
69
  transforms:
70
  - _target_: torchvision.transforms.Normalize
71
+ mean: [-123.0, -117.0, -104.0]
72
+ std: [255.0, 255.0, 255.0]
 
 
 
 
 
 
73
  - _target_: torchvision.transforms.Resize
74
+ size: [380, 380]
 
 
75
  device:
76
  _target_: torch.device
77
  type: ${analyzer.device}
 
80
  embed:
81
  _target_: facetorch.analyzer.predictor.FacePredictor
82
  downloader:
83
+ _target_: facetorch.downloader.DownloaderHuggingFace
84
  file_id: 19h3kqar1wlELAmM5hDyj9tlrUh8yjrCl
85
+ path_local: /opt/facetorch/models/torchscript/predictor/embed/1/model.pt
86
+ repo_id: tomas-gajarsky/facetorch-embed-resnet-50
87
+ filename: model.pt
88
  device:
89
  _target_: torch.device
90
  type: ${analyzer.device}
 
94
  _target_: torchvision.transforms.Compose
95
  transforms:
96
  - _target_: torchvision.transforms.Resize
97
+ size: [244, 244]
 
 
98
  - _target_: torchvision.transforms.Normalize
99
+ mean: [0.485, 0.456, 0.406]
100
+ std: [0.228, 0.224, 0.225]
 
 
 
 
 
 
101
  device:
102
  _target_: torch.device
103
  type: ${analyzer.predictor.fer.device.type}
 
110
  _target_: torch.device
111
  type: ${analyzer.predictor.fer.device.type}
112
  optimize_transform: ${analyzer.optimize_transforms}
113
+ labels: [abstract]
 
114
  verify:
115
  _target_: facetorch.analyzer.predictor.FacePredictor
116
  downloader:
117
+ _target_: facetorch.downloader.DownloaderHuggingFace
118
  file_id: 1H-aPtFd9C5D7y1vzoWsObKAxeIBE9QPd
119
+ path_local: /opt/facetorch/models/torchscript/predictor/verify/1/model.pt
120
+ repo_id: tomas-gajarsky/facetorch-verify-adaface
121
+ filename: model.pt
122
  device:
123
  _target_: torch.device
124
  type: ${analyzer.device}
 
128
  _target_: torchvision.transforms.Compose
129
  transforms:
130
  - _target_: torchvision.transforms.Resize
131
+ size: [112, 112]
 
 
132
  - _target_: torchvision.transforms.Normalize
133
+ mean: [0.485, 0.456, 0.406]
134
+ std: [0.229, 0.224, 0.225]
 
 
 
 
 
 
135
  device:
136
  _target_: torch.device
137
  type: ${analyzer.predictor.verify.device.type}
 
144
  _target_: torch.device
145
  type: ${analyzer.predictor.verify.device.type}
146
  optimize_transform: ${analyzer.optimize_transforms}
147
+ labels: [abstract]
 
148
  fer:
149
  _target_: facetorch.analyzer.predictor.FacePredictor
150
  downloader:
151
+ _target_: facetorch.downloader.DownloaderHuggingFace
152
  file_id: 1xoB5VYOd0XLjb-rQqqHWCkQvma4NytEd
153
+ path_local: /opt/facetorch/models/torchscript/predictor/fer/2/model.pt
154
+ repo_id: tomas-gajarsky/facetorch-fer-efficientnet-b2
155
+ filename: model.pt
156
  device:
157
  _target_: torch.device
158
  type: ${analyzer.device}
 
162
  _target_: torchvision.transforms.Compose
163
  transforms:
164
  - _target_: torchvision.transforms.Resize
165
+ size: [260, 260]
 
 
166
  - _target_: torchvision.transforms.Normalize
167
+ mean: [0.485, 0.456, 0.406]
168
+ std: [0.229, 0.224, 0.225]
 
 
 
 
 
 
169
  device:
170
  _target_: torch.device
171
  type: ${analyzer.predictor.fer.device.type}
 
191
  au:
192
  _target_: facetorch.analyzer.predictor.FacePredictor
193
  downloader:
194
+ _target_: facetorch.downloader.DownloaderHuggingFace
195
  file_id: 1uoVX9suSA5JVWTms3hEtJKzwO-CUR_jV
196
+ path_local: /opt/facetorch/models/torchscript/predictor/au/1/model.pt
197
+ repo_id: tomas-gajarsky/facetorch-au-opengraph
198
+ filename: model.pt
199
  device:
200
  _target_: torch.device
201
+ type: ${analyzer.device} # str
202
  preprocessor:
203
  _target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor
204
  transform:
205
  _target_: torchvision.transforms.Compose
206
  transforms:
207
+ - _target_: torchvision.transforms.Resize
208
+ size: [224, 224] # List[int]
209
+ - _target_: torchvision.transforms.Normalize
210
+ mean: [0.485, 0.456, 0.406] # List[float]
211
+ std: [0.229, 0.224, 0.225] # List[float]
212
+ device:
213
  _target_: torch.device
214
  type: ${analyzer.predictor.au.device.type}
215
  optimize_transform: ${analyzer.optimize_transforms}
216
+ reverse_colors: False # bool
217
  postprocessor:
218
  _target_: facetorch.analyzer.predictor.post.PostMultiLabel
219
  transform: None
220
+ device:
221
  _target_: torch.device
222
  type: ${analyzer.predictor.au.device.type}
223
  optimize_transform: ${analyzer.optimize_transforms}
 
268
  va:
269
  _target_: facetorch.analyzer.predictor.FacePredictor
270
  downloader:
271
+ _target_: facetorch.downloader.DownloaderHuggingFace
272
  file_id: 1Xl4ilNCU_DgKNhITrXb3UyQUUdm3VTKS
273
+ path_local: /opt/facetorch/models/torchscript/predictor/va/1/model.pt
274
+ repo_id: tomas-gajarsky/facetorch-va-elim
275
+ filename: model.pt
276
  device:
277
  _target_: torch.device
278
  type: ${analyzer.device}
 
282
  _target_: torchvision.transforms.Compose
283
  transforms:
284
  - _target_: torchvision.transforms.Resize
285
+ size: [224, 224]
 
 
286
  antialias: true
287
  - _target_: torchvision.transforms.Normalize
288
+ mean: [0.485, 0.456, 0.406]
289
+ std: [0.229, 0.224, 0.225]
 
 
 
 
 
 
290
  device:
291
  _target_: torch.device
292
  type: ${analyzer.predictor.va.device.type}
 
299
  _target_: torch.device
300
  type: ${analyzer.predictor.va.device.type}
301
  optimize_transform: ${analyzer.optimize_transforms}
302
+ labels: [valence, arousal]
 
 
303
  deepfake:
304
  _target_: facetorch.analyzer.predictor.FacePredictor
305
  downloader:
306
+ _target_: facetorch.downloader.DownloaderHuggingFace
307
  file_id: 1GjDTwQpvrkCjXOdiBy1oMkzm7nt-bXFg
308
+ path_local: /opt/facetorch/models/torchscript/predictor/deepfake/1/model.pt
309
+ repo_id: tomas-gajarsky/facetorch-deepfake-efficientnet-b7
310
+ filename: model.pt
311
  device:
312
  _target_: torch.device
313
  type: ${analyzer.device}
 
317
  _target_: torchvision.transforms.Compose
318
  transforms:
319
  - _target_: torchvision.transforms.Resize
320
+ size: [380, 380]
 
 
321
  - _target_: torchvision.transforms.Normalize
322
+ mean: [0.485, 0.456, 0.406]
323
+ std: [0.229, 0.224, 0.225]
 
 
 
 
 
 
324
  device:
325
  _target_: torch.device
326
  type: ${analyzer.device}
 
333
  _target_: torch.device
334
  type: ${analyzer.device}
335
  optimize_transform: ${analyzer.optimize_transforms}
336
+ labels: [Real, Fake]
 
 
337
  threshold: 0.7
338
  align:
339
  _target_: facetorch.analyzer.predictor.FacePredictor
340
  downloader:
341
+ _target_: facetorch.downloader.DownloaderHuggingFace
342
  file_id: 16gNFQdEH2nWvW3zTbdIAniKIbPAp6qBA
343
+ path_local: /opt/facetorch/models/torchscript/predictor/align/1/model.pt
344
+ repo_id: tomas-gajarsky/facetorch-align-synergynet
345
+ filename: model.pt
346
  device:
347
  _target_: torch.device
348
  type: ${analyzer.device}
 
352
  _target_: torchvision.transforms.Compose
353
  transforms:
354
  - _target_: torchvision.transforms.Resize
355
+ size: [120, 120]
 
 
356
  device:
357
  _target_: torch.device
358
  type: ${analyzer.predictor.align.device.type}
 
365
  _target_: torch.device
366
  type: ${analyzer.predictor.align.device.type}
367
  optimize_transform: ${analyzer.optimize_transforms}
368
+ labels: [abstract]
 
369
  utilizer:
370
  align:
371
  _target_: facetorch.analyzer.utilizer.align.Lmk3DMeshPose
 
375
  type: ${analyzer.device}
376
  optimize_transform: false
377
  downloader_meta:
378
+ _target_: facetorch.downloader.DownloaderGDrive # (unchanged)
379
  file_id: 11tdAcFuSXqCCf58g52WT1Rpa8KuQwe2o
380
+ path_local: /opt/facetorch/data/3dmm/meta.pt
381
  image_size: 120
382
  draw_boxes:
383
  _target_: facetorch.analyzer.utilizer.draw.BoxDrawer