Spaces:
Running
Running
analyzer: | |
device: cpu | |
optimize_transforms: true | |
reader: | |
_target_: facetorch.analyzer.reader.ImageReader | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: ${analyzer.optimize_transforms} | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: facetorch.transforms.SquarePad | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 1080 | |
detector: | |
_target_: facetorch.analyzer.detector.FaceDetector | |
downloader: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 154x2VjmTQVqmowB0yZw4Uck7uQs2vVBs | |
path_local: /opt/facetorch/models/torchscript/detector/1/model.pt | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
reverse_colors: true | |
preprocessor: | |
_target_: facetorch.analyzer.detector.pre.DetectorPreProcessor | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Normalize | |
mean: | |
- 104.0 | |
- 117.0 | |
- 123.0 | |
std: | |
- 1.0 | |
- 1.0 | |
- 1.0 | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: ${analyzer.optimize_transforms} | |
reverse_colors: ${analyzer.detector.reverse_colors} | |
postprocessor: | |
_target_: facetorch.analyzer.detector.post.PostRetFace | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: ${analyzer.optimize_transforms} | |
confidence_threshold: 0.02 | |
top_k: 5000 | |
nms_threshold: 0.4 | |
keep_top_k: 750 | |
score_threshold: 0.6 | |
prior_box: | |
_target_: facetorch.analyzer.detector.post.PriorBox | |
min_sizes: | |
- - 16 | |
- 32 | |
- - 64 | |
- 128 | |
- - 256 | |
- 512 | |
steps: | |
- 8 | |
- 16 | |
- 32 | |
clip: false | |
variance: | |
- 0.1 | |
- 0.2 | |
reverse_colors: ${analyzer.detector.reverse_colors} | |
expand_pixels: 4 | |
unifier: | |
_target_: facetorch.analyzer.unifier.FaceUnifier | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Normalize | |
mean: | |
- -123.0 | |
- -117.0 | |
- -104.0 | |
std: | |
- 255.0 | |
- 255.0 | |
- 255.0 | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 380 | |
- 380 | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: ${analyzer.optimize_transforms} | |
predictor: | |
embed: | |
_target_: facetorch.analyzer.predictor.FacePredictor | |
downloader: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 19h3kqar1wlELAmM5hDyj9tlrUh8yjrCl | |
path_local: /opt/facetorch/models/torchscript/predictor/embed/1/model.pt | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
preprocessor: | |
_target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 244 | |
- 244 | |
- _target_: torchvision.transforms.Normalize | |
mean: | |
- 0.485 | |
- 0.456 | |
- 0.406 | |
std: | |
- 0.228 | |
- 0.224 | |
- 0.225 | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.fer.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
reverse_colors: false | |
postprocessor: | |
_target_: facetorch.analyzer.predictor.post.PostEmbedder | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.fer.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
labels: | |
- abstract | |
verify: | |
_target_: facetorch.analyzer.predictor.FacePredictor | |
downloader: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 1H-aPtFd9C5D7y1vzoWsObKAxeIBE9QPd | |
path_local: /opt/facetorch/models/torchscript/predictor/verify/1/model.pt | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
preprocessor: | |
_target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 112 | |
- 112 | |
- _target_: torchvision.transforms.Normalize | |
mean: | |
- 0.485 | |
- 0.456 | |
- 0.406 | |
std: | |
- 0.229 | |
- 0.224 | |
- 0.225 | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.fer.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
reverse_colors: false | |
postprocessor: | |
_target_: facetorch.analyzer.predictor.post.PostEmbedder | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.fer.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
labels: | |
- abstract | |
fer: | |
_target_: facetorch.analyzer.predictor.FacePredictor | |
downloader: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 1xoB5VYOd0XLjb-rQqqHWCkQvma4NytEd | |
path_local: /opt/facetorch/models/torchscript/predictor/fer/2/model.pt | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
preprocessor: | |
_target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 260 | |
- 260 | |
- _target_: torchvision.transforms.Normalize | |
mean: | |
- 0.485 | |
- 0.456 | |
- 0.406 | |
std: | |
- 0.229 | |
- 0.224 | |
- 0.225 | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.fer.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
reverse_colors: false | |
postprocessor: | |
_target_: facetorch.analyzer.predictor.post.PostArgMax | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.fer.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
dim: 1 | |
labels: | |
- Anger | |
- Contempt | |
- Disgust | |
- Fear | |
- Happiness | |
- Neutral | |
- Sadness | |
- Surprise | |
deepfake: | |
_target_: facetorch.analyzer.predictor.FacePredictor | |
downloader: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 1GjDTwQpvrkCjXOdiBy1oMkzm7nt-bXFg | |
path_local: /opt/facetorch/models/torchscript/predictor/deepfake/1/model.pt | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
preprocessor: | |
_target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 380 | |
- 380 | |
- _target_: torchvision.transforms.Normalize | |
mean: | |
- 0.485 | |
- 0.456 | |
- 0.406 | |
std: | |
- 0.229 | |
- 0.224 | |
- 0.225 | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: ${analyzer.optimize_transforms} | |
reverse_colors: false | |
postprocessor: | |
_target_: facetorch.analyzer.predictor.post.PostSigmoidBinary | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: ${analyzer.optimize_transforms} | |
labels: | |
- Real | |
- Fake | |
threshold: 0.7 | |
align: | |
_target_: facetorch.analyzer.predictor.FacePredictor | |
downloader: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 16gNFQdEH2nWvW3zTbdIAniKIbPAp6qBA | |
path_local: /opt/facetorch/models/torchscript/predictor/align/1/model.pt | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
preprocessor: | |
_target_: facetorch.analyzer.predictor.pre.PredictorPreProcessor | |
transform: | |
_target_: torchvision.transforms.Compose | |
transforms: | |
- _target_: torchvision.transforms.Resize | |
size: | |
- 120 | |
- 120 | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.align.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
reverse_colors: false | |
postprocessor: | |
_target_: facetorch.analyzer.predictor.post.PostEmbedder | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.predictor.align.device.type} | |
optimize_transform: ${analyzer.optimize_transforms} | |
labels: | |
- abstract | |
utilizer: | |
align: | |
_target_: facetorch.analyzer.utilizer.align.Lmk3DMeshPose | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: false | |
downloader_meta: | |
_target_: facetorch.downloader.DownloaderGDrive | |
file_id: 11tdAcFuSXqCCf58g52WT1Rpa8KuQwe2o | |
path_local: /opt/facetorch/data/3dmm/meta.pt | |
image_size: 120 | |
draw_boxes: | |
_target_: facetorch.analyzer.utilizer.draw.BoxDrawer | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: false | |
color: green | |
line_width: 3 | |
draw_landmarks: | |
_target_: facetorch.analyzer.utilizer.draw.LandmarkDrawer | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: false | |
marker: o | |
markersize: 0.9 | |
opacity: 0.8 | |
line_width: 0.8 | |
color: g | |
markeredgecolor: green | |
save: | |
_target_: facetorch.analyzer.utilizer.save.ImageSaver | |
transform: None | |
device: | |
_target_: torch.device | |
type: ${analyzer.device} | |
optimize_transform: false | |
logger: | |
_target_: facetorch.logger.LoggerJsonFile | |
name: facetorch | |
level: 20 | |
path_file: /opt/logs/facetorch/main.log | |
json_format: '%(asctime)s %(levelname)s %(message)s' | |
main: | |
sleep: 3 | |
debug: true | |
path_image: /opt/facetorch/data/input/test.jpg | |
batch_size: 8 | |
fix_img_size: true | |
return_img_data: false | |
include_tensors: false | |
path_output: /opt/facetorch/data/output/test.png | |
path_image_2: /opt/facetorch/data/input/test3.jpg | |