ibug-fpage / app.py
hysts's picture
hysts HF staff
Update
6c048e6
#!/usr/bin/env python
from __future__ import annotations
import os
import pathlib
import sys
import cv2
import gradio as gr
import huggingface_hub
import numpy as np
import spaces
import torch
from huggingface_hub import hf_hub_download
torch.jit.script = lambda f: f
sys.path.insert(0, "face_detection")
sys.path.insert(0, "face_parsing")
sys.path.insert(0, "fpage")
sys.path.insert(0, "roi_tanh_warping")
from ibug.age_estimation import AgeEstimator
from ibug.face_detection import RetinaFacePredictor
from ibug.face_parsing.utils import label_colormap
DESCRIPTION = "# [FP-Age](https://github.com/ibug-group/fpage)"
def is_lfs_pointer_file(path: pathlib.Path) -> bool:
try:
with open(path, "r") as f:
# Git LFS pointer files usually start with version line
version_line = f.readline()
if version_line.startswith("version https://git-lfs.github.com/spec/"):
# Check for the presence of oid and size lines
oid_line = f.readline()
size_line = f.readline()
if oid_line.startswith("oid sha256:") and size_line.startswith("size "):
return True
except Exception as e:
print(f"Error reading file {path}: {e}")
return False
lfs_model_paths = sorted(pathlib.Path("face_parsing").rglob("*.torch"))
for lfs_model_path in lfs_model_paths:
if is_lfs_pointer_file(lfs_model_path):
os.remove(lfs_model_path)
out_path = hf_hub_download(
"public-data/ibug-face-parsing",
filename=lfs_model_path.name,
repo_type="model",
subfolder=lfs_model_path.parts[-3],
)
os.symlink(out_path, lfs_model_path)
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
detector = RetinaFacePredictor(threshold=0.8, device="cpu", model=RetinaFacePredictor.get_model("mobilenet0.25"))
detector.device = device
detector.net.to(device)
model = AgeEstimator(
device=device,
ckpt=huggingface_hub.hf_hub_download("hysts/ibug", "fpage/models/fpage-resnet50-fcn-14-97.torch"),
encoder="resnet50",
decoder="fcn",
age_classes=97,
face_classes=14,
)
@spaces.GPU
def predict(image: np.ndarray, max_num_faces: int) -> np.ndarray:
colormap = label_colormap(14)
# RGB -> BGR
image = image[:, :, ::-1]
faces = detector(image, rgb=False)
if len(faces) == 0:
raise RuntimeError("No face was found.")
faces = sorted(list(faces), key=lambda x: -x[4])[:max_num_faces][::-1]
ages, masks = model.predict_img(image, faces, rgb=False)
mask_image = np.zeros_like(image)
for mask in masks:
temp = colormap[mask]
mask_image[temp > 0] = temp[temp > 0]
res = image.astype(float) * 0.5 + mask_image[:, :, ::-1] * 0.5
res = np.clip(np.round(res), 0, 255).astype(np.uint8)
for face, age in zip(faces, ages):
bbox = np.round(face[:4]).astype(int)
cv2.rectangle(
res,
(bbox[0], bbox[1]),
(bbox[2], bbox[3]),
color=(0, 255, 0),
thickness=2,
)
text_content = f"Age: ({age: .1f})"
cv2.putText(
res,
text_content,
(bbox[0], bbox[1] - 10),
cv2.FONT_HERSHEY_DUPLEX,
1,
(255, 255, 255),
lineType=cv2.LINE_AA,
)
return res[:, :, ::-1]
with gr.Blocks(css="style.css") as demo:
gr.Markdown(DESCRIPTION)
with gr.Row():
with gr.Column():
image = gr.Image(label="Input")
max_num_faces = gr.Slider(label="Max Number of Faces", minimum=1, maximum=20, step=1, value=5)
run_button = gr.Button()
with gr.Column():
result = gr.Image(label="Output")
gr.Examples(
examples=[[path.as_posix(), 5] for path in sorted(pathlib.Path("images").rglob("*.jpg"))],
inputs=[image, max_num_faces],
outputs=result,
fn=predict,
)
run_button.click(
fn=predict,
inputs=[image, max_num_faces],
outputs=result,
api_name="predict",
)
if __name__ == "__main__":
demo.queue(max_size=20).launch()