mafqoud / modules /streaming.py
Dev-mohamed's picture
init
677c57e
raw
history blame
No virus
31.4 kB
# built-in dependencies
import os
import time
from typing import List, Tuple, Optional
# 3rd party dependencies
import numpy as np
import pandas as pd
import cv2
# project dependencies
from deepface import DeepFace
from deepface.commons import logger as log
logger = log.get_singletonish_logger()
# dependency configuration
os.environ["TF_CPP_MIN_LOG_LEVEL"] = "2"
IDENTIFIED_IMG_SIZE = 112
TEXT_COLOR = (255, 255, 255)
def analysis(
db_path: str,
model_name="VGG-Face",
detector_backend="opencv",
distance_metric="cosine",
enable_face_analysis=True,
source=0,
time_threshold=5,
frame_threshold=5,
):
"""
Run real time face recognition and facial attribute analysis
Args:
db_path (string): Path to the folder containing image files. All detected faces
in the database will be considered in the decision-making process.
model_name (str): Model for face recognition. Options: VGG-Face, Facenet, Facenet512,
OpenFace, DeepFace, DeepID, Dlib, ArcFace, SFace and GhostFaceNet (default is VGG-Face).
detector_backend (string): face detector backend. Options: 'opencv', 'retinaface',
'mtcnn', 'ssd', 'dlib', 'mediapipe', 'yolov8', 'centerface' or 'skip'
(default is opencv).
distance_metric (string): Metric for measuring similarity. Options: 'cosine',
'euclidean', 'euclidean_l2' (default is cosine).
enable_face_analysis (bool): Flag to enable face analysis (default is True).
source (Any): The source for the video stream (default is 0, which represents the
default camera).
time_threshold (int): The time threshold (in seconds) for face recognition (default is 5).
frame_threshold (int): The frame threshold for face recognition (default is 5).
Returns:
None
"""
# initialize models
build_demography_models(enable_face_analysis=enable_face_analysis)
build_facial_recognition_model(model_name=model_name)
# call a dummy find function for db_path once to create embeddings before starting webcam
_ = search_identity(
detected_face=np.zeros([224, 224, 3]),
db_path=db_path,
detector_backend=detector_backend,
distance_metric=distance_metric,
model_name=model_name,
)
freezed_img = None
freeze = False
num_frames_with_faces = 0
tic = time.time()
cap = cv2.VideoCapture(source) # webcam
while True:
has_frame, img = cap.read()
if not has_frame:
break
# we are adding some figures into img such as identified facial image, age, gender
# that is why, we need raw image itself to make analysis
raw_img = img.copy()
faces_coordinates = []
if freeze is False:
faces_coordinates = grab_facial_areas(img=img, detector_backend=detector_backend)
# we will pass img to analyze modules (identity, demography) and add some illustrations
# that is why, we will not be able to extract detected face from img clearly
detected_faces = extract_facial_areas(img=img, faces_coordinates=faces_coordinates)
img = highlight_facial_areas(img=img, faces_coordinates=faces_coordinates)
img = countdown_to_freeze(
img=img,
faces_coordinates=faces_coordinates,
frame_threshold=frame_threshold,
num_frames_with_faces=num_frames_with_faces,
)
num_frames_with_faces = num_frames_with_faces + 1 if len(faces_coordinates) else 0
freeze = num_frames_with_faces > 0 and num_frames_with_faces % frame_threshold == 0
if freeze:
# add analyze results into img - derive from raw_img
img = highlight_facial_areas(img=raw_img, faces_coordinates=faces_coordinates)
# age, gender and emotion analysis
img = perform_demography_analysis(
enable_face_analysis=enable_face_analysis,
img=raw_img,
faces_coordinates=faces_coordinates,
detected_faces=detected_faces,
)
# facial recogntion analysis
img = perform_facial_recognition(
img=img,
faces_coordinates=faces_coordinates,
detected_faces=detected_faces,
db_path=db_path,
detector_backend=detector_backend,
distance_metric=distance_metric,
model_name=model_name,
)
# freeze the img after analysis
freezed_img = img.copy()
# start counter for freezing
tic = time.time()
logger.info("freezed")
elif freeze is True and time.time() - tic > time_threshold:
freeze = False
freezed_img = None
# reset counter for freezing
tic = time.time()
logger.info("freeze released")
freezed_img = countdown_to_release(img=freezed_img, tic=tic, time_threshold=time_threshold)
cv2.imshow("img", img if freezed_img is None else freezed_img)
if cv2.waitKey(1) & 0xFF == ord("q"): # press q to quit
break
# kill open cv things
cap.release()
cv2.destroyAllWindows()
def build_facial_recognition_model(model_name: str) -> None:
"""
Build facial recognition model
Args:
model_name (str): Model for face recognition. Options: VGG-Face, Facenet, Facenet512,
OpenFace, DeepFace, DeepID, Dlib, ArcFace, SFace and GhostFaceNet (default is VGG-Face).
Returns
input_shape (tuple): input shape of given facial recognitio n model.
"""
_ = DeepFace.build_model(model_name=model_name)
logger.info(f"{model_name} is built")
def search_identity(
detected_face: np.ndarray,
db_path: str,
model_name: str,
detector_backend: str,
distance_metric: str,
) -> Tuple[Optional[str], Optional[np.ndarray]]:
"""
Search an identity in facial database.
Args:
detected_face (np.ndarray): extracted individual facial image
db_path (string): Path to the folder containing image files. All detected faces
in the database will be considered in the decision-making process.
model_name (str): Model for face recognition. Options: VGG-Face, Facenet, Facenet512,
OpenFace, DeepFace, DeepID, Dlib, ArcFace, SFace and GhostFaceNet (default is VGG-Face).
detector_backend (string): face detector backend. Options: 'opencv', 'retinaface',
'mtcnn', 'ssd', 'dlib', 'mediapipe', 'yolov8', 'centerface' or 'skip'
(default is opencv).
distance_metric (string): Metric for measuring similarity. Options: 'cosine',
'euclidean', 'euclidean_l2' (default is cosine).
Returns:
result (tuple): result consisting of following objects
identified image path (str)
identified image itself (np.ndarray)
"""
target_path = None
try:
dfs = DeepFace.find(
img_path=detected_face,
db_path=db_path,
model_name=model_name,
detector_backend=detector_backend,
distance_metric=distance_metric,
enforce_detection=False,
silent=True,
)
except ValueError as err:
if f"No item found in {db_path}" in str(err):
logger.warn(
f"No item is found in {db_path}."
"So, no facial recognition analysis will be performed."
)
dfs = []
else:
raise err
if len(dfs) == 0:
# you may consider to return unknown person's image here
return None, None
# detected face is coming from parent, safe to access 1st index
df = dfs[0]
if df.shape[0] == 0:
return None, None
candidate = df.iloc[0]
target_path = candidate["identity"]
logger.info(f"Hello, {target_path}")
# load found identity image - extracted if possible
target_objs = DeepFace.extract_faces(
img_path=target_path,
detector_backend=detector_backend,
enforce_detection=False,
align=True,
)
# extract facial area of the identified image if and only if it has one face
# otherwise, show image as is
if len(target_objs) == 1:
# extract 1st item directly
target_obj = target_objs[0]
target_img = target_obj["face"]
target_img = cv2.resize(target_img, (IDENTIFIED_IMG_SIZE, IDENTIFIED_IMG_SIZE))
target_img *= 255
target_img = target_img[:, :, ::-1]
else:
target_img = cv2.imread(target_path)
return target_path.split("/")[-1], target_img
def build_demography_models(enable_face_analysis: bool) -> None:
"""
Build demography analysis models
Args:
enable_face_analysis (bool): Flag to enable face analysis (default is True).
Returns:
None
"""
if enable_face_analysis is False:
return
DeepFace.build_model(model_name="Age")
logger.info("Age model is just built")
DeepFace.build_model(model_name="Gender")
logger.info("Gender model is just built")
DeepFace.build_model(model_name="Emotion")
logger.info("Emotion model is just built")
def highlight_facial_areas(
img: np.ndarray, faces_coordinates: List[Tuple[int, int, int, int]]
) -> np.ndarray:
"""
Highlight detected faces with rectangles in the given image
Args:
img (np.ndarray): image itself
faces_coordinates (list): list of face coordinates as tuple with x, y, w and h
Returns:
img (np.ndarray): image with highlighted facial areas
"""
for x, y, w, h in faces_coordinates:
# highlight facial area with rectangle
cv2.rectangle(img, (x, y), (x + w, y + h), (67, 67, 67), 1)
return img
def countdown_to_freeze(
img: np.ndarray,
faces_coordinates: List[Tuple[int, int, int, int]],
frame_threshold: int,
num_frames_with_faces: int,
) -> np.ndarray:
"""
Highlight time to freeze in the image's facial areas
Args:
img (np.ndarray): image itself
faces_coordinates (list): list of face coordinates as tuple with x, y, w and h
frame_threshold (int): how many sequantial frames required with face(s) to freeze
num_frames_with_faces (int): how many sequantial frames do we have with face(s)
Returns:
img (np.ndarray): image with counter values
"""
for x, y, w, h in faces_coordinates:
cv2.putText(
img,
str(frame_threshold - (num_frames_with_faces % frame_threshold)),
(int(x + w / 4), int(y + h / 1.5)),
cv2.FONT_HERSHEY_SIMPLEX,
4,
(255, 255, 255),
2,
)
return img
def countdown_to_release(
img: Optional[np.ndarray], tic: float, time_threshold: int
) -> Optional[np.ndarray]:
"""
Highlight time to release the freezing in the image top left area
Args:
img (np.ndarray): image itself
tic (float): time specifying when freezing started
time_threshold (int): freeze time threshold
Returns:
img (np.ndarray): image with time to release the freezing
"""
# do not take any action if it is not frozen yet
if img is None:
return img
toc = time.time()
time_left = int(time_threshold - (toc - tic) + 1)
cv2.rectangle(img, (10, 10), (90, 50), (67, 67, 67), -10)
cv2.putText(
img,
str(time_left),
(40, 40),
cv2.FONT_HERSHEY_SIMPLEX,
1,
(255, 255, 255),
1,
)
return img
def grab_facial_areas(
img: np.ndarray, detector_backend: str, threshold: int = 130
) -> List[Tuple[int, int, int, int]]:
"""
Find facial area coordinates in the given image
Args:
img (np.ndarray): image itself
detector_backend (string): face detector backend. Options: 'opencv', 'retinaface',
'mtcnn', 'ssd', 'dlib', 'mediapipe', 'yolov8', 'centerface' or 'skip'
(default is opencv).
threshold (int): threshold for facial area, discard smaller ones
Returns
result (list): list of tuple with x, y, w and h coordinates
"""
try:
face_objs = DeepFace.extract_faces(
img_path=img,
detector_backend=detector_backend,
# you may consider to extract with larger expanding value
expand_percentage=0,
)
faces = [
(
face_obj["facial_area"]["x"],
face_obj["facial_area"]["y"],
face_obj["facial_area"]["w"],
face_obj["facial_area"]["h"],
)
for face_obj in face_objs
if face_obj["facial_area"]["w"] > threshold
]
return faces
except: # to avoid exception if no face detected
return []
def extract_facial_areas(
img: np.ndarray, faces_coordinates: List[Tuple[int, int, int, int]]
) -> List[np.ndarray]:
"""
Extract facial areas as numpy array from given image
Args:
img (np.ndarray): image itself
faces_coordinates (list): list of facial area coordinates as tuple with
x, y, w and h values
Returns:
detected_faces (list): list of detected facial area images
"""
detected_faces = []
for x, y, w, h in faces_coordinates:
detected_face = img[int(y) : int(y + h), int(x) : int(x + w)]
detected_faces.append(detected_face)
return detected_faces
def perform_facial_recognition(
img: np.ndarray,
detected_faces: List[np.ndarray],
faces_coordinates: List[Tuple[int, int, int, int]],
db_path: str,
detector_backend: str,
distance_metric: str,
model_name: str,
) -> np.ndarray:
"""
Perform facial recognition
Args:
img (np.ndarray): image itself
detected_faces (list): list of extracted detected face images as numpy
faces_coordinates (list): list of facial area coordinates as tuple with
x, y, w and h values
db_path (string): Path to the folder containing image files. All detected faces
in the database will be considered in the decision-making process.
detector_backend (string): face detector backend. Options: 'opencv', 'retinaface',
'mtcnn', 'ssd', 'dlib', 'mediapipe', 'yolov8', 'centerface' or 'skip'
(default is opencv).
distance_metric (string): Metric for measuring similarity. Options: 'cosine',
'euclidean', 'euclidean_l2' (default is cosine).
model_name (str): Model for face recognition. Options: VGG-Face, Facenet, Facenet512,
OpenFace, DeepFace, DeepID, Dlib, ArcFace, SFace and GhostFaceNet (default is VGG-Face).
Returns:
img (np.ndarray): image with identified face informations
"""
for idx, (x, y, w, h) in enumerate(faces_coordinates):
detected_face = detected_faces[idx]
target_label, target_img = search_identity(
detected_face=detected_face,
db_path=db_path,
detector_backend=detector_backend,
distance_metric=distance_metric,
model_name=model_name,
)
if target_label is None:
continue
img = overlay_identified_face(
img=img,
target_img=target_img,
label=target_label,
x=x,
y=y,
w=w,
h=h,
)
return img
def perform_demography_analysis(
enable_face_analysis: bool,
img: np.ndarray,
faces_coordinates: List[Tuple[int, int, int, int]],
detected_faces: List[np.ndarray],
) -> np.ndarray:
"""
Perform demography analysis on given image
Args:
enable_face_analysis (bool): Flag to enable face analysis.
img (np.ndarray): image itself
faces_coordinates (list): list of face coordinates as tuple with
x, y, w and h values
detected_faces (list): list of extracted detected face images as numpy
Returns:
img (np.ndarray): image with analyzed demography information
"""
if enable_face_analysis is False:
return img
for idx, (x, y, w, h) in enumerate(faces_coordinates):
detected_face = detected_faces[idx]
demographies = DeepFace.analyze(
img_path=detected_face,
actions=("age", "gender", "emotion"),
detector_backend="skip",
enforce_detection=False,
silent=True,
)
if len(demographies) == 0:
continue
# safe to access 1st index because detector backend is skip
demography = demographies[0]
img = overlay_emotion(img=img, emotion_probas=demography["emotion"], x=x, y=y, w=w, h=h)
img = overlay_age_gender(
img=img,
apparent_age=demography["age"],
gender=demography["dominant_gender"][0:1], # M or W
x=x,
y=y,
w=w,
h=h,
)
return img
def overlay_identified_face(
img: np.ndarray,
target_img: np.ndarray,
label: str,
x: int,
y: int,
w: int,
h: int,
) -> np.ndarray:
"""
Overlay the identified face onto image itself
Args:
img (np.ndarray): image itself
target_img (np.ndarray): identified face's image
label (str): name of the identified face
x (int): x coordinate of the face on the given image
y (int): y coordinate of the face on the given image
w (int): w coordinate of the face on the given image
h (int): h coordinate of the face on the given image
Returns:
img (np.ndarray): image with overlayed identity
"""
try:
if y - IDENTIFIED_IMG_SIZE > 0 and x + w + IDENTIFIED_IMG_SIZE < img.shape[1]:
# top right
img[
y - IDENTIFIED_IMG_SIZE : y,
x + w : x + w + IDENTIFIED_IMG_SIZE,
] = target_img
overlay = img.copy()
opacity = 0.4
cv2.rectangle(
img,
(x + w, y),
(x + w + IDENTIFIED_IMG_SIZE, y + 20),
(46, 200, 255),
cv2.FILLED,
)
cv2.addWeighted(
overlay,
opacity,
img,
1 - opacity,
0,
img,
)
cv2.putText(
img,
label,
(x + w, y + 10),
cv2.FONT_HERSHEY_SIMPLEX,
0.5,
TEXT_COLOR,
1,
)
# connect face and text
cv2.line(
img,
(x + int(w / 2), y),
(x + 3 * int(w / 4), y - int(IDENTIFIED_IMG_SIZE / 2)),
(67, 67, 67),
1,
)
cv2.line(
img,
(x + 3 * int(w / 4), y - int(IDENTIFIED_IMG_SIZE / 2)),
(x + w, y - int(IDENTIFIED_IMG_SIZE / 2)),
(67, 67, 67),
1,
)
elif y + h + IDENTIFIED_IMG_SIZE < img.shape[0] and x - IDENTIFIED_IMG_SIZE > 0:
# bottom left
img[
y + h : y + h + IDENTIFIED_IMG_SIZE,
x - IDENTIFIED_IMG_SIZE : x,
] = target_img
overlay = img.copy()
opacity = 0.4
cv2.rectangle(
img,
(x - IDENTIFIED_IMG_SIZE, y + h - 20),
(x, y + h),
(46, 200, 255),
cv2.FILLED,
)
cv2.addWeighted(
overlay,
opacity,
img,
1 - opacity,
0,
img,
)
cv2.putText(
img,
label,
(x - IDENTIFIED_IMG_SIZE, y + h - 10),
cv2.FONT_HERSHEY_SIMPLEX,
0.5,
TEXT_COLOR,
1,
)
# connect face and text
cv2.line(
img,
(x + int(w / 2), y + h),
(
x + int(w / 2) - int(w / 4),
y + h + int(IDENTIFIED_IMG_SIZE / 2),
),
(67, 67, 67),
1,
)
cv2.line(
img,
(
x + int(w / 2) - int(w / 4),
y + h + int(IDENTIFIED_IMG_SIZE / 2),
),
(x, y + h + int(IDENTIFIED_IMG_SIZE / 2)),
(67, 67, 67),
1,
)
elif y - IDENTIFIED_IMG_SIZE > 0 and x - IDENTIFIED_IMG_SIZE > 0:
# top left
img[y - IDENTIFIED_IMG_SIZE : y, x - IDENTIFIED_IMG_SIZE : x] = target_img
overlay = img.copy()
opacity = 0.4
cv2.rectangle(
img,
(x - IDENTIFIED_IMG_SIZE, y),
(x, y + 20),
(46, 200, 255),
cv2.FILLED,
)
cv2.addWeighted(
overlay,
opacity,
img,
1 - opacity,
0,
img,
)
cv2.putText(
img,
label,
(x - IDENTIFIED_IMG_SIZE, y + 10),
cv2.FONT_HERSHEY_SIMPLEX,
0.5,
TEXT_COLOR,
1,
)
# connect face and text
cv2.line(
img,
(x + int(w / 2), y),
(
x + int(w / 2) - int(w / 4),
y - int(IDENTIFIED_IMG_SIZE / 2),
),
(67, 67, 67),
1,
)
cv2.line(
img,
(
x + int(w / 2) - int(w / 4),
y - int(IDENTIFIED_IMG_SIZE / 2),
),
(x, y - int(IDENTIFIED_IMG_SIZE / 2)),
(67, 67, 67),
1,
)
elif (
x + w + IDENTIFIED_IMG_SIZE < img.shape[1]
and y + h + IDENTIFIED_IMG_SIZE < img.shape[0]
):
# bottom righ
img[
y + h : y + h + IDENTIFIED_IMG_SIZE,
x + w : x + w + IDENTIFIED_IMG_SIZE,
] = target_img
overlay = img.copy()
opacity = 0.4
cv2.rectangle(
img,
(x + w, y + h - 20),
(x + w + IDENTIFIED_IMG_SIZE, y + h),
(46, 200, 255),
cv2.FILLED,
)
cv2.addWeighted(
overlay,
opacity,
img,
1 - opacity,
0,
img,
)
cv2.putText(
img,
label,
(x + w, y + h - 10),
cv2.FONT_HERSHEY_SIMPLEX,
0.5,
TEXT_COLOR,
1,
)
# connect face and text
cv2.line(
img,
(x + int(w / 2), y + h),
(
x + int(w / 2) + int(w / 4),
y + h + int(IDENTIFIED_IMG_SIZE / 2),
),
(67, 67, 67),
1,
)
cv2.line(
img,
(
x + int(w / 2) + int(w / 4),
y + h + int(IDENTIFIED_IMG_SIZE / 2),
),
(x + w, y + h + int(IDENTIFIED_IMG_SIZE / 2)),
(67, 67, 67),
1,
)
else:
logger.info("cannot put facial recognition info on the image")
except Exception as err: # pylint: disable=broad-except
logger.error(str(err))
return img
def overlay_emotion(
img: np.ndarray, emotion_probas: dict, x: int, y: int, w: int, h: int
) -> np.ndarray:
"""
Overlay the analyzed emotion of face onto image itself
Args:
img (np.ndarray): image itself
emotion_probas (dict): probability of different emotionas dictionary
x (int): x coordinate of the face on the given image
y (int): y coordinate of the face on the given image
w (int): w coordinate of the face on the given image
h (int): h coordinate of the face on the given image
Returns:
img (np.ndarray): image with overlay emotion analsis results
"""
emotion_df = pd.DataFrame(emotion_probas.items(), columns=["emotion", "score"])
emotion_df = emotion_df.sort_values(by=["score"], ascending=False).reset_index(drop=True)
# background of mood box
# transparency
overlay = img.copy()
opacity = 0.4
# put gray background to the right of the detected image
if x + w + IDENTIFIED_IMG_SIZE < img.shape[1]:
cv2.rectangle(
img,
(x + w, y),
(x + w + IDENTIFIED_IMG_SIZE, y + h),
(64, 64, 64),
cv2.FILLED,
)
cv2.addWeighted(overlay, opacity, img, 1 - opacity, 0, img)
# put gray background to the left of the detected image
elif x - IDENTIFIED_IMG_SIZE > 0:
cv2.rectangle(
img,
(x - IDENTIFIED_IMG_SIZE, y),
(x, y + h),
(64, 64, 64),
cv2.FILLED,
)
cv2.addWeighted(overlay, opacity, img, 1 - opacity, 0, img)
for index, instance in emotion_df.iterrows():
current_emotion = instance["emotion"]
emotion_label = f"{current_emotion} "
emotion_score = instance["score"] / 100
filled_bar_x = 35 # this is the size if an emotion is 100%
bar_x = int(filled_bar_x * emotion_score)
if x + w + IDENTIFIED_IMG_SIZE < img.shape[1]:
text_location_y = y + 20 + (index + 1) * 20
text_location_x = x + w
if text_location_y < y + h:
cv2.putText(
img,
emotion_label,
(text_location_x, text_location_y),
cv2.FONT_HERSHEY_SIMPLEX,
0.5,
(255, 255, 255),
1,
)
cv2.rectangle(
img,
(x + w + 70, y + 13 + (index + 1) * 20),
(
x + w + 70 + bar_x,
y + 13 + (index + 1) * 20 + 5,
),
(255, 255, 255),
cv2.FILLED,
)
elif x - IDENTIFIED_IMG_SIZE > 0:
text_location_y = y + 20 + (index + 1) * 20
text_location_x = x - IDENTIFIED_IMG_SIZE
if text_location_y <= y + h:
cv2.putText(
img,
emotion_label,
(text_location_x, text_location_y),
cv2.FONT_HERSHEY_SIMPLEX,
0.5,
(255, 255, 255),
1,
)
cv2.rectangle(
img,
(
x - IDENTIFIED_IMG_SIZE + 70,
y + 13 + (index + 1) * 20,
),
(
x - IDENTIFIED_IMG_SIZE + 70 + bar_x,
y + 13 + (index + 1) * 20 + 5,
),
(255, 255, 255),
cv2.FILLED,
)
return img
def overlay_age_gender(
img: np.ndarray, apparent_age: float, gender: str, x: int, y: int, w: int, h: int
) -> np.ndarray:
"""
Overlay the analyzed age and gender of face onto image itself
Args:
img (np.ndarray): image itself
apparent_age (float): analyzed apparent age
gender (str): analyzed gender
x (int): x coordinate of the face on the given image
y (int): y coordinate of the face on the given image
w (int): w coordinate of the face on the given image
h (int): h coordinate of the face on the given image
Returns:
img (np.ndarray): image with overlay age and gender analsis results
"""
logger.debug(f"{apparent_age} years old {gender}")
analysis_report = f"{int(apparent_age)} {gender}"
info_box_color = (46, 200, 255)
# show its age and gender on the top of the image
if y - IDENTIFIED_IMG_SIZE + int(IDENTIFIED_IMG_SIZE / 5) > 0:
triangle_coordinates = np.array(
[
(x + int(w / 2), y),
(
x + int(w / 2) - int(w / 10),
y - int(IDENTIFIED_IMG_SIZE / 3),
),
(
x + int(w / 2) + int(w / 10),
y - int(IDENTIFIED_IMG_SIZE / 3),
),
]
)
cv2.drawContours(
img,
[triangle_coordinates],
0,
info_box_color,
-1,
)
cv2.rectangle(
img,
(
x + int(w / 5),
y - IDENTIFIED_IMG_SIZE + int(IDENTIFIED_IMG_SIZE / 5),
),
(x + w - int(w / 5), y - int(IDENTIFIED_IMG_SIZE / 3)),
info_box_color,
cv2.FILLED,
)
cv2.putText(
img,
analysis_report,
(x + int(w / 3.5), y - int(IDENTIFIED_IMG_SIZE / 2.1)),
cv2.FONT_HERSHEY_SIMPLEX,
1,
(0, 111, 255),
2,
)
# show its age and gender on the top of the image
elif y + h + IDENTIFIED_IMG_SIZE - int(IDENTIFIED_IMG_SIZE / 5) < img.shape[0]:
triangle_coordinates = np.array(
[
(x + int(w / 2), y + h),
(
x + int(w / 2) - int(w / 10),
y + h + int(IDENTIFIED_IMG_SIZE / 3),
),
(
x + int(w / 2) + int(w / 10),
y + h + int(IDENTIFIED_IMG_SIZE / 3),
),
]
)
cv2.drawContours(
img,
[triangle_coordinates],
0,
info_box_color,
-1,
)
cv2.rectangle(
img,
(x + int(w / 5), y + h + int(IDENTIFIED_IMG_SIZE / 3)),
(
x + w - int(w / 5),
y + h + IDENTIFIED_IMG_SIZE - int(IDENTIFIED_IMG_SIZE / 5),
),
info_box_color,
cv2.FILLED,
)
cv2.putText(
img,
analysis_report,
(x + int(w / 3.5), y + h + int(IDENTIFIED_IMG_SIZE / 1.5)),
cv2.FONT_HERSHEY_SIMPLEX,
1,
(0, 111, 255),
2,
)
return img