| | import math |
| | import numpy as np |
| | import cv2 |
| | from typing import List, Tuple, Union |
| |
|
| | from .body import BodyResult, Keypoint |
| |
|
| | eps = 0.01 |
| |
|
| |
|
| | def smart_resize(x, s): |
| | Ht, Wt = s |
| | if x.ndim == 2: |
| | Ho, Wo = x.shape |
| | Co = 1 |
| | else: |
| | Ho, Wo, Co = x.shape |
| | if Co == 3 or Co == 1: |
| | k = float(Ht + Wt) / float(Ho + Wo) |
| | return cv2.resize(x, (int(Wt), int(Ht)), interpolation=cv2.INTER_AREA if k < 1 else cv2.INTER_LANCZOS4) |
| | else: |
| | return np.stack([smart_resize(x[:, :, i], s) for i in range(Co)], axis=2) |
| |
|
| |
|
| | def smart_resize_k(x, fx, fy): |
| | if x.ndim == 2: |
| | Ho, Wo = x.shape |
| | Co = 1 |
| | else: |
| | Ho, Wo, Co = x.shape |
| | Ht, Wt = Ho * fy, Wo * fx |
| | if Co == 3 or Co == 1: |
| | k = float(Ht + Wt) / float(Ho + Wo) |
| | return cv2.resize(x, (int(Wt), int(Ht)), interpolation=cv2.INTER_AREA if k < 1 else cv2.INTER_LANCZOS4) |
| | else: |
| | return np.stack([smart_resize_k(x[:, :, i], fx, fy) for i in range(Co)], axis=2) |
| |
|
| |
|
| | def padRightDownCorner(img, stride, padValue): |
| | h = img.shape[0] |
| | w = img.shape[1] |
| |
|
| | pad = 4 * [None] |
| | pad[0] = 0 |
| | pad[1] = 0 |
| | pad[2] = 0 if (h % stride == 0) else stride - (h % stride) |
| | pad[3] = 0 if (w % stride == 0) else stride - (w % stride) |
| |
|
| | img_padded = img |
| | pad_up = np.tile(img_padded[0:1, :, :]*0 + padValue, (pad[0], 1, 1)) |
| | img_padded = np.concatenate((pad_up, img_padded), axis=0) |
| | pad_left = np.tile(img_padded[:, 0:1, :]*0 + padValue, (1, pad[1], 1)) |
| | img_padded = np.concatenate((pad_left, img_padded), axis=1) |
| | pad_down = np.tile(img_padded[-2:-1, :, :]*0 + padValue, (pad[2], 1, 1)) |
| | img_padded = np.concatenate((img_padded, pad_down), axis=0) |
| | pad_right = np.tile(img_padded[:, -2:-1, :]*0 + padValue, (1, pad[3], 1)) |
| | img_padded = np.concatenate((img_padded, pad_right), axis=1) |
| |
|
| | return img_padded, pad |
| |
|
| |
|
| | def transfer(model, model_weights): |
| | transfered_model_weights = {} |
| | for weights_name in model.state_dict().keys(): |
| | transfered_model_weights[weights_name] = model_weights['.'.join(weights_name.split('.')[1:])] |
| | return transfered_model_weights |
| |
|
| |
|
| | def draw_bodypose(canvas: np.ndarray, keypoints: List[Keypoint]) -> np.ndarray: |
| | """ |
| | Draw keypoints and limbs representing body pose on a given canvas. |
| | |
| | Args: |
| | canvas (np.ndarray): A 3D numpy array representing the canvas (image) on which to draw the body pose. |
| | keypoints (List[Keypoint]): A list of Keypoint objects representing the body keypoints to be drawn. |
| | |
| | Returns: |
| | np.ndarray: A 3D numpy array representing the modified canvas with the drawn body pose. |
| | |
| | Note: |
| | The function expects the x and y coordinates of the keypoints to be normalized between 0 and 1. |
| | """ |
| | H, W, C = canvas.shape |
| |
|
| | |
| | if max(W, H) < 500: |
| | ratio = 1.0 |
| | elif max(W, H) >= 500 and max(W, H) < 1000: |
| | ratio = 2.0 |
| | elif max(W, H) >= 1000 and max(W, H) < 2000: |
| | ratio = 3.0 |
| | elif max(W, H) >= 2000 and max(W, H) < 3000: |
| | ratio = 4.0 |
| | elif max(W, H) >= 3000 and max(W, H) < 4000: |
| | ratio = 5.0 |
| | elif max(W, H) >= 4000 and max(W, H) < 5000: |
| | ratio = 6.0 |
| | else: |
| | ratio = 7.0 |
| |
|
| | stickwidth = 4 |
| |
|
| | limbSeq = [ |
| | [2, 3], [2, 6], [3, 4], [4, 5], |
| | [6, 7], [7, 8], [2, 9], [9, 10], |
| | [10, 11], [2, 12], [12, 13], [13, 14], |
| | [2, 1], [1, 15], [15, 17], [1, 16], |
| | [16, 18], |
| | ] |
| |
|
| | colors = [[255, 0, 0], [255, 85, 0], [255, 170, 0], [255, 255, 0], [170, 255, 0], [85, 255, 0], [0, 255, 0], \ |
| | [0, 255, 85], [0, 255, 170], [0, 255, 255], [0, 170, 255], [0, 85, 255], [0, 0, 255], [85, 0, 255], \ |
| | [170, 0, 255], [255, 0, 255], [255, 0, 170], [255, 0, 85]] |
| |
|
| | for (k1_index, k2_index), color in zip(limbSeq, colors): |
| | keypoint1 = keypoints[k1_index - 1] |
| | keypoint2 = keypoints[k2_index - 1] |
| |
|
| | if keypoint1 is None or keypoint2 is None: |
| | continue |
| |
|
| | Y = np.array([keypoint1.x, keypoint2.x]) * float(W) |
| | X = np.array([keypoint1.y, keypoint2.y]) * float(H) |
| | mX = np.mean(X) |
| | mY = np.mean(Y) |
| | length = ((X[0] - X[1]) ** 2 + (Y[0] - Y[1]) ** 2) ** 0.5 |
| | angle = math.degrees(math.atan2(X[0] - X[1], Y[0] - Y[1])) |
| | polygon = cv2.ellipse2Poly((int(mY), int(mX)), (int(length / 2), int(stickwidth * ratio)), int(angle), 0, 360, 1) |
| | cv2.fillConvexPoly(canvas, polygon, [int(float(c) * 0.6) for c in color]) |
| |
|
| | for keypoint, color in zip(keypoints, colors): |
| | if keypoint is None: |
| | continue |
| |
|
| | x, y = keypoint.x, keypoint.y |
| | x = int(x * W) |
| | y = int(y * H) |
| | cv2.circle(canvas, (int(x), int(y)), int(4 * ratio), color, thickness=-1) |
| |
|
| | return canvas |
| |
|
| |
|
| |
|
| | def draw_handpose(canvas: np.ndarray, keypoints: Union[List[Keypoint], None]) -> np.ndarray: |
| | import matplotlib |
| | """ |
| | Draw keypoints and connections representing hand pose on a given canvas. |
| | |
| | Args: |
| | canvas (np.ndarray): A 3D numpy array representing the canvas (image) on which to draw the hand pose. |
| | keypoints (List[Keypoint]| None): A list of Keypoint objects representing the hand keypoints to be drawn |
| | or None if no keypoints are present. |
| | |
| | Returns: |
| | np.ndarray: A 3D numpy array representing the modified canvas with the drawn hand pose. |
| | |
| | Note: |
| | The function expects the x and y coordinates of the keypoints to be normalized between 0 and 1. |
| | """ |
| | if not keypoints: |
| | return canvas |
| | |
| | H, W, C = canvas.shape |
| |
|
| | edges = [[0, 1], [1, 2], [2, 3], [3, 4], [0, 5], [5, 6], [6, 7], [7, 8], [0, 9], [9, 10], \ |
| | [10, 11], [11, 12], [0, 13], [13, 14], [14, 15], [15, 16], [0, 17], [17, 18], [18, 19], [19, 20]] |
| |
|
| | for ie, (e1, e2) in enumerate(edges): |
| | k1 = keypoints[e1] |
| | k2 = keypoints[e2] |
| | if k1 is None or k2 is None: |
| | continue |
| | |
| | x1 = int(k1.x * W) |
| | y1 = int(k1.y * H) |
| | x2 = int(k2.x * W) |
| | y2 = int(k2.y * H) |
| | if x1 > eps and y1 > eps and x2 > eps and y2 > eps: |
| | cv2.line(canvas, (x1, y1), (x2, y2), matplotlib.colors.hsv_to_rgb([ie / float(len(edges)), 1.0, 1.0]) * 255, thickness=2) |
| |
|
| | for keypoint in keypoints: |
| | x, y = keypoint.x, keypoint.y |
| | x = int(x * W) |
| | y = int(y * H) |
| | if x > eps and y > eps: |
| | cv2.circle(canvas, (x, y), 4, (0, 0, 255), thickness=-1) |
| | return canvas |
| |
|
| |
|
| | def draw_facepose(canvas: np.ndarray, keypoints: Union[List[Keypoint], None]) -> np.ndarray: |
| | """ |
| | Draw keypoints representing face pose on a given canvas. |
| | |
| | Args: |
| | canvas (np.ndarray): A 3D numpy array representing the canvas (image) on which to draw the face pose. |
| | keypoints (List[Keypoint]| None): A list of Keypoint objects representing the face keypoints to be drawn |
| | or None if no keypoints are present. |
| | |
| | Returns: |
| | np.ndarray: A 3D numpy array representing the modified canvas with the drawn face pose. |
| | |
| | Note: |
| | The function expects the x and y coordinates of the keypoints to be normalized between 0 and 1. |
| | """ |
| | if not keypoints: |
| | return canvas |
| | |
| | H, W, C = canvas.shape |
| | for keypoint in keypoints: |
| | x, y = keypoint.x, keypoint.y |
| | x = int(x * W) |
| | y = int(y * H) |
| | if x > eps and y > eps: |
| | cv2.circle(canvas, (x, y), 3, (255, 255, 255), thickness=-1) |
| | return canvas |
| |
|
| |
|
| | |
| | |
| | def handDetect(body: BodyResult, oriImg) -> List[Tuple[int, int, int, bool]]: |
| | """ |
| | Detect hands in the input body pose keypoints and calculate the bounding box for each hand. |
| | |
| | Args: |
| | body (BodyResult): A BodyResult object containing the detected body pose keypoints. |
| | oriImg (numpy.ndarray): A 3D numpy array representing the original input image. |
| | |
| | Returns: |
| | List[Tuple[int, int, int, bool]]: A list of tuples, each containing the coordinates (x, y) of the top-left |
| | corner of the bounding box, the width (height) of the bounding box, and |
| | a boolean flag indicating whether the hand is a left hand (True) or a |
| | right hand (False). |
| | |
| | Notes: |
| | - The width and height of the bounding boxes are equal since the network requires squared input. |
| | - The minimum bounding box size is 20 pixels. |
| | """ |
| | ratioWristElbow = 0.33 |
| | detect_result = [] |
| | image_height, image_width = oriImg.shape[0:2] |
| | |
| | keypoints = body.keypoints |
| | |
| | |
| | left_shoulder = keypoints[5] |
| | left_elbow = keypoints[6] |
| | left_wrist = keypoints[7] |
| | right_shoulder = keypoints[2] |
| | right_elbow = keypoints[3] |
| | right_wrist = keypoints[4] |
| |
|
| | |
| | has_left = all(keypoint is not None for keypoint in (left_shoulder, left_elbow, left_wrist)) |
| | has_right = all(keypoint is not None for keypoint in (right_shoulder, right_elbow, right_wrist)) |
| | if not (has_left or has_right): |
| | return [] |
| | |
| | hands = [] |
| | |
| | if has_left: |
| | hands.append([ |
| | left_shoulder.x, left_shoulder.y, |
| | left_elbow.x, left_elbow.y, |
| | left_wrist.x, left_wrist.y, |
| | True |
| | ]) |
| | |
| | if has_right: |
| | hands.append([ |
| | right_shoulder.x, right_shoulder.y, |
| | right_elbow.x, right_elbow.y, |
| | right_wrist.x, right_wrist.y, |
| | False |
| | ]) |
| |
|
| | for x1, y1, x2, y2, x3, y3, is_left in hands: |
| | |
| | |
| | |
| | |
| | |
| | |
| | x = x3 + ratioWristElbow * (x3 - x2) |
| | y = y3 + ratioWristElbow * (y3 - y2) |
| | distanceWristElbow = math.sqrt((x3 - x2) ** 2 + (y3 - y2) ** 2) |
| | distanceElbowShoulder = math.sqrt((x2 - x1) ** 2 + (y2 - y1) ** 2) |
| | width = 1.5 * max(distanceWristElbow, 0.9 * distanceElbowShoulder) |
| | |
| | |
| | |
| | x -= width / 2 |
| | y -= width / 2 |
| | |
| | if x < 0: x = 0 |
| | if y < 0: y = 0 |
| | width1 = width |
| | width2 = width |
| | if x + width > image_width: width1 = image_width - x |
| | if y + width > image_height: width2 = image_height - y |
| | width = min(width1, width2) |
| | |
| | if width >= 20: |
| | detect_result.append((int(x), int(y), int(width), is_left)) |
| |
|
| | ''' |
| | return value: [[x, y, w, True if left hand else False]]. |
| | width=height since the network require squared input. |
| | x, y is the coordinate of top left |
| | ''' |
| | return detect_result |
| |
|
| |
|
| | |
| | def faceDetect(body: BodyResult, oriImg) -> Union[Tuple[int, int, int], None]: |
| | """ |
| | Detect the face in the input body pose keypoints and calculate the bounding box for the face. |
| | |
| | Args: |
| | body (BodyResult): A BodyResult object containing the detected body pose keypoints. |
| | oriImg (numpy.ndarray): A 3D numpy array representing the original input image. |
| | |
| | Returns: |
| | Tuple[int, int, int] | None: A tuple containing the coordinates (x, y) of the top-left corner of the |
| | bounding box and the width (height) of the bounding box, or None if the |
| | face is not detected or the bounding box width is less than 20 pixels. |
| | |
| | Notes: |
| | - The width and height of the bounding box are equal. |
| | - The minimum bounding box size is 20 pixels. |
| | """ |
| | |
| | image_height, image_width = oriImg.shape[0:2] |
| | |
| | keypoints = body.keypoints |
| | head = keypoints[0] |
| | left_eye = keypoints[14] |
| | right_eye = keypoints[15] |
| | left_ear = keypoints[16] |
| | right_ear = keypoints[17] |
| | |
| | if head is None or all(keypoint is None for keypoint in (left_eye, right_eye, left_ear, right_ear)): |
| | return None |
| |
|
| | width = 0.0 |
| | x0, y0 = head.x, head.y |
| |
|
| | if left_eye is not None: |
| | x1, y1 = left_eye.x, left_eye.y |
| | d = max(abs(x0 - x1), abs(y0 - y1)) |
| | width = max(width, d * 3.0) |
| |
|
| | if right_eye is not None: |
| | x1, y1 = right_eye.x, right_eye.y |
| | d = max(abs(x0 - x1), abs(y0 - y1)) |
| | width = max(width, d * 3.0) |
| |
|
| | if left_ear is not None: |
| | x1, y1 = left_ear.x, left_ear.y |
| | d = max(abs(x0 - x1), abs(y0 - y1)) |
| | width = max(width, d * 1.5) |
| |
|
| | if right_ear is not None: |
| | x1, y1 = right_ear.x, right_ear.y |
| | d = max(abs(x0 - x1), abs(y0 - y1)) |
| | width = max(width, d * 1.5) |
| |
|
| | x, y = x0, y0 |
| |
|
| | x -= width |
| | y -= width |
| |
|
| | if x < 0: |
| | x = 0 |
| |
|
| | if y < 0: |
| | y = 0 |
| |
|
| | width1 = width * 2 |
| | width2 = width * 2 |
| |
|
| | if x + width > image_width: |
| | width1 = image_width - x |
| |
|
| | if y + width > image_height: |
| | width2 = image_height - y |
| |
|
| | width = min(width1, width2) |
| |
|
| | if width >= 20: |
| | return int(x), int(y), int(width) |
| | else: |
| | return None |
| |
|
| |
|
| | |
| | def npmax(array): |
| | arrayindex = array.argmax(1) |
| | arrayvalue = array.max(1) |
| | i = arrayvalue.argmax() |
| | j = arrayindex[i] |
| | return i, j |
| |
|