|
import os |
|
import argparse |
|
|
|
import onnxruntime as ort |
|
from utils import * |
|
|
|
|
|
CFG = { |
|
"name": "mobilenet0.25", |
|
"min_sizes": [[16, 32], [64, 128], [256, 512]], |
|
"steps": [8, 16, 32], |
|
"variance": [0.1, 0.2], |
|
"clip": False, |
|
} |
|
INPUT_SIZE = [608, 640] |
|
DEVICE = torch.device("cpu") |
|
|
|
|
|
def vis(img_raw, dets, vis_thres): |
|
"""Visualization original image |
|
Args: |
|
img_raw: origin image |
|
dets: detections |
|
vis_thres: visualization threshold |
|
Returns: |
|
visualization results |
|
""" |
|
for b in dets: |
|
if b[4] < vis_thres: |
|
continue |
|
text = "{:.4f}".format(b[4]) |
|
b = list(map(int, b)) |
|
cv2.rectangle(img_raw, (b[0], b[1]), (b[2], b[3]), (0, 0, 255), 2) |
|
cx = b[0] |
|
cy = b[1] + 12 |
|
cv2.putText(img_raw, text, (cx, cy), cv2.FONT_HERSHEY_DUPLEX, 0.5, (255, 255, 255),) |
|
|
|
|
|
cv2.circle(img_raw, (b[5], b[6]), 1, (0, 0, 255), 4) |
|
cv2.circle(img_raw, (b[7], b[8]), 1, (0, 255, 255), 4) |
|
cv2.circle(img_raw, (b[9], b[10]), 1, (255, 0, 255), 4) |
|
cv2.circle(img_raw, (b[11], b[12]), 1, (0, 255, 0), 4) |
|
cv2.circle(img_raw, (b[13], b[14]), 1, (255, 0, 0), 4) |
|
|
|
if not os.path.exists("./results/"): |
|
os.makedirs("./results/") |
|
name = "./results/" + 'result' + ".jpg" |
|
cv2.imwrite(name, img_raw) |
|
|
|
|
|
def Retinaface_inference(run_ort, args): |
|
"""Infer an image with onnx seession |
|
Args: |
|
run_ort: Onnx session |
|
args: including image path and hyperparameters |
|
Returns: boxes_list, confidence_list, landm_list |
|
boxes_list = [[left, top, right, bottom]...] |
|
confidence_list = [[confidence]...] |
|
landm_list = [[landms(dim=10)]...] |
|
""" |
|
img_raw = cv2.imread(args.image_path, cv2.IMREAD_COLOR) |
|
|
|
img, scale, resize = preprocess(img_raw, INPUT_SIZE, DEVICE) |
|
|
|
outputs = run_ort.run(None, {run_ort.get_inputs()[0].name: img}) |
|
|
|
dets = postprocess(CFG, img, outputs, scale, resize, args.confidence_threshold, args.nms_threshold, DEVICE) |
|
|
|
|
|
boxes = dets[:, :4] |
|
confidences = dets[:, 4:5] |
|
landms = dets[:, 5:] |
|
boxes_list = [box.tolist() for box in boxes] |
|
confidence_list = [confidence.tolist() for confidence in confidences] |
|
landm_list = [landm.tolist() for landm in landms] |
|
|
|
|
|
if args.save_image: |
|
vis(img_raw, dets, args.vis_thres) |
|
|
|
return boxes_list, confidence_list, landm_list |
|
|
|
|
|
if __name__ == '__main__': |
|
parser = argparse.ArgumentParser(description="Retinaface") |
|
parser.add_argument( |
|
"-m", |
|
"--trained_model", |
|
default="./weights/RetinaFace_int.onnx", |
|
type=str, |
|
help="Trained state_dict file path to open", |
|
) |
|
parser.add_argument( |
|
"--image_path", |
|
default="./data/widerface/val/images/18--Concerts/18_Concerts_Concerts_18_38.jpg", |
|
type=str, |
|
help="image path", |
|
) |
|
parser.add_argument( |
|
"--confidence_threshold", |
|
default=0.4, |
|
type=float, |
|
help="confidence_threshold" |
|
) |
|
parser.add_argument( |
|
"--nms_threshold", |
|
default=0.4, |
|
type=float, |
|
help="nms_threshold" |
|
) |
|
parser.add_argument( |
|
"-s", |
|
"--save_image", |
|
action="store_true", |
|
default=False, |
|
help="show detection results", |
|
) |
|
parser.add_argument( |
|
"--vis_thres", |
|
default=0.5, |
|
type=float, |
|
help="visualization_threshold" |
|
) |
|
parser.add_argument( |
|
"--ipu", |
|
action="store_true", |
|
help="Use IPU for inference.", |
|
) |
|
parser.add_argument( |
|
"--provider_config", |
|
type=str, |
|
default="vaip_config.json", |
|
help="Path of the config file for seting provider_options.", |
|
) |
|
|
|
args = parser.parse_args() |
|
|
|
if args.ipu: |
|
providers = ["VitisAIExecutionProvider"] |
|
provider_options = [{"config_file": args.provider_config}] |
|
else: |
|
providers = ['CUDAExecutionProvider', 'CPUExecutionProvider'] |
|
provider_options = None |
|
|
|
print("Loading pretrained model from {}".format(args.trained_model)) |
|
run_ort = ort.InferenceSession(args.trained_model, providers=providers, provider_options=provider_options) |
|
|
|
boxes_list, confidence_list, landm_list = Retinaface_inference(run_ort, args) |
|
print('inference done!') |
|
print(boxes_list, confidence_list, landm_list) |
|
|