|
"""
|
|
WiderFace evaluation code
|
|
author: wondervictor
|
|
mail: tianhengcheng@gmail.com
|
|
copyright@wondervictor
|
|
"""
|
|
|
|
import os
|
|
import tqdm
|
|
import pickle
|
|
import argparse
|
|
import numpy as np
|
|
from scipy.io import loadmat
|
|
from bbox import bbox_overlaps
|
|
from IPython import embed
|
|
|
|
|
|
def get_gt_boxes(gt_dir):
|
|
""" gt dir: (wider_face_val.mat, wider_easy_val.mat, wider_medium_val.mat, wider_hard_val.mat)"""
|
|
|
|
gt_mat = loadmat(os.path.join(gt_dir, 'wider_face_val.mat'))
|
|
hard_mat = loadmat(os.path.join(gt_dir, 'wider_hard_val.mat'))
|
|
medium_mat = loadmat(os.path.join(gt_dir, 'wider_medium_val.mat'))
|
|
easy_mat = loadmat(os.path.join(gt_dir, 'wider_easy_val.mat'))
|
|
|
|
facebox_list = gt_mat['face_bbx_list']
|
|
event_list = gt_mat['event_list']
|
|
file_list = gt_mat['file_list']
|
|
|
|
hard_gt_list = hard_mat['gt_list']
|
|
medium_gt_list = medium_mat['gt_list']
|
|
easy_gt_list = easy_mat['gt_list']
|
|
|
|
return facebox_list, event_list, file_list, hard_gt_list, medium_gt_list, easy_gt_list
|
|
|
|
|
|
def get_gt_boxes_from_txt(gt_path, cache_dir):
|
|
|
|
cache_file = os.path.join(cache_dir, 'gt_cache.pkl')
|
|
if os.path.exists(cache_file):
|
|
f = open(cache_file, 'rb')
|
|
boxes = pickle.load(f)
|
|
f.close()
|
|
return boxes
|
|
|
|
f = open(gt_path, 'r')
|
|
state = 0
|
|
lines = f.readlines()
|
|
lines = list(map(lambda x: x.rstrip('\r\n'), lines))
|
|
boxes = {}
|
|
print(len(lines))
|
|
f.close()
|
|
current_boxes = []
|
|
current_name = None
|
|
for line in lines:
|
|
if state == 0 and '--' in line:
|
|
state = 1
|
|
current_name = line
|
|
continue
|
|
if state == 1:
|
|
state = 2
|
|
continue
|
|
|
|
if state == 2 and '--' in line:
|
|
state = 1
|
|
boxes[current_name] = np.array(current_boxes).astype('float32')
|
|
current_name = line
|
|
current_boxes = []
|
|
continue
|
|
|
|
if state == 2:
|
|
box = [float(x) for x in line.split(' ')[:4]]
|
|
current_boxes.append(box)
|
|
continue
|
|
|
|
f = open(cache_file, 'wb')
|
|
pickle.dump(boxes, f)
|
|
f.close()
|
|
return boxes
|
|
|
|
|
|
def read_pred_file(filepath):
|
|
|
|
with open(filepath, 'r') as f:
|
|
lines = f.readlines()
|
|
img_file = lines[0].rstrip('\n\r')
|
|
lines = lines[2:]
|
|
|
|
|
|
|
|
|
|
boxes = []
|
|
for line in lines:
|
|
line = line.rstrip('\r\n').split(' ')
|
|
if line[0] is '':
|
|
continue
|
|
|
|
boxes.append([float(line[0]), float(line[1]), float(line[2]), float(line[3]), float(line[4])])
|
|
boxes = np.array(boxes)
|
|
|
|
return img_file.split('/')[-1], boxes
|
|
|
|
|
|
def get_preds(pred_dir):
|
|
events = os.listdir(pred_dir)
|
|
boxes = dict()
|
|
pbar = tqdm.tqdm(events)
|
|
|
|
for event in pbar:
|
|
pbar.set_description('Reading Predictions ')
|
|
event_dir = os.path.join(pred_dir, event)
|
|
event_images = os.listdir(event_dir)
|
|
current_event = dict()
|
|
for imgtxt in event_images:
|
|
imgname, _boxes = read_pred_file(os.path.join(event_dir, imgtxt))
|
|
current_event[imgname.rstrip('.jpg')] = _boxes
|
|
boxes[event] = current_event
|
|
return boxes
|
|
|
|
|
|
def norm_score(pred):
|
|
""" norm score
|
|
pred {key: [[x1,y1,x2,y2,s]]}
|
|
"""
|
|
|
|
max_score = 0
|
|
min_score = 1
|
|
|
|
for _, k in pred.items():
|
|
for _, v in k.items():
|
|
if len(v) == 0:
|
|
continue
|
|
_min = np.min(v[:, -1])
|
|
_max = np.max(v[:, -1])
|
|
max_score = max(_max, max_score)
|
|
min_score = min(_min, min_score)
|
|
|
|
diff = max_score - min_score
|
|
for _, k in pred.items():
|
|
for _, v in k.items():
|
|
if len(v) == 0:
|
|
continue
|
|
v[:, -1] = (v[:, -1] - min_score)/diff
|
|
|
|
|
|
def image_eval(pred, gt, ignore, iou_thresh):
|
|
""" single image evaluation
|
|
pred: Nx5
|
|
gt: Nx4
|
|
ignore:
|
|
"""
|
|
|
|
_pred = pred.copy()
|
|
_gt = gt.copy()
|
|
pred_recall = np.zeros(_pred.shape[0])
|
|
recall_list = np.zeros(_gt.shape[0])
|
|
proposal_list = np.ones(_pred.shape[0])
|
|
|
|
_pred[:, 2] = _pred[:, 2] + _pred[:, 0]
|
|
_pred[:, 3] = _pred[:, 3] + _pred[:, 1]
|
|
_gt[:, 2] = _gt[:, 2] + _gt[:, 0]
|
|
_gt[:, 3] = _gt[:, 3] + _gt[:, 1]
|
|
|
|
overlaps = bbox_overlaps(_pred[:, :4], _gt)
|
|
|
|
for h in range(_pred.shape[0]):
|
|
|
|
gt_overlap = overlaps[h]
|
|
max_overlap, max_idx = gt_overlap.max(), gt_overlap.argmax()
|
|
if max_overlap >= iou_thresh:
|
|
if ignore[max_idx] == 0:
|
|
recall_list[max_idx] = -1
|
|
proposal_list[h] = -1
|
|
elif recall_list[max_idx] == 0:
|
|
recall_list[max_idx] = 1
|
|
|
|
r_keep_index = np.where(recall_list == 1)[0]
|
|
pred_recall[h] = len(r_keep_index)
|
|
return pred_recall, proposal_list
|
|
|
|
|
|
def img_pr_info(thresh_num, pred_info, proposal_list, pred_recall):
|
|
pr_info = np.zeros((thresh_num, 2)).astype('float')
|
|
for t in range(thresh_num):
|
|
|
|
thresh = 1 - (t+1)/thresh_num
|
|
r_index = np.where(pred_info[:, 4] >= thresh)[0]
|
|
if len(r_index) == 0:
|
|
pr_info[t, 0] = 0
|
|
pr_info[t, 1] = 0
|
|
else:
|
|
r_index = r_index[-1]
|
|
p_index = np.where(proposal_list[:r_index+1] == 1)[0]
|
|
pr_info[t, 0] = len(p_index)
|
|
pr_info[t, 1] = pred_recall[r_index]
|
|
return pr_info
|
|
|
|
|
|
def dataset_pr_info(thresh_num, pr_curve, count_face):
|
|
_pr_curve = np.zeros((thresh_num, 2))
|
|
for i in range(thresh_num):
|
|
_pr_curve[i, 0] = pr_curve[i, 1] / pr_curve[i, 0]
|
|
_pr_curve[i, 1] = pr_curve[i, 1] / count_face
|
|
return _pr_curve
|
|
|
|
|
|
def voc_ap(rec, prec):
|
|
|
|
|
|
|
|
mrec = np.concatenate(([0.], rec, [1.]))
|
|
mpre = np.concatenate(([0.], prec, [0.]))
|
|
|
|
|
|
for i in range(mpre.size - 1, 0, -1):
|
|
mpre[i - 1] = np.maximum(mpre[i - 1], mpre[i])
|
|
|
|
|
|
|
|
i = np.where(mrec[1:] != mrec[:-1])[0]
|
|
|
|
|
|
ap = np.sum((mrec[i + 1] - mrec[i]) * mpre[i + 1])
|
|
return ap
|
|
|
|
|
|
def evaluation(pred, gt_path, iou_thresh=0.5):
|
|
pred = get_preds(pred)
|
|
norm_score(pred)
|
|
facebox_list, event_list, file_list, hard_gt_list, medium_gt_list, easy_gt_list = get_gt_boxes(gt_path)
|
|
event_num = len(event_list)
|
|
thresh_num = 1000
|
|
settings = ['easy', 'medium', 'hard']
|
|
setting_gts = [easy_gt_list, medium_gt_list, hard_gt_list]
|
|
aps = []
|
|
for setting_id in range(3):
|
|
|
|
gt_list = setting_gts[setting_id]
|
|
count_face = 0
|
|
pr_curve = np.zeros((thresh_num, 2)).astype('float')
|
|
|
|
pbar = tqdm.tqdm(range(event_num))
|
|
for i in pbar:
|
|
pbar.set_description('Processing {}'.format(settings[setting_id]))
|
|
event_name = str(event_list[i][0][0])
|
|
img_list = file_list[i][0]
|
|
pred_list = pred[event_name]
|
|
sub_gt_list = gt_list[i][0]
|
|
|
|
gt_bbx_list = facebox_list[i][0]
|
|
|
|
for j in range(len(img_list)):
|
|
pred_info = pred_list[str(img_list[j][0][0])]
|
|
|
|
gt_boxes = gt_bbx_list[j][0].astype('float')
|
|
keep_index = sub_gt_list[j][0]
|
|
count_face += len(keep_index)
|
|
|
|
if len(gt_boxes) == 0 or len(pred_info) == 0:
|
|
continue
|
|
ignore = np.zeros(gt_boxes.shape[0])
|
|
if len(keep_index) != 0:
|
|
ignore[keep_index-1] = 1
|
|
pred_recall, proposal_list = image_eval(pred_info, gt_boxes, ignore, iou_thresh)
|
|
|
|
_img_pr_info = img_pr_info(thresh_num, pred_info, proposal_list, pred_recall)
|
|
|
|
pr_curve += _img_pr_info
|
|
pr_curve = dataset_pr_info(thresh_num, pr_curve, count_face)
|
|
|
|
propose = pr_curve[:, 0]
|
|
recall = pr_curve[:, 1]
|
|
|
|
ap = voc_ap(recall, propose)
|
|
aps.append(ap)
|
|
|
|
print("==================== Results ====================")
|
|
print("Easy Val AP: {}".format(aps[0]))
|
|
print("Medium Val AP: {}".format(aps[1]))
|
|
print("Hard Val AP: {}".format(aps[2]))
|
|
print("=================================================")
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument('-p', '--pred', default="./widerface_txt/")
|
|
parser.add_argument('-g', '--gt', default='./ground_truth/')
|
|
|
|
args = parser.parse_args()
|
|
evaluation(args.pred, args.gt)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|