evp / depth /utils_depth /metrics.py
nick_93
init
bcec54e
raw
history blame
2.95 kB
# ------------------------------------------------------------------------------
# The code is from GLPDepth (https://github.com/vinvino02/GLPDepth).
# For non-commercial purpose only (research, evaluation etc).
# ------------------------------------------------------------------------------
import torch
def eval_depth(pred, target):
assert pred.shape == target.shape
thresh = torch.max((target / pred), (pred / target))
d1 = torch.sum(thresh < 1.25).float() / len(thresh)
d2 = torch.sum(thresh < 1.25 ** 2).float() / len(thresh)
d3 = torch.sum(thresh < 1.25 ** 3).float() / len(thresh)
diff = pred - target
diff_log = torch.log(pred) - torch.log(target)
abs_rel = torch.mean(torch.abs(diff) / target)
sq_rel = torch.mean(torch.pow(diff, 2) / target)
rmse = torch.sqrt(torch.mean(torch.pow(diff, 2)))
rmse_log = torch.sqrt(torch.mean(torch.pow(diff_log , 2)))
log10 = torch.mean(torch.abs(torch.log10(pred) - torch.log10(target)))
silog = torch.sqrt(torch.pow(diff_log, 2).mean() - 0.5 * torch.pow(diff_log.mean(), 2))
return {'d1': d1.item(), 'd2': d2.item(), 'd3': d3.item(), 'abs_rel': abs_rel.item(),
'sq_rel': sq_rel.item(), 'rmse': rmse.item(), 'rmse_log': rmse_log.item(),
'log10':log10.item(), 'silog':silog.item()}
def cropping_img(args, pred, gt_depth):
min_depth_eval = args.min_depth_eval
max_depth_eval = args.max_depth_eval
pred[torch.isinf(pred)] = max_depth_eval
pred[torch.isnan(pred)] = min_depth_eval
valid_mask = torch.logical_and(
gt_depth > min_depth_eval, gt_depth < max_depth_eval)
if args.dataset == 'kitti':
if args.do_kb_crop:
height, width = gt_depth.shape
top_margin = int(height - 352)
left_margin = int((width - 1216) / 2)
gt_depth = gt_depth[top_margin:top_margin +
352, left_margin:left_margin + 1216]
if args.kitti_crop:
gt_height, gt_width = gt_depth.shape
eval_mask = torch.zeros(valid_mask.shape).to(
device=valid_mask.device)
if args.kitti_crop == 'garg_crop':
eval_mask[int(0.40810811 * gt_height):int(0.99189189 * gt_height),
int(0.03594771 * gt_width):int(0.96405229 * gt_width)] = 1
elif args.kitti_crop == 'eigen_crop':
eval_mask[int(0.3324324 * gt_height):int(0.91351351 * gt_height),
int(0.0359477 * gt_width):int(0.96405229 * gt_width)] = 1
else:
eval_mask = valid_mask
elif args.dataset == 'nyudepthv2':
eval_mask = torch.zeros(valid_mask.shape).to(device=valid_mask.device)
eval_mask[45:471, 41:601] = 1
else:
eval_mask = valid_mask
valid_mask = torch.logical_and(valid_mask, eval_mask)
return pred[valid_mask], gt_depth[valid_mask]