Vincentqyw
update: features and matchers
404d2af
raw
history blame
1.72 kB
# Copyright 2019-present NAVER Corp.
# CC BY-NC-SA 3.0
# Available only for non-commercial use
import pdb
import torch
import torch.nn as nn
import torch.nn.functional as F
from nets.sampler import *
from nets.repeatability_loss import *
from nets.reliability_loss import *
class MultiLoss (nn.Module):
""" Combines several loss functions for convenience.
*args: [loss weight (float), loss creator, ... ]
Example:
loss = MultiLoss( 1, MyFirstLoss(), 0.5, MySecondLoss() )
"""
def __init__(self, *args, dbg=()):
nn.Module.__init__(self)
assert len(args) % 2 == 0, 'args must be a list of (float, loss)'
self.weights = []
self.losses = nn.ModuleList()
for i in range(len(args)//2):
weight = float(args[2*i+0])
loss = args[2*i+1]
assert isinstance(loss, nn.Module), "%s is not a loss!" % loss
self.weights.append(weight)
self.losses.append(loss)
def forward(self, select=None, **variables):
assert not select or all(1<=n<=len(self.losses) for n in select)
d = dict()
cum_loss = 0
for num, (weight, loss_func) in enumerate(zip(self.weights, self.losses),1):
if select is not None and num not in select: continue
l = loss_func(**{k:v for k,v in variables.items()})
if isinstance(l, tuple):
assert len(l) == 2 and isinstance(l[1], dict)
else:
l = l, {loss_func.name:l}
cum_loss = cum_loss + weight * l[0]
for key,val in l[1].items():
d['loss_'+key] = float(val)
d['loss'] = float(cum_loss)
return cum_loss, d