Vincentqyw
fix: roma
c74a070
raw
history blame
1.75 kB
# Copyright 2019-present NAVER Corp.
# CC BY-NC-SA 3.0
# Available only for non-commercial use
import pdb
import torch
import torch.nn as nn
import torch.nn.functional as F
from nets.sampler import *
from nets.repeatability_loss import *
from nets.reliability_loss import *
class MultiLoss(nn.Module):
"""Combines several loss functions for convenience.
*args: [loss weight (float), loss creator, ... ]
Example:
loss = MultiLoss( 1, MyFirstLoss(), 0.5, MySecondLoss() )
"""
def __init__(self, *args, dbg=()):
nn.Module.__init__(self)
assert len(args) % 2 == 0, "args must be a list of (float, loss)"
self.weights = []
self.losses = nn.ModuleList()
for i in range(len(args) // 2):
weight = float(args[2 * i + 0])
loss = args[2 * i + 1]
assert isinstance(loss, nn.Module), "%s is not a loss!" % loss
self.weights.append(weight)
self.losses.append(loss)
def forward(self, select=None, **variables):
assert not select or all(1 <= n <= len(self.losses) for n in select)
d = dict()
cum_loss = 0
for num, (weight, loss_func) in enumerate(zip(self.weights, self.losses), 1):
if select is not None and num not in select:
continue
l = loss_func(**{k: v for k, v in variables.items()})
if isinstance(l, tuple):
assert len(l) == 2 and isinstance(l[1], dict)
else:
l = l, {loss_func.name: l}
cum_loss = cum_loss + weight * l[0]
for key, val in l[1].items():
d["loss_" + key] = float(val)
d["loss"] = float(cum_loss)
return cum_loss, d