github-actions[bot]
Sync to HuggingFace Spaces
123489f
import torch
import torch.nn as nn
import torch.nn.functional as F
from collections import defaultdict
def dice_loss(input_mask, cls_gt):
num_objects = input_mask.shape[1]
losses = []
for i in range(num_objects):
mask = input_mask[:, i].flatten(start_dim=1)
# background not in mask, so we add one to cls_gt
gt = (cls_gt == (i + 1)).float().flatten(start_dim=1)
numerator = 2 * (mask * gt).sum(-1)
denominator = mask.sum(-1) + gt.sum(-1)
loss = 1 - (numerator + 1) / (denominator + 1)
losses.append(loss)
return torch.cat(losses).mean()
# https://stackoverflow.com/questions/63735255/how-do-i-compute-bootstrapped-cross-entropy-loss-in-pytorch
class BootstrappedCE(nn.Module):
def __init__(self, start_warm, end_warm, top_p=0.15):
super().__init__()
self.start_warm = start_warm
self.end_warm = end_warm
self.top_p = top_p
def forward(self, input, target, it):
if it < self.start_warm:
return F.cross_entropy(input, target), 1.0
raw_loss = F.cross_entropy(input, target, reduction="none").view(-1)
num_pixels = raw_loss.numel()
if it > self.end_warm:
this_p = self.top_p
else:
this_p = self.top_p + (1 - self.top_p) * (
(self.end_warm - it) / (self.end_warm - self.start_warm)
)
loss, _ = torch.topk(raw_loss, int(num_pixels * this_p), sorted=False)
return loss.mean(), this_p
class LossComputer:
def __init__(self, config):
super().__init__()
self.config = config
self.bce = BootstrappedCE(config["start_warm"], config["end_warm"])
def compute(self, data, num_objects, it):
losses = defaultdict(int)
b, t = data["rgb"].shape[:2]
losses["total_loss"] = 0
for ti in range(1, t):
for bi in range(b):
loss, p = self.bce(
data[f"logits_{ti}"][bi : bi + 1, : num_objects[bi] + 1],
data["cls_gt"][bi : bi + 1, ti, 0],
it,
)
losses["p"] += p / b / (t - 1)
losses[f"ce_loss_{ti}"] += loss / b
losses["total_loss"] += losses["ce_loss_%d" % ti]
losses[f"dice_loss_{ti}"] = dice_loss(
data[f"masks_{ti}"], data["cls_gt"][:, ti, 0]
)
losses["total_loss"] += losses[f"dice_loss_{ti}"]
return losses