FashionGAN / netdissect /dissection.py
fiesty-bear
Initial Commit
6064c9d
raw
history blame
73.5 kB
'''
To run dissection:
1. Load up the convolutional model you wish to dissect, and wrap it in
an InstrumentedModel; then call imodel.retain_layers([layernames,..])
to instrument the layers of interest.
2. Load the segmentation dataset using the BrodenDataset class;
use the transform_image argument to normalize images to be
suitable for the model, or the size argument to truncate the dataset.
3. Choose a directory in which to write the output, and call
dissect(outdir, model, dataset).
Example:
from dissect import InstrumentedModel, dissect
from broden import BrodenDataset
model = InstrumentedModel(load_my_model())
model.eval()
model.cuda()
model.retain_layers(['conv1', 'conv2', 'conv3', 'conv4', 'conv5'])
bds = BrodenDataset('dataset/broden1_227',
transform_image=transforms.Compose([
transforms.ToTensor(),
transforms.Normalize(IMAGE_MEAN, IMAGE_STDEV)]),
size=1000)
dissect('result/dissect', model, bds,
examples_per_unit=10)
'''
import torch, numpy, os, re, json, shutil, types, tempfile, torchvision
# import warnings
# warnings.simplefilter('error', UserWarning)
from PIL import Image
from xml.etree import ElementTree as et
from collections import OrderedDict, defaultdict
from .progress import verbose_progress, default_progress, print_progress
from .progress import desc_progress
from .runningstats import RunningQuantile, RunningTopK
from .runningstats import RunningCrossCovariance, RunningConditionalQuantile
from .sampler import FixedSubsetSampler
from .actviz import activation_visualization
from .segviz import segment_visualization, high_contrast
from .workerpool import WorkerBase, WorkerPool
from .segmenter import UnifiedParsingSegmenter
def dissect(outdir, model, dataset,
segrunner=None,
train_dataset=None,
model_segmenter=None,
quantile_threshold=0.005,
iou_threshold=0.05,
iqr_threshold=0.01,
examples_per_unit=100,
batch_size=100,
num_workers=24,
seg_batch_size=5,
make_images=True,
make_labels=True,
make_maxiou=False,
make_covariance=False,
make_report=True,
make_row_images=True,
make_single_images=False,
rank_all_labels=False,
netname=None,
meta=None,
merge=None,
settings=None,
):
'''
Runs net dissection in-memory, using pytorch, and saves visualizations
and metadata into outdir.
'''
assert not model.training, 'Run model.eval() before dissection'
if netname is None:
netname = type(model).__name__
if segrunner is None:
segrunner = ClassifierSegRunner(dataset)
if train_dataset is None:
train_dataset = dataset
make_iqr = (quantile_threshold == 'iqr')
with torch.no_grad():
device = next(model.parameters()).device
levels = None
labelnames, catnames = None, None
maxioudata, iqrdata = None, None
labeldata = None
iqrdata, cov = None, None
labelnames, catnames = segrunner.get_label_and_category_names()
label_category = [catnames.index(c) if c in catnames else 0
for l, c in labelnames]
# First, always collect qunatiles and topk information.
segloader = torch.utils.data.DataLoader(dataset,
batch_size=batch_size, num_workers=num_workers,
pin_memory=(device.type == 'cuda'))
quantiles, topk = collect_quantiles_and_topk(outdir, model,
segloader, segrunner, k=examples_per_unit)
# Thresholds can be automatically chosen by maximizing iqr
if make_iqr:
# Get thresholds based on an IQR optimization
segloader = torch.utils.data.DataLoader(train_dataset,
batch_size=1, num_workers=num_workers,
pin_memory=(device.type == 'cuda'))
iqrdata = collect_iqr(outdir, model, segloader, segrunner)
max_iqr, full_iqr_levels = iqrdata[:2]
max_iqr_agreement = iqrdata[4]
# qualified_iqr[max_iqr_quantile[layer] > 0.5] = 0
levels = {layer: full_iqr_levels[layer][
max_iqr[layer].max(0)[1],
torch.arange(max_iqr[layer].shape[1])].to(device)
for layer in full_iqr_levels}
else:
levels = {k: qc.quantiles([1.0 - quantile_threshold])[:,0]
for k, qc in quantiles.items()}
quantiledata = (topk, quantiles, levels, quantile_threshold)
if make_images:
segloader = torch.utils.data.DataLoader(dataset,
batch_size=batch_size, num_workers=num_workers,
pin_memory=(device.type == 'cuda'))
generate_images(outdir, model, dataset, topk, levels, segrunner,
row_length=examples_per_unit, batch_size=seg_batch_size,
row_images=make_row_images,
single_images=make_single_images,
num_workers=num_workers)
if make_maxiou:
assert train_dataset, "Need training dataset for maxiou."
segloader = torch.utils.data.DataLoader(train_dataset,
batch_size=1, num_workers=num_workers,
pin_memory=(device.type == 'cuda'))
maxioudata = collect_maxiou(outdir, model, segloader,
segrunner)
if make_labels:
segloader = torch.utils.data.DataLoader(dataset,
batch_size=1, num_workers=num_workers,
pin_memory=(device.type == 'cuda'))
iou_scores, iqr_scores, tcs, lcs, ccs, ics = (
collect_bincounts(outdir, model, segloader,
levels, segrunner))
labeldata = (iou_scores, iqr_scores, lcs, ccs, ics, iou_threshold,
iqr_threshold)
if make_covariance:
segloader = torch.utils.data.DataLoader(dataset,
batch_size=seg_batch_size,
num_workers=num_workers,
pin_memory=(device.type == 'cuda'))
cov = collect_covariance(outdir, model, segloader, segrunner)
if make_report:
generate_report(outdir,
quantiledata=quantiledata,
labelnames=labelnames,
catnames=catnames,
labeldata=labeldata,
maxioudata=maxioudata,
iqrdata=iqrdata,
covariancedata=cov,
rank_all_labels=rank_all_labels,
netname=netname,
meta=meta,
mergedata=merge,
settings=settings)
return quantiledata, labeldata
def generate_report(outdir, quantiledata, labelnames=None, catnames=None,
labeldata=None, maxioudata=None, iqrdata=None, covariancedata=None,
rank_all_labels=False, netname='Model', meta=None, settings=None,
mergedata=None):
'''
Creates dissection.json reports and summary bargraph.svg files in the
specified output directory, and copies a dissection.html interface
to go along with it.
'''
all_layers = []
# Current source code directory, for html to copy.
srcdir = os.path.realpath(
os.path.join(os.getcwd(), os.path.dirname(__file__)))
# Unpack arguments
topk, quantiles, levels, quantile_threshold = quantiledata
top_record = dict(
netname=netname,
meta=meta,
default_ranking='unit',
quantile_threshold=quantile_threshold)
if settings is not None:
top_record['settings'] = settings
if labeldata is not None:
iou_scores, iqr_scores, lcs, ccs, ics, iou_threshold, iqr_threshold = (
labeldata)
catorder = {'object': -7, 'scene': -6, 'part': -5,
'piece': -4,
'material': -3, 'texture': -2, 'color': -1}
for i, cat in enumerate(c for c in catnames if c not in catorder):
catorder[cat] = i
catnumber = {n: i for i, n in enumerate(catnames)}
catnumber['-'] = 0
top_record['default_ranking'] = 'label'
top_record['iou_threshold'] = iou_threshold
top_record['iqr_threshold'] = iqr_threshold
labelnumber = dict((name[0], num)
for num, name in enumerate(labelnames))
# Make a segmentation color dictionary
segcolors = {}
for i, name in enumerate(labelnames):
key = ','.join(str(s) for s in high_contrast[i % len(high_contrast)])
if key in segcolors:
segcolors[key] += '/' + name[0]
else:
segcolors[key] = name[0]
top_record['segcolors'] = segcolors
for layer in topk.keys():
units, rankings = [], []
record = dict(layer=layer, units=units, rankings=rankings)
# For every unit, we always have basic visualization information.
topa, topi = topk[layer].result()
lev = levels[layer]
for u in range(len(topa)):
units.append(dict(
unit=u,
interp=True,
level=lev[u].item(),
top=[dict(imgnum=i.item(), maxact=a.item())
for i, a in zip(topi[u], topa[u])],
))
rankings.append(dict(name="unit", score=list([
u for u in range(len(topa))])))
# TODO: consider including stats and ranking based on quantiles,
# variance, connectedness here.
# if we have labeldata, then every unit also gets a bunch of other info
if labeldata is not None:
lscore, qscore, cc, ic = [dat[layer]
for dat in [iou_scores, iqr_scores, ccs, ics]]
if iqrdata is not None:
# If we have IQR thresholds, assign labels based on that
max_iqr, max_iqr_level = iqrdata[:2]
best_label = max_iqr[layer].max(0)[1]
best_score = lscore[best_label, torch.arange(lscore.shape[1])]
best_qscore = qscore[best_label, torch.arange(lscore.shape[1])]
else:
# Otherwise, assign labels based on max iou
best_score, best_label = lscore.max(0)
best_qscore = qscore[best_label, torch.arange(qscore.shape[1])]
record['iou_threshold'] = iou_threshold,
for u, urec in enumerate(units):
score, qscore, label = (
best_score[u], best_qscore[u], best_label[u])
urec.update(dict(
iou=score.item(),
iou_iqr=qscore.item(),
lc=lcs[label].item(),
cc=cc[catnumber[labelnames[label][1]], u].item(),
ic=ic[label, u].item(),
interp=(qscore.item() > iqr_threshold and
score.item() > iou_threshold),
iou_labelnum=label.item(),
iou_label=labelnames[label.item()][0],
iou_cat=labelnames[label.item()][1],
))
if maxioudata is not None:
max_iou, max_iou_level, max_iou_quantile = maxioudata
qualified_iou = max_iou[layer].clone()
# qualified_iou[max_iou_quantile[layer] > 0.75] = 0
best_score, best_label = qualified_iou.max(0)
for u, urec in enumerate(units):
urec.update(dict(
maxiou=best_score[u].item(),
maxiou_label=labelnames[best_label[u].item()][0],
maxiou_cat=labelnames[best_label[u].item()][1],
maxiou_level=max_iou_level[layer][best_label[u], u].item(),
maxiou_quantile=max_iou_quantile[layer][
best_label[u], u].item()))
if iqrdata is not None:
[max_iqr, max_iqr_level, max_iqr_quantile,
max_iqr_iou, max_iqr_agreement] = iqrdata
qualified_iqr = max_iqr[layer].clone()
qualified_iqr[max_iqr_quantile[layer] > 0.5] = 0
best_score, best_label = qualified_iqr.max(0)
for u, urec in enumerate(units):
urec.update(dict(
iqr=best_score[u].item(),
iqr_label=labelnames[best_label[u].item()][0],
iqr_cat=labelnames[best_label[u].item()][1],
iqr_level=max_iqr_level[layer][best_label[u], u].item(),
iqr_quantile=max_iqr_quantile[layer][
best_label[u], u].item(),
iqr_iou=max_iqr_iou[layer][best_label[u], u].item()
))
if covariancedata is not None:
score = covariancedata[layer].correlation()
best_score, best_label = score.max(1)
for u, urec in enumerate(units):
urec.update(dict(
cor=best_score[u].item(),
cor_label=labelnames[best_label[u].item()][0],
cor_cat=labelnames[best_label[u].item()][1]
))
if mergedata is not None:
# Final step: if the user passed any data to merge into the
# units, merge them now. This can be used, for example, to
# indiate that a unit is not interpretable based on some
# outside analysis of unit statistics.
for lrec in mergedata.get('layers', []):
if lrec['layer'] == layer:
break
else:
lrec = None
for u, urec in enumerate(lrec.get('units', []) if lrec else []):
units[u].update(urec)
# After populating per-unit info, populate per-layer ranking info
if labeldata is not None:
# Collect all labeled units
labelunits = defaultdict(list)
all_labelunits = defaultdict(list)
for u, urec in enumerate(units):
if urec['interp']:
labelunits[urec['iou_labelnum']].append(u)
all_labelunits[urec['iou_labelnum']].append(u)
# Sort all units in order with most popular label first.
label_ordering = sorted(units,
# Sort by:
key=lambda r: (-1 if r['interp'] else 0, # interpretable
-len(labelunits[r['iou_labelnum']]), # label freq, score
-max([units[u]['iou']
for u in labelunits[r['iou_labelnum']]], default=0),
r['iou_labelnum'], # label
-r['iou'])) # unit score
# Add label and iou ranking.
rankings.append(dict(name="label", score=(numpy.argsort(list(
ur['unit'] for ur in label_ordering))).tolist()))
rankings.append(dict(name="max iou", metric="iou", score=list(
-ur['iou'] for ur in units)))
# Add ranking for top labels
# for labelnum in [n for n in sorted(
# all_labelunits.keys(), key=lambda x:
# -len(all_labelunits[x])) if len(all_labelunits[n])]:
# label = labelnames[labelnum][0]
# rankings.append(dict(name="%s-iou" % label,
# concept=label, metric='iou',
# score=(-lscore[labelnum, :]).tolist()))
# Collate labels by category then frequency.
record['labels'] = [dict(
label=labelnames[label][0],
labelnum=label,
units=labelunits[label],
cat=labelnames[label][1])
for label in (sorted(labelunits.keys(),
# Sort by:
key=lambda l: (catorder.get( # category
labelnames[l][1], 0),
-len(labelunits[l]), # label freq
-max([units[u]['iou'] for u in labelunits[l]],
default=0) # score
))) if len(labelunits[label])]
# Total number of interpretable units.
record['interpretable'] = sum(len(group['units'])
for group in record['labels'])
# Make a bargraph of labels
os.makedirs(os.path.join(outdir, safe_dir_name(layer)),
exist_ok=True)
catgroups = OrderedDict()
for _, cat in sorted([(v, k) for k, v in catorder.items()]):
catgroups[cat] = []
for rec in record['labels']:
if rec['cat'] not in catgroups:
catgroups[rec['cat']] = []
catgroups[rec['cat']].append(rec['label'])
make_svg_bargraph(
[rec['label'] for rec in record['labels']],
[len(rec['units']) for rec in record['labels']],
[(cat, len(group)) for cat, group in catgroups.items()],
filename=os.path.join(outdir, safe_dir_name(layer),
'bargraph.svg'))
# Only show the bargraph if it is non-empty.
if len(record['labels']):
record['bargraph'] = 'bargraph.svg'
if maxioudata is not None:
rankings.append(dict(name="max maxiou", metric="maxiou", score=list(
-ur['maxiou'] for ur in units)))
if iqrdata is not None:
rankings.append(dict(name="max iqr", metric="iqr", score=list(
-ur['iqr'] for ur in units)))
if covariancedata is not None:
rankings.append(dict(name="max cor", metric="cor", score=list(
-ur['cor'] for ur in units)))
all_layers.append(record)
# Now add the same rankings to every layer...
all_labels = None
if rank_all_labels:
all_labels = [name for name, cat in labelnames]
if labeldata is not None:
# Count layers+quadrants with a given label, and sort by freq
counted_labels = defaultdict(int)
for label in [
re.sub(r'-(?:t|b|l|r|tl|tr|bl|br)$', '', unitrec['iou_label'])
for record in all_layers for unitrec in record['units']]:
counted_labels[label] += 1
if all_labels is None:
all_labels = [label for count, label in sorted((-v, k)
for k, v in counted_labels.items())]
for record in all_layers:
layer = record['layer']
for label in all_labels:
labelnum = labelnumber[label]
record['rankings'].append(dict(name="%s-iou" % label,
concept=label, metric='iou',
score=(-iou_scores[layer][labelnum, :]).tolist()))
if maxioudata is not None:
if all_labels is None:
counted_labels = defaultdict(int)
for label in [
re.sub(r'-(?:t|b|l|r|tl|tr|bl|br)$', '',
unitrec['maxiou_label'])
for record in all_layers for unitrec in record['units']]:
counted_labels[label] += 1
all_labels = [label for count, label in sorted((-v, k)
for k, v in counted_labels.items())]
qualified_iou = max_iou[layer].clone()
qualified_iou[max_iou_quantile[layer] > 0.5] = 0
for record in all_layers:
layer = record['layer']
for label in all_labels:
labelnum = labelnumber[label]
record['rankings'].append(dict(name="%s-maxiou" % label,
concept=label, metric='maxiou',
score=(-qualified_iou[labelnum, :]).tolist()))
if iqrdata is not None:
if all_labels is None:
counted_labels = defaultdict(int)
for label in [
re.sub(r'-(?:t|b|l|r|tl|tr|bl|br)$', '',
unitrec['iqr_label'])
for record in all_layers for unitrec in record['units']]:
counted_labels[label] += 1
all_labels = [label for count, label in sorted((-v, k)
for k, v in counted_labels.items())]
# qualified_iqr[max_iqr_quantile[layer] > 0.5] = 0
for record in all_layers:
layer = record['layer']
qualified_iqr = max_iqr[layer].clone()
for label in all_labels:
labelnum = labelnumber[label]
record['rankings'].append(dict(name="%s-iqr" % label,
concept=label, metric='iqr',
score=(-qualified_iqr[labelnum, :]).tolist()))
if covariancedata is not None:
if all_labels is None:
counted_labels = defaultdict(int)
for label in [
re.sub(r'-(?:t|b|l|r|tl|tr|bl|br)$', '',
unitrec['cor_label'])
for record in all_layers for unitrec in record['units']]:
counted_labels[label] += 1
all_labels = [label for count, label in sorted((-v, k)
for k, v in counted_labels.items())]
for record in all_layers:
layer = record['layer']
score = covariancedata[layer].correlation()
for label in all_labels:
labelnum = labelnumber[label]
record['rankings'].append(dict(name="%s-cor" % label,
concept=label, metric='cor',
score=(-score[:, labelnum]).tolist()))
for record in all_layers:
layer = record['layer']
# Dump per-layer json inside per-layer directory
record['dirname'] = '.'
with open(os.path.join(outdir, safe_dir_name(layer), 'dissect.json'),
'w') as jsonfile:
top_record['layers'] = [record]
json.dump(top_record, jsonfile, indent=1)
# Copy the per-layer html
shutil.copy(os.path.join(srcdir, 'dissect.html'),
os.path.join(outdir, safe_dir_name(layer), 'dissect.html'))
record['dirname'] = safe_dir_name(layer)
# Dump all-layer json in parent directory
with open(os.path.join(outdir, 'dissect.json'), 'w') as jsonfile:
top_record['layers'] = all_layers
json.dump(top_record, jsonfile, indent=1)
# Copy the all-layer html
shutil.copy(os.path.join(srcdir, 'dissect.html'),
os.path.join(outdir, 'dissect.html'))
shutil.copy(os.path.join(srcdir, 'edit.html'),
os.path.join(outdir, 'edit.html'))
def generate_images(outdir, model, dataset, topk, levels,
segrunner, row_length=None, gap_pixels=5,
row_images=True, single_images=False, prefix='',
batch_size=100, num_workers=24):
'''
Creates an image strip file for every unit of every retained layer
of the model, in the format [outdir]/[layername]/[unitnum]-top.jpg.
Assumes that the indexes of topk refer to the indexes of dataset.
Limits each strip to the top row_length images.
'''
progress = default_progress()
needed_images = {}
if row_images is False:
row_length = 1
# Pass 1: needed_images lists all images that are topk for some unit.
for layer in topk:
topresult = topk[layer].result()[1].cpu()
for unit, row in enumerate(topresult):
for rank, imgnum in enumerate(row[:row_length]):
imgnum = imgnum.item()
if imgnum not in needed_images:
needed_images[imgnum] = []
needed_images[imgnum].append((layer, unit, rank))
levels = {k: v.cpu().numpy() for k, v in levels.items()}
row_length = len(row[:row_length])
needed_sample = FixedSubsetSampler(sorted(needed_images.keys()))
device = next(model.parameters()).device
segloader = torch.utils.data.DataLoader(dataset,
batch_size=batch_size, num_workers=num_workers,
pin_memory=(device.type == 'cuda'),
sampler=needed_sample)
vizgrid, maskgrid, origrid, seggrid = [{} for _ in range(4)]
# Pass 2: populate vizgrid with visualizations of top units.
pool = None
for i, batch in enumerate(
progress(segloader, desc='Making images')):
# Reverse transformation to get the image in byte form.
seg, _, byte_im, _ = segrunner.run_and_segment_batch(batch, model,
want_rgb=True)
torch_features = model.retained_features()
scale_offset = getattr(model, 'scale_offset', None)
if pool is None:
# Distribute the work across processes: create shared mmaps.
for layer, tf in torch_features.items():
[vizgrid[layer], maskgrid[layer], origrid[layer],
seggrid[layer]] = [
create_temp_mmap_grid((tf.shape[1],
byte_im.shape[1], row_length,
byte_im.shape[2] + gap_pixels, depth),
dtype='uint8',
fill=255)
for depth in [3, 4, 3, 3]]
# Pass those mmaps to worker processes.
pool = WorkerPool(worker=VisualizeImageWorker,
memmap_grid_info=[
{layer: (g.filename, g.shape, g.dtype)
for layer, g in grid.items()}
for grid in [vizgrid, maskgrid, origrid, seggrid]])
byte_im = byte_im.cpu().numpy()
numpy_seg = seg.cpu().numpy()
features = {}
for index in range(len(byte_im)):
imgnum = needed_sample.samples[index + i*segloader.batch_size]
for layer, unit, rank in needed_images[imgnum]:
if layer not in features:
features[layer] = torch_features[layer].cpu().numpy()
pool.add(layer, unit, rank,
byte_im[index],
features[layer][index, unit],
levels[layer][unit],
scale_offset[layer] if scale_offset else None,
numpy_seg[index])
pool.join()
# Pass 3: save image strips as [outdir]/[layer]/[unitnum]-[top/orig].jpg
pool = WorkerPool(worker=SaveImageWorker)
for layer, vg in progress(vizgrid.items(), desc='Saving images'):
os.makedirs(os.path.join(outdir, safe_dir_name(layer),
prefix + 'image'), exist_ok=True)
if single_images:
os.makedirs(os.path.join(outdir, safe_dir_name(layer),
prefix + 's-image'), exist_ok=True)
og, sg, mg = origrid[layer], seggrid[layer], maskgrid[layer]
for unit in progress(range(len(vg)), desc='Units'):
for suffix, grid in [('top.jpg', vg), ('orig.jpg', og),
('seg.png', sg), ('mask.png', mg)]:
strip = grid[unit].reshape(
(grid.shape[1], grid.shape[2] * grid.shape[3],
grid.shape[4]))
if row_images:
filename = os.path.join(outdir, safe_dir_name(layer),
prefix + 'image', '%d-%s' % (unit, suffix))
pool.add(strip[:,:-gap_pixels,:].copy(), filename)
# Image.fromarray(strip[:,:-gap_pixels,:]).save(filename,
# optimize=True, quality=80)
if single_images:
single_filename = os.path.join(outdir, safe_dir_name(layer),
prefix + 's-image', '%d-%s' % (unit, suffix))
pool.add(strip[:,:strip.shape[1] // row_length
- gap_pixels,:].copy(), single_filename)
# Image.fromarray(strip[:,:strip.shape[1] // row_length
# - gap_pixels,:]).save(single_filename,
# optimize=True, quality=80)
pool.join()
# Delete the shared memory map files
clear_global_shared_files([g.filename
for grid in [vizgrid, maskgrid, origrid, seggrid]
for g in grid.values()])
global_shared_files = {}
def create_temp_mmap_grid(shape, dtype, fill):
dtype = numpy.dtype(dtype)
filename = os.path.join(tempfile.mkdtemp(), 'temp-%s-%s.mmap' %
('x'.join('%d' % s for s in shape), dtype.name))
fid = open(filename, mode='w+b')
original = numpy.memmap(fid, dtype=dtype, mode='w+', shape=shape)
original.fid = fid
original[...] = fill
global_shared_files[filename] = original
return original
def shared_temp_mmap_grid(filename, shape, dtype):
if filename not in global_shared_files:
global_shared_files[filename] = numpy.memmap(
filename, dtype=dtype, mode='r+', shape=shape)
return global_shared_files[filename]
def clear_global_shared_files(filenames):
for fn in filenames:
if fn in global_shared_files:
del global_shared_files[fn]
try:
os.unlink(fn)
except OSError:
pass
class VisualizeImageWorker(WorkerBase):
def setup(self, memmap_grid_info):
self.vizgrid, self.maskgrid, self.origrid, self.seggrid = [
{layer: shared_temp_mmap_grid(*info)
for layer, info in grid.items()}
for grid in memmap_grid_info]
def work(self, layer, unit, rank,
byte_im, acts, level, scale_offset, seg):
self.origrid[layer][unit,:,rank,:byte_im.shape[0],:] = byte_im
[self.vizgrid[layer][unit,:,rank,:byte_im.shape[0],:],
self.maskgrid[layer][unit,:,rank,:byte_im.shape[0],:]] = (
activation_visualization(
byte_im,
acts,
level,
scale_offset=scale_offset,
return_mask=True))
self.seggrid[layer][unit,:,rank,:byte_im.shape[0],:] = (
segment_visualization(seg, byte_im.shape[0:2]))
class SaveImageWorker(WorkerBase):
def work(self, data, filename):
Image.fromarray(data).save(filename, optimize=True, quality=80)
def score_tally_stats(label_category, tc, truth, cc, ic):
pred = cc[label_category]
total = tc[label_category][:, None]
truth = truth[:, None]
epsilon = 1e-20 # avoid division-by-zero
union = pred + truth - ic
iou = ic.double() / (union.double() + epsilon)
arr = torch.empty(size=(2, 2) + ic.shape, dtype=ic.dtype, device=ic.device)
arr[0, 0] = ic
arr[0, 1] = pred - ic
arr[1, 0] = truth - ic
arr[1, 1] = total - union
arr = arr.double() / total.double()
mi = mutual_information(arr)
je = joint_entropy(arr)
iqr = mi / je
iqr[torch.isnan(iqr)] = 0 # Zero out any 0/0
return iou, iqr
def collect_quantiles_and_topk(outdir, model, segloader,
segrunner, k=100, resolution=1024):
'''
Collects (estimated) quantile information and (exact) sorted top-K lists
for every channel in the retained layers of the model. Returns
a map of quantiles (one RunningQuantile for each layer) along with
a map of topk (one RunningTopK for each layer).
'''
device = next(model.parameters()).device
features = model.retained_features()
cached_quantiles = {
layer: load_quantile_if_present(os.path.join(outdir,
safe_dir_name(layer)), 'quantiles.npz',
device=torch.device('cpu'))
for layer in features }
cached_topks = {
layer: load_topk_if_present(os.path.join(outdir,
safe_dir_name(layer)), 'topk.npz',
device=torch.device('cpu'))
for layer in features }
if (all(value is not None for value in cached_quantiles.values()) and
all(value is not None for value in cached_topks.values())):
return cached_quantiles, cached_topks
layer_batch_size = 8
all_layers = list(features.keys())
layer_batches = [all_layers[i:i+layer_batch_size]
for i in range(0, len(all_layers), layer_batch_size)]
quantiles, topks = {}, {}
progress = default_progress()
for layer_batch in layer_batches:
for i, batch in enumerate(progress(segloader, desc='Quantiles')):
# We don't actually care about the model output.
model(batch[0].to(device))
features = model.retained_features()
# We care about the retained values
for key in layer_batch:
value = features[key]
if topks.get(key, None) is None:
topks[key] = RunningTopK(k)
if quantiles.get(key, None) is None:
quantiles[key] = RunningQuantile(resolution=resolution)
topvalue = value
if len(value.shape) > 2:
topvalue, _ = value.view(*(value.shape[:2] + (-1,))).max(2)
# Put the channel index last.
value = value.permute(
(0,) + tuple(range(2, len(value.shape))) + (1,)
).contiguous().view(-1, value.shape[1])
quantiles[key].add(value)
topks[key].add(topvalue)
# Save GPU memory
for key in layer_batch:
quantiles[key].to_(torch.device('cpu'))
topks[key].to_(torch.device('cpu'))
for layer in quantiles:
save_state_dict(quantiles[layer],
os.path.join(outdir, safe_dir_name(layer), 'quantiles.npz'))
save_state_dict(topks[layer],
os.path.join(outdir, safe_dir_name(layer), 'topk.npz'))
return quantiles, topks
def collect_bincounts(outdir, model, segloader, levels, segrunner):
'''
Returns label_counts, category_activation_counts, and intersection_counts,
across the data set, counting the pixels of intersection between upsampled,
thresholded model featuremaps, with segmentation classes in the segloader.
label_counts (independent of model): pixels across the data set that
are labeled with the given label.
category_activation_counts (one per layer): for each feature channel,
pixels across the dataset where the channel exceeds the level
threshold. There is one count per category: activations only
contribute to the categories for which any category labels are
present on the images.
intersection_counts (one per layer): for each feature channel and
label, pixels across the dataset where the channel exceeds
the level, and the labeled segmentation class is also present.
This is a performance-sensitive function. Best performance is
achieved with a counting scheme which assumes a segloader with
batch_size 1.
'''
# Load cached data if present
(iou_scores, iqr_scores,
total_counts, label_counts, category_activation_counts,
intersection_counts) = {}, {}, None, None, {}, {}
found_all = True
for layer in model.retained_features():
filename = os.path.join(outdir, safe_dir_name(layer), 'bincounts.npz')
if os.path.isfile(filename):
data = numpy.load(filename)
iou_scores[layer] = torch.from_numpy(data['iou_scores'])
iqr_scores[layer] = torch.from_numpy(data['iqr_scores'])
total_counts = torch.from_numpy(data['total_counts'])
label_counts = torch.from_numpy(data['label_counts'])
category_activation_counts[layer] = torch.from_numpy(
data['category_activation_counts'])
intersection_counts[layer] = torch.from_numpy(
data['intersection_counts'])
else:
found_all = False
if found_all:
return (iou_scores, iqr_scores,
total_counts, label_counts, category_activation_counts,
intersection_counts)
device = next(model.parameters()).device
labelcat, categories = segrunner.get_label_and_category_names()
label_category = [categories.index(c) if c in categories else 0
for l, c in labelcat]
num_labels, num_categories = (len(n) for n in [labelcat, categories])
# One-hot vector of category for each label
labelcat = torch.zeros(num_labels, num_categories,
dtype=torch.long, device=device)
labelcat.scatter_(1, torch.from_numpy(numpy.array(label_category,
dtype='int64')).to(device)[:,None], 1)
# Running bincounts
# activation_counts = {}
assert segloader.batch_size == 1 # category_activation_counts needs this.
category_activation_counts = {}
intersection_counts = {}
label_counts = torch.zeros(num_labels, dtype=torch.long, device=device)
total_counts = torch.zeros(num_categories, dtype=torch.long, device=device)
progress = default_progress()
scale_offset_map = getattr(model, 'scale_offset', None)
upsample_grids = {}
# total_batch_categories = torch.zeros(
# labelcat.shape[1], dtype=torch.long, device=device)
for i, batch in enumerate(progress(segloader, desc='Bincounts')):
seg, batch_label_counts, _, imshape = segrunner.run_and_segment_batch(
batch, model, want_bincount=True, want_rgb=True)
bc = batch_label_counts.cpu()
batch_label_counts = batch_label_counts.to(device)
seg = seg.to(device)
features = model.retained_features()
# Accumulate bincounts and identify nonzeros
label_counts += batch_label_counts[0]
batch_labels = bc[0].nonzero()[:,0]
batch_categories = labelcat[batch_labels].max(0)[0]
total_counts += batch_categories * (
seg.shape[0] * seg.shape[2] * seg.shape[3])
for key, value in features.items():
if key not in upsample_grids:
upsample_grids[key] = upsample_grid(value.shape[2:],
seg.shape[2:], imshape,
scale_offset=scale_offset_map.get(key, None)
if scale_offset_map is not None else None,
dtype=value.dtype, device=value.device)
upsampled = torch.nn.functional.grid_sample(value,
upsample_grids[key], padding_mode='border')
amask = (upsampled > levels[key][None,:,None,None].to(
upsampled.device))
ac = amask.int().view(amask.shape[1], -1).sum(1)
# if key not in activation_counts:
# activation_counts[key] = ac
# else:
# activation_counts[key] += ac
# The fastest approach: sum over each label separately!
for label in batch_labels.tolist():
if label == 0:
continue # ignore the background label
imask = amask * ((seg == label).max(dim=1, keepdim=True)[0])
ic = imask.int().view(imask.shape[1], -1).sum(1)
if key not in intersection_counts:
intersection_counts[key] = torch.zeros(num_labels,
amask.shape[1], dtype=torch.long, device=device)
intersection_counts[key][label] += ic
# Count activations within images that have category labels.
# Note: This only makes sense with batch-size one
# total_batch_categories += batch_categories
cc = batch_categories[:,None] * ac[None,:]
if key not in category_activation_counts:
category_activation_counts[key] = cc
else:
category_activation_counts[key] += cc
iou_scores = {}
iqr_scores = {}
for k in intersection_counts:
iou_scores[k], iqr_scores[k] = score_tally_stats(
label_category, total_counts, label_counts,
category_activation_counts[k], intersection_counts[k])
for k in intersection_counts:
numpy.savez(os.path.join(outdir, safe_dir_name(k), 'bincounts.npz'),
iou_scores=iou_scores[k].cpu().numpy(),
iqr_scores=iqr_scores[k].cpu().numpy(),
total_counts=total_counts.cpu().numpy(),
label_counts=label_counts.cpu().numpy(),
category_activation_counts=category_activation_counts[k]
.cpu().numpy(),
intersection_counts=intersection_counts[k].cpu().numpy(),
levels=levels[k].cpu().numpy())
return (iou_scores, iqr_scores,
total_counts, label_counts, category_activation_counts,
intersection_counts)
def collect_cond_quantiles(outdir, model, segloader, segrunner):
'''
Returns maxiou and maxiou_level across the data set, one per layer.
This is a performance-sensitive function. Best performance is
achieved with a counting scheme which assumes a segloader with
batch_size 1.
'''
device = next(model.parameters()).device
cached_cond_quantiles = {
layer: load_conditional_quantile_if_present(os.path.join(outdir,
safe_dir_name(layer)), 'cond_quantiles.npz') # on cpu
for layer in model.retained_features() }
label_fracs = load_npy_if_present(outdir, 'label_fracs.npy', 'cpu')
if label_fracs is not None and all(
value is not None for value in cached_cond_quantiles.values()):
return cached_cond_quantiles, label_fracs
labelcat, categories = segrunner.get_label_and_category_names()
label_category = [categories.index(c) if c in categories else 0
for l, c in labelcat]
num_labels, num_categories = (len(n) for n in [labelcat, categories])
# One-hot vector of category for each label
labelcat = torch.zeros(num_labels, num_categories,
dtype=torch.long, device=device)
labelcat.scatter_(1, torch.from_numpy(numpy.array(label_category,
dtype='int64')).to(device)[:,None], 1)
# Running maxiou
assert segloader.batch_size == 1 # category_activation_counts needs this.
conditional_quantiles = {}
label_counts = torch.zeros(num_labels, dtype=torch.long, device=device)
pixel_count = 0
progress = default_progress()
scale_offset_map = getattr(model, 'scale_offset', None)
upsample_grids = {}
common_conditions = set()
if label_fracs is None or label_fracs is 0:
for i, batch in enumerate(progress(segloader, desc='label fracs')):
seg, batch_label_counts, im, _ = segrunner.run_and_segment_batch(
batch, model, want_bincount=True, want_rgb=True)
batch_label_counts = batch_label_counts.to(device)
features = model.retained_features()
# Accumulate bincounts and identify nonzeros
label_counts += batch_label_counts[0]
pixel_count += seg.shape[2] * seg.shape[3]
label_fracs = (label_counts.cpu().float() / pixel_count)[:, None, None]
numpy.save(os.path.join(outdir, 'label_fracs.npy'), label_fracs)
skip_threshold = 1e-4
skip_labels = set(i.item()
for i in (label_fracs.view(-1) < skip_threshold).nonzero().view(-1))
for layer in progress(model.retained_features().keys(), desc='CQ layers'):
if cached_cond_quantiles.get(layer, None) is not None:
conditional_quantiles[layer] = cached_cond_quantiles[layer]
continue
for i, batch in enumerate(progress(segloader, desc='Condquant')):
seg, batch_label_counts, _, imshape = (
segrunner.run_and_segment_batch(
batch, model, want_bincount=True, want_rgb=True))
bc = batch_label_counts.cpu()
batch_label_counts = batch_label_counts.to(device)
features = model.retained_features()
# Accumulate bincounts and identify nonzeros
label_counts += batch_label_counts[0]
pixel_count += seg.shape[2] * seg.shape[3]
batch_labels = bc[0].nonzero()[:,0]
batch_categories = labelcat[batch_labels].max(0)[0]
cpu_seg = None
value = features[layer]
if layer not in upsample_grids:
upsample_grids[layer] = upsample_grid(value.shape[2:],
seg.shape[2:], imshape,
scale_offset=scale_offset_map.get(layer, None)
if scale_offset_map is not None else None,
dtype=value.dtype, device=value.device)
if layer not in conditional_quantiles:
conditional_quantiles[layer] = RunningConditionalQuantile(
resolution=2048)
upsampled = torch.nn.functional.grid_sample(value,
upsample_grids[layer], padding_mode='border').view(
value.shape[1], -1)
conditional_quantiles[layer].add(('all',), upsampled.t())
cpu_upsampled = None
for label in batch_labels.tolist():
if label in skip_labels:
continue
label_key = ('label', label)
if label_key in common_conditions:
imask = (seg == label).max(dim=1)[0].view(-1)
intersected = upsampled[:, imask]
conditional_quantiles[layer].add(('label', label),
intersected.t())
else:
if cpu_seg is None:
cpu_seg = seg.cpu()
if cpu_upsampled is None:
cpu_upsampled = upsampled.cpu()
imask = (cpu_seg == label).max(dim=1)[0].view(-1)
intersected = cpu_upsampled[:, imask]
conditional_quantiles[layer].add(('label', label),
intersected.t())
if num_categories > 1:
for cat in batch_categories.nonzero()[:,0]:
conditional_quantiles[layer].add(('cat', cat.item()),
upsampled.t())
# Move the most common conditions to the GPU.
if i and not i & (i - 1): # if i is a power of 2:
cq = conditional_quantiles[layer]
common_conditions = set(cq.most_common_conditions(64))
cq.to_('cpu', [k for k in cq.running_quantiles.keys()
if k not in common_conditions])
# When a layer is done, get it off the GPU
conditional_quantiles[layer].to_('cpu')
label_fracs = (label_counts.cpu().float() / pixel_count)[:, None, None]
for cq in conditional_quantiles.values():
cq.to_('cpu')
for layer in conditional_quantiles:
save_state_dict(conditional_quantiles[layer],
os.path.join(outdir, safe_dir_name(layer), 'cond_quantiles.npz'))
numpy.save(os.path.join(outdir, 'label_fracs.npy'), label_fracs)
return conditional_quantiles, label_fracs
def collect_maxiou(outdir, model, segloader, segrunner):
'''
Returns maxiou and maxiou_level across the data set, one per layer.
This is a performance-sensitive function. Best performance is
achieved with a counting scheme which assumes a segloader with
batch_size 1.
'''
device = next(model.parameters()).device
conditional_quantiles, label_fracs = collect_cond_quantiles(
outdir, model, segloader, segrunner)
labelcat, categories = segrunner.get_label_and_category_names()
label_category = [categories.index(c) if c in categories else 0
for l, c in labelcat]
num_labels, num_categories = (len(n) for n in [labelcat, categories])
label_list = [('label', i) for i in range(num_labels)]
category_list = [('all',)] if num_categories <= 1 else (
[('cat', i) for i in range(num_categories)])
max_iou, max_iou_level, max_iou_quantile = {}, {}, {}
fracs = torch.logspace(-3, 0, 100)
progress = default_progress()
for layer, cq in progress(conditional_quantiles.items(), desc='Maxiou'):
levels = cq.conditional(('all',)).quantiles(1 - fracs)
denoms = 1 - cq.collected_normalize(category_list, levels)
isects = (1 - cq.collected_normalize(label_list, levels)) * label_fracs
unions = label_fracs + denoms[label_category, :, :] - isects
iou = isects / unions
# TODO: erase any for which threshold is bad
max_iou[layer], level_bucket = iou.max(2)
max_iou_level[layer] = levels[
torch.arange(levels.shape[0])[None,:], level_bucket]
max_iou_quantile[layer] = fracs[level_bucket]
for layer in model.retained_features():
numpy.savez(os.path.join(outdir, safe_dir_name(layer), 'max_iou.npz'),
max_iou=max_iou[layer].cpu().numpy(),
max_iou_level=max_iou_level[layer].cpu().numpy(),
max_iou_quantile=max_iou_quantile[layer].cpu().numpy())
return (max_iou, max_iou_level, max_iou_quantile)
def collect_iqr(outdir, model, segloader, segrunner):
'''
Returns iqr and iqr_level.
This is a performance-sensitive function. Best performance is
achieved with a counting scheme which assumes a segloader with
batch_size 1.
'''
max_iqr, max_iqr_level, max_iqr_quantile, max_iqr_iou = {}, {}, {}, {}
max_iqr_agreement = {}
found_all = True
for layer in model.retained_features():
filename = os.path.join(outdir, safe_dir_name(layer), 'iqr.npz')
if os.path.isfile(filename):
data = numpy.load(filename)
max_iqr[layer] = torch.from_numpy(data['max_iqr'])
max_iqr_level[layer] = torch.from_numpy(data['max_iqr_level'])
max_iqr_quantile[layer] = torch.from_numpy(data['max_iqr_quantile'])
max_iqr_iou[layer] = torch.from_numpy(data['max_iqr_iou'])
max_iqr_agreement[layer] = torch.from_numpy(
data['max_iqr_agreement'])
else:
found_all = False
if found_all:
return (max_iqr, max_iqr_level, max_iqr_quantile, max_iqr_iou,
max_iqr_agreement)
device = next(model.parameters()).device
conditional_quantiles, label_fracs = collect_cond_quantiles(
outdir, model, segloader, segrunner)
labelcat, categories = segrunner.get_label_and_category_names()
label_category = [categories.index(c) if c in categories else 0
for l, c in labelcat]
num_labels, num_categories = (len(n) for n in [labelcat, categories])
label_list = [('label', i) for i in range(num_labels)]
category_list = [('all',)] if num_categories <= 1 else (
[('cat', i) for i in range(num_categories)])
full_mi, full_je, full_iqr = {}, {}, {}
fracs = torch.logspace(-3, 0, 100)
progress = default_progress()
for layer, cq in progress(conditional_quantiles.items(), desc='IQR'):
levels = cq.conditional(('all',)).quantiles(1 - fracs)
truth = label_fracs.to(device)
preds = (1 - cq.collected_normalize(category_list, levels)
)[label_category, :, :].to(device)
cond_isects = 1 - cq.collected_normalize(label_list, levels).to(device)
isects = cond_isects * truth
unions = truth + preds - isects
arr = torch.empty(size=(2, 2) + isects.shape, dtype=isects.dtype,
device=device)
arr[0, 0] = isects
arr[0, 1] = preds - isects
arr[1, 0] = truth - isects
arr[1, 1] = 1 - unions
arr.clamp_(0, 1)
mi = mutual_information(arr)
mi[:,:,-1] = 0 # at the 1.0 quantile should be no MI.
# Don't trust mi when less than label_frac is less than 1e-3,
# because our samples are too small.
mi[label_fracs.view(-1) < 1e-3, :, :] = 0
je = joint_entropy(arr)
iqr = mi / je
iqr[torch.isnan(iqr)] = 0 # Zero out any 0/0
full_mi[layer] = mi.cpu()
full_je[layer] = je.cpu()
full_iqr[layer] = iqr.cpu()
del mi, je
agreement = isects + arr[1, 1]
# When optimizing, maximize only over those pairs where the
# unit is positively correlated with the label, and where the
# threshold level is positive
positive_iqr = iqr
positive_iqr[agreement <= 0.8] = 0
positive_iqr[(levels <= 0.0)[None, :, :].expand(positive_iqr.shape)] = 0
# TODO: erase any for which threshold is bad
maxiqr, level_bucket = positive_iqr.max(2)
max_iqr[layer] = maxiqr.cpu()
max_iqr_level[layer] = levels.to(device)[
torch.arange(levels.shape[0])[None,:], level_bucket].cpu()
max_iqr_quantile[layer] = fracs.to(device)[level_bucket].cpu()
max_iqr_agreement[layer] = agreement[
torch.arange(agreement.shape[0])[:, None],
torch.arange(agreement.shape[1])[None, :],
level_bucket].cpu()
# Compute the iou that goes with each maximized iqr
matching_iou = (isects[
torch.arange(isects.shape[0])[:, None],
torch.arange(isects.shape[1])[None, :],
level_bucket] /
unions[
torch.arange(unions.shape[0])[:, None],
torch.arange(unions.shape[1])[None, :],
level_bucket])
matching_iou[torch.isnan(matching_iou)] = 0
max_iqr_iou[layer] = matching_iou.cpu()
for layer in model.retained_features():
numpy.savez(os.path.join(outdir, safe_dir_name(layer), 'iqr.npz'),
max_iqr=max_iqr[layer].cpu().numpy(),
max_iqr_level=max_iqr_level[layer].cpu().numpy(),
max_iqr_quantile=max_iqr_quantile[layer].cpu().numpy(),
max_iqr_iou=max_iqr_iou[layer].cpu().numpy(),
max_iqr_agreement=max_iqr_agreement[layer].cpu().numpy(),
full_mi=full_mi[layer].cpu().numpy(),
full_je=full_je[layer].cpu().numpy(),
full_iqr=full_iqr[layer].cpu().numpy())
return (max_iqr, max_iqr_level, max_iqr_quantile, max_iqr_iou,
max_iqr_agreement)
def mutual_information(arr):
total = 0
for j in range(arr.shape[0]):
for k in range(arr.shape[1]):
joint = arr[j,k]
ind = arr[j,:].sum(dim=0) * arr[:,k].sum(dim=0)
term = joint * (joint / ind).log()
term[torch.isnan(term)] = 0
total += term
return total.clamp_(0)
def joint_entropy(arr):
total = 0
for j in range(arr.shape[0]):
for k in range(arr.shape[1]):
joint = arr[j,k]
term = joint * joint.log()
term[torch.isnan(term)] = 0
total += term
return (-total).clamp_(0)
def information_quality_ratio(arr):
iqr = mutual_information(arr) / joint_entropy(arr)
iqr[torch.isnan(iqr)] = 0
return iqr
def collect_covariance(outdir, model, segloader, segrunner):
'''
Returns label_mean, label_variance, unit_mean, unit_variance,
and cross_covariance across the data set.
label_mean, label_variance (independent of model):
treating the label as a one-hot, each label's mean and variance.
unit_mean, unit_variance (one per layer): for each feature channel,
the mean and variance of the activations in that channel.
cross_covariance (one per layer): the cross covariance between the
labels and the units in the layer.
'''
device = next(model.parameters()).device
cached_covariance = {
layer: load_covariance_if_present(os.path.join(outdir,
safe_dir_name(layer)), 'covariance.npz', device=device)
for layer in model.retained_features() }
if all(value is not None for value in cached_covariance.values()):
return cached_covariance
labelcat, categories = segrunner.get_label_and_category_names()
label_category = [categories.index(c) if c in categories else 0
for l, c in labelcat]
num_labels, num_categories = (len(n) for n in [labelcat, categories])
# Running covariance
cov = {}
progress = default_progress()
scale_offset_map = getattr(model, 'scale_offset', None)
upsample_grids = {}
for i, batch in enumerate(progress(segloader, desc='Covariance')):
seg, _, _, imshape = segrunner.run_and_segment_batch(batch, model,
want_rgb=True)
features = model.retained_features()
ohfeats = multilabel_onehot(seg, num_labels, ignore_index=0)
# Accumulate bincounts and identify nonzeros
for key, value in features.items():
if key not in upsample_grids:
upsample_grids[key] = upsample_grid(value.shape[2:],
seg.shape[2:], imshape,
scale_offset=scale_offset_map.get(key, None)
if scale_offset_map is not None else None,
dtype=value.dtype, device=value.device)
upsampled = torch.nn.functional.grid_sample(value,
upsample_grids[key].expand(
(value.shape[0],) + upsample_grids[key].shape[1:]),
padding_mode='border')
if key not in cov:
cov[key] = RunningCrossCovariance()
cov[key].add(upsampled, ohfeats)
for layer in cov:
save_state_dict(cov[layer],
os.path.join(outdir, safe_dir_name(layer), 'covariance.npz'))
return cov
def multilabel_onehot(labels, num_labels, dtype=None, ignore_index=None):
'''
Converts a multilabel tensor into a onehot tensor.
The input labels is a tensor of shape (samples, multilabels, y, x).
The output is a tensor of shape (samples, num_labels, y, x).
If ignore_index is specified, labels with that index are ignored.
Each x in labels should be 0 <= x < num_labels, or x == ignore_index.
'''
assert ignore_index is None or ignore_index <= 0
if dtype is None:
dtype = torch.float
device = labels.device
chans = num_labels + (-ignore_index if ignore_index else 0)
outshape = (labels.shape[0], chans) + labels.shape[2:]
result = torch.zeros(outshape, device=device, dtype=dtype)
if ignore_index and ignore_index < 0:
labels = labels + (-ignore_index)
result.scatter_(1, labels, 1)
if ignore_index and ignore_index < 0:
result = result[:, -ignore_index:]
elif ignore_index is not None:
result[:, ignore_index] = 0
return result
def load_npy_if_present(outdir, filename, device):
filepath = os.path.join(outdir, filename)
if os.path.isfile(filepath):
data = numpy.load(filepath)
return torch.from_numpy(data).to(device)
return 0
def load_npz_if_present(outdir, filename, varnames, device):
filepath = os.path.join(outdir, filename)
if os.path.isfile(filepath):
data = numpy.load(filepath)
numpy_result = [data[n] for n in varnames]
return tuple(torch.from_numpy(data).to(device) for data in numpy_result)
return None
def load_quantile_if_present(outdir, filename, device):
filepath = os.path.join(outdir, filename)
if os.path.isfile(filepath):
data = numpy.load(filepath)
result = RunningQuantile(state=data)
result.to_(device)
return result
return None
def load_conditional_quantile_if_present(outdir, filename):
filepath = os.path.join(outdir, filename)
if os.path.isfile(filepath):
data = numpy.load(filepath)
result = RunningConditionalQuantile(state=data)
return result
return None
def load_topk_if_present(outdir, filename, device):
filepath = os.path.join(outdir, filename)
if os.path.isfile(filepath):
data = numpy.load(filepath)
result = RunningTopK(state=data)
result.to_(device)
return result
return None
def load_covariance_if_present(outdir, filename, device):
filepath = os.path.join(outdir, filename)
if os.path.isfile(filepath):
data = numpy.load(filepath)
result = RunningCrossCovariance(state=data)
result.to_(device)
return result
return None
def save_state_dict(obj, filepath):
dirname = os.path.dirname(filepath)
os.makedirs(dirname, exist_ok=True)
dic = obj.state_dict()
numpy.savez(filepath, **dic)
def upsample_grid(data_shape, target_shape, input_shape=None,
scale_offset=None, dtype=torch.float, device=None):
'''Prepares a grid to use with grid_sample to upsample a batch of
features in data_shape to the target_shape. Can use scale_offset
and input_shape to center the grid in a nondefault way: scale_offset
maps feature pixels to input_shape pixels, and it is assumed that
the target_shape is a uniform downsampling of input_shape.'''
# Default is that nothing is resized.
if target_shape is None:
target_shape = data_shape
# Make a default scale_offset to fill the image if there isn't one
if scale_offset is None:
scale = tuple(float(ts) / ds
for ts, ds in zip(target_shape, data_shape))
offset = tuple(0.5 * s - 0.5 for s in scale)
else:
scale, offset = (v for v in zip(*scale_offset))
# Handle downsampling for different input vs target shape.
if input_shape is not None:
scale = tuple(s * (ts - 1) / (ns - 1)
for s, ns, ts in zip(scale, input_shape, target_shape))
offset = tuple(o * (ts - 1) / (ns - 1)
for o, ns, ts in zip(offset, input_shape, target_shape))
# Pytorch needs target coordinates in terms of source coordinates [-1..1]
ty, tx = (((torch.arange(ts, dtype=dtype, device=device) - o)
* (2 / (s * (ss - 1))) - 1)
for ts, ss, s, o, in zip(target_shape, data_shape, scale, offset))
# Whoa, note that grid_sample reverses the order y, x -> x, y.
grid = torch.stack(
(tx[None,:].expand(target_shape), ty[:,None].expand(target_shape)),2
)[None,:,:,:].expand((1, target_shape[0], target_shape[1], 2))
return grid
def safe_dir_name(filename):
keepcharacters = (' ','.','_','-')
return ''.join(c
for c in filename if c.isalnum() or c in keepcharacters).rstrip()
bargraph_palette = [
('#4B4CBF', '#B6B6F2'),
('#55B05B', '#B6F2BA'),
('#50BDAC', '#A5E5DB'),
('#81C679', '#C0FF9B'),
('#F0883B', '#F2CFB6'),
('#D4CF24', '#F2F1B6'),
('#D92E2B', '#F2B6B6'),
('#AB6BC6', '#CFAAFF'),
]
def make_svg_bargraph(labels, heights, categories,
barheight=100, barwidth=12, show_labels=True, filename=None):
# if len(labels) == 0:
# return # Nothing to do
unitheight = float(barheight) / max(max(heights, default=1), 1)
textheight = barheight if show_labels else 0
labelsize = float(barwidth)
gap = float(barwidth) / 4
textsize = barwidth + gap
rollup = max(heights, default=1)
textmargin = float(labelsize) * 2 / 3
leftmargin = 32
rightmargin = 8
svgwidth = len(heights) * (barwidth + gap) + 2 * leftmargin + rightmargin
svgheight = barheight + textheight
# create an SVG XML element
svg = et.Element('svg', width=str(svgwidth), height=str(svgheight),
version='1.1', xmlns='http://www.w3.org/2000/svg')
# Draw the bar graph
basey = svgheight - textheight
x = leftmargin
# Add units scale on left
if len(heights):
for h in [1, (max(heights) + 1) // 2, max(heights)]:
et.SubElement(svg, 'text', x='0', y='0',
style=('font-family:sans-serif;font-size:%dpx;' +
'text-anchor:end;alignment-baseline:hanging;' +
'transform:translate(%dpx, %dpx);') %
(textsize, x - gap, basey - h * unitheight)).text = str(h)
et.SubElement(svg, 'text', x='0', y='0',
style=('font-family:sans-serif;font-size:%dpx;' +
'text-anchor:middle;' +
'transform:translate(%dpx, %dpx) rotate(-90deg)') %
(textsize, x - gap - textsize, basey - h * unitheight / 2)
).text = 'units'
# Draw big category background rectangles
for catindex, (cat, catcount) in enumerate(categories):
if not catcount:
continue
et.SubElement(svg, 'rect', x=str(x), y=str(basey - rollup * unitheight),
width=(str((barwidth + gap) * catcount - gap)),
height = str(rollup*unitheight),
fill=bargraph_palette[catindex % len(bargraph_palette)][1])
x += (barwidth + gap) * catcount
# Draw small bars as well as 45degree text labels
x = leftmargin
catindex = -1
catcount = 0
for label, height in zip(labels, heights):
while not catcount and catindex <= len(categories):
catindex += 1
catcount = categories[catindex][1]
color = bargraph_palette[catindex % len(bargraph_palette)][0]
et.SubElement(svg, 'rect', x=str(x), y=str(basey-(height * unitheight)),
width=str(barwidth), height=str(height * unitheight),
fill=color)
x += barwidth
if show_labels:
et.SubElement(svg, 'text', x='0', y='0',
style=('font-family:sans-serif;font-size:%dpx;text-anchor:end;'+
'transform:translate(%dpx, %dpx) rotate(-45deg);') %
(labelsize, x, basey + textmargin)).text = readable(label)
x += gap
catcount -= 1
# Text labels for each category
x = leftmargin
for cat, catcount in categories:
if not catcount:
continue
et.SubElement(svg, 'text', x='0', y='0',
style=('font-family:sans-serif;font-size:%dpx;text-anchor:end;'+
'transform:translate(%dpx, %dpx) rotate(-90deg);') %
(textsize, x + (barwidth + gap) * catcount - gap,
basey - rollup * unitheight + gap)).text = '%d %s' % (
catcount, readable(cat + ('s' if catcount != 1 else '')))
x += (barwidth + gap) * catcount
# Output - this is the bare svg.
result = et.tostring(svg)
if filename:
f = open(filename, 'wb')
# When writing to a file a special header is needed.
f.write(''.join([
'<?xml version=\"1.0\" standalone=\"no\"?>\n',
'<!DOCTYPE svg PUBLIC \"-//W3C//DTD SVG 1.1//EN\"\n',
'\"http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd\">\n']
).encode('utf-8'))
f.write(result)
f.close()
return result
readable_replacements = [(re.compile(r[0]), r[1]) for r in [
(r'-[sc]$', ''),
(r'_', ' '),
]]
def readable(label):
for pattern, subst in readable_replacements:
label= re.sub(pattern, subst, label)
return label
def reverse_normalize_from_transform(transform):
'''
Crawl around the transforms attached to a dataset looking for a
Normalize transform, and return it a corresponding ReverseNormalize,
or None if no normalization is found.
'''
if isinstance(transform, torchvision.transforms.Normalize):
return ReverseNormalize(transform.mean, transform.std)
t = getattr(transform, 'transform', None)
if t is not None:
return reverse_normalize_from_transform(t)
transforms = getattr(transform, 'transforms', None)
if transforms is not None:
for t in reversed(transforms):
result = reverse_normalize_from_transform(t)
if result is not None:
return result
return None
class ReverseNormalize:
'''
Applies the reverse of torchvision.transforms.Normalize.
'''
def __init__(self, mean, stdev):
mean = numpy.array(mean)
stdev = numpy.array(stdev)
self.mean = torch.from_numpy(mean)[None,:,None,None].float()
self.stdev = torch.from_numpy(stdev)[None,:,None,None].float()
def __call__(self, data):
device = data.device
return data.mul(self.stdev.to(device)).add_(self.mean.to(device))
class ImageOnlySegRunner:
def __init__(self, dataset, recover_image=None):
if recover_image is None:
recover_image = reverse_normalize_from_transform(dataset)
self.recover_image = recover_image
self.dataset = dataset
def get_label_and_category_names(self):
return [('-', '-')], ['-']
def run_and_segment_batch(self, batch, model,
want_bincount=False, want_rgb=False):
[im] = batch
device = next(model.parameters()).device
if want_rgb:
rgb = self.recover_image(im.clone()
).permute(0, 2, 3, 1).mul_(255).clamp(0, 255).byte()
else:
rgb = None
# Stubs for seg and bc
seg = torch.zeros(im.shape[0], 1, 1, 1, dtype=torch.long)
bc = torch.ones(im.shape[0], 1, dtype=torch.long)
# Run the model.
model(im.to(device))
return seg, bc, rgb, im.shape[2:]
class ClassifierSegRunner:
def __init__(self, dataset, recover_image=None):
# The dataset contains explicit segmentations
if recover_image is None:
recover_image = reverse_normalize_from_transform(dataset)
self.recover_image = recover_image
self.dataset = dataset
def get_label_and_category_names(self):
catnames = self.dataset.categories
label_and_cat_names = [(readable(label),
catnames[self.dataset.label_category[i]])
for i, label in enumerate(self.dataset.labels)]
return label_and_cat_names, catnames
def run_and_segment_batch(self, batch, model,
want_bincount=False, want_rgb=False):
'''
Runs the dissected model on one batch of the dataset, and
returns a multilabel semantic segmentation for the data.
Given a batch of size (n, c, y, x) the segmentation should
be a (long integer) tensor of size (n, d, y//r, x//r) where
d is the maximum number of simultaneous labels given to a pixel,
and where r is some (optional) resolution reduction factor.
In the segmentation returned, the label `0` is reserved for
the background "no-label".
In addition to the segmentation, bc, rgb, and shape are returned
where bc is a per-image bincount counting returned label pixels,
rgb is a viewable (n, y, x, rgb) byte image tensor for the data
for visualizations (reversing normalizations, for example), and
shape is the (y, x) size of the data. If want_bincount or
want_rgb are False, those return values may be None.
'''
im, seg, bc = batch
device = next(model.parameters()).device
if want_rgb:
rgb = self.recover_image(im.clone()
).permute(0, 2, 3, 1).mul_(255).clamp(0, 255).byte()
else:
rgb = None
# Run the model.
model(im.to(device))
return seg, bc, rgb, im.shape[2:]
class GeneratorSegRunner:
def __init__(self, segmenter):
# The segmentations are given by an algorithm
if segmenter is None:
segmenter = UnifiedParsingSegmenter(segsizes=[256], segdiv='quad')
self.segmenter = segmenter
self.num_classes = len(segmenter.get_label_and_category_names()[0])
def get_label_and_category_names(self):
return self.segmenter.get_label_and_category_names()
def run_and_segment_batch(self, batch, model,
want_bincount=False, want_rgb=False):
'''
Runs the dissected model on one batch of the dataset, and
returns a multilabel semantic segmentation for the data.
Given a batch of size (n, c, y, x) the segmentation should
be a (long integer) tensor of size (n, d, y//r, x//r) where
d is the maximum number of simultaneous labels given to a pixel,
and where r is some (optional) resolution reduction factor.
In the segmentation returned, the label `0` is reserved for
the background "no-label".
In addition to the segmentation, bc, rgb, and shape are returned
where bc is a per-image bincount counting returned label pixels,
rgb is a viewable (n, y, x, rgb) byte image tensor for the data
for visualizations (reversing normalizations, for example), and
shape is the (y, x) size of the data. If want_bincount or
want_rgb are False, those return values may be None.
'''
device = next(model.parameters()).device
z_batch = batch[0]
tensor_images = model(z_batch.to(device))
seg = self.segmenter.segment_batch(tensor_images, downsample=2)
if want_bincount:
index = torch.arange(z_batch.shape[0],
dtype=torch.long, device=device)
bc = (seg + index[:, None, None, None] * self.num_classes).view(-1
).bincount(minlength=z_batch.shape[0] * self.num_classes)
bc = bc.view(z_batch.shape[0], self.num_classes)
else:
bc = None
if want_rgb:
images = ((tensor_images + 1) / 2 * 255)
rgb = images.permute(0, 2, 3, 1).clamp(0, 255).byte()
else:
rgb = None
return seg, bc, rgb, tensor_images.shape[2:]