| from typing import List, Dict, Tuple, Callable |
| import torch |
| import os |
| import torch.nn as nn |
| from pathlib import Path |
| from torch.utils.data import DataLoader |
| import torchvision.transforms as transforms |
| import torchvision.datasets as datasets |
| import shutil |
|
|
| IMAGENET_PCA = { |
| 'eigval': torch.Tensor([0.2175, 0.0188, 0.0045]), |
| 'eigvec': torch.Tensor([ |
| [-0.5675, 0.7192, 0.4009], |
| [-0.5808, -0.0045, -0.8140], |
| [-0.5836, -0.6948, 0.4203], |
| ]) |
| } |
|
|
| class Lighting(object): |
| """ |
| Lighting noise (see https://git.io/fhBOc) |
| https://github.com/MadryLab/robustness/blob/a9541241defd9972e9334bfcdb804f6aefe24dc7/robustness/data_augmentation.py#L18 |
| """ |
| def __init__(self, alphastd, eigval, eigvec): |
| self.alphastd = alphastd |
| self.eigval = eigval |
| self.eigvec = eigvec |
|
|
| def __call__(self, img): |
| if self.alphastd == 0: |
| return img |
|
|
| alpha = img.new().resize_(3).normal_(0, self.alphastd) |
| rgb = self.eigvec.type_as(img).clone()\ |
| .mul(alpha.view(1, 3).expand(3, 3))\ |
| .mul(self.eigval.view(1, 3).expand(3, 3))\ |
| .sum(1).squeeze() |
|
|
| return img.add(rgb.view(3, 1, 1).expand_as(img)) |
|
|
| class PSiLU(nn.Module): |
| def __init__(self, alpha: float = 1.0, device=None, dtype=None) -> None: |
| super().__init__() |
| factory_kwargs = {'device': device, 'dtype': dtype} |
| self.alpha = nn.Parameter(torch.empty(1, **factory_kwargs).fill_(alpha)) |
|
|
| def forward(self, x: torch.Tensor) -> torch.Tensor: |
| return x * torch.sigmoid(self.alpha * x) |
|
|
| class PSSiLU(nn.Module): |
| def __init__(self, alpha: float = 1.0, beta: float = 1e-4, device=None, dtype=None) -> None: |
| super().__init__() |
| factory_kwargs = {'device': device, 'dtype': dtype} |
| self.alpha = nn.Parameter(torch.empty(1, **factory_kwargs).fill_(alpha)) |
| self.beta = nn.Parameter(torch.empty(1, **factory_kwargs).fill_(beta)) |
|
|
| def forward(self, x: torch.Tensor) -> torch.Tensor: |
| self.beta.data = torch.clamp(self.beta.data, 0., 1.) |
| return x * (torch.sigmoid(self.alpha * x) - self.beta) / (1. - self.beta + 1e-6) |
| class AverageMeter(object): |
| """Computes and stores the average and current value""" |
|
|
| def __init__(self): |
| self.reset() |
|
|
| def reset(self): |
| self.val = 0 |
| self.avg = 0 |
| self.sum = 0 |
| self.count = 0 |
|
|
| def update(self, val: float, n: int = 1): |
| self.val = val |
| self.sum += val * n |
| self.count += n |
| self.avg = self.sum / self.count |
|
|
| def accuracy(output: torch.Tensor, target: torch.Tensor, topk: tuple = (1, 5)) -> List[float]: |
| """Computes the accuracy over the k top predictions for the specified values of k""" |
|
|
| with torch.no_grad(): |
| maxk = max(topk) |
| batch_size = target.shape[0] |
|
|
| _, pred = output.topk(k=maxk, dim=1, largest=True, sorted=True) |
| pred = pred.t() |
| correct = pred.eq(target.view(1, -1).expand_as(pred)) |
|
|
| res = [] |
| for k in topk: |
| correct_k = correct[:k].flatten().float().sum(0) |
| res.append(correct_k.mul_(100.0 / batch_size)) |
| return res |
|
|
| def configure_optimizers(model: nn.Module) -> List[Dict]: |
| """ |
| all normalization params and all bias shouldn't be weight decayed |
| code adapted from https://github.com/karpathy/minGPT/blob/3ed14b2cec0dfdad3f4b2831f2b4a86d11aef150/mingpt/model.py#L136 |
| """ |
| decay = set() |
| no_decay = set() |
| weight_decay_blacklist = (nn.BatchNorm2d, nn.LayerNorm, nn.InstanceNorm2d, nn.PReLU, PSiLU, PSSiLU) |
| for mn, m in model.named_modules(): |
| for pn, p in m.named_parameters(recurse=False): |
| full_name = f"{mn}.{pn}" if mn else pn |
| if pn.endswith("bias"): |
| |
| no_decay.add(full_name) |
| elif pn.endswith("weight") and isinstance(m, weight_decay_blacklist): |
| |
| no_decay.add(full_name) |
| else: |
| decay.add(full_name) |
|
|
| |
| param_dict = {pn: p for pn, p in model.named_parameters()} |
| inter_params = decay & no_decay |
| union_params = decay | no_decay |
| assert len(inter_params) == 0, f"parameters {str(inter_params)} made it into both decay/no_decay sets!" |
| assert len(param_dict.keys() - union_params) == 0, f"parameters {str(param_dict.keys() - union_params)} were not separated into either decay/no_decay set!" |
|
|
| optim_group = list([ |
| {"params": [param_dict[fpn] for fpn in sorted(list(decay))]}, |
| {"params": [param_dict[fpn] for fpn in sorted(list(no_decay))], "weight_decay": 0.0}, |
| ]) |
|
|
| return optim_group |
|
|
| def get_cifar10(root_dir: str, batch_size: int, workers: int, crop_size: int) -> List[DataLoader]: |
| """ |
| The transforms follow the robustness library: https://github.com/MadryLab/robustness/blob/a9541241defd9972e9334bfcdb804f6aefe24dc7/robustness/data_augmentation.py#L68 |
| """ |
| train_transform = transforms.Compose([ |
| transforms.RandomCrop(crop_size, padding=4), |
| transforms.RandomHorizontalFlip(), |
| transforms.ColorJitter(.25, .25, .25), |
| transforms.RandomRotation(2), |
| transforms.ToTensor(), |
| ]) |
|
|
| test_transform = transforms.Compose([ |
| transforms.Resize(crop_size), |
| transforms.CenterCrop(crop_size), |
| transforms.ToTensor() |
| ]) |
|
|
| train_dataset = datasets.CIFAR10(root=root_dir, train=True, download=True, transform=train_transform) |
| test_dataset = datasets.CIFAR10(root=root_dir, train=False, download=True, transform=test_transform) |
|
|
| train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True, num_workers=workers, pin_memory=True) |
| test_loader = DataLoader(test_dataset, batch_size=batch_size, shuffle=True, num_workers=workers, pin_memory=True) |
|
|
| return train_loader, test_loader |
|
|
| def get_imagenet(data_dir: str, batch_size: int, workers: int, crop_size: int, color_jitter: float = 0., use_lighting: bool = False, mode: str = "fat") -> List[DataLoader]: |
| train_aug = list([ |
| transforms.RandomResizedCrop(crop_size), |
| transforms.RandomHorizontalFlip(), |
| ]) |
|
|
| if color_jitter > 0: |
| cj = (float(color_jitter),) * 3 |
| train_aug.append(transforms.ColorJitter(*cj)) |
|
|
| train_aug.append(transforms.ToTensor()) |
|
|
| if use_lighting: |
| train_aug.append(Lighting(0.05, IMAGENET_PCA['eigval'], IMAGENET_PCA['eigvec'])) |
|
|
| train_transform = transforms.Compose(train_aug) |
|
|
| test_aug = list() |
|
|
| |
| if mode == "at_pgd": |
| test_aug.append(transforms.Resize(256)) |
| test_aug.extend([ |
| transforms.CenterCrop(crop_size), |
| transforms.ToTensor(), |
| ]) |
|
|
| test_transform = transforms.Compose(test_aug) |
|
|
| train_path = os.path.join(data_dir, 'train') |
| test_path = os.path.join(data_dir, 'val') |
|
|
| train_dataset = datasets.ImageFolder(root=train_path, transform=train_transform) |
| test_dataset = datasets.ImageFolder(root=test_path, transform=test_transform) |
|
|
| train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True, num_workers=workers, pin_memory=True) |
| test_loader = DataLoader(test_dataset, batch_size=batch_size, shuffle=False, num_workers=workers, pin_memory=True) |
|
|
| return train_loader, test_loader |
|
|
|
|
| def get_datasets(cfg): |
| if cfg.dataset.dataset == "imagenet": |
| return get_imagenet( |
| data_dir=cfg.dataset.data_dir, |
| batch_size=cfg.train_test.batch_size, |
| workers=cfg.train_test.workers, |
| crop_size=cfg.train_test.crop_size, |
| color_jitter=cfg.train_test.color_jitter, |
| use_lighting=cfg.train_test.lighting, |
| mode=cfg.train_test.mode, |
| ) |
| elif cfg.dataset.dataset == "cifar10": |
| return get_cifar10( |
| root_dir=cfg.dataset.data_dir, |
| batch_size=cfg.train_test.batch_size, |
| workers=cfg.train_test.workers, |
| crop_size=cfg.train_test.crop_size, |
| ) |
|
|
| def pad_str(msg: str, total_len: int=80) -> str: |
| rem_len = total_len - len(msg) |
| return f"{'*' * (rem_len // 2)}{msg}{'*' * (rem_len // 2)}" |
|
|
| def save_checkpoint(model_info: dict, filepath: Path, is_best: bool): |
| filename = filepath / f"ckpt_epoch{model_info['epoch']}.pt" |
| torch.save(model_info, filename) |
| if is_best: |
| shutil.copyfile(filename, filepath / "model_best.pt") |
|
|
| def compute_total_parameters(model: nn.Module) -> float: |
| |
| pytorch_total_params = sum(p.numel() for p in model.parameters()) |
| return pytorch_total_params / 1e6 |
|
|
| |
| def make_linear_schedule(final: float, warmup: int, zero_eps_epochs: int) -> Callable[[int], float]: |
| def linear_schedule(step: int) -> float: |
| if step < zero_eps_epochs: |
| return 0.0 |
| if step < warmup: |
| return (step - zero_eps_epochs) / warmup * final |
| return final |
| return linear_schedule |