Samuel Mueller
init
e487255
import torch
from .utils import get_batch_to_dataloader
from utils import default_device
def get_batch(batch_size, seq_len, num_features, device=default_device
, hyperparameters=None, batch_size_per_gp_sample=None, **kwargs):
batch_size_per_gp_sample = batch_size_per_gp_sample or (min(64, batch_size))
num_models = batch_size // batch_size_per_gp_sample
assert num_models * batch_size_per_gp_sample == batch_size, f'Batch size ({batch_size}) not divisible by batch_size_per_gp_sample ({batch_size_per_gp_sample})'
args = {'device': device, 'seq_len': seq_len, 'num_features': num_features, 'batch_size': batch_size_per_gp_sample}
prior_bag_priors_get_batch = hyperparameters['prior_bag_get_batch']
prior_bag_priors_p = [1.0] + [hyperparameters[f'prior_bag_exp_weights_{i}'] for i in range(1, len(prior_bag_priors_get_batch))]
weights = torch.tensor(prior_bag_priors_p, dtype=torch.float) # create a tensor of weights
batch_assignments = torch.multinomial(torch.softmax(weights, 0), num_models, replacement=True).numpy()
if 'verbose' in hyperparameters and hyperparameters['verbose']:
print('PRIOR_BAG:', weights, batch_assignments)
sample = sum([[prior_bag_priors_get_batch[int(prior_idx)](hyperparameters=hyperparameters, **args)] for prior_idx in batch_assignments], [])
x, y, y_ = zip(*sample)
x, y, y_ = (torch.cat(x, 1).detach()
, torch.cat(y, 1).detach()
, torch.cat(y_, 1).detach())
return x, y, y_
DataLoader = get_batch_to_dataloader(get_batch)
DataLoader.num_outputs = 1