| from math import sin, cos, sqrt, atan2, radians, asin |
| import numpy as np |
| import torch |
| import os |
| from torch.distributed import init_process_group, destroy_process_group |
| import torch.nn.functional as F |
| import random |
| def resample_trajectory(x, length=200): |
| """ |
| Resamples a trajectory to a new length. |
| |
| Parameters: |
| x (np.ndarray): original trajectory, shape (N, 2) |
| length (int): length of resampled trajectory |
| |
| Returns: |
| np.ndarray: resampled trajectory, shape (length, 2) |
| """ |
| len_x = len(x) |
| time_steps = np.arange(length) * (len_x - 1) / (length - 1) |
| x = x.T |
| resampled_trajectory = np.zeros((2, length)) |
| for i in range(2): |
| resampled_trajectory[i] = np.interp(time_steps, np.arange(len_x), x[i]) |
| return resampled_trajectory.T |
|
|
|
|
| def time_warping(x, length=200): |
| """ |
| Resamples a trajectory to a new length. |
| """ |
| len_x = len(x) |
| time_steps = np.arange(length) * (len_x - 1) / (length - 1) |
| x = x.T |
| warped_trajectory = np.zeros((2, length)) |
| for i in range(2): |
| warped_trajectory[i] = np.interp(time_steps, np.arange(len_x), x[i]) |
| return warped_trajectory.T |
|
|
|
|
| def gather(consts: torch.Tensor, t: torch.Tensor): |
| """ |
| Gather consts for $t$ and reshape to feature map shape |
| :param consts: (N, 1, 1) |
| :param t: (N, H, W) |
| :return: (N, H, W) |
| """ |
| c = consts.gather(-1, t) |
| return c.reshape(-1, 1, 1) |
|
|
|
|
| def q_xt_x0(x0, t, alpha_bar): |
| |
| mean = gather(alpha_bar, t) ** 0.5 * x0 |
| var = 1 - gather(alpha_bar, t) |
| |
| eps = torch.randn_like(x0).to(x0.device) |
| xt = mean + (var ** 0.5) * eps |
| return xt, eps |
|
|
|
|
| def compute_alpha(beta, t): |
| beta = torch.cat([torch.zeros(1).to(beta.device), beta], dim=0) |
| a = (1 - beta).cumprod(dim=0).index_select(0, t + 1).view(-1, 1, 1) |
| return a |
|
|
|
|
| def p_xt(xt, noise, t, next_t, beta, eta=0): |
| at = compute_alpha(beta.cuda(), t.long()) |
| at_next = compute_alpha(beta, next_t.long()) |
| x0_t = (xt - noise * (1 - at).sqrt()) / at.sqrt() |
| c1 = (eta * ((1 - at / at_next) * (1 - at_next) / (1 - at)).sqrt()) |
| c2 = ((1 - at_next) - c1 ** 2).sqrt() |
| eps = torch.randn(xt.shape, device=xt.device) |
| xt_next = at_next.sqrt() * x0_t + c1 * eps + c2 * noise |
| return xt_next |
|
|
|
|
| def divide_grids(boundary, grids_num): |
| lati_min, lati_max = boundary['lati_min'], boundary['lati_max'] |
| long_min, long_max = boundary['long_min'], boundary['long_max'] |
| |
| lati_interval = (lati_max - lati_min) / grids_num |
| long_interval = (long_max - long_min) / grids_num |
| |
| latgrids = np.arange(lati_min, lati_max, lati_interval) |
| longrids = np.arange(long_min, long_max, long_interval) |
| return latgrids, longrids |
|
|
|
|
| |
| def distance(lat1, lon1, lat2, lon2): |
| """ |
| Calculate the great circle distance between two points |
| on the earth (specified in decimal degrees) |
| """ |
| |
| lon1, lat1, lon2, lat2 = map(radians, [lon1, lat1, lon2, lat2]) |
| |
| dlon = lon2 - lon1 |
| dlat = lat2 - lat1 |
| a = sin(dlat/2)**2 + cos(lat1) * cos(lat2) * sin(dlon/2)**2 |
| c = 2 * asin(sqrt(a)) |
| r = 6371 |
| return c * r * 1000 |
|
|
| def set_seed(seed): |
| random.seed(seed) |
| np.random.seed(seed) |
| torch.manual_seed(seed) |
| if torch.cuda.is_available(): |
| torch.cuda.manual_seed(seed) |
| torch.cuda.manual_seed_all(seed) |
| torch.backends.cudnn.deterministic = True |
| torch.backends.cudnn.benchmark = False |
| |
| def ddp_setup(): |
| init_process_group(backend="nccl") |
| torch.cuda.set_device(int(os.environ['LOCAL_RANK'])) |
|
|
|
|
| def destroy_process_group(): |
| destroy_process_group() |
|
|
|
|
| import torch |
|
|
| class IterativeKMeans: |
| def __init__(self, num_clusters, device, num_iters=100, tol=1e-4): |
| self.num_clusters = num_clusters |
| self.num_iters = num_iters |
| self.tol = tol |
| self.cluster_centers = None |
| self.labels = None |
| self.device = device |
|
|
| def fit(self, X): |
| |
| X = X.clone().detach().to(self.device) |
| num_samples, num_features = X.shape |
| indices = torch.randperm(num_samples)[:self.num_clusters] |
| self.cluster_centers = X[indices].clone().detach() |
| self.labels = torch.argmin(torch.cdist(X, self.cluster_centers), dim=1).cpu().numpy() |
|
|
| for _ in range(self.num_iters): |
| distances = torch.cdist(X, self.cluster_centers) |
| labels = torch.argmin(distances, dim=1) |
| |
| new_cluster_centers = torch.stack([X[labels == i].mean(dim=0) if (labels == i).sum() > 0 else self.cluster_centers[i] for i in range(self.num_clusters)]) |
| center_shift = torch.norm(new_cluster_centers - self.cluster_centers, dim=1).sum().item() |
| if center_shift < self.tol: |
| break |
| self.cluster_centers = new_cluster_centers |
|
|
| self.labels = labels.cpu().numpy() |
| return self.cluster_centers, self.labels |
|
|
| def update(self, new_X, original_X): |
| combined_X = torch.cat([original_X, new_X], dim=0) |
| combined_X = combined_X.clone().detach().to(self.device) |
|
|
| for _ in range(self.num_iters): |
| distances = torch.cdist(combined_X, self.cluster_centers) |
| labels = torch.argmin(distances, dim=1) |
| new_cluster_centers = torch.stack([combined_X[labels == i].mean(dim=0) if (labels == i).sum() > 0 else self.cluster_centers[i] for i in range(self.num_clusters)]) |
| center_shift = torch.norm(new_cluster_centers - self.cluster_centers, dim=1).sum().item() |
| if center_shift < self.tol: |
| break |
| self.cluster_centers = new_cluster_centers |
|
|
| self.labels = labels.cpu().numpy() |
| return self.cluster_centers, self.labels |
|
|
| def predict(self, X): |
| |
| X = X.clone().detach().to(self.device) |
| distances = torch.cdist(X, self.cluster_centers) |
| labels = torch.argmin(distances, dim=1) |
| return labels |
|
|
| def to(self, device): |
| self.device = device |
| if self.cluster_centers is not None: |
| self.cluster_centers = self.cluster_centers.to(device) |
| return self |
|
|
|
|
| def assign_labels(prototypes, features): |
| |
| distances = F.pairwise_distance(features.unsqueeze(1), prototypes.unsqueeze(0)) |
| |
| labels = torch.argmin(distances, dim=-1) |
|
|
| return labels |
|
|
|
|
| def get_positive_negative_pairs(prototypes, samples): |
| positive_pairs = [] |
| negative_pairs = [] |
| for sample in samples: |
| distances = F.pairwise_distance(sample.unsqueeze(0), prototypes) |
| pos_idx = torch.argmin(distances).item() |
| neg_idx = torch.argmax(distances).item() |
| positive_pairs.append(prototypes[pos_idx]) |
| negative_pairs.append(prototypes[neg_idx]) |
| return torch.stack(positive_pairs), torch.stack(negative_pairs) |
|
|
|
|
|
|
| def mask_data_general(x: torch.Tensor): |
| """Mask the input data""" |
| mask = torch.ones_like(x) |
| mask[:, :, 1:-1] = 0 |
| return x * mask.float() |
|
|
|
|
| def continuous_mask_data(x: torch.Tensor, mask_ratio: float = 0.5): |
| """ |
| Mask a continuous block of the input data. |
| It keeps the first and last elements unmasked. |
| """ |
| mask = torch.ones_like(x) |
| |
| traj_length = x.shape[2] |
| if traj_length <= 2: |
| return x * mask.float() |
|
|
| masked_length = int((traj_length - 2) * mask_ratio) |
| if masked_length == 0: |
| return x * mask.float() |
|
|
| |
| |
| mask_start = random.randint(1, traj_length - 2 - masked_length) |
| mask_end = mask_start + masked_length |
|
|
| mask[:, :, mask_start:mask_end] = 0 |
| return x * mask.float() |
|
|
|
|
| def update_npy(file_path, data): |
| if os.path.exists(file_path): |
| existing_data = np.load(file_path, allow_pickle=True).item() |
| existing_data.update(data) |
| else: |
| existing_data = data |
| np.save(file_path, existing_data) |
|
|
|
|
| def haversine(lat1, lon1, lat2, lon2): |
| |
| lat1, lon1, lat2, lon2 = map(np.radians, [lat1, lon1, lat2, lon2]) |
|
|
| |
| dlat = lat2 - lat1 |
| dlon = lon2 - lon1 |
| a = np.sin(dlat / 2.0)**2 + np.cos(lat1) * np.cos(lat2) * np.sin(dlon / 2.0)**2 |
| c = 2 * np.arcsin(np.sqrt(a)) |
| r = 6371 |
| return c * r * 1000 |
|
|
|
|
| def continuous_time_based_mask(x: torch.Tensor, points_to_mask: int): |
| """ |
| Mask a continuous block of the input data based on a fixed number of points. |
| It keeps the first and last elements unmasked. |
| |
| Args: |
| x (torch.Tensor): Input tensor of shape (batch, features, length). |
| points_to_mask (int): The number of continuous points to mask. |
| """ |
| mask = torch.ones_like(x) |
| |
| traj_length = x.shape[2] |
| |
| if traj_length <= points_to_mask + 2: |
| |
| return x * mask.float() |
|
|
| |
| |
| mask_start = random.randint(1, traj_length - 1 - points_to_mask) |
| mask_end = mask_start + points_to_mask |
|
|
| mask[:, :, mask_start:mask_end] = 0 |
| return x * mask.float() |
|
|
|
|
| def mask_multiple_segments(x: torch.Tensor, points_per_segment: list): |
| """ |
| Mask multiple non-overlapping continuous segments in the input data. |
| Keeps the first and last elements unmasked. |
| |
| Args: |
| x (torch.Tensor): Input tensor of shape (batch, features, length). |
| points_per_segment (list of int): List containing the length of each segment to mask. |
| """ |
| mask = torch.ones_like(x) |
| traj_length = x.shape[2] |
| |
| |
| segments = sorted(points_per_segment, reverse=True) |
| total_mask_points = sum(segments) |
|
|
| |
| if traj_length < total_mask_points + 2: |
| return x * mask.float() |
|
|
| |
| |
| possible_indices = list(range(1, traj_length - 1)) |
| |
| masked_intervals = [] |
|
|
| for seg_length in segments: |
| |
| valid_starts = [] |
| for i in possible_indices: |
| |
| |
| if i + seg_length > traj_length - 1: |
| continue |
| |
| is_valid = True |
| for start, end in masked_intervals: |
| |
| if not (i + seg_length <= start or i >= end): |
| is_valid = False |
| break |
| if is_valid: |
| valid_starts.append(i) |
| |
| if not valid_starts: |
| |
| continue |
|
|
| |
| start_pos = random.choice(valid_starts) |
| end_pos = start_pos + seg_length |
| mask[:, :, start_pos:end_pos] = 0 |
| |
| |
| masked_intervals.append((start_pos, end_pos)) |
| |
| |
| indices_to_remove = set(range(start_pos, end_pos)) |
| possible_indices = [idx for idx in possible_indices if idx not in indices_to_remove] |
|
|
| return x * mask.float() |
|
|
| def get_data_paths(data_config, for_train=True): |
| """Get the file paths for training or testing data for TKY-like structure. |
| Assumes data_config.traj_path1 points to a directory containing train.h5 and test.h5. |
| """ |
|
|
| base_path = data_config.traj_path1 |
| if not isinstance(base_path, str): |
| base_path = str(base_path) |
|
|
| |
| if hasattr(data_config, 'dataset') and 'temporal' in data_config.dataset: |
| if for_train: |
| file_path = os.path.join(base_path, "train_temporal.h5") |
| else: |
| file_path = os.path.join(base_path, "test_temporal.h5") |
| else: |
| |
| if for_train: |
| file_path = os.path.join(base_path, "train.h5") |
| else: |
| file_path = os.path.join(base_path, "test.h5") |
| |
|
|
| return [file_path] |