| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| A collection of "functional" transforms for spatial operations. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import math |
| import warnings |
| from enum import Enum |
|
|
| import numpy as np |
| import torch |
|
|
| import monai |
| from monai.config import USE_COMPILED |
| from monai.config.type_definitions import NdarrayOrTensor |
| from monai.data.meta_obj import get_track_meta |
| from monai.data.meta_tensor import MetaTensor |
| from monai.data.utils import AFFINE_TOL, compute_shape_offset, to_affine_nd |
| from monai.networks.layers import AffineTransform |
| from monai.transforms.croppad.array import ResizeWithPadOrCrop |
| from monai.transforms.intensity.array import GaussianSmooth |
| from monai.transforms.inverse import TraceableTransform |
| from monai.transforms.utils import create_rotate, create_translate, resolves_modes, scale_affine |
| from monai.transforms.utils_pytorch_numpy_unification import allclose |
| from monai.utils import ( |
| LazyAttr, |
| TraceKeys, |
| convert_to_dst_type, |
| convert_to_numpy, |
| convert_to_tensor, |
| ensure_tuple, |
| ensure_tuple_rep, |
| fall_back_tuple, |
| optional_import, |
| ) |
|
|
| nib, has_nib = optional_import("nibabel") |
| cupy, _ = optional_import("cupy") |
| cupy_ndi, _ = optional_import("cupyx.scipy.ndimage") |
| np_ndi, _ = optional_import("scipy.ndimage") |
|
|
| __all__ = ["spatial_resample", "orientation", "flip", "resize", "rotate", "zoom", "rotate90", "affine_func"] |
|
|
|
|
| def _maybe_new_metatensor(img, dtype=None, device=None): |
| """create a metatensor with fresh metadata if track_meta is True otherwise convert img into a torch tensor""" |
| return convert_to_tensor( |
| img.as_tensor() if isinstance(img, MetaTensor) else img, |
| dtype=dtype, |
| device=device, |
| track_meta=get_track_meta(), |
| wrap_sequence=True, |
| ) |
|
|
|
|
| def spatial_resample( |
| img, dst_affine, spatial_size, mode, padding_mode, align_corners, dtype_pt, lazy, transform_info |
| ) -> torch.Tensor: |
| """ |
| Functional implementation of resampling the input image to the specified ``dst_affine`` matrix and ``spatial_size``. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be resampled, assuming `img` is channel-first. |
| dst_affine: target affine matrix, if None, use the input affine matrix, effectively no resampling. |
| spatial_size: output spatial size, if the component is ``-1``, use the corresponding input spatial size. |
| mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| Interpolation mode to calculate output values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| and the value represents the order of the spline interpolation. |
| See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| dtype_pt: data `dtype` for resampling computation. |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| """ |
| original_spatial_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| src_affine: torch.Tensor = img.peek_pending_affine() if isinstance(img, MetaTensor) else torch.eye(4) |
| img = convert_to_tensor(data=img, track_meta=get_track_meta()) |
| |
| spatial_rank = min(len(img.shape) - 1, src_affine.shape[0] - 1, 3) |
| if (not isinstance(spatial_size, int) or spatial_size != -1) and spatial_size is not None: |
| spatial_rank = min(len(ensure_tuple(spatial_size)), 3) |
| src_affine = to_affine_nd(spatial_rank, src_affine).to(torch.float64) |
| dst_affine = to_affine_nd(spatial_rank, dst_affine) if dst_affine is not None else src_affine |
| dst_affine = convert_to_dst_type(dst_affine, src_affine)[0] |
| if not isinstance(dst_affine, torch.Tensor): |
| raise ValueError(f"dst_affine should be a torch.Tensor, got {type(dst_affine)}") |
|
|
| in_spatial_size = torch.tensor(original_spatial_shape[:spatial_rank]) |
| if isinstance(spatial_size, int) and (spatial_size == -1): |
| spatial_size = in_spatial_size |
| elif spatial_size is None and spatial_rank > 1: |
| spatial_size, _ = compute_shape_offset(in_spatial_size, src_affine, dst_affine) |
| spatial_size = torch.tensor( |
| fall_back_tuple(ensure_tuple(spatial_size)[:spatial_rank], in_spatial_size, lambda x: x >= 0) |
| ) |
| extra_info = { |
| "dtype": str(dtype_pt)[6:], |
| "mode": mode.value if isinstance(mode, Enum) else mode, |
| "padding_mode": padding_mode.value if isinstance(padding_mode, Enum) else padding_mode, |
| "align_corners": align_corners if align_corners is not None else TraceKeys.NONE, |
| "src_affine": src_affine, |
| } |
| try: |
| _s = convert_to_numpy(src_affine) |
| _d = convert_to_numpy(dst_affine) |
| xform = np.eye(spatial_rank + 1) if spatial_rank < 2 else np.linalg.solve(_s, _d) |
| except (np.linalg.LinAlgError, RuntimeError) as e: |
| raise ValueError(f"src affine is not invertible {_s}, {_d}.") from e |
| xform = convert_to_tensor(to_affine_nd(spatial_rank, xform)).to(device=img.device, dtype=torch.float64) |
| affine_unchanged = ( |
| allclose(src_affine, dst_affine, atol=AFFINE_TOL) and allclose(spatial_size, in_spatial_size) |
| ) or (allclose(xform, np.eye(len(xform)), atol=AFFINE_TOL) and allclose(spatial_size, in_spatial_size)) |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=spatial_size, |
| affine=None if affine_unchanged and not lazy else xform, |
| extra_info=extra_info, |
| orig_size=original_spatial_shape, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| if lazy: |
| out = _maybe_new_metatensor(img) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| if affine_unchanged: |
| |
| out = _maybe_new_metatensor(img, dtype=torch.float32) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
| |
| img = img.as_tensor() if isinstance(img, MetaTensor) else img |
| im_size = list(img.shape) |
| chns, in_sp_size, additional_dims = im_size[0], im_size[1 : spatial_rank + 1], im_size[spatial_rank + 1 :] |
|
|
| if additional_dims: |
| xform_shape = [-1] + in_sp_size |
| img = img.reshape(xform_shape) |
| img = img.to(dtype_pt) |
| if isinstance(mode, int) or USE_COMPILED: |
| dst_xform = create_translate(spatial_rank, [float(d - 1) / 2 for d in spatial_size]) |
| xform = xform @ convert_to_dst_type(dst_xform, xform)[0] |
| affine_xform = monai.transforms.Affine( |
| affine=xform, |
| spatial_size=spatial_size, |
| normalized=True, |
| image_only=True, |
| dtype=dtype_pt, |
| align_corners=align_corners, |
| ) |
| with affine_xform.trace_transform(False): |
| img = affine_xform(img, mode=mode, padding_mode=padding_mode) |
| else: |
| _, _m, _p, _ = resolves_modes(mode, padding_mode) |
| affine_xform = AffineTransform( |
| normalized=False, mode=_m, padding_mode=_p, align_corners=align_corners, reverse_indexing=True |
| ) |
| img = affine_xform(img.unsqueeze(0), theta=xform.to(img), spatial_size=spatial_size).squeeze(0) |
| if additional_dims: |
| full_shape = (chns, *spatial_size, *additional_dims) |
| img = img.reshape(full_shape) |
| out = _maybe_new_metatensor(img, dtype=torch.float32) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
|
|
|
|
| def orientation(img, original_affine, spatial_ornt, lazy, transform_info) -> torch.Tensor: |
| """ |
| Functional implementation of changing the input image's orientation into the specified based on `spatial_ornt`. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| original_affine: original affine of the input image. |
| spatial_ornt: orientations of the spatial axes, |
| see also https://nipy.org/nibabel/reference/nibabel.orientations.html |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| """ |
| spatial_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| xform = nib.orientations.inv_ornt_aff(spatial_ornt, spatial_shape) |
| img = convert_to_tensor(img, track_meta=get_track_meta()) |
|
|
| spatial_ornt[:, 0] += 1 |
| spatial_ornt = np.concatenate([np.array([[0, 1]]), spatial_ornt]) |
| axes = [ax for ax, flip in enumerate(spatial_ornt[:, 1]) if flip == -1] |
| full_transpose = np.arange(len(spatial_shape) + 1) |
| full_transpose[: len(spatial_ornt)] = np.argsort(spatial_ornt[:, 0]) |
| extra_info = {"original_affine": original_affine} |
|
|
| shape_np = convert_to_numpy(spatial_shape, wrap_sequence=True) |
| shape_np = shape_np[[i - 1 for i in full_transpose if i > 0]] |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=shape_np, |
| affine=xform, |
| extra_info=extra_info, |
| orig_size=spatial_shape, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| out = _maybe_new_metatensor(img) |
| if lazy: |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| if axes: |
| out = torch.flip(out, dims=axes) |
| if not np.all(full_transpose == np.arange(len(out.shape))): |
| out = out.permute(full_transpose.tolist()) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
|
|
|
|
| def flip(img, sp_axes, lazy, transform_info): |
| """ |
| Functional implementation of flip. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| sp_axes: spatial axes along which to flip over. |
| If None, will flip over all of the axes of the input array. |
| If axis is negative it counts from the last to the first axis. |
| If axis is a tuple of ints, flipping is performed on all of the axes |
| specified in the tuple. |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| """ |
| sp_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| sp_size = convert_to_numpy(sp_size, wrap_sequence=True).tolist() |
| extra_info = {"axes": sp_axes} |
| axes = monai.transforms.utils.map_spatial_axes(img.ndim, sp_axes) |
| rank = img.peek_pending_rank() if isinstance(img, MetaTensor) else torch.tensor(3.0, dtype=torch.double) |
| |
| xform = torch.eye(int(rank) + 1, dtype=torch.double) |
| for axis in axes: |
| sp = axis - 1 |
| xform[sp, sp], xform[sp, -1] = xform[sp, sp] * -1, sp_size[sp] - 1 |
| meta_info = TraceableTransform.track_transform_meta( |
| img, sp_size=sp_size, affine=xform, extra_info=extra_info, transform_info=transform_info, lazy=lazy |
| ) |
| out = _maybe_new_metatensor(img) |
| if lazy: |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| out = torch.flip(out, axes) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
|
|
|
|
| def resize( |
| img, out_size, mode, align_corners, dtype, input_ndim, anti_aliasing, anti_aliasing_sigma, lazy, transform_info |
| ): |
| """ |
| Functional implementation of resize. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| out_size: expected shape of spatial dimensions after resize operation. |
| mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, |
| ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. |
| dtype: data type for resampling computation. If None, use the data type of input data. |
| input_ndim: number of spatial dimensions. |
| anti_aliasing: whether to apply a Gaussian filter to smooth the image prior |
| to downsampling. It is crucial to filter when downsampling |
| the image to avoid aliasing artifacts. See also ``skimage.transform.resize`` |
| anti_aliasing_sigma: {float, tuple of floats}, optional |
| Standard deviation for Gaussian filtering used when anti-aliasing. |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| """ |
| img = convert_to_tensor(img, track_meta=get_track_meta()) |
| orig_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| extra_info = { |
| "mode": mode, |
| "align_corners": align_corners if align_corners is not None else TraceKeys.NONE, |
| "dtype": str(dtype)[6:], |
| "new_dim": len(orig_size) - input_ndim, |
| } |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=out_size, |
| affine=scale_affine(orig_size, out_size), |
| extra_info=extra_info, |
| orig_size=orig_size, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| if lazy: |
| if anti_aliasing and lazy: |
| warnings.warn("anti-aliasing is not compatible with lazy evaluation.") |
| out = _maybe_new_metatensor(img) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| if tuple(convert_to_numpy(orig_size)) == out_size: |
| out = _maybe_new_metatensor(img, dtype=torch.float32) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
| out = _maybe_new_metatensor(img) |
| img_ = convert_to_tensor(out, dtype=dtype, track_meta=False) |
| if anti_aliasing and any(x < y for x, y in zip(out_size, img_.shape[1:])): |
| factors = torch.div(torch.Tensor(list(img_.shape[1:])), torch.Tensor(out_size)) |
| if anti_aliasing_sigma is None: |
| |
| anti_aliasing_sigma = torch.maximum(torch.zeros(factors.shape), (factors - 1) / 2).tolist() |
| else: |
| |
| anti_aliasing_sigma = list(ensure_tuple_rep(anti_aliasing_sigma, len(out_size))) |
| for axis in range(len(out_size)): |
| anti_aliasing_sigma[axis] = anti_aliasing_sigma[axis] * int(factors[axis] > 1) |
| anti_aliasing_filter = GaussianSmooth(sigma=anti_aliasing_sigma) |
| img_ = convert_to_tensor(anti_aliasing_filter(img_), track_meta=False) |
| _, _m, _, _ = resolves_modes(mode, torch_interpolate_spatial_nd=len(img_.shape) - 1) |
| resized = torch.nn.functional.interpolate( |
| input=img_.unsqueeze(0), size=out_size, mode=_m, align_corners=align_corners |
| ) |
| out, *_ = convert_to_dst_type(resized.squeeze(0), out, dtype=torch.float32) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
|
|
|
|
| def rotate(img, angle, output_shape, mode, padding_mode, align_corners, dtype, lazy, transform_info): |
| """ |
| Functional implementation of rotate. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| angle: Rotation angle(s) in radians. should a float for 2D, three floats for 3D. |
| output_shape: output shape of the rotated data. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| align_corners: See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| dtype: data type for resampling computation. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``float32``. |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| |
| """ |
|
|
| im_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| input_ndim = len(im_shape) |
| if input_ndim not in (2, 3): |
| raise ValueError(f"Unsupported image dimension: {input_ndim}, available options are [2, 3].") |
| _angle = ensure_tuple_rep(angle, 1 if input_ndim == 2 else 3) |
| transform = create_rotate(input_ndim, _angle) |
| if output_shape is None: |
| corners = np.asarray(np.meshgrid(*[(0, dim) for dim in im_shape], indexing="ij")).reshape((len(im_shape), -1)) |
| corners = transform[:-1, :-1] @ corners |
| output_shape = np.asarray(corners.ptp(axis=1) + 0.5, dtype=int) |
| else: |
| output_shape = np.asarray(output_shape, dtype=int) |
| shift = create_translate(input_ndim, ((np.array(im_shape) - 1) / 2).tolist()) |
| shift_1 = create_translate(input_ndim, (-(np.asarray(output_shape, dtype=int) - 1) / 2).tolist()) |
| transform = shift @ transform @ shift_1 |
| extra_info = { |
| "rot_mat": transform, |
| "mode": mode, |
| "padding_mode": padding_mode, |
| "align_corners": align_corners if align_corners is not None else TraceKeys.NONE, |
| "dtype": str(dtype)[6:], |
| } |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=output_shape, |
| affine=transform, |
| extra_info=extra_info, |
| orig_size=im_shape, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| out = _maybe_new_metatensor(img) |
| if lazy: |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| _, _m, _p, _ = resolves_modes(mode, padding_mode) |
| xform = AffineTransform( |
| normalized=False, mode=_m, padding_mode=_p, align_corners=align_corners, reverse_indexing=True |
| ) |
| img_t = out.to(dtype) |
| transform_t, *_ = convert_to_dst_type(transform, img_t) |
| output: torch.Tensor = xform(img_t.unsqueeze(0), transform_t, spatial_size=tuple(int(i) for i in output_shape)) |
| output = output.float().squeeze(0) |
| out, *_ = convert_to_dst_type(output, dst=out, dtype=torch.float32) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
|
|
|
|
| def zoom(img, scale_factor, keep_size, mode, padding_mode, align_corners, dtype, lazy, transform_info): |
| """ |
| Functional implementation of zoom. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| scale_factor: The zoom factor along the spatial axes. |
| If a float, zoom is the same for each spatial axis. |
| If a sequence, zoom should contain one value for each spatial axis. |
| keep_size: Whether keep original size (padding/slicing if needed). |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| align_corners: See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| dtype: data type for resampling computation. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``float32``. |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| |
| """ |
| im_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| output_size = [int(math.floor(float(i) * z)) for i, z in zip(im_shape, scale_factor)] |
| xform = scale_affine(im_shape, output_size) |
| extra_info = { |
| "mode": mode, |
| "align_corners": align_corners if align_corners is not None else TraceKeys.NONE, |
| "dtype": str(dtype)[6:], |
| "do_padcrop": False, |
| "padcrop": {}, |
| } |
| if keep_size: |
| do_pad_crop = not np.allclose(output_size, im_shape) |
| if do_pad_crop and lazy: |
| _pad_crop = ResizeWithPadOrCrop(spatial_size=im_shape, mode=padding_mode) |
| _pad_crop.lazy = True |
| _tmp_img = MetaTensor([], affine=torch.eye(len(output_size) + 1)) |
| _tmp_img.push_pending_operation({LazyAttr.SHAPE: list(output_size), LazyAttr.AFFINE: xform}) |
| lazy_cropped = _pad_crop(_tmp_img) |
| if isinstance(lazy_cropped, MetaTensor): |
| xform = lazy_cropped.peek_pending_affine() |
| extra_info["padcrop"] = lazy_cropped.pending_operations[-1] |
| extra_info["do_padcrop"] = do_pad_crop |
| output_size = [int(i) for i in im_shape] |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=output_size, |
| affine=xform, |
| extra_info=extra_info, |
| orig_size=im_shape, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| out = _maybe_new_metatensor(img) |
| if lazy: |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| img_t = out.to(dtype) |
| _, _m, _, _ = resolves_modes(mode, torch_interpolate_spatial_nd=len(img_t.shape) - 1) |
| zoomed: NdarrayOrTensor = torch.nn.functional.interpolate( |
| recompute_scale_factor=True, |
| input=img_t.unsqueeze(0), |
| scale_factor=list(scale_factor), |
| mode=_m, |
| align_corners=align_corners, |
| ).squeeze(0) |
| out, *_ = convert_to_dst_type(zoomed, dst=out, dtype=torch.float32) |
| if isinstance(out, MetaTensor): |
| out = out.copy_meta_from(meta_info) |
| do_pad_crop = not np.allclose(output_size, zoomed.shape[1:]) |
| if do_pad_crop: |
| _pad_crop = ResizeWithPadOrCrop(spatial_size=img_t.shape[1:], mode=padding_mode) |
| out = _pad_crop(out) |
| if get_track_meta() and do_pad_crop: |
| padcrop_xform = out.applied_operations.pop() |
| out.applied_operations[-1]["extra_info"]["do_padcrop"] = True |
| out.applied_operations[-1]["extra_info"]["padcrop"] = padcrop_xform |
| return out |
|
|
|
|
| def rotate90(img, axes, k, lazy, transform_info): |
| """ |
| Functional implementation of rotate90. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| axes: 2 int numbers, defines the plane to rotate with 2 spatial axes. |
| If axis is negative it counts from the last to the first axis. |
| k: number of times to rotate by 90 degrees. |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| """ |
| extra_info = {"axes": [d - 1 for d in axes], "k": k} |
| ori_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| sp_shape = list(ori_shape) |
| if k in (1, 3): |
| a_0, a_1 = axes[0] - 1, axes[1] - 1 |
| sp_shape[a_0], sp_shape[a_1] = ori_shape[a_1], ori_shape[a_0] |
| rank = img.peek_pending_rank() if isinstance(img, MetaTensor) else torch.tensor(3.0, dtype=torch.double) |
| r, sp_r = int(rank), len(ori_shape) |
| xform = to_affine_nd(r, create_translate(sp_r, [-float(d - 1) / 2 for d in sp_shape])) |
| s = -1.0 if int(axes[0]) - int(axes[1]) in (-1, 2) else 1.0 |
| if sp_r == 2: |
| rot90 = to_affine_nd(r, create_rotate(sp_r, [s * np.pi / 2])) |
| else: |
| idx = {1, 2, 3} - set(axes) |
| angle: list[float] = [0, 0, 0] |
| angle[idx.pop() - 1] = s * np.pi / 2 |
| rot90 = to_affine_nd(r, create_rotate(sp_r, angle)) |
| for _ in range(k): |
| xform = rot90 @ xform |
| xform = to_affine_nd(r, create_translate(sp_r, [float(d - 1) / 2 for d in ori_shape])) @ xform |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=sp_shape, |
| affine=xform, |
| extra_info=extra_info, |
| orig_size=ori_shape, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| out = _maybe_new_metatensor(img) |
| if lazy: |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| out = torch.rot90(out, k, axes) |
| return out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
|
|
|
|
| def affine_func( |
| img, affine, grid, resampler, sp_size, mode, padding_mode, do_resampling, image_only, lazy, transform_info |
| ): |
| """ |
| Functional implementation of affine. |
| This function operates eagerly or lazily according to |
| ``lazy`` (default ``False``). |
| |
| Args: |
| img: data to be changed, assuming `img` is channel-first. |
| affine: the affine transformation to be applied, it can be a 3x3 or 4x4 matrix. This should be defined |
| for the voxel space spatial centers (``float(size - 1)/2``). |
| grid: used in non-lazy mode to pre-compute the grid to do the resampling. |
| resampler: the resampler function, see also: :py:class:`monai.transforms.Resample`. |
| sp_size: output image spatial size. |
| mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| Interpolation mode to calculate output values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| and the value represents the order of the spline interpolation. |
| See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. |
| See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| do_resampling: whether to do the resampling, this is a flag for the use case of updating metadata but |
| skipping the actual (potentially heavy) resampling operation. |
| image_only: if True return only the image volume, otherwise return (image, affine). |
| lazy: a flag that indicates whether the operation should be performed lazily or not |
| transform_info: a dictionary with the relevant information pertaining to an applied transform. |
| |
| """ |
|
|
| |
| img_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| rank = img.peek_pending_rank() if isinstance(img, MetaTensor) else torch.tensor(3.0, dtype=torch.double) |
| extra_info = { |
| "affine": affine, |
| "mode": mode, |
| "padding_mode": padding_mode, |
| "do_resampling": do_resampling, |
| "align_corners": resampler.align_corners, |
| } |
| affine = monai.transforms.Affine.compute_w_affine(rank, affine, img_size, sp_size) |
| meta_info = TraceableTransform.track_transform_meta( |
| img, |
| sp_size=sp_size, |
| affine=affine, |
| extra_info=extra_info, |
| orig_size=img_size, |
| transform_info=transform_info, |
| lazy=lazy, |
| ) |
| if lazy: |
| out = _maybe_new_metatensor(img) |
| out = out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else meta_info |
| return out if image_only else (out, affine) |
| if do_resampling: |
| out = resampler(img=img, grid=grid, mode=mode, padding_mode=padding_mode) |
| out = _maybe_new_metatensor(out) |
| else: |
| out = _maybe_new_metatensor(img, dtype=torch.float32, device=resampler.device) |
| out = out.copy_meta_from(meta_info) if isinstance(out, MetaTensor) else out |
| return out if image_only else (out, affine) |
|
|