| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | A collection of "vanilla" transforms for spatial operations. |
| | """ |
| |
|
| | from __future__ import annotations |
| |
|
| | import warnings |
| | from collections.abc import Callable |
| | from copy import deepcopy |
| | from itertools import zip_longest |
| | from typing import Any, Optional, Sequence, Tuple, Union, cast |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai.config import USE_COMPILED, DtypeLike |
| | from monai.config.type_definitions import NdarrayOrTensor |
| | from monai.data.meta_obj import get_track_meta, set_track_meta |
| | from monai.data.meta_tensor import MetaTensor |
| | from monai.data.utils import AFFINE_TOL, affine_to_spacing, compute_shape_offset, iter_patch, to_affine_nd, zoom_affine |
| | from monai.networks.layers import AffineTransform, GaussianFilter, grid_pull |
| | from monai.networks.utils import meshgrid_ij |
| | from monai.transforms.croppad.array import CenterSpatialCrop, ResizeWithPadOrCrop |
| | from monai.transforms.inverse import InvertibleTransform |
| | from monai.transforms.spatial.functional import ( |
| | affine_func, |
| | flip, |
| | orientation, |
| | resize, |
| | rotate, |
| | rotate90, |
| | spatial_resample, |
| | zoom, |
| | ) |
| | from monai.transforms.traits import MultiSampleTrait |
| | from monai.transforms.transform import LazyTransform, Randomizable, RandomizableTransform, Transform |
| | from monai.transforms.utils import ( |
| | create_control_grid, |
| | create_grid, |
| | create_rotate, |
| | create_scale, |
| | create_shear, |
| | create_translate, |
| | map_spatial_axes, |
| | resolves_modes, |
| | scale_affine, |
| | ) |
| | from monai.transforms.utils_pytorch_numpy_unification import argsort, argwhere, linalg_inv, moveaxis |
| | from monai.utils import ( |
| | GridSampleMode, |
| | GridSamplePadMode, |
| | InterpolateMode, |
| | NumpyPadMode, |
| | convert_to_cupy, |
| | convert_to_dst_type, |
| | convert_to_numpy, |
| | convert_to_tensor, |
| | ensure_tuple, |
| | ensure_tuple_rep, |
| | ensure_tuple_size, |
| | fall_back_tuple, |
| | issequenceiterable, |
| | optional_import, |
| | ) |
| | from monai.utils.enums import GridPatchSort, PatchKeys, TraceKeys, TransformBackends |
| | from monai.utils.misc import ImageMetaKey as Key |
| | from monai.utils.module import look_up_option |
| | from monai.utils.type_conversion import convert_data_type, get_equivalent_dtype, get_torch_dtype_from_string |
| |
|
| | nib, has_nib = optional_import("nibabel") |
| | cupy, _ = optional_import("cupy") |
| | cupy_ndi, _ = optional_import("cupyx.scipy.ndimage") |
| | np_ndi, _ = optional_import("scipy.ndimage") |
| |
|
| | __all__ = [ |
| | "SpatialResample", |
| | "ResampleToMatch", |
| | "Spacing", |
| | "Orientation", |
| | "Flip", |
| | "GridDistortion", |
| | "GridSplit", |
| | "GridPatch", |
| | "RandGridPatch", |
| | "Resize", |
| | "Rotate", |
| | "Zoom", |
| | "Rotate90", |
| | "RandRotate90", |
| | "RandRotate", |
| | "RandFlip", |
| | "RandGridDistortion", |
| | "RandAxisFlip", |
| | "RandZoom", |
| | "AffineGrid", |
| | "RandAffineGrid", |
| | "RandDeformGrid", |
| | "Resample", |
| | "Affine", |
| | "RandAffine", |
| | "Rand2DElastic", |
| | "Rand3DElastic", |
| | "RandSimulateLowResolution", |
| | ] |
| |
|
| | RandRange = Optional[Union[Sequence[Union[Tuple[float, float], float]], float]] |
| |
|
| |
|
| | class SpatialResample(InvertibleTransform, LazyTransform): |
| | """ |
| | Resample input image from the orientation/spacing defined by ``src_affine`` affine matrix into |
| | the ones specified by ``dst_affine`` affine matrix. |
| | |
| | Internally this transform computes the affine transform matrix from ``src_affine`` to ``dst_affine``, |
| | by ``xform = linalg.solve(src_affine, dst_affine)``, and call ``monai.transforms.Affine`` with ``xform``. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY, TransformBackends.CUPY] |
| |
|
| | def __init__( |
| | self, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | align_corners: bool = False, |
| | dtype: DtypeLike = np.float64, |
| | lazy: bool = False, |
| | ): |
| | """ |
| | Args: |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | dtype: data type for resampling computation. Defaults to ``float64`` for best precision. |
| | If ``None``, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.mode = mode |
| | self.padding_mode = padding_mode |
| | self.align_corners = align_corners |
| | self.dtype = dtype |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | dst_affine: torch.Tensor | None = None, |
| | spatial_size: Sequence[int] | torch.Tensor | int | None = None, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: input image to be resampled. It currently supports channel-first arrays with |
| | at most three spatial dimensions. |
| | dst_affine: destination affine matrix. Defaults to ``None``, which means the same as `img.affine`. |
| | the shape should be `(r+1, r+1)` where `r` is the spatial rank of ``img``. |
| | when `dst_affine` and `spatial_size` are None, the input will be returned without resampling, |
| | but the data type will be `float32`. |
| | spatial_size: output image spatial size. |
| | if `spatial_size` and `self.spatial_size` are not defined, |
| | the transform will compute a spatial size automatically containing the previous field of view. |
| | if `spatial_size` is ``-1`` are the transform will use the corresponding input img size. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | Defaults to ``None``, effectively using the value of `self.align_corners`. |
| | dtype: data type for resampling computation. Defaults to ``self.dtype`` or |
| | ``np.float64`` (for best precision). If ``None``, use the data type of input data. |
| | To be compatible with other modules, the output data type is always `float32`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | The spatial rank is determined by the smallest among ``img.ndim -1``, ``len(src_affine) - 1``, and ``3``. |
| | |
| | When both ``monai.config.USE_COMPILED`` and ``align_corners`` are set to ``True``, |
| | MONAI's resampling implementation will be used. |
| | Set `dst_affine` and `spatial_size` to `None` to turn off the resampling step. |
| | """ |
| | |
| | dtype_pt = get_equivalent_dtype(dtype or self.dtype or img.dtype, torch.Tensor) |
| | align_corners = align_corners if align_corners is not None else self.align_corners |
| | mode = mode if mode is not None else self.mode |
| | padding_mode = padding_mode if padding_mode is not None else self.padding_mode |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return spatial_resample( |
| | img, |
| | dst_affine, |
| | spatial_size, |
| | mode, |
| | padding_mode, |
| | align_corners, |
| | dtype_pt, |
| | lazy=lazy_, |
| | transform_info=self.get_transform_info(), |
| | ) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | |
| | kw_args = transform[TraceKeys.EXTRA_INFO] |
| | |
| | kw_args["dtype"] = get_torch_dtype_from_string(kw_args["dtype"]) |
| | |
| | kw_args["dst_affine"] = kw_args.pop("src_affine") |
| | kw_args["spatial_size"] = transform[TraceKeys.ORIG_SIZE] |
| | if kw_args.get("align_corners") == TraceKeys.NONE: |
| | kw_args["align_corners"] = False |
| | with self.trace_transform(False): |
| | |
| | out: torch.Tensor = SpatialResample.__call__(self, data, **kw_args) |
| | kw_args["src_affine"] = kw_args.get("dst_affine") |
| | return out |
| |
|
| |
|
| | class ResampleToMatch(SpatialResample): |
| | """ |
| | Resample an image to match given metadata. The affine matrix will be aligned, |
| | and the size of the output image will match. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | img_dst: torch.Tensor, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: input image to be resampled to match ``img_dst``. It currently supports channel-first arrays with |
| | at most three spatial dimensions. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| | Defaults to ``None``, effectively using the value of `self.align_corners`. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``self.dtype`` or |
| | ``np.float64`` (for best precision). If ``None``, use the data type of input data. |
| | To be compatible with other modules, the output data type is always `float32`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | |
| | Raises: |
| | ValueError: When the affine matrix of the source image is not invertible. |
| | Returns: |
| | Resampled input tensor or MetaTensor. |
| | """ |
| | if img_dst is None: |
| | raise RuntimeError("`img_dst` is missing.") |
| | dst_affine = img_dst.peek_pending_affine() if isinstance(img_dst, MetaTensor) else torch.eye(4) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | img = super().__call__( |
| | img=img, |
| | dst_affine=dst_affine, |
| | spatial_size=img_dst.peek_pending_shape() if isinstance(img_dst, MetaTensor) else img_dst.shape[1:], |
| | mode=mode, |
| | padding_mode=padding_mode, |
| | align_corners=align_corners, |
| | dtype=dtype, |
| | lazy=lazy_, |
| | ) |
| | if not lazy_: |
| | if isinstance(img, MetaTensor): |
| | img.affine = dst_affine |
| | if isinstance(img_dst, MetaTensor): |
| | original_fname = img.meta.get(Key.FILENAME_OR_OBJ, "resample_to_match_source") |
| | img.meta = deepcopy(img_dst.meta) |
| | img.meta[Key.FILENAME_OR_OBJ] = original_fname |
| | else: |
| | if isinstance(img, MetaTensor) and isinstance(img_dst, MetaTensor): |
| | original_fname = img.meta.get(Key.FILENAME_OR_OBJ, "resample_to_match_source") |
| | meta_dict = deepcopy(img_dst.meta) |
| | for k in ("affine", "spatial_shape"): |
| | meta_dict.pop(k, None) |
| | img.meta.update(meta_dict) |
| | img.meta[Key.FILENAME_OR_OBJ] = original_fname |
| | return img |
| |
|
| |
|
| | class Spacing(InvertibleTransform, LazyTransform): |
| | """ |
| | Resample input image into the specified `pixdim`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = SpatialResample.backend |
| |
|
| | def __init__( |
| | self, |
| | pixdim: Sequence[float] | float | np.ndarray, |
| | diagonal: bool = False, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | align_corners: bool = False, |
| | dtype: DtypeLike = np.float64, |
| | scale_extent: bool = False, |
| | recompute_affine: bool = False, |
| | min_pixdim: Sequence[float] | float | np.ndarray | None = None, |
| | max_pixdim: Sequence[float] | float | np.ndarray | None = None, |
| | lazy: bool = False, |
| | ) -> None: |
| | """ |
| | Args: |
| | pixdim: output voxel spacing. if providing a single number, will use it for the first dimension. |
| | items of the pixdim sequence map to the spatial dimensions of input image, if length |
| | of pixdim sequence is longer than image spatial dimensions, will ignore the longer part, |
| | if shorter, will pad with the last value. For example, for 3D image if pixdim is [1.0, 2.0] it |
| | will be padded to [1.0, 2.0, 2.0] |
| | if the components of the `pixdim` are non-positive values, the transform will use the |
| | corresponding components of the original pixdim, which is computed from the `affine` |
| | matrix of input image. |
| | diagonal: whether to resample the input to have a diagonal affine matrix. |
| | If True, the input data is resampled to the following affine:: |
| | |
| | np.diag((pixdim_0, pixdim_1, ..., pixdim_n, 1)) |
| | |
| | This effectively resets the volume to the world coordinate system (RAS+ in nibabel). |
| | The original orientation, rotation, shearing are not preserved. |
| | |
| | If False, this transform preserves the axes orientation, orthogonal rotation and |
| | translation components from the original affine. This option will not flip/swap axes |
| | of the original data. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``float64`` for best precision. |
| | If None, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | scale_extent: whether the scale is computed based on the spacing or the full extent of voxels, |
| | default False. The option is ignored if output spatial size is specified when calling this transform. |
| | See also: :py:func:`monai.data.utils.compute_shape_offset`. When this is True, `align_corners` |
| | should be `True` because `compute_shape_offset` already provides the corner alignment shift/scaling. |
| | recompute_affine: whether to recompute affine based on the output shape. The affine computed |
| | analytically does not reflect the potential quantization errors in terms of the output shape. |
| | Set this flag to True to recompute the output affine based on the actual pixdim. Default to ``False``. |
| | min_pixdim: minimal input spacing to be resampled. If provided, input image with a larger spacing than this |
| | value will be kept in its original spacing (not be resampled to `pixdim`). Set it to `None` to use the |
| | value of `pixdim`. Default to `None`. |
| | max_pixdim: maximal input spacing to be resampled. If provided, input image with a smaller spacing than this |
| | value will be kept in its original spacing (not be resampled to `pixdim`). Set it to `None` to use the |
| | value of `pixdim`. Default to `None`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.pixdim = np.array(ensure_tuple(pixdim), dtype=np.float64) |
| | self.min_pixdim = np.array(ensure_tuple(min_pixdim), dtype=np.float64) |
| | self.max_pixdim = np.array(ensure_tuple(max_pixdim), dtype=np.float64) |
| | self.diagonal = diagonal |
| | self.scale_extent = scale_extent |
| | self.recompute_affine = recompute_affine |
| |
|
| | for mn, mx in zip(self.min_pixdim, self.max_pixdim): |
| | if (not np.isnan(mn)) and (not np.isnan(mx)) and ((mx < mn) or (mn < 0)): |
| | raise ValueError(f"min_pixdim {self.min_pixdim} must be positive, smaller than max {self.max_pixdim}.") |
| |
|
| | self.sp_resample = SpatialResample( |
| | mode=mode, padding_mode=padding_mode, align_corners=align_corners, dtype=dtype, lazy=lazy |
| | ) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, val: bool) -> None: |
| | self._lazy = val |
| | self.sp_resample.lazy = val |
| |
|
| | def __call__( |
| | self, |
| | data_array: torch.Tensor, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike = None, |
| | scale_extent: bool | None = None, |
| | output_spatial_shape: Sequence[int] | np.ndarray | int | None = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | data_array: in shape (num_channels, H[, W, ...]). |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"self.mode"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"self.padding_mode"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | Defaults to ``None``, effectively using the value of `self.align_corners`. |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | If None, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | scale_extent: whether the scale is computed based on the spacing or the full extent of voxels, |
| | The option is ignored if output spatial size is specified when calling this transform. |
| | See also: :py:func:`monai.data.utils.compute_shape_offset`. When this is True, `align_corners` |
| | should be `True` because `compute_shape_offset` already provides the corner alignment shift/scaling. |
| | output_spatial_shape: specify the shape of the output data_array. This is typically useful for |
| | the inverse of `Spacingd` where sometimes we could not compute the exact shape due to the quantization |
| | error with the affine. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | |
| | Raises: |
| | ValueError: When ``data_array`` has no spatial dimensions. |
| | ValueError: When ``pixdim`` is nonpositive. |
| | |
| | Returns: |
| | data tensor or MetaTensor (resampled into `self.pixdim`). |
| | |
| | """ |
| | original_spatial_shape = ( |
| | data_array.peek_pending_shape() if isinstance(data_array, MetaTensor) else data_array.shape[1:] |
| | ) |
| | sr = len(original_spatial_shape) |
| | if sr <= 0: |
| | raise ValueError(f"data_array must have at least one spatial dimension, got {original_spatial_shape}.") |
| | affine_: np.ndarray |
| | input_affine = data_array.peek_pending_affine() if isinstance(data_array, MetaTensor) else None |
| | if input_affine is None: |
| | warnings.warn("`data_array` is not of type MetaTensor, assuming affine to be identity.") |
| | |
| | input_affine = np.eye(sr + 1, dtype=np.float64) |
| | affine_ = to_affine_nd(sr, convert_data_type(input_affine, np.ndarray)[0]) |
| |
|
| | out_d = self.pixdim[:sr].copy() |
| | if out_d.size < sr: |
| | out_d = np.append(out_d, [out_d[-1]] * (sr - out_d.size)) |
| |
|
| | orig_d = affine_to_spacing(affine_, sr, out_d.dtype) |
| | for idx, (_d, mn, mx) in enumerate( |
| | zip_longest(orig_d, self.min_pixdim[:sr], self.max_pixdim[:sr], fillvalue=np.nan) |
| | ): |
| | target = out_d[idx] |
| | mn = target if np.isnan(mn) else min(mn, target) |
| | mx = target if np.isnan(mx) else max(mx, target) |
| | if mn > mx: |
| | raise ValueError(f"min_pixdim is larger than max_pixdim at dim {idx}: min {mn} max {mx} out {target}.") |
| | out_d[idx] = _d if (mn - AFFINE_TOL) <= _d <= (mx + AFFINE_TOL) else target |
| |
|
| | if not align_corners and scale_extent: |
| | warnings.warn("align_corners=False is not compatible with scale_extent=True.") |
| |
|
| | |
| | new_affine = zoom_affine(affine_, out_d, diagonal=self.diagonal) |
| | scale_extent = self.scale_extent if scale_extent is None else scale_extent |
| | output_shape, offset = compute_shape_offset(original_spatial_shape, affine_, new_affine, scale_extent) |
| | new_affine[:sr, -1] = offset[:sr] |
| |
|
| | actual_shape = list(output_shape) if output_spatial_shape is None else output_spatial_shape |
| | lazy_ = self.lazy if lazy is None else lazy |
| | data_array = self.sp_resample( |
| | data_array, |
| | dst_affine=torch.as_tensor(new_affine), |
| | spatial_size=actual_shape, |
| | mode=mode, |
| | padding_mode=padding_mode, |
| | align_corners=align_corners, |
| | dtype=dtype, |
| | lazy=lazy_, |
| | ) |
| | if self.recompute_affine and isinstance(data_array, MetaTensor): |
| | if lazy_: |
| | raise NotImplementedError("recompute_affine is not supported with lazy evaluation.") |
| | a = scale_affine(original_spatial_shape, actual_shape) |
| | data_array.affine = convert_to_dst_type(a, affine_)[0] |
| | return data_array |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | return self.sp_resample.inverse(data) |
| |
|
| |
|
| | class Orientation(InvertibleTransform, LazyTransform): |
| | """ |
| | Change the input image's orientation into the specified based on `axcodes`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = [TransformBackends.NUMPY, TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | axcodes: str | None = None, |
| | as_closest_canonical: bool = False, |
| | labels: Sequence[tuple[str, str]] | None = (("L", "R"), ("P", "A"), ("I", "S")), |
| | lazy: bool = False, |
| | ) -> None: |
| | """ |
| | Args: |
| | axcodes: N elements sequence for spatial ND input's orientation. |
| | e.g. axcodes='RAS' represents 3D orientation: |
| | (Left, Right), (Posterior, Anterior), (Inferior, Superior). |
| | default orientation labels options are: 'L' and 'R' for the first dimension, |
| | 'P' and 'A' for the second, 'I' and 'S' for the third. |
| | as_closest_canonical: if True, load the image as closest to canonical axis format. |
| | labels: optional, None or sequence of (2,) sequences |
| | (2,) sequences are labels for (beginning, end) of output axis. |
| | Defaults to ``(('L', 'R'), ('P', 'A'), ('I', 'S'))``. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | |
| | Raises: |
| | ValueError: When ``axcodes=None`` and ``as_closest_canonical=True``. Incompatible values. |
| | |
| | See Also: `nibabel.orientations.ornt2axcodes`. |
| | |
| | """ |
| | LazyTransform.__init__(self, lazy=lazy) |
| | if axcodes is None and not as_closest_canonical: |
| | raise ValueError("Incompatible values: axcodes=None and as_closest_canonical=True.") |
| | if axcodes is not None and as_closest_canonical: |
| | warnings.warn("using as_closest_canonical=True, axcodes ignored.") |
| | self.axcodes = axcodes |
| | self.as_closest_canonical = as_closest_canonical |
| | self.labels = labels |
| |
|
| | def __call__(self, data_array: torch.Tensor, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | If input type is `MetaTensor`, original affine is extracted with `data_array.affine`. |
| | If input type is `torch.Tensor`, original affine is assumed to be identity. |
| | |
| | Args: |
| | data_array: in shape (num_channels, H[, W, ...]). |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | |
| | Raises: |
| | ValueError: When ``data_array`` has no spatial dimensions. |
| | ValueError: When ``axcodes`` spatiality differs from ``data_array``. |
| | |
| | Returns: |
| | data_array [reoriented in `self.axcodes`]. Output type will be `MetaTensor` |
| | unless `get_track_meta() == False`, in which case it will be |
| | `torch.Tensor`. |
| | |
| | """ |
| | spatial_shape = data_array.peek_pending_shape() if isinstance(data_array, MetaTensor) else data_array.shape[1:] |
| | sr = len(spatial_shape) |
| | if sr <= 0: |
| | raise ValueError(f"data_array must have at least one spatial dimension, got {spatial_shape}.") |
| | affine_: np.ndarray |
| | affine_np: np.ndarray |
| | if isinstance(data_array, MetaTensor): |
| | affine_np, *_ = convert_data_type(data_array.peek_pending_affine(), np.ndarray) |
| | affine_ = to_affine_nd(sr, affine_np) |
| | else: |
| | warnings.warn("`data_array` is not of type `MetaTensor, assuming affine to be identity.") |
| | |
| | affine_np = np.eye(sr + 1, dtype=np.float64) |
| | affine_ = np.eye(sr + 1, dtype=np.float64) |
| |
|
| | src = nib.io_orientation(affine_) |
| | if self.as_closest_canonical: |
| | spatial_ornt = src |
| | else: |
| | if self.axcodes is None: |
| | raise ValueError("Incompatible values: axcodes=None and as_closest_canonical=True.") |
| | if sr < len(self.axcodes): |
| | warnings.warn( |
| | f"axcodes ('{self.axcodes}') length is smaller than number of input spatial dimensions D={sr}.\n" |
| | f"{self.__class__.__name__}: spatial shape = {spatial_shape}, channels = {data_array.shape[0]}," |
| | "please make sure the input is in the channel-first format." |
| | ) |
| | dst = nib.orientations.axcodes2ornt(self.axcodes[:sr], labels=self.labels) |
| | if len(dst) < sr: |
| | raise ValueError( |
| | f"axcodes must match data_array spatially, got axcodes={len(self.axcodes)}D data_array={sr}D" |
| | ) |
| | spatial_ornt = nib.orientations.ornt_transform(src, dst) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return orientation(data_array, affine_np, spatial_ornt, lazy=lazy_, transform_info=self.get_transform_info()) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | |
| | orig_affine = transform[TraceKeys.EXTRA_INFO]["original_affine"] |
| | orig_axcodes = nib.orientations.aff2axcodes(orig_affine) |
| | inverse_transform = Orientation(axcodes=orig_axcodes, as_closest_canonical=False, labels=self.labels) |
| | |
| | with inverse_transform.trace_transform(False): |
| | data = inverse_transform(data) |
| |
|
| | return data |
| |
|
| |
|
| | class Flip(InvertibleTransform, LazyTransform): |
| | """ |
| | Reverses the order of elements along the given spatial axis. Preserves shape. |
| | See `torch.flip` documentation for additional details: |
| | https://pytorch.org/docs/stable/generated/torch.flip.html |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_axis: spatial axes along which to flip over. Default is None. |
| | The default `axis=None` will flip over all of the axes of the input array. |
| | If axis is negative it counts from the last to the first axis. |
| | If axis is a tuple of ints, flipping is performed on all of the axes |
| | specified in the tuple. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__(self, spatial_axis: Sequence[int] | int | None = None, lazy: bool = False) -> None: |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.spatial_axis = spatial_axis |
| |
|
| | def __call__(self, img: torch.Tensor, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]) |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return flip(img, self.spatial_axis, lazy=lazy_, transform_info=self.get_transform_info()) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | self.pop_transform(data) |
| | flipper = Flip(spatial_axis=self.spatial_axis) |
| | with flipper.trace_transform(False): |
| | return flipper(data) |
| |
|
| |
|
| | class Resize(InvertibleTransform, LazyTransform): |
| | """ |
| | Resize the input image to given spatial size (with scaling, not cropping/padding). |
| | Implemented using :py:class:`torch.nn.functional.interpolate`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_size: expected shape of spatial dimensions after resize operation. |
| | if some components of the `spatial_size` are non-positive values, the transform will use the |
| | corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| | to `(32, 64)` if the second spatial dimension size of img is `64`. |
| | size_mode: should be "all" or "longest", if "all", will use `spatial_size` for all the spatial dims, |
| | if "longest", rescale the image so that only the longest side is equal to specified `spatial_size`, |
| | which must be an int number in this case, keeping the aspect ratio of the initial image, refer to: |
| | https://albumentations.ai/docs/api_reference/augmentations/geometric/resize/ |
| | #albumentations.augmentations.geometric.resize.LongestMaxSize. |
| | mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| | The interpolation mode. Defaults to ``"area"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | align_corners: This only has an effect when mode is |
| | 'linear', 'bilinear', 'bicubic' or 'trilinear'. Default: None. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | anti_aliasing: bool |
| | Whether to apply a Gaussian filter to smooth the image prior |
| | to downsampling. It is crucial to filter when downsampling |
| | the image to avoid aliasing artifacts. See also ``skimage.transform.resize`` |
| | anti_aliasing_sigma: {float, tuple of floats}, optional |
| | Standard deviation for Gaussian filtering used when anti-aliasing. |
| | By default, this value is chosen as (s - 1) / 2 where s is the |
| | downsampling factor, where s > 1. For the up-size case, s < 1, no |
| | anti-aliasing is performed prior to rescaling. |
| | dtype: data type for resampling computation. Defaults to ``float32``. |
| | If None, use the data type of input data. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | spatial_size: Sequence[int] | int, |
| | size_mode: str = "all", |
| | mode: str = InterpolateMode.AREA, |
| | align_corners: bool | None = None, |
| | anti_aliasing: bool = False, |
| | anti_aliasing_sigma: Sequence[float] | float | None = None, |
| | dtype: DtypeLike | torch.dtype = torch.float32, |
| | lazy: bool = False, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.size_mode = look_up_option(size_mode, ["all", "longest"]) |
| | self.spatial_size = spatial_size |
| | self.mode = mode |
| | self.align_corners = align_corners |
| | self.anti_aliasing = anti_aliasing |
| | self.anti_aliasing_sigma = anti_aliasing_sigma |
| | self.dtype = dtype |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | mode: str | None = None, |
| | align_corners: bool | None = None, |
| | anti_aliasing: bool | None = None, |
| | anti_aliasing_sigma: Sequence[float] | float | None = None, |
| | dtype: DtypeLike | torch.dtype = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]). |
| | mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, |
| | ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| | The interpolation mode. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | align_corners: This only has an effect when mode is |
| | 'linear', 'bilinear', 'bicubic' or 'trilinear'. Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | anti_aliasing: bool, optional |
| | Whether to apply a Gaussian filter to smooth the image prior |
| | to downsampling. It is crucial to filter when downsampling |
| | the image to avoid aliasing artifacts. See also ``skimage.transform.resize`` |
| | anti_aliasing_sigma: {float, tuple of floats}, optional |
| | Standard deviation for Gaussian filtering used when anti-aliasing. |
| | By default, this value is chosen as (s - 1) / 2 where s is the |
| | downsampling factor, where s > 1. For the up-size case, s < 1, no |
| | anti-aliasing is performed prior to rescaling. |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | If None, use the data type of input data. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | Raises: |
| | ValueError: When ``self.spatial_size`` length is less than ``img`` spatial dimensions. |
| | |
| | """ |
| | anti_aliasing = self.anti_aliasing if anti_aliasing is None else anti_aliasing |
| | anti_aliasing_sigma = self.anti_aliasing_sigma if anti_aliasing_sigma is None else anti_aliasing_sigma |
| |
|
| | input_ndim = img.ndim - 1 |
| | if self.size_mode == "all": |
| | output_ndim = len(ensure_tuple(self.spatial_size)) |
| | if output_ndim > input_ndim: |
| | input_shape = ensure_tuple_size(img.shape, output_ndim + 1, 1) |
| | img = img.reshape(input_shape) |
| | elif output_ndim < input_ndim: |
| | raise ValueError( |
| | "len(spatial_size) must be greater or equal to img spatial dimensions, " |
| | f"got spatial_size={output_ndim} img={input_ndim}." |
| | ) |
| | _sp = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | sp_size = fall_back_tuple(self.spatial_size, _sp) |
| | else: |
| | img_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | if not isinstance(self.spatial_size, int): |
| | raise ValueError("spatial_size must be an int number if size_mode is 'longest'.") |
| | scale = self.spatial_size / max(img_size) |
| | sp_size = tuple(int(round(s * scale)) for s in img_size) |
| |
|
| | _mode = self.mode if mode is None else mode |
| | _align_corners = self.align_corners if align_corners is None else align_corners |
| | _dtype = get_equivalent_dtype(dtype or self.dtype or img.dtype, torch.Tensor) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return resize( |
| | img, |
| | tuple(int(_s) for _s in sp_size), |
| | _mode, |
| | _align_corners, |
| | _dtype, |
| | input_ndim, |
| | anti_aliasing, |
| | anti_aliasing_sigma, |
| | lazy_, |
| | self.get_transform_info(), |
| | ) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | return self.inverse_transform(data, transform) |
| |
|
| | def inverse_transform(self, data: torch.Tensor, transform) -> torch.Tensor: |
| | orig_size = transform[TraceKeys.ORIG_SIZE] |
| | mode = transform[TraceKeys.EXTRA_INFO]["mode"] |
| | align_corners = transform[TraceKeys.EXTRA_INFO]["align_corners"] |
| | dtype = transform[TraceKeys.EXTRA_INFO]["dtype"] |
| | xform = Resize( |
| | spatial_size=orig_size, |
| | mode=mode, |
| | align_corners=None if align_corners == TraceKeys.NONE else align_corners, |
| | dtype=dtype, |
| | ) |
| | with xform.trace_transform(False): |
| | data = xform(data) |
| | for _ in range(transform[TraceKeys.EXTRA_INFO]["new_dim"]): |
| | data = data.squeeze(-1) |
| | return data |
| |
|
| |
|
| | class Rotate(InvertibleTransform, LazyTransform): |
| | """ |
| | Rotates an input image by given angle using :py:class:`monai.networks.layers.AffineTransform`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | angle: Rotation angle(s) in radians. should a float for 2D, three floats for 3D. |
| | keep_size: If it is True, the output shape is kept the same as the input. |
| | If it is False, the output shape is adapted so that the |
| | input array is contained completely in the output. Default is True. |
| | mode: {``"bilinear"``, ``"nearest"``} |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | align_corners: Defaults to False. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``float32``. |
| | If None, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | angle: Sequence[float] | float, |
| | keep_size: bool = True, |
| | mode: str = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | align_corners: bool = False, |
| | dtype: DtypeLike | torch.dtype = torch.float32, |
| | lazy: bool = False, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.angle = angle |
| | self.keep_size = keep_size |
| | self.mode: str = mode |
| | self.padding_mode: str = padding_mode |
| | self.align_corners = align_corners |
| | self.dtype = dtype |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | mode: str | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike | torch.dtype = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: [chns, H, W] or [chns, H, W, D]. |
| | mode: {``"bilinear"``, ``"nearest"``} |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | align_corners: Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | align_corners: Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | If None, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | |
| | Raises: |
| | ValueError: When ``img`` spatially is not one of [2D, 3D]. |
| | |
| | """ |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | _dtype = get_equivalent_dtype(dtype or self.dtype or img.dtype, torch.Tensor) |
| | _mode = mode or self.mode |
| | _padding_mode = padding_mode or self.padding_mode |
| | _align_corners = self.align_corners if align_corners is None else align_corners |
| | im_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | output_shape = im_shape if self.keep_size else None |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return rotate( |
| | img, |
| | self.angle, |
| | output_shape, |
| | _mode, |
| | _padding_mode, |
| | _align_corners, |
| | _dtype, |
| | lazy=lazy_, |
| | transform_info=self.get_transform_info(), |
| | ) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | return self.inverse_transform(data, transform) |
| |
|
| | def inverse_transform(self, data: torch.Tensor, transform) -> torch.Tensor: |
| | fwd_rot_mat = transform[TraceKeys.EXTRA_INFO]["rot_mat"] |
| | mode = transform[TraceKeys.EXTRA_INFO]["mode"] |
| | padding_mode = transform[TraceKeys.EXTRA_INFO]["padding_mode"] |
| | align_corners = transform[TraceKeys.EXTRA_INFO]["align_corners"] |
| | dtype = transform[TraceKeys.EXTRA_INFO]["dtype"] |
| | inv_rot_mat = linalg_inv(convert_to_numpy(fwd_rot_mat)) |
| |
|
| | _, _m, _p, _ = resolves_modes(mode, padding_mode) |
| | xform = AffineTransform( |
| | normalized=False, |
| | mode=_m, |
| | padding_mode=_p, |
| | align_corners=False if align_corners == TraceKeys.NONE else align_corners, |
| | reverse_indexing=True, |
| | ) |
| | img_t: torch.Tensor = convert_data_type(data, MetaTensor, dtype=dtype)[0] |
| | transform_t, *_ = convert_to_dst_type(inv_rot_mat, img_t) |
| | sp_size = transform[TraceKeys.ORIG_SIZE] |
| | out: torch.Tensor = xform(img_t.unsqueeze(0), transform_t, spatial_size=sp_size).float().squeeze(0) |
| | out = convert_to_dst_type(out, dst=data, dtype=out.dtype)[0] |
| | if isinstance(out, MetaTensor): |
| | affine = convert_to_tensor(out.peek_pending_affine(), track_meta=False) |
| | mat = to_affine_nd(len(affine) - 1, transform_t) |
| | out.affine @= convert_to_dst_type(mat, affine)[0] |
| | return out |
| |
|
| |
|
| | class Zoom(InvertibleTransform, LazyTransform): |
| | """ |
| | Zooms an ND image using :py:class:`torch.nn.functional.interpolate`. |
| | For details, please see https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html. |
| | |
| | Different from :py:class:`monai.transforms.resize`, this transform takes scaling factors |
| | as input, and provides an option of preserving the input spatial size. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | zoom: The zoom factor along the spatial axes. |
| | If a float, zoom is the same for each spatial axis. |
| | If a sequence, zoom should contain one value for each spatial axis. |
| | mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| | The interpolation mode. Defaults to ``"area"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | padding_mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"edge"``. |
| | The mode to pad data after zooming. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | align_corners: This only has an effect when mode is |
| | 'linear', 'bilinear', 'bicubic' or 'trilinear'. Default: None. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | dtype: data type for resampling computation. Defaults to ``float32``. |
| | If None, use the data type of input data. |
| | keep_size: Should keep original size (padding/slicing if needed), default is True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | zoom: Sequence[float] | float, |
| | mode: str = InterpolateMode.AREA, |
| | padding_mode: str = NumpyPadMode.EDGE, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike | torch.dtype = torch.float32, |
| | keep_size: bool = True, |
| | lazy: bool = False, |
| | **kwargs, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.zoom = zoom |
| | self.mode = mode |
| | self.padding_mode = padding_mode |
| | self.align_corners = align_corners |
| | self.dtype = dtype |
| | self.keep_size = keep_size |
| | self.kwargs = kwargs |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | mode: str | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike | torch.dtype = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]). |
| | mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, |
| | ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| | The interpolation mode. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | padding_mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"edge"``. |
| | The mode to pad data after zooming. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | align_corners: This only has an effect when mode is |
| | 'linear', 'bilinear', 'bicubic' or 'trilinear'. Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | If None, use the data type of input data. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | _zoom = ensure_tuple_rep(self.zoom, img.ndim - 1) |
| | _mode = self.mode if mode is None else mode |
| | _padding_mode = padding_mode or self.padding_mode |
| | _align_corners = self.align_corners if align_corners is None else align_corners |
| | _dtype = get_equivalent_dtype(dtype or self.dtype or img.dtype, torch.Tensor) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return zoom( |
| | img, |
| | _zoom, |
| | self.keep_size, |
| | _mode, |
| | _padding_mode, |
| | _align_corners, |
| | _dtype, |
| | lazy=lazy_, |
| | transform_info=self.get_transform_info(), |
| | ) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | return self.inverse_transform(data, transform) |
| |
|
| | def inverse_transform(self, data: torch.Tensor, transform) -> torch.Tensor: |
| | if transform[TraceKeys.EXTRA_INFO]["do_padcrop"]: |
| | orig_size = transform[TraceKeys.ORIG_SIZE] |
| | pad_or_crop = ResizeWithPadOrCrop(spatial_size=orig_size, mode="edge") |
| | padcrop_xform = transform[TraceKeys.EXTRA_INFO]["padcrop"] |
| | padcrop_xform[TraceKeys.EXTRA_INFO]["pad_info"][TraceKeys.ID] = TraceKeys.NONE |
| | padcrop_xform[TraceKeys.EXTRA_INFO]["crop_info"][TraceKeys.ID] = TraceKeys.NONE |
| | |
| | data = pad_or_crop.inverse_transform(data, padcrop_xform) |
| | |
| | mode = transform[TraceKeys.EXTRA_INFO]["mode"] |
| | align_corners = transform[TraceKeys.EXTRA_INFO]["align_corners"] |
| | dtype = transform[TraceKeys.EXTRA_INFO]["dtype"] |
| | inverse_transform = Resize(spatial_size=transform[TraceKeys.ORIG_SIZE]) |
| | |
| | with inverse_transform.trace_transform(False): |
| | out = inverse_transform( |
| | data, mode=mode, align_corners=None if align_corners == TraceKeys.NONE else align_corners, dtype=dtype |
| | ) |
| | return out |
| |
|
| |
|
| | class Rotate90(InvertibleTransform, LazyTransform): |
| | """ |
| | Rotate an array by 90 degrees in the plane specified by `axes`. |
| | See `torch.rot90` for additional details: |
| | https://pytorch.org/docs/stable/generated/torch.rot90.html#torch-rot90. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__(self, k: int = 1, spatial_axes: tuple[int, int] = (0, 1), lazy: bool = False) -> None: |
| | """ |
| | Args: |
| | k: number of times to rotate by 90 degrees. |
| | spatial_axes: 2 int numbers, defines the plane to rotate with 2 spatial axes. |
| | Default: (0, 1), this is the first two axis in spatial dimensions. |
| | If axis is negative it counts from the last to the first axis. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.k = (4 + (k % 4)) % 4 |
| | spatial_axes_: tuple[int, int] = ensure_tuple(spatial_axes) |
| | if len(spatial_axes_) != 2: |
| | raise ValueError(f"spatial_axes must be 2 numbers to define the plane to rotate, got {spatial_axes_}.") |
| | self.spatial_axes = spatial_axes_ |
| |
|
| | def __call__(self, img: torch.Tensor, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | axes = map_spatial_axes(img.ndim, self.spatial_axes) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return rotate90(img, axes, self.k, lazy=lazy_, transform_info=self.get_transform_info()) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | return self.inverse_transform(data, transform) |
| |
|
| | def inverse_transform(self, data: torch.Tensor, transform) -> torch.Tensor: |
| | axes = transform[TraceKeys.EXTRA_INFO]["axes"] |
| | k = transform[TraceKeys.EXTRA_INFO]["k"] |
| | inv_k = 4 - k % 4 |
| | xform = Rotate90(k=inv_k, spatial_axes=axes) |
| | with xform.trace_transform(False): |
| | return xform(data) |
| |
|
| |
|
| | class RandRotate90(RandomizableTransform, InvertibleTransform, LazyTransform): |
| | """ |
| | With probability `prob`, input arrays are rotated by 90 degrees |
| | in the plane specified by `spatial_axes`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = Rotate90.backend |
| |
|
| | def __init__( |
| | self, prob: float = 0.1, max_k: int = 3, spatial_axes: tuple[int, int] = (0, 1), lazy: bool = False |
| | ) -> None: |
| | """ |
| | Args: |
| | prob: probability of rotating. |
| | (Default 0.1, with 10% probability it returns a rotated array) |
| | max_k: number of rotations will be sampled from `np.random.randint(max_k) + 1`, (Default 3). |
| | spatial_axes: 2 int numbers, defines the plane to rotate with 2 spatial axes. |
| | Default: (0, 1), this is the first two axis in spatial dimensions. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| | RandomizableTransform.__init__(self, prob) |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.max_k = max_k |
| | self.spatial_axes = spatial_axes |
| |
|
| | self._rand_k = 0 |
| |
|
| | def randomize(self, data: Any | None = None) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self._rand_k = self.R.randint(self.max_k) + 1 |
| |
|
| | def __call__(self, img: torch.Tensor, randomize: bool = True, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| |
|
| | if randomize: |
| | self.randomize() |
| |
|
| | lazy_ = self.lazy if lazy is None else lazy |
| | if self._do_transform: |
| | xform = Rotate90(self._rand_k, self.spatial_axes, lazy=lazy_) |
| | out = xform(img) |
| | else: |
| | out = convert_to_tensor(img, track_meta=get_track_meta()) |
| |
|
| | self.push_transform(out, replace=True, lazy=lazy_) |
| | return out |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | xform_info = self.pop_transform(data) |
| | if not xform_info[TraceKeys.DO_TRANSFORM]: |
| | return data |
| | rotate_xform = xform_info[TraceKeys.EXTRA_INFO] |
| | return Rotate90().inverse_transform(data, rotate_xform) |
| |
|
| |
|
| | class RandRotate(RandomizableTransform, InvertibleTransform, LazyTransform): |
| | """ |
| | Randomly rotate the input arrays. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | range_x: Range of rotation angle in radians in the plane defined by the first and second axes. |
| | If single number, angle is uniformly sampled from (-range_x, range_x). |
| | range_y: Range of rotation angle in radians in the plane defined by the first and third axes. |
| | If single number, angle is uniformly sampled from (-range_y, range_y). only work for 3D data. |
| | range_z: Range of rotation angle in radians in the plane defined by the second and third axes. |
| | If single number, angle is uniformly sampled from (-range_z, range_z). only work for 3D data. |
| | prob: Probability of rotation. |
| | keep_size: If it is False, the output shape is adapted so that the |
| | input array is contained completely in the output. |
| | If it is True, the output shape is the same as the input. Default is True. |
| | mode: {``"bilinear"``, ``"nearest"``} |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | align_corners: Defaults to False. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``float32``. |
| | If None, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| |
|
| | backend = Rotate.backend |
| |
|
| | def __init__( |
| | self, |
| | range_x: tuple[float, float] | float = 0.0, |
| | range_y: tuple[float, float] | float = 0.0, |
| | range_z: tuple[float, float] | float = 0.0, |
| | prob: float = 0.1, |
| | keep_size: bool = True, |
| | mode: str = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | align_corners: bool = False, |
| | dtype: DtypeLike | torch.dtype = np.float32, |
| | lazy: bool = False, |
| | ) -> None: |
| | RandomizableTransform.__init__(self, prob) |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.range_x = ensure_tuple(range_x) |
| | if len(self.range_x) == 1: |
| | self.range_x = tuple(sorted([-self.range_x[0], self.range_x[0]])) |
| | self.range_y = ensure_tuple(range_y) |
| | if len(self.range_y) == 1: |
| | self.range_y = tuple(sorted([-self.range_y[0], self.range_y[0]])) |
| | self.range_z = ensure_tuple(range_z) |
| | if len(self.range_z) == 1: |
| | self.range_z = tuple(sorted([-self.range_z[0], self.range_z[0]])) |
| |
|
| | self.keep_size = keep_size |
| | self.mode: str = mode |
| | self.padding_mode: str = padding_mode |
| | self.align_corners = align_corners |
| | self.dtype = dtype |
| |
|
| | self.x = 0.0 |
| | self.y = 0.0 |
| | self.z = 0.0 |
| |
|
| | def randomize(self, data: Any | None = None) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self.x = self.R.uniform(low=self.range_x[0], high=self.range_x[1]) |
| | self.y = self.R.uniform(low=self.range_y[0], high=self.range_y[1]) |
| | self.z = self.R.uniform(low=self.range_z[0], high=self.range_z[1]) |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | mode: str | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike | torch.dtype = None, |
| | randomize: bool = True, |
| | lazy: bool | None = None, |
| | ): |
| | """ |
| | Args: |
| | img: channel first array, must have shape 2D: (nchannels, H, W), or 3D: (nchannels, H, W, D). |
| | mode: {``"bilinear"``, ``"nearest"``} |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | align_corners: Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | If None, use the data type of input data. To be compatible with other modules, |
| | the output data type is always ``float32``. |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | if randomize: |
| | self.randomize() |
| |
|
| | lazy_ = self.lazy if lazy is None else lazy |
| | if self._do_transform: |
| | ndim = len(img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:]) |
| | rotator = Rotate( |
| | angle=self.x if ndim == 2 else (self.x, self.y, self.z), |
| | keep_size=self.keep_size, |
| | mode=mode or self.mode, |
| | padding_mode=padding_mode or self.padding_mode, |
| | align_corners=self.align_corners if align_corners is None else align_corners, |
| | dtype=dtype or self.dtype or img.dtype, |
| | lazy=lazy_, |
| | ) |
| | out = rotator(img) |
| | else: |
| | out = convert_to_tensor(img, track_meta=get_track_meta(), dtype=torch.float32) |
| | self.push_transform(out, replace=True, lazy=lazy_) |
| | return out |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | xform_info = self.pop_transform(data) |
| | if not xform_info[TraceKeys.DO_TRANSFORM]: |
| | return data |
| | return Rotate(0).inverse_transform(data, xform_info[TraceKeys.EXTRA_INFO]) |
| |
|
| |
|
| | class RandFlip(RandomizableTransform, InvertibleTransform, LazyTransform): |
| | """ |
| | Randomly flips the image along axes. Preserves shape. |
| | See numpy.flip for additional details. |
| | https://docs.scipy.org/doc/numpy/reference/generated/numpy.flip.html |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | prob: Probability of flipping. |
| | spatial_axis: Spatial axes along which to flip over. Default is None. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| |
|
| | backend = Flip.backend |
| |
|
| | def __init__(self, prob: float = 0.1, spatial_axis: Sequence[int] | int | None = None, lazy: bool = False) -> None: |
| | RandomizableTransform.__init__(self, prob) |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.flipper = Flip(spatial_axis=spatial_axis, lazy=lazy) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, val: bool): |
| | self.flipper.lazy = val |
| | self._lazy = val |
| |
|
| | def __call__(self, img: torch.Tensor, randomize: bool = True, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | if randomize: |
| | self.randomize(None) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | out = self.flipper(img, lazy=lazy_) if self._do_transform else img |
| | out = convert_to_tensor(out, track_meta=get_track_meta()) |
| | self.push_transform(out, replace=True, lazy=lazy_) |
| | return out |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | if not transform[TraceKeys.DO_TRANSFORM]: |
| | return data |
| | data.applied_operations.append(transform[TraceKeys.EXTRA_INFO]) |
| | return self.flipper.inverse(data) |
| |
|
| |
|
| | class RandAxisFlip(RandomizableTransform, InvertibleTransform, LazyTransform): |
| | """ |
| | Randomly select a spatial axis and flip along it. |
| | See numpy.flip for additional details. |
| | https://docs.scipy.org/doc/numpy/reference/generated/numpy.flip.html |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | prob: Probability of flipping. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| |
|
| | backend = Flip.backend |
| |
|
| | def __init__(self, prob: float = 0.1, lazy: bool = False) -> None: |
| | RandomizableTransform.__init__(self, prob) |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self._axis: int | None = None |
| | self.flipper = Flip(spatial_axis=self._axis) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, val: bool): |
| | self.flipper.lazy = val |
| | self._lazy = val |
| |
|
| | def randomize(self, data: NdarrayOrTensor) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self._axis = self.R.randint(data.ndim - 1) |
| |
|
| | def __call__(self, img: torch.Tensor, randomize: bool = True, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]) |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | if randomize: |
| | self.randomize(data=img) |
| |
|
| | lazy_ = self.lazy if lazy is None else lazy |
| | if self._do_transform: |
| | self.flipper.spatial_axis = self._axis |
| | out = self.flipper(img, lazy=lazy_) |
| | else: |
| | out = convert_to_tensor(img, track_meta=get_track_meta()) |
| | self.push_transform(out, replace=True, lazy=lazy_) |
| | return out |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | if not transform[TraceKeys.DO_TRANSFORM]: |
| | return data |
| | flipper = Flip(spatial_axis=transform[TraceKeys.EXTRA_INFO][TraceKeys.EXTRA_INFO]["axes"]) |
| | with flipper.trace_transform(False): |
| | return flipper(data) |
| |
|
| |
|
| | class RandZoom(RandomizableTransform, InvertibleTransform, LazyTransform): |
| | """ |
| | Randomly zooms input arrays with given probability within given zoom range. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | prob: Probability of zooming. |
| | min_zoom: Min zoom factor. Can be float or sequence same size as image. |
| | If a float, select a random factor from `[min_zoom, max_zoom]` then apply to all spatial dims |
| | to keep the original spatial shape ratio. |
| | If a sequence, min_zoom should contain one value for each spatial axis. |
| | If 2 values provided for 3D data, use the first value for both H & W dims to keep the same zoom ratio. |
| | max_zoom: Max zoom factor. Can be float or sequence same size as image. |
| | If a float, select a random factor from `[min_zoom, max_zoom]` then apply to all spatial dims |
| | to keep the original spatial shape ratio. |
| | If a sequence, max_zoom should contain one value for each spatial axis. |
| | If 2 values provided for 3D data, use the first value for both H & W dims to keep the same zoom ratio. |
| | mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| | The interpolation mode. Defaults to ``"area"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | padding_mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | The mode to pad data after zooming. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | align_corners: This only has an effect when mode is |
| | 'linear', 'bilinear', 'bicubic' or 'trilinear'. Default: None. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | dtype: data type for resampling computation. Defaults to ``float32``. |
| | If None, use the data type of input data. |
| | keep_size: Should keep original size (pad if needed), default is True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| |
|
| | backend = Zoom.backend |
| |
|
| | def __init__( |
| | self, |
| | prob: float = 0.1, |
| | min_zoom: Sequence[float] | float = 0.9, |
| | max_zoom: Sequence[float] | float = 1.1, |
| | mode: str = InterpolateMode.AREA, |
| | padding_mode: str = NumpyPadMode.EDGE, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike | torch.dtype = torch.float32, |
| | keep_size: bool = True, |
| | lazy: bool = False, |
| | **kwargs, |
| | ) -> None: |
| | RandomizableTransform.__init__(self, prob) |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.min_zoom = ensure_tuple(min_zoom) |
| | self.max_zoom = ensure_tuple(max_zoom) |
| | if len(self.min_zoom) != len(self.max_zoom): |
| | raise ValueError( |
| | f"min_zoom and max_zoom must have same length, got {len(self.min_zoom)} and {len(self.max_zoom)}." |
| | ) |
| | self.mode = mode |
| | self.padding_mode = padding_mode |
| | self.align_corners = align_corners |
| | self.dtype = dtype |
| | self.keep_size = keep_size |
| | self.kwargs = kwargs |
| |
|
| | self._zoom: Sequence[float] = [1.0] |
| |
|
| | def randomize(self, img: NdarrayOrTensor) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self._zoom = [self.R.uniform(l, h) for l, h in zip(self.min_zoom, self.max_zoom)] |
| | if len(self._zoom) == 1: |
| | |
| | self._zoom = ensure_tuple_rep(self._zoom[0], img.ndim - 1) |
| | elif len(self._zoom) == 2 and img.ndim > 3: |
| | |
| | self._zoom = ensure_tuple_rep(self._zoom[0], img.ndim - 2) + ensure_tuple(self._zoom[-1]) |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | mode: str | None = None, |
| | padding_mode: str | None = None, |
| | align_corners: bool | None = None, |
| | dtype: DtypeLike | torch.dtype = None, |
| | randomize: bool = True, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape 2D: (nchannels, H, W), or 3D: (nchannels, H, W, D). |
| | mode: {``"nearest"``, ``"nearest-exact"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, |
| | ``"area"``}, the interpolation mode. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | padding_mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | The mode to pad data after zooming. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | align_corners: This only has an effect when mode is |
| | 'linear', 'bilinear', 'bicubic' or 'trilinear'. Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | If None, use the data type of input data. |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | |
| | if randomize: |
| | self.randomize(img=img) |
| |
|
| | lazy_ = self.lazy if lazy is None else lazy |
| | if not self._do_transform: |
| | out = convert_to_tensor(img, track_meta=get_track_meta(), dtype=torch.float32) |
| | else: |
| | xform = Zoom( |
| | self._zoom, |
| | keep_size=self.keep_size, |
| | mode=mode or self.mode, |
| | padding_mode=padding_mode or self.padding_mode, |
| | align_corners=self.align_corners if align_corners is None else align_corners, |
| | dtype=dtype or self.dtype, |
| | lazy=lazy_, |
| | **self.kwargs, |
| | ) |
| | out = xform(img) |
| | self.push_transform(out, replace=True, lazy=lazy_) |
| | return out |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | xform_info = self.pop_transform(data) |
| | if not xform_info[TraceKeys.DO_TRANSFORM]: |
| | return data |
| | return Zoom(self._zoom).inverse_transform(data, xform_info[TraceKeys.EXTRA_INFO]) |
| |
|
| |
|
| | class AffineGrid(LazyTransform): |
| | """ |
| | Affine transforms on the coordinates. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | rotate_params: a rotation angle in radians, a scalar for 2D image, a tuple of 3 floats for 3D. |
| | Defaults to no rotation. |
| | shear_params: shearing factors for affine matrix, take a 3D affine as example:: |
| | |
| | [ |
| | [1.0, params[0], params[1], 0.0], |
| | [params[2], 1.0, params[3], 0.0], |
| | [params[4], params[5], 1.0, 0.0], |
| | [0.0, 0.0, 0.0, 1.0], |
| | ] |
| | |
| | a tuple of 2 floats for 2D, a tuple of 6 floats for 3D. Defaults to no shearing. |
| | translate_params: a tuple of 2 floats for 2D, a tuple of 3 floats for 3D. Translation is in |
| | pixel/voxel relative to the center of the input image. Defaults to no translation. |
| | scale_params: scale factor for every spatial dims. a tuple of 2 floats for 2D, |
| | a tuple of 3 floats for 3D. Defaults to `1.0`. |
| | dtype: data type for the grid computation. Defaults to ``float32``. |
| | If ``None``, use the data type of input data (if `grid` is provided). |
| | device: device on which the tensor will be allocated, if a new grid is generated. |
| | align_corners: Defaults to False. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | affine: If applied, ignore the params (`rotate_params`, etc.) and use the |
| | supplied matrix. Should be square with each side = num of image spatial |
| | dimensions + 1. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | rotate_params: Sequence[float] | float | None = None, |
| | shear_params: Sequence[float] | float | None = None, |
| | translate_params: Sequence[float] | float | None = None, |
| | scale_params: Sequence[float] | float | None = None, |
| | device: torch.device | None = None, |
| | dtype: DtypeLike = np.float32, |
| | align_corners: bool = False, |
| | affine: NdarrayOrTensor | None = None, |
| | lazy: bool = False, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.rotate_params = rotate_params |
| | self.shear_params = shear_params |
| | self.translate_params = translate_params |
| | self.scale_params = scale_params |
| | self.device = device |
| | _dtype = get_equivalent_dtype(dtype, torch.Tensor) |
| | self.dtype = _dtype if _dtype in (torch.float16, torch.float64, None) else torch.float32 |
| | self.align_corners = align_corners |
| | self.affine = affine |
| |
|
| | def __call__( |
| | self, spatial_size: Sequence[int] | None = None, grid: torch.Tensor | None = None, lazy: bool | None = None |
| | ) -> tuple[torch.Tensor | None, torch.Tensor]: |
| | """ |
| | The grid can be initialized with a `spatial_size` parameter, or provided directly as `grid`. |
| | Therefore, either `spatial_size` or `grid` must be provided. |
| | When initialising from `spatial_size`, the backend "torch" will be used. |
| | |
| | Args: |
| | spatial_size: output grid size. |
| | grid: grid to be transformed. Shape must be (3, H, W) for 2D or (4, H, W, D) for 3D. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | Raises: |
| | ValueError: When ``grid=None`` and ``spatial_size=None``. Incompatible values. |
| | |
| | """ |
| | lazy_ = self.lazy if lazy is None else lazy |
| | if not lazy_: |
| | if grid is None: |
| | if spatial_size is None: |
| | raise ValueError("Incompatible values: grid=None and spatial_size=None.") |
| | grid_ = create_grid(spatial_size, device=self.device, backend="torch", dtype=self.dtype) |
| | else: |
| | grid_ = grid |
| | _dtype = self.dtype or grid_.dtype |
| | grid_: torch.Tensor = convert_to_tensor(grid_, dtype=_dtype, track_meta=get_track_meta()) |
| | _device = grid_.device |
| | spatial_dims = len(grid_.shape) - 1 |
| | else: |
| | _device = self.device |
| | spatial_dims = len(spatial_size) |
| | _b = TransformBackends.TORCH |
| | affine: torch.Tensor |
| | if self.affine is None: |
| | affine = torch.eye(spatial_dims + 1, device=_device) |
| | if self.rotate_params: |
| | affine @= create_rotate(spatial_dims, self.rotate_params, device=_device, backend=_b) |
| | if self.shear_params: |
| | affine @= create_shear(spatial_dims, self.shear_params, device=_device, backend=_b) |
| | if self.translate_params: |
| | affine @= create_translate(spatial_dims, self.translate_params, device=_device, backend=_b) |
| | if self.scale_params: |
| | affine @= create_scale(spatial_dims, self.scale_params, device=_device, backend=_b) |
| | else: |
| | affine = self.affine |
| | affine = to_affine_nd(spatial_dims, affine) |
| | if lazy_: |
| | return None, affine |
| |
|
| | affine = convert_to_tensor(affine, device=grid_.device, dtype=grid_.dtype, track_meta=False) |
| | if self.align_corners: |
| | sc = create_scale( |
| | spatial_dims, [max(d, 2) / (max(d, 2) - 1) for d in grid_.shape[1:]], device=_device, backend=_b |
| | ) |
| | sc = convert_to_dst_type(sc, affine)[0] |
| | grid_ = ((affine @ sc) @ grid_.view((grid_.shape[0], -1))).view([-1] + list(grid_.shape[1:])) |
| | else: |
| | grid_ = (affine @ grid_.view((grid_.shape[0], -1))).view([-1] + list(grid_.shape[1:])) |
| | return grid_, affine |
| |
|
| |
|
| | class RandAffineGrid(Randomizable, LazyTransform): |
| | """ |
| | Generate randomised affine grid. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = AffineGrid.backend |
| |
|
| | def __init__( |
| | self, |
| | rotate_range: RandRange = None, |
| | shear_range: RandRange = None, |
| | translate_range: RandRange = None, |
| | scale_range: RandRange = None, |
| | device: torch.device | None = None, |
| | dtype: DtypeLike = np.float32, |
| | lazy: bool = False, |
| | ) -> None: |
| | """ |
| | Args: |
| | rotate_range: angle range in radians. If element `i` is a pair of (min, max) values, then |
| | `uniform[-rotate_range[i][0], rotate_range[i][1])` will be used to generate the rotation parameter |
| | for the `i`th spatial dimension. If not, `uniform[-rotate_range[i], rotate_range[i])` will be used. |
| | This can be altered on a per-dimension basis. E.g., `((0,3), 1, ...)`: for dim0, rotation will be |
| | in range `[0, 3]`, and for dim1 `[-1, 1]` will be used. Setting a single value will use `[-x, x]` |
| | for dim0 and nothing for the remaining dimensions. |
| | shear_range: shear range with format matching `rotate_range`, it defines the range to randomly select |
| | shearing factors(a tuple of 2 floats for 2D, a tuple of 6 floats for 3D) for affine matrix, |
| | take a 3D affine as example:: |
| | |
| | [ |
| | [1.0, params[0], params[1], 0.0], |
| | [params[2], 1.0, params[3], 0.0], |
| | [params[4], params[5], 1.0, 0.0], |
| | [0.0, 0.0, 0.0, 1.0], |
| | ] |
| | |
| | translate_range: translate range with format matching `rotate_range`, it defines the range to randomly |
| | select voxels to translate for every spatial dims. |
| | scale_range: scaling range with format matching `rotate_range`. it defines the range to randomly select |
| | the scale factor to translate for every spatial dims. A value of 1.0 is added to the result. |
| | This allows 0 to correspond to no change (i.e., a scaling of 1.0). |
| | device: device to store the output grid data. |
| | dtype: data type for the grid computation. Defaults to ``np.float32``. |
| | If ``None``, use the data type of input data (if `grid` is provided). |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | |
| | See also: |
| | - :py:meth:`monai.transforms.utils.create_rotate` |
| | - :py:meth:`monai.transforms.utils.create_shear` |
| | - :py:meth:`monai.transforms.utils.create_translate` |
| | - :py:meth:`monai.transforms.utils.create_scale` |
| | |
| | """ |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.rotate_range = ensure_tuple(rotate_range) |
| | self.shear_range = ensure_tuple(shear_range) |
| | self.translate_range = ensure_tuple(translate_range) |
| | self.scale_range = ensure_tuple(scale_range) |
| |
|
| | self.rotate_params: list[float] | None = None |
| | self.shear_params: list[float] | None = None |
| | self.translate_params: list[float] | None = None |
| | self.scale_params: list[float] | None = None |
| |
|
| | self.device = device |
| | self.dtype = dtype |
| | self.affine: torch.Tensor | None = torch.eye(4, dtype=torch.float64) |
| |
|
| | def _get_rand_param(self, param_range, add_scalar: float = 0.0): |
| | out_param = [] |
| | for f in param_range: |
| | if issequenceiterable(f): |
| | if len(f) != 2: |
| | raise ValueError(f"If giving range as [min,max], should have 2 elements per dim, got {f}.") |
| | out_param.append(self.R.uniform(f[0], f[1]) + add_scalar) |
| | elif f is not None: |
| | out_param.append(self.R.uniform(-f, f) + add_scalar) |
| | return out_param |
| |
|
| | def randomize(self, data: Any | None = None) -> None: |
| | self.rotate_params = self._get_rand_param(self.rotate_range) |
| | self.shear_params = self._get_rand_param(self.shear_range) |
| | self.translate_params = self._get_rand_param(self.translate_range) |
| | self.scale_params = self._get_rand_param(self.scale_range, 1.0) |
| |
|
| | def __call__( |
| | self, |
| | spatial_size: Sequence[int] | None = None, |
| | grid: NdarrayOrTensor | None = None, |
| | randomize: bool = True, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | spatial_size: output grid size. |
| | grid: grid to be transformed. Shape must be (3, H, W) for 2D or (4, H, W, D) for 3D. |
| | randomize: boolean as to whether the grid parameters governing the grid should be randomized. |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | |
| | Returns: |
| | a 2D (3xHxW) or 3D (4xHxWxD) grid. |
| | """ |
| | if randomize: |
| | self.randomize() |
| | lazy_ = self.lazy if lazy is None else lazy |
| | affine_grid = AffineGrid( |
| | rotate_params=self.rotate_params, |
| | shear_params=self.shear_params, |
| | translate_params=self.translate_params, |
| | scale_params=self.scale_params, |
| | device=self.device, |
| | dtype=self.dtype, |
| | lazy=lazy_, |
| | ) |
| | if lazy_: |
| | self.affine = affine_grid(spatial_size, grid)[1] |
| | return None |
| | _grid: torch.Tensor |
| | _grid, self.affine = affine_grid(spatial_size, grid) |
| | return _grid |
| |
|
| | def get_transformation_matrix(self) -> torch.Tensor | None: |
| | """Get the most recently applied transformation matrix""" |
| | return self.affine |
| |
|
| |
|
| | class RandDeformGrid(Randomizable, Transform): |
| | """ |
| | Generate random deformation grid. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, spacing: Sequence[float] | float, magnitude_range: tuple[float, float], device: torch.device | None = None |
| | ) -> None: |
| | """ |
| | Args: |
| | spacing: spacing of the grid in 2D or 3D. |
| | e.g., spacing=(1, 1) indicates pixel-wise deformation in 2D, |
| | spacing=(1, 1, 1) indicates voxel-wise deformation in 3D, |
| | spacing=(2, 2) indicates deformation field defined on every other pixel in 2D. |
| | magnitude_range: the random offsets will be generated from |
| | `uniform[magnitude[0], magnitude[1])`. |
| | device: device to store the output grid data. |
| | """ |
| | self.spacing = spacing |
| | self.magnitude = magnitude_range |
| |
|
| | self.rand_mag = 1.0 |
| | self.random_offset: np.ndarray |
| | self.device = device |
| |
|
| | def randomize(self, grid_size: Sequence[int]) -> None: |
| | self.random_offset = self.R.normal(size=([len(grid_size)] + list(grid_size))).astype(np.float32, copy=False) |
| | self.rand_mag = self.R.uniform(self.magnitude[0], self.magnitude[1]) |
| |
|
| | def __call__(self, spatial_size: Sequence[int]) -> torch.Tensor: |
| | """ |
| | Args: |
| | spatial_size: spatial size of the grid. |
| | """ |
| | self.spacing = fall_back_tuple(self.spacing, (1.0,) * len(spatial_size)) |
| | control_grid = create_control_grid(spatial_size, self.spacing, device=self.device, backend="torch") |
| | self.randomize(control_grid.shape[1:]) |
| | _offset, *_ = convert_to_dst_type(self.rand_mag * self.random_offset, control_grid) |
| | control_grid[: len(spatial_size)] += _offset |
| | return control_grid |
| |
|
| |
|
| | class Resample(Transform): |
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | norm_coords: bool = True, |
| | device: torch.device | None = None, |
| | align_corners: bool = False, |
| | dtype: DtypeLike = np.float64, |
| | ) -> None: |
| | """ |
| | computes output image using values from `img`, locations from `grid` using pytorch. |
| | supports spatially 2D or 3D (num_channels, H, W[, D]). |
| | |
| | Args: |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `USE_COMPILED` is `True`, this argument uses |
| | ``"nearest"``, ``"bilinear"``, ``"bicubic"`` to indicate 0, 1, 3 order interpolations. |
| | See also: https://docs.monai.io/en/stable/networks.html#grid-pull (experimental). |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `USE_COMPILED` is `True`, this argument uses an integer to represent the padding mode. |
| | See also: https://docs.monai.io/en/stable/networks.html#grid-pull (experimental). |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | norm_coords: whether to normalize the coordinates from `[-(size-1)/2, (size-1)/2]` to |
| | `[0, size - 1]` (for ``monai/csrc`` implementation) or |
| | `[-1, 1]` (for torch ``grid_sample`` implementation) to be compatible with the underlying |
| | resampling API. |
| | device: device on which the tensor will be allocated. |
| | align_corners: Defaults to False. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | dtype: data type for resampling computation. Defaults to ``float64`` for best precision. |
| | If ``None``, use the data type of input data. To be compatible with other modules, |
| | the output data type is always `float32`. |
| | |
| | """ |
| | self.mode = mode |
| | self.padding_mode = padding_mode |
| | self.norm_coords = norm_coords |
| | self.device = device |
| | self.align_corners = align_corners |
| | self.dtype = dtype |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | grid: torch.Tensor | None = None, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | dtype: DtypeLike = None, |
| | align_corners: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W[, D]). |
| | grid: shape must be (3, H, W) for 2D or (4, H, W, D) for 3D. |
| | if ``norm_coords`` is True, the grid values must be in `[-(size-1)/2, (size-1)/2]`. |
| | if ``USE_COMPILED=True`` and ``norm_coords=False``, grid values must be in `[0, size-1]`. |
| | if ``USE_COMPILED=False`` and ``norm_coords=False``, grid values must be in `[-1, 1]`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `USE_COMPILED` is `True`, this argument uses |
| | ``"nearest"``, ``"bilinear"``, ``"bicubic"`` to indicate 0, 1, 3 order interpolations. |
| | See also: https://docs.monai.io/en/stable/networks.html#grid-pull (experimental). |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `USE_COMPILED` is `True`, this argument uses an integer to represent the padding mode. |
| | See also: https://docs.monai.io/en/stable/networks.html#grid-pull (experimental). |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| | To be compatible with other modules, the output data type is always `float32`. |
| | align_corners: Defaults to ``self.align_corners``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | |
| | See also: |
| | :py:const:`monai.config.USE_COMPILED` |
| | """ |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | if grid is None: |
| | return img |
| |
|
| | _device = img.device if isinstance(img, torch.Tensor) else self.device |
| | _dtype = dtype or self.dtype or img.dtype |
| | _align_corners = self.align_corners if align_corners is None else align_corners |
| | img_t, *_ = convert_data_type(img, torch.Tensor, dtype=_dtype, device=_device) |
| | sr = min(len(img_t.peek_pending_shape() if isinstance(img_t, MetaTensor) else img_t.shape[1:]), 3) |
| | backend, _interp_mode, _padding_mode, _ = resolves_modes( |
| | self.mode if mode is None else mode, |
| | self.padding_mode if padding_mode is None else padding_mode, |
| | backend=None, |
| | use_compiled=USE_COMPILED, |
| | ) |
| |
|
| | if USE_COMPILED or backend == TransformBackends.NUMPY: |
| | grid_t, *_ = convert_to_dst_type(grid[:sr], img_t, dtype=grid.dtype, wrap_sequence=True) |
| | if isinstance(grid, torch.Tensor) and grid_t.data_ptr() == grid.data_ptr(): |
| | grid_t = grid_t.clone(memory_format=torch.contiguous_format) |
| | for i, dim in enumerate(img_t.shape[1 : 1 + sr]): |
| | _dim = max(2, dim) |
| | t = (_dim - 1) / 2.0 |
| | if self.norm_coords: |
| | grid_t[i] = ((_dim - 1) / _dim) * grid_t[i] + t if _align_corners else grid_t[i] + t |
| | elif _align_corners: |
| | grid_t[i] = ((_dim - 1) / _dim) * (grid_t[i] + 0.5) |
| | if USE_COMPILED and backend == TransformBackends.TORCH: |
| | grid_t = moveaxis(grid_t, 0, -1) |
| | out = grid_pull( |
| | img_t.unsqueeze(0), |
| | grid_t.unsqueeze(0).to(img_t), |
| | bound=_padding_mode, |
| | extrapolate=True, |
| | interpolation=_interp_mode, |
| | )[0] |
| | elif backend == TransformBackends.NUMPY: |
| | is_cuda = img_t.is_cuda |
| | img_np = (convert_to_cupy if is_cuda else convert_to_numpy)(img_t, wrap_sequence=True) |
| | grid_np, *_ = convert_to_dst_type(grid_t, img_np, dtype=grid_t.dtype, wrap_sequence=True) |
| | _map_coord = (cupy_ndi if is_cuda else np_ndi).map_coordinates |
| | out = (cupy if is_cuda else np).stack( |
| | [_map_coord(c, grid_np, order=_interp_mode, mode=_padding_mode) for c in img_np] |
| | ) |
| | out = convert_to_dst_type(out, img_t)[0] |
| | else: |
| | grid_t = moveaxis(grid[list(range(sr - 1, -1, -1))], 0, -1) |
| | grid_t = convert_to_dst_type(grid_t, img_t, wrap_sequence=True)[0].unsqueeze(0) |
| | if isinstance(grid, torch.Tensor) and grid_t.data_ptr() == grid.data_ptr(): |
| | grid_t = grid_t.clone(memory_format=torch.contiguous_format) |
| | if self.norm_coords: |
| | for i, dim in enumerate(img_t.shape[sr + 1 : 0 : -1]): |
| | grid_t[0, ..., i] *= 2.0 / max(2, dim) |
| | out = torch.nn.functional.grid_sample( |
| | img_t.unsqueeze(0), |
| | grid_t, |
| | mode=_interp_mode, |
| | padding_mode=_padding_mode, |
| | align_corners=None if _align_corners == TraceKeys.NONE else _align_corners, |
| | )[0] |
| | out_val, *_ = convert_to_dst_type(out, dst=img, dtype=np.float32) |
| | return out_val |
| |
|
| |
|
| | class Affine(InvertibleTransform, LazyTransform): |
| | """ |
| | Transform ``img`` given the affine parameters. |
| | A tutorial is available: https://github.com/Project-MONAI/tutorials/blob/0.6.0/modules/transforms_demo_2d.ipynb. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = list(set(AffineGrid.backend) & set(Resample.backend)) |
| |
|
| | def __init__( |
| | self, |
| | rotate_params: Sequence[float] | float | None = None, |
| | shear_params: Sequence[float] | float | None = None, |
| | translate_params: Sequence[float] | float | None = None, |
| | scale_params: Sequence[float] | float | None = None, |
| | affine: NdarrayOrTensor | None = None, |
| | spatial_size: Sequence[int] | int | None = None, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.REFLECTION, |
| | normalized: bool = False, |
| | device: torch.device | None = None, |
| | dtype: DtypeLike = np.float32, |
| | align_corners: bool = False, |
| | image_only: bool = False, |
| | lazy: bool = False, |
| | ) -> None: |
| | """ |
| | The affine transformations are applied in rotate, shear, translate, scale order. |
| | |
| | Args: |
| | rotate_params: a rotation angle in radians, a scalar for 2D image, a tuple of 3 floats for 3D. |
| | Defaults to no rotation. |
| | shear_params: shearing factors for affine matrix, take a 3D affine as example:: |
| | |
| | [ |
| | [1.0, params[0], params[1], 0.0], |
| | [params[2], 1.0, params[3], 0.0], |
| | [params[4], params[5], 1.0, 0.0], |
| | [0.0, 0.0, 0.0, 1.0], |
| | ] |
| | |
| | a tuple of 2 floats for 2D, a tuple of 6 floats for 3D. Defaults to no shearing. |
| | translate_params: a tuple of 2 floats for 2D, a tuple of 3 floats for 3D. Translation is in |
| | pixel/voxel relative to the center of the input image. Defaults to no translation. |
| | scale_params: scale factor for every spatial dims. a tuple of 2 floats for 2D, |
| | a tuple of 3 floats for 3D. Defaults to `1.0`. |
| | affine: If applied, ignore the params (`rotate_params`, etc.) and use the |
| | supplied matrix. Should be square with each side = num of image spatial |
| | dimensions + 1. |
| | spatial_size: output image spatial size. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | if some components of the `spatial_size` are non-positive values, the transform will use the |
| | corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| | to `(32, 64)` if the second spatial dimension size of img is `64`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"reflection"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | normalized: indicating whether the provided `affine` is defined to include a normalization |
| | transform converting the coordinates from `[-(size-1)/2, (size-1)/2]` (defined in ``create_grid``) to |
| | `[0, size - 1]` or `[-1, 1]` in order to be compatible with the underlying resampling API. |
| | If `normalized=False`, additional coordinate normalization will be applied before resampling. |
| | See also: :py:func:`monai.networks.utils.normalize_transform`. |
| | device: device on which the tensor will be allocated. |
| | dtype: data type for resampling computation. Defaults to ``float32``. |
| | If ``None``, use the data type of input data. To be compatible with other modules, |
| | the output data type is always `float32`. |
| | align_corners: Defaults to False. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | image_only: if True return only the image volume, otherwise return (image, affine). |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | """ |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.affine_grid = AffineGrid( |
| | rotate_params=rotate_params, |
| | shear_params=shear_params, |
| | translate_params=translate_params, |
| | scale_params=scale_params, |
| | affine=affine, |
| | dtype=dtype, |
| | align_corners=align_corners, |
| | device=device, |
| | lazy=lazy, |
| | ) |
| | self.image_only = image_only |
| | self.norm_coord = not normalized |
| | self.resampler = Resample(norm_coords=self.norm_coord, device=device, dtype=dtype, align_corners=align_corners) |
| | self.spatial_size = spatial_size |
| | self.mode = mode |
| | self.padding_mode: str = padding_mode |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, val: bool) -> None: |
| | self.affine_grid.lazy = val |
| | self._lazy = val |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | spatial_size: Sequence[int] | int | None = None, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor | tuple[torch.Tensor, NdarrayOrTensor]: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W[, D]), |
| | spatial_size: output image spatial size. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | if `img` has two spatial dimensions, `spatial_size` should have 2 elements [h, w]. |
| | if `img` has three spatial dimensions, `spatial_size` should have 3 elements [h, w, d]. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | img_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | sp_size = fall_back_tuple(self.spatial_size if spatial_size is None else spatial_size, img_size) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | _mode = mode if mode is not None else self.mode |
| | _padding_mode = padding_mode if padding_mode is not None else self.padding_mode |
| | grid, affine = self.affine_grid(spatial_size=sp_size, lazy=lazy_) |
| |
|
| | return affine_func( |
| | img, |
| | affine, |
| | grid, |
| | self.resampler, |
| | sp_size, |
| | _mode, |
| | _padding_mode, |
| | True, |
| | self.image_only, |
| | lazy=lazy_, |
| | transform_info=self.get_transform_info(), |
| | ) |
| |
|
| | @classmethod |
| | def compute_w_affine(cls, spatial_rank, mat, img_size, sp_size): |
| | r = int(spatial_rank) |
| | mat = to_affine_nd(r, mat) |
| | shift_1 = create_translate(r, [float(d - 1) / 2 for d in img_size[:r]]) |
| | shift_2 = create_translate(r, [-float(d - 1) / 2 for d in sp_size[:r]]) |
| | mat = shift_1 @ convert_data_type(mat, np.ndarray)[0] @ shift_2 |
| | return mat |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | orig_size = transform[TraceKeys.ORIG_SIZE] |
| | |
| | fwd_affine = transform[TraceKeys.EXTRA_INFO]["affine"] |
| | mode = transform[TraceKeys.EXTRA_INFO]["mode"] |
| | padding_mode = transform[TraceKeys.EXTRA_INFO]["padding_mode"] |
| | align_corners = transform[TraceKeys.EXTRA_INFO]["align_corners"] |
| | inv_affine = linalg_inv(convert_to_numpy(fwd_affine)) |
| | inv_affine = convert_to_dst_type(inv_affine, data, dtype=inv_affine.dtype)[0] |
| |
|
| | affine_grid = AffineGrid(affine=inv_affine, align_corners=align_corners) |
| | grid, _ = affine_grid(orig_size) |
| | |
| | out = self.resampler(data, grid, mode, padding_mode, align_corners=align_corners) |
| | if not isinstance(out, MetaTensor): |
| | out = MetaTensor(out) |
| | out.meta = data.meta |
| | affine = convert_data_type(out.peek_pending_affine(), torch.Tensor)[0] |
| | xform, *_ = convert_to_dst_type( |
| | Affine.compute_w_affine(len(affine) - 1, inv_affine, data.shape[1:], orig_size), affine |
| | ) |
| | out.affine @= xform |
| | return out |
| |
|
| |
|
| | class RandAffine(RandomizableTransform, InvertibleTransform, LazyTransform): |
| | """ |
| | Random affine transform. |
| | A tutorial is available: https://github.com/Project-MONAI/tutorials/blob/0.6.0/modules/transforms_demo_2d.ipynb. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = Affine.backend |
| |
|
| | def __init__( |
| | self, |
| | prob: float = 0.1, |
| | rotate_range: RandRange = None, |
| | shear_range: RandRange = None, |
| | translate_range: RandRange = None, |
| | scale_range: RandRange = None, |
| | spatial_size: Sequence[int] | int | None = None, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.REFLECTION, |
| | cache_grid: bool = False, |
| | device: torch.device | None = None, |
| | lazy: bool = False, |
| | ) -> None: |
| | """ |
| | Args: |
| | prob: probability of returning a randomized affine grid. |
| | defaults to 0.1, with 10% chance returns a randomized grid. |
| | rotate_range: angle range in radians. If element `i` is a pair of (min, max) values, then |
| | `uniform[-rotate_range[i][0], rotate_range[i][1])` will be used to generate the rotation parameter |
| | for the `i`th spatial dimension. If not, `uniform[-rotate_range[i], rotate_range[i])` will be used. |
| | This can be altered on a per-dimension basis. E.g., `((0,3), 1, ...)`: for dim0, rotation will be |
| | in range `[0, 3]`, and for dim1 `[-1, 1]` will be used. Setting a single value will use `[-x, x]` |
| | for dim0 and nothing for the remaining dimensions. |
| | shear_range: shear range with format matching `rotate_range`, it defines the range to randomly select |
| | shearing factors(a tuple of 2 floats for 2D, a tuple of 6 floats for 3D) for affine matrix, |
| | take a 3D affine as example:: |
| | |
| | [ |
| | [1.0, params[0], params[1], 0.0], |
| | [params[2], 1.0, params[3], 0.0], |
| | [params[4], params[5], 1.0, 0.0], |
| | [0.0, 0.0, 0.0, 1.0], |
| | ] |
| | |
| | translate_range: translate range with format matching `rotate_range`, it defines the range to randomly |
| | select pixel/voxel to translate for every spatial dims. |
| | scale_range: scaling range with format matching `rotate_range`. it defines the range to randomly select |
| | the scale factor to translate for every spatial dims. A value of 1.0 is added to the result. |
| | This allows 0 to correspond to no change (i.e., a scaling of 1.0). |
| | spatial_size: output image spatial size. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | if some components of the `spatial_size` are non-positive values, the transform will use the |
| | corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| | to `(32, 64)` if the second spatial dimension size of img is `64`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``bilinear``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``reflection``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | cache_grid: whether to cache the identity sampling grid. |
| | If the spatial size is not dynamically defined by input image, enabling this option could |
| | accelerate the transform. |
| | device: device on which the tensor will be allocated. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. |
| | Defaults to False |
| | |
| | See also: |
| | - :py:class:`RandAffineGrid` for the random affine parameters configurations. |
| | - :py:class:`Affine` for the affine transformation parameters configurations. |
| | |
| | """ |
| | RandomizableTransform.__init__(self, prob) |
| | LazyTransform.__init__(self, lazy=lazy) |
| | self.rand_affine_grid = RandAffineGrid( |
| | rotate_range=rotate_range, |
| | shear_range=shear_range, |
| | translate_range=translate_range, |
| | scale_range=scale_range, |
| | device=device, |
| | lazy=lazy, |
| | ) |
| | self.resampler = Resample(device=device) |
| |
|
| | self.spatial_size = spatial_size |
| | self.cache_grid = cache_grid |
| | self._cached_grid = self._init_identity_cache(lazy) |
| | self.mode = mode |
| | self.padding_mode: str = padding_mode |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, val: bool) -> None: |
| | self._lazy = val |
| | self.rand_affine_grid.lazy = val |
| |
|
| | def _init_identity_cache(self, lazy: bool): |
| | """ |
| | Create cache of the identity grid if cache_grid=True and spatial_size is known. |
| | """ |
| | if lazy: |
| | return None |
| | if self.spatial_size is None: |
| | if self.cache_grid: |
| | warnings.warn( |
| | "cache_grid=True is not compatible with the dynamic spatial_size, please specify 'spatial_size'." |
| | ) |
| | return None |
| | _sp_size = ensure_tuple(self.spatial_size) |
| | _ndim = len(_sp_size) |
| | if _sp_size != fall_back_tuple(_sp_size, [1] * _ndim) or _sp_size != fall_back_tuple(_sp_size, [2] * _ndim): |
| | |
| | if self.cache_grid: |
| | warnings.warn( |
| | "cache_grid=True is not compatible with the dynamic spatial_size " |
| | f"'spatial_size={self.spatial_size}', please specify 'spatial_size'." |
| | ) |
| | return None |
| | return create_grid(spatial_size=_sp_size, device=self.rand_affine_grid.device, backend="torch") |
| |
|
| | def get_identity_grid(self, spatial_size: Sequence[int], lazy: bool): |
| | """ |
| | Return a cached or new identity grid depends on the availability. |
| | |
| | Args: |
| | spatial_size: non-dynamic spatial size |
| | """ |
| | if lazy: |
| | return None |
| | ndim = len(spatial_size) |
| | if spatial_size != fall_back_tuple(spatial_size, [1] * ndim) or spatial_size != fall_back_tuple( |
| | spatial_size, [2] * ndim |
| | ): |
| | raise RuntimeError(f"spatial_size should not be dynamic, got {spatial_size}.") |
| | return ( |
| | create_grid(spatial_size=spatial_size, device=self.rand_affine_grid.device, backend="torch") |
| | if self._cached_grid is None |
| | else self._cached_grid |
| | ) |
| |
|
| | def set_random_state(self, seed: int | None = None, state: np.random.RandomState | None = None) -> RandAffine: |
| | self.rand_affine_grid.set_random_state(seed, state) |
| | super().set_random_state(seed, state) |
| | return self |
| |
|
| | def randomize(self, data: Any | None = None) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self.rand_affine_grid.randomize() |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | spatial_size: Sequence[int] | int | None = None, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | randomize: bool = True, |
| | grid=None, |
| | lazy: bool | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W[, D]), |
| | spatial_size: output image spatial size. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | if `img` has two spatial dimensions, `spatial_size` should have 2 elements [h, w]. |
| | if `img` has three spatial dimensions, `spatial_size` should have 3 elements [h, w, d]. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | grid: precomputed grid to be used (mainly to accelerate `RandAffined`). |
| | lazy: a flag to indicate whether this transform should execute lazily or not |
| | during this call. Setting this to False or True overrides the ``lazy`` flag set |
| | during initialization for this call. Defaults to None. |
| | """ |
| | if randomize: |
| | self.randomize() |
| | |
| | |
| | ori_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | sp_size = fall_back_tuple(self.spatial_size if spatial_size is None else spatial_size, ori_size) |
| | do_resampling = self._do_transform or (sp_size != ensure_tuple(ori_size)) |
| | _mode = mode if mode is not None else self.mode |
| | _padding_mode = padding_mode if padding_mode is not None else self.padding_mode |
| | lazy_ = self.lazy if lazy is None else lazy |
| | img = convert_to_tensor(img, track_meta=get_track_meta()) |
| | if lazy_: |
| | if self._do_transform: |
| | if grid is None: |
| | self.rand_affine_grid(sp_size, randomize=randomize, lazy=True) |
| | affine = self.rand_affine_grid.get_transformation_matrix() |
| | else: |
| | affine = convert_to_dst_type(torch.eye(len(sp_size) + 1), img, dtype=self.rand_affine_grid.dtype)[0] |
| | else: |
| | if grid is None: |
| | grid = self.get_identity_grid(sp_size, lazy_) |
| | if self._do_transform: |
| | grid = self.rand_affine_grid(grid=grid, randomize=randomize, lazy=lazy_) |
| | affine = self.rand_affine_grid.get_transformation_matrix() |
| | return affine_func( |
| | img, |
| | affine, |
| | grid, |
| | self.resampler, |
| | sp_size, |
| | _mode, |
| | _padding_mode, |
| | do_resampling, |
| | True, |
| | lazy=lazy_, |
| | transform_info=self.get_transform_info(), |
| | ) |
| |
|
| | def inverse(self, data: torch.Tensor) -> torch.Tensor: |
| | transform = self.pop_transform(data) |
| | |
| | if not transform[TraceKeys.EXTRA_INFO]["do_resampling"]: |
| | return data |
| | orig_size = transform[TraceKeys.ORIG_SIZE] |
| | orig_size = fall_back_tuple(orig_size, data.shape[1:]) |
| | |
| | fwd_affine = transform[TraceKeys.EXTRA_INFO]["affine"] |
| | mode = transform[TraceKeys.EXTRA_INFO]["mode"] |
| | padding_mode = transform[TraceKeys.EXTRA_INFO]["padding_mode"] |
| | inv_affine = linalg_inv(convert_to_numpy(fwd_affine)) |
| | inv_affine = convert_to_dst_type(inv_affine, data, dtype=inv_affine.dtype)[0] |
| | affine_grid = AffineGrid(affine=inv_affine) |
| | grid, _ = affine_grid(orig_size) |
| |
|
| | |
| | out = self.resampler(data, grid, mode, padding_mode) |
| | if not isinstance(out, MetaTensor): |
| | out = MetaTensor(out) |
| | out.meta = data.meta |
| | affine = convert_data_type(out.peek_pending_affine(), torch.Tensor)[0] |
| | xform, *_ = convert_to_dst_type( |
| | Affine.compute_w_affine(len(affine) - 1, inv_affine, data.shape[1:], orig_size), affine |
| | ) |
| | out.affine @= xform |
| | return out |
| |
|
| |
|
| | class Rand2DElastic(RandomizableTransform): |
| | """ |
| | Random elastic deformation and affine in 2D. |
| | A tutorial is available: https://github.com/Project-MONAI/tutorials/blob/0.6.0/modules/transforms_demo_2d.ipynb. |
| | |
| | """ |
| |
|
| | backend = Resample.backend |
| |
|
| | def __init__( |
| | self, |
| | spacing: tuple[float, float] | float, |
| | magnitude_range: tuple[float, float], |
| | prob: float = 0.1, |
| | rotate_range: RandRange = None, |
| | shear_range: RandRange = None, |
| | translate_range: RandRange = None, |
| | scale_range: RandRange = None, |
| | spatial_size: tuple[int, int] | int | None = None, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.REFLECTION, |
| | device: torch.device | None = None, |
| | ) -> None: |
| | """ |
| | Args: |
| | spacing : distance in between the control points. |
| | magnitude_range: the random offsets will be generated from ``uniform[magnitude[0], magnitude[1])``. |
| | prob: probability of returning a randomized elastic transform. |
| | defaults to 0.1, with 10% chance returns a randomized elastic transform, |
| | otherwise returns a ``spatial_size`` centered area extracted from the input image. |
| | rotate_range: angle range in radians. If element `i` is a pair of (min, max) values, then |
| | `uniform[-rotate_range[i][0], rotate_range[i][1])` will be used to generate the rotation parameter |
| | for the `i`th spatial dimension. If not, `uniform[-rotate_range[i], rotate_range[i])` will be used. |
| | This can be altered on a per-dimension basis. E.g., `((0,3), 1, ...)`: for dim0, rotation will be |
| | in range `[0, 3]`, and for dim1 `[-1, 1]` will be used. Setting a single value will use `[-x, x]` |
| | for dim0 and nothing for the remaining dimensions. |
| | shear_range: shear range with format matching `rotate_range`, it defines the range to randomly select |
| | shearing factors(a tuple of 2 floats for 2D) for affine matrix, take a 2D affine as example:: |
| | |
| | [ |
| | [1.0, params[0], 0.0], |
| | [params[1], 1.0, 0.0], |
| | [0.0, 0.0, 1.0], |
| | ] |
| | |
| | translate_range: translate range with format matching `rotate_range`, it defines the range to randomly |
| | select pixel to translate for every spatial dims. |
| | scale_range: scaling range with format matching `rotate_range`. it defines the range to randomly select |
| | the scale factor to translate for every spatial dims. A value of 1.0 is added to the result. |
| | This allows 0 to correspond to no change (i.e., a scaling of 1.0). |
| | spatial_size: specifying output image spatial size [h, w]. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | if some components of the `spatial_size` are non-positive values, the transform will use the |
| | corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| | to `(32, 64)` if the second spatial dimension size of img is `64`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"reflection"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | device: device on which the tensor will be allocated. |
| | |
| | See also: |
| | - :py:class:`RandAffineGrid` for the random affine parameters configurations. |
| | - :py:class:`Affine` for the affine transformation parameters configurations. |
| | |
| | """ |
| | RandomizableTransform.__init__(self, prob) |
| | self.deform_grid = RandDeformGrid(spacing=spacing, magnitude_range=magnitude_range, device=device) |
| | self.rand_affine_grid = RandAffineGrid( |
| | rotate_range=rotate_range, |
| | shear_range=shear_range, |
| | translate_range=translate_range, |
| | scale_range=scale_range, |
| | device=device, |
| | lazy=False, |
| | ) |
| | self.resampler = Resample(device=device) |
| |
|
| | self.device = device |
| | self.spatial_size = spatial_size |
| | self.mode = mode |
| | self.padding_mode: str = padding_mode |
| |
|
| | def set_random_state(self, seed: int | None = None, state: np.random.RandomState | None = None) -> Rand2DElastic: |
| | self.deform_grid.set_random_state(seed, state) |
| | self.rand_affine_grid.set_random_state(seed, state) |
| | super().set_random_state(seed, state) |
| | return self |
| |
|
| | def set_device(self, device): |
| | self.deform_grid.device = device |
| | self.rand_affine_grid.device = device |
| | self.resampler.device = device |
| | self.device = device |
| |
|
| | def randomize(self, spatial_size: Sequence[int]) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self.deform_grid.randomize(spatial_size) |
| | self.rand_affine_grid.randomize() |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | spatial_size: tuple[int, int] | int | None = None, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | randomize: bool = True, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W), |
| | spatial_size: specifying output image spatial size [h, w]. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | """ |
| | sp_size = fall_back_tuple(self.spatial_size if spatial_size is None else spatial_size, img.shape[1:]) |
| | if randomize: |
| | self.randomize(spatial_size=sp_size) |
| |
|
| | if self._do_transform: |
| | grid = self.deform_grid(spatial_size=sp_size) |
| | grid = self.rand_affine_grid(grid=grid) |
| | grid = torch.nn.functional.interpolate( |
| | recompute_scale_factor=True, |
| | input=grid.unsqueeze(0), |
| | scale_factor=list(ensure_tuple(self.deform_grid.spacing)), |
| | mode=InterpolateMode.BICUBIC.value, |
| | align_corners=False, |
| | ) |
| | grid = CenterSpatialCrop(roi_size=sp_size)(grid[0]) |
| | else: |
| | _device = img.device if isinstance(img, torch.Tensor) else self.device |
| | grid = cast(torch.Tensor, create_grid(spatial_size=sp_size, device=_device, backend="torch")) |
| | out: torch.Tensor = self.resampler( |
| | img, |
| | grid, |
| | mode=mode if mode is not None else self.mode, |
| | padding_mode=padding_mode if padding_mode is not None else self.padding_mode, |
| | ) |
| | return out |
| |
|
| |
|
| | class Rand3DElastic(RandomizableTransform): |
| | """ |
| | Random elastic deformation and affine in 3D. |
| | A tutorial is available: https://github.com/Project-MONAI/tutorials/blob/0.6.0/modules/transforms_demo_2d.ipynb. |
| | |
| | """ |
| |
|
| | backend = Resample.backend |
| |
|
| | def __init__( |
| | self, |
| | sigma_range: tuple[float, float], |
| | magnitude_range: tuple[float, float], |
| | prob: float = 0.1, |
| | rotate_range: RandRange = None, |
| | shear_range: RandRange = None, |
| | translate_range: RandRange = None, |
| | scale_range: RandRange = None, |
| | spatial_size: tuple[int, int, int] | int | None = None, |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.REFLECTION, |
| | device: torch.device | None = None, |
| | ) -> None: |
| | """ |
| | Args: |
| | sigma_range: a Gaussian kernel with standard deviation sampled from |
| | ``uniform[sigma_range[0], sigma_range[1])`` will be used to smooth the random offset grid. |
| | magnitude_range: the random offsets on the grid will be generated from |
| | ``uniform[magnitude[0], magnitude[1])``. |
| | prob: probability of returning a randomized elastic transform. |
| | defaults to 0.1, with 10% chance returns a randomized elastic transform, |
| | otherwise returns a ``spatial_size`` centered area extracted from the input image. |
| | rotate_range: angle range in radians. If element `i` is a pair of (min, max) values, then |
| | `uniform[-rotate_range[i][0], rotate_range[i][1])` will be used to generate the rotation parameter |
| | for the `i`th spatial dimension. If not, `uniform[-rotate_range[i], rotate_range[i])` will be used. |
| | This can be altered on a per-dimension basis. E.g., `((0,3), 1, ...)`: for dim0, rotation will be |
| | in range `[0, 3]`, and for dim1 `[-1, 1]` will be used. Setting a single value will use `[-x, x]` |
| | for dim0 and nothing for the remaining dimensions. |
| | shear_range: shear range with format matching `rotate_range`, it defines the range to randomly select |
| | shearing factors(a tuple of 6 floats for 3D) for affine matrix, take a 3D affine as example:: |
| | |
| | [ |
| | [1.0, params[0], params[1], 0.0], |
| | [params[2], 1.0, params[3], 0.0], |
| | [params[4], params[5], 1.0, 0.0], |
| | [0.0, 0.0, 0.0, 1.0], |
| | ] |
| | |
| | translate_range: translate range with format matching `rotate_range`, it defines the range to randomly |
| | select voxel to translate for every spatial dims. |
| | scale_range: scaling range with format matching `rotate_range`. it defines the range to randomly select |
| | the scale factor to translate for every spatial dims. A value of 1.0 is added to the result. |
| | This allows 0 to correspond to no change (i.e., a scaling of 1.0). |
| | spatial_size: specifying output image spatial size [h, w, d]. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | if some components of the `spatial_size` are non-positive values, the transform will use the |
| | corresponding components of img size. For example, `spatial_size=(32, 32, -1)` will be adapted |
| | to `(32, 32, 64)` if the third spatial dimension size of img is `64`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"reflection"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | device: device on which the tensor will be allocated. |
| | |
| | See also: |
| | - :py:class:`RandAffineGrid` for the random affine parameters configurations. |
| | - :py:class:`Affine` for the affine transformation parameters configurations. |
| | |
| | """ |
| | RandomizableTransform.__init__(self, prob) |
| | self.rand_affine_grid = RandAffineGrid( |
| | rotate_range=rotate_range, |
| | shear_range=shear_range, |
| | translate_range=translate_range, |
| | scale_range=scale_range, |
| | device=device, |
| | lazy=False, |
| | ) |
| | self.resampler = Resample(device=device) |
| |
|
| | self.sigma_range = sigma_range |
| | self.magnitude_range = magnitude_range |
| | self.spatial_size = spatial_size |
| | self.mode = mode |
| | self.padding_mode: str = padding_mode |
| | self.device = device |
| |
|
| | self.rand_offset: np.ndarray |
| | self.magnitude = 1.0 |
| | self.sigma = 1.0 |
| |
|
| | def set_random_state(self, seed: int | None = None, state: np.random.RandomState | None = None) -> Rand3DElastic: |
| | self.rand_affine_grid.set_random_state(seed, state) |
| | super().set_random_state(seed, state) |
| | return self |
| |
|
| | def set_device(self, device): |
| | self.rand_affine_grid.device = device |
| | self.resampler.device = device |
| | self.device = device |
| |
|
| | def randomize(self, grid_size: Sequence[int]) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return None |
| | self.rand_offset = self.R.uniform(-1.0, 1.0, [3] + list(grid_size)).astype(np.float32, copy=False) |
| | self.magnitude = self.R.uniform(self.magnitude_range[0], self.magnitude_range[1]) |
| | self.sigma = self.R.uniform(self.sigma_range[0], self.sigma_range[1]) |
| | self.rand_affine_grid.randomize() |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | spatial_size: tuple[int, int, int] | int | None = None, |
| | mode: str | int | None = None, |
| | padding_mode: str | None = None, |
| | randomize: bool = True, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W, D), |
| | spatial_size: specifying spatial 3D output image spatial size [h, w, d]. |
| | if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| | the transform will use the spatial size of `img`. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | randomize: whether to execute `randomize()` function first, default to True. |
| | """ |
| | sp_size = fall_back_tuple(self.spatial_size if spatial_size is None else spatial_size, img.shape[1:]) |
| | if randomize: |
| | self.randomize(grid_size=sp_size) |
| |
|
| | _device = img.device if isinstance(img, torch.Tensor) else self.device |
| | grid = create_grid(spatial_size=sp_size, device=_device, backend="torch") |
| | if self._do_transform: |
| | if self.rand_offset is None: |
| | raise RuntimeError("rand_offset is not initialized.") |
| | gaussian = GaussianFilter(3, self.sigma, 3.0).to(device=_device) |
| | offset = torch.as_tensor(self.rand_offset, device=_device).unsqueeze(0) |
| | grid[:3] += gaussian(offset)[0] * self.magnitude |
| | grid = self.rand_affine_grid(grid=grid) |
| | out: torch.Tensor = self.resampler( |
| | img, |
| | grid, |
| | mode=mode if mode is not None else self.mode, |
| | padding_mode=padding_mode if padding_mode is not None else self.padding_mode, |
| | ) |
| | return out |
| |
|
| |
|
| | class GridDistortion(Transform): |
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | num_cells: tuple[int] | int, |
| | distort_steps: Sequence[Sequence[float]], |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | device: torch.device | None = None, |
| | ) -> None: |
| | """ |
| | Grid distortion transform. Refer to: |
| | https://github.com/albumentations-team/albumentations/blob/master/albumentations/augmentations/transforms.py |
| | |
| | Args: |
| | num_cells: number of grid cells on each dimension. |
| | distort_steps: This argument is a list of tuples, where each tuple contains the distort steps of the |
| | corresponding dimensions (in the order of H, W[, D]). The length of each tuple equals to `num_cells + 1`. |
| | Each value in the tuple represents the distort step of the related cell. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | device: device on which the tensor will be allocated. |
| | |
| | """ |
| | self.resampler = Resample(mode=mode, padding_mode=padding_mode, device=device) |
| | self.num_cells = num_cells |
| | self.distort_steps = distort_steps |
| | self.device = device |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | distort_steps: Sequence[Sequence] | None = None, |
| | mode: str | None = None, |
| | padding_mode: str | None = None, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W[, D]). |
| | distort_steps: This argument is a list of tuples, where each tuple contains the distort steps of the |
| | corresponding dimensions (in the order of H, W[, D]). The length of each tuple equals to `num_cells + 1`. |
| | Each value in the tuple represents the distort step of the related cell. |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | |
| | """ |
| | distort_steps = self.distort_steps if distort_steps is None else distort_steps |
| | if len(img.shape) != len(distort_steps) + 1: |
| | raise ValueError("the spatial size of `img` does not match with the length of `distort_steps`") |
| |
|
| | all_ranges = [] |
| | num_cells = ensure_tuple_rep(self.num_cells, len(img.shape) - 1) |
| | if isinstance(img, MetaTensor) and img.pending_operations: |
| | warnings.warn("MetaTensor img has pending operations, transform may return incorrect results.") |
| | for dim_idx, dim_size in enumerate(img.shape[1:]): |
| | dim_distort_steps = distort_steps[dim_idx] |
| | ranges = torch.zeros(dim_size, dtype=torch.float32) |
| | cell_size = dim_size // num_cells[dim_idx] |
| | prev = 0 |
| | for idx in range(num_cells[dim_idx] + 1): |
| | start = int(idx * cell_size) |
| | end = start + cell_size |
| | if end > dim_size: |
| | end = dim_size |
| | cur = dim_size |
| | else: |
| | cur = prev + cell_size * dim_distort_steps[idx] |
| | ranges[start:end] = torch.linspace(prev, cur, end - start) |
| | prev = cur |
| | ranges = ranges - (dim_size - 1.0) / 2.0 |
| | all_ranges.append(ranges) |
| |
|
| | coords = meshgrid_ij(*all_ranges) |
| | grid = torch.stack([*coords, torch.ones_like(coords[0])]) |
| |
|
| | return self.resampler(img, grid=grid, mode=mode, padding_mode=padding_mode) |
| |
|
| |
|
| | class RandGridDistortion(RandomizableTransform): |
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__( |
| | self, |
| | num_cells: tuple[int] | int = 5, |
| | prob: float = 0.1, |
| | distort_limit: tuple[float, float] | float = (-0.03, 0.03), |
| | mode: str | int = GridSampleMode.BILINEAR, |
| | padding_mode: str = GridSamplePadMode.BORDER, |
| | device: torch.device | None = None, |
| | ) -> None: |
| | """ |
| | Random grid distortion transform. Refer to: |
| | https://github.com/albumentations-team/albumentations/blob/master/albumentations/augmentations/transforms.py |
| | |
| | Args: |
| | num_cells: number of grid cells on each dimension. |
| | prob: probability of returning a randomized grid distortion transform. Defaults to 0.1. |
| | distort_limit: range to randomly distort. |
| | If single number, distort_limit is picked from (-distort_limit, distort_limit). |
| | Defaults to (-0.03, 0.03). |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``"border"``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | device: device on which the tensor will be allocated. |
| | |
| | """ |
| | RandomizableTransform.__init__(self, prob) |
| | self.num_cells = num_cells |
| | if isinstance(distort_limit, (int, float)): |
| | self.distort_limit = (min(-distort_limit, distort_limit), max(-distort_limit, distort_limit)) |
| | else: |
| | self.distort_limit = (min(distort_limit), max(distort_limit)) |
| | self.distort_steps: Sequence[Sequence[float]] = ((1.0,),) |
| | self.grid_distortion = GridDistortion( |
| | num_cells=num_cells, distort_steps=self.distort_steps, mode=mode, padding_mode=padding_mode, device=device |
| | ) |
| |
|
| | def randomize(self, spatial_shape: Sequence[int]) -> None: |
| | super().randomize(None) |
| | if not self._do_transform: |
| | return |
| | self.distort_steps = tuple( |
| | tuple(1.0 + self.R.uniform(low=self.distort_limit[0], high=self.distort_limit[1], size=n_cells + 1)) |
| | for n_cells in ensure_tuple_rep(self.num_cells, len(spatial_shape)) |
| | ) |
| |
|
| | def __call__( |
| | self, img: torch.Tensor, mode: str | None = None, padding_mode: str | None = None, randomize: bool = True |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W[, D]). |
| | mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order 0-5 (integers). |
| | Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When it's an integer, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used |
| | and the value represents the order of the spline interpolation. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| | Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
| | When `mode` is an integer, using numpy/cupy backends, this argument accepts |
| | {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. |
| | See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html |
| | randomize: whether to shuffle the random factors using `randomize()`, default to True. |
| | """ |
| | if randomize: |
| | if isinstance(img, MetaTensor) and img.pending_operations: |
| | warnings.warn("MetaTensor img has pending operations, transform may return incorrect results.") |
| | self.randomize(img.shape[1:]) |
| | if not self._do_transform: |
| | return convert_to_tensor(img, track_meta=get_track_meta()) |
| | return self.grid_distortion(img, distort_steps=self.distort_steps, mode=mode, padding_mode=padding_mode) |
| |
|
| |
|
| | class GridSplit(Transform, MultiSampleTrait): |
| | """ |
| | Split the image into patches based on the provided grid in 2D. |
| | |
| | Args: |
| | grid: a tuple define the shape of the grid upon which the image is split. Defaults to (2, 2) |
| | size: a tuple or an integer that defines the output patch sizes. |
| | If it's an integer, the value will be repeated for each dimension. |
| | The default is None, where the patch size will be inferred from the grid shape. |
| | |
| | Example: |
| | Given an image (torch.Tensor or numpy.ndarray) with size of (3, 10, 10) and a grid of (2, 2), |
| | it will return a Tensor or array with the size of (4, 3, 5, 5). |
| | Here, if the `size` is provided, the returned shape will be (4, 3, size, size) |
| | |
| | Note: This transform currently support only image with two spatial dimensions. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, grid: tuple[int, int] = (2, 2), size: int | tuple[int, int] | None = None): |
| | |
| | self.grid = grid |
| |
|
| | |
| | self.size = None if size is None else ensure_tuple_rep(size, len(self.grid)) |
| |
|
| | def __call__( |
| | self, image: NdarrayOrTensor, size: int | tuple[int, int] | np.ndarray | None = None |
| | ) -> list[NdarrayOrTensor]: |
| | input_size = self.size if size is None else ensure_tuple_rep(size, len(self.grid)) |
| |
|
| | if self.grid == (1, 1) and input_size is None: |
| | return [image] |
| | if isinstance(image, MetaTensor) and image.pending_operations: |
| | warnings.warn("MetaTensor img has pending operations, transform may return incorrect results.") |
| | split_size, steps = self._get_params(image.shape[1:], input_size) |
| | patches: list[NdarrayOrTensor] |
| | as_strided_func: Callable |
| | if isinstance(image, torch.Tensor): |
| | as_strided_func = torch.as_strided |
| | c_stride, x_stride, y_stride = image.stride() |
| | elif isinstance(image, np.ndarray): |
| | as_strided_func = np.lib.stride_tricks.as_strided |
| | c_stride, x_stride, y_stride = image.strides |
| | else: |
| | raise ValueError(f"Input type [{type(image)}] is not supported.") |
| |
|
| | x_step, y_step = steps |
| | n_channels = image.shape[0] |
| | strided_image = as_strided_func( |
| | image, |
| | (*self.grid, n_channels, split_size[0], split_size[1]), |
| | (x_stride * x_step, y_stride * y_step, c_stride, x_stride, y_stride), |
| | ) |
| | |
| | strided_image = strided_image.reshape(-1, *strided_image.shape[2:]) |
| | |
| | if isinstance(image, torch.Tensor): |
| | patches = [p.contiguous() for p in strided_image] |
| | elif isinstance(image, np.ndarray): |
| | patches = [np.ascontiguousarray(p) for p in strided_image] |
| |
|
| | return patches |
| |
|
| | def _get_params(self, image_size: Sequence[int] | np.ndarray, size: Sequence[int] | np.ndarray | None = None): |
| | """ |
| | Calculate the size and step required for splitting the image |
| | Args: |
| | The size of the input image |
| | """ |
| | if size is None: |
| | |
| | size = tuple(image_size[i] // self.grid[i] for i in range(len(self.grid))) |
| |
|
| | if any(size[i] > image_size[i] for i in range(len(self.grid))): |
| | raise ValueError(f"The image size ({image_size})is smaller than the requested split size ({size})") |
| |
|
| | steps = tuple( |
| | (image_size[i] - size[i]) // (self.grid[i] - 1) if self.grid[i] > 1 else image_size[i] |
| | for i in range(len(self.grid)) |
| | ) |
| |
|
| | return size, steps |
| |
|
| |
|
| | class GridPatch(Transform, MultiSampleTrait): |
| | """ |
| | Extract all the patches sweeping the entire image in a row-major sliding-window manner with possible overlaps. |
| | It can sort the patches and return all or a subset of them. |
| | |
| | Args: |
| | patch_size: size of patches to generate slices for, 0 or None selects whole dimension |
| | offset: offset of starting position in the array, default is 0 for each dimension. |
| | num_patches: number of patches (or maximum number of patches) to return. |
| | If the requested number of patches is greater than the number of available patches, |
| | padding will be applied to provide exactly `num_patches` patches unless `threshold` is set. |
| | When `threshold` is set, this value is treated as the maximum number of patches. |
| | Defaults to None, which does not limit number of the patches. |
| | overlap: the amount of overlap of neighboring patches in each dimension (a value between 0.0 and 1.0). |
| | If only one float number is given, it will be applied to all dimensions. Defaults to 0.0. |
| | sort_fn: when `num_patches` is provided, it determines if keep patches with highest values (`"max"`), |
| | lowest values (`"min"`), or in their default order (`None`). Default to None. |
| | threshold: a value to keep only the patches whose sum of intensities are less than the threshold. |
| | Defaults to no filtering. |
| | pad_mode: the mode for padding the input image by `patch_size` to include patches that cross boundaries. |
| | Available modes: (Numpy) {``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | (PyTorch) {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. |
| | Defaults to `None`, which means no padding will be applied. |
| | See also: https://numpy.org/doc/stable/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | requires pytorch >= 1.10 for best compatibility. |
| | pad_kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | Returns: |
| | MetaTensor: the extracted patches as a single tensor (with patch dimension as the first dimension), |
| | with following metadata: |
| | |
| | - `PatchKeys.LOCATION`: the starting location of the patch in the image, |
| | - `PatchKeys.COUNT`: total number of patches in the image, |
| | - "spatial_shape": spatial size of the extracted patch, and |
| | - "offset": the amount of offset for the patches in the image (starting position of the first patch) |
| | |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | patch_size: Sequence[int], |
| | offset: Sequence[int] | None = None, |
| | num_patches: int | None = None, |
| | overlap: Sequence[float] | float = 0.0, |
| | sort_fn: str | None = None, |
| | threshold: float | None = None, |
| | pad_mode: str | None = None, |
| | **pad_kwargs, |
| | ): |
| | self.patch_size = ensure_tuple(patch_size) |
| | self.offset = ensure_tuple(offset) if offset else (0,) * len(self.patch_size) |
| | self.pad_mode = pad_mode |
| | self.pad_kwargs = pad_kwargs |
| | self.overlap = overlap |
| | self.num_patches = num_patches |
| | self.sort_fn = sort_fn.lower() if sort_fn else None |
| | self.threshold = threshold |
| |
|
| | def filter_threshold(self, image_np: NdarrayOrTensor, locations: np.ndarray) -> tuple[NdarrayOrTensor, np.ndarray]: |
| | """ |
| | Filter the patches and their locations according to a threshold. |
| | |
| | Args: |
| | image_np: a numpy.ndarray or torch.Tensor representing a stack of patches. |
| | locations: a numpy.ndarray representing the stack of location of each patch. |
| | |
| | Returns: |
| | tuple[NdarrayOrTensor, numpy.ndarray]: tuple of filtered patches and locations. |
| | """ |
| | n_dims = len(image_np.shape) |
| | idx = argwhere(image_np.sum(tuple(range(1, n_dims))) < self.threshold).reshape(-1) |
| | idx_np = convert_data_type(idx, np.ndarray)[0] |
| | return image_np[idx], locations[idx_np] |
| |
|
| | def filter_count(self, image_np: NdarrayOrTensor, locations: np.ndarray) -> tuple[NdarrayOrTensor, np.ndarray]: |
| | """ |
| | Sort the patches based on the sum of their intensity, and just keep `self.num_patches` of them. |
| | |
| | Args: |
| | image_np: a numpy.ndarray or torch.Tensor representing a stack of patches. |
| | locations: a numpy.ndarray representing the stack of location of each patch. |
| | """ |
| | if self.sort_fn is None: |
| | image_np = image_np[: self.num_patches] |
| | locations = locations[: self.num_patches] |
| | elif self.num_patches is not None: |
| | n_dims = len(image_np.shape) |
| | if self.sort_fn == GridPatchSort.MIN: |
| | idx = argsort(image_np.sum(tuple(range(1, n_dims)))) |
| | elif self.sort_fn == GridPatchSort.MAX: |
| | idx = argsort(-image_np.sum(tuple(range(1, n_dims)))) |
| | else: |
| | raise ValueError(f'`sort_fn` should be either "min", "max", or None! {self.sort_fn} provided!') |
| | idx = idx[: self.num_patches] |
| | idx_np = convert_data_type(idx, np.ndarray)[0] |
| | image_np = image_np[idx] |
| | locations = locations[idx_np] |
| | return image_np, locations |
| |
|
| | def __call__(self, array: NdarrayOrTensor) -> MetaTensor: |
| | """ |
| | Extract the patches (sweeping the entire image in a row-major sliding-window manner with possible overlaps). |
| | |
| | Args: |
| | array: a input image as `numpy.ndarray` or `torch.Tensor` |
| | |
| | Return: |
| | MetaTensor: the extracted patches as a single tensor (with patch dimension as the first dimension), |
| | with defined `PatchKeys.LOCATION` and `PatchKeys.COUNT` metadata. |
| | """ |
| | |
| | patch_iterator = iter_patch( |
| | array, |
| | patch_size=(None,) + self.patch_size, |
| | start_pos=(0,) + self.offset, |
| | overlap=self.overlap, |
| | copy_back=False, |
| | mode=self.pad_mode, |
| | **self.pad_kwargs, |
| | ) |
| | patches = list(zip(*patch_iterator)) |
| | patched_image = np.stack(patches[0]) if isinstance(array, np.ndarray) else torch.stack(patches[0]) |
| | locations = np.stack(patches[1])[:, 1:, 0] |
| |
|
| | |
| | if self.threshold is not None: |
| | patched_image, locations = self.filter_threshold(patched_image, locations) |
| |
|
| | |
| | if self.num_patches: |
| | |
| | patched_image, locations = self.filter_count(patched_image, locations) |
| | |
| | if self.threshold is None: |
| | padding = self.num_patches - len(patched_image) |
| | if padding > 0: |
| | |
| | constant_values = self.pad_kwargs.get("constant_values", 0) |
| | padding_shape = (padding, *list(patched_image.shape)[1:]) |
| | constant_padding: NdarrayOrTensor |
| | if isinstance(patched_image, np.ndarray): |
| | constant_padding = np.full(padding_shape, constant_values, dtype=patched_image.dtype) |
| | patched_image = np.concatenate([patched_image, constant_padding], axis=0) |
| | else: |
| | constant_padding = torch.full( |
| | padding_shape, |
| | constant_values, |
| | dtype=patched_image.dtype, |
| | layout=patched_image.layout, |
| | device=patched_image.device, |
| | ) |
| | patched_image = torch.cat([patched_image, constant_padding], dim=0) |
| | locations = np.pad(locations, [[0, padding], [0, 0]], constant_values=0) |
| |
|
| | |
| | metadata = array.meta if isinstance(array, MetaTensor) else MetaTensor.get_default_meta() |
| | metadata[PatchKeys.LOCATION] = locations.T |
| | metadata[PatchKeys.COUNT] = len(locations) |
| | metadata["spatial_shape"] = np.tile(np.array(self.patch_size), (len(locations), 1)).T |
| | metadata["offset"] = self.offset |
| | output = MetaTensor(x=patched_image, meta=metadata) |
| | output.is_batch = True |
| |
|
| | return output |
| |
|
| |
|
| | class RandGridPatch(GridPatch, RandomizableTransform, MultiSampleTrait): |
| | """ |
| | Extract all the patches sweeping the entire image in a row-major sliding-window manner with possible overlaps, |
| | and with random offset for the minimal corner of the image, (0,0) for 2D and (0,0,0) for 3D. |
| | It can sort the patches and return all or a subset of them. |
| | |
| | Args: |
| | patch_size: size of patches to generate slices for, 0 or None selects whole dimension |
| | min_offset: the minimum range of offset to be selected randomly. Defaults to 0. |
| | max_offset: the maximum range of offset to be selected randomly. |
| | Defaults to image size modulo patch size. |
| | num_patches: number of patches (or maximum number of patches) to return. |
| | If the requested number of patches is greater than the number of available patches, |
| | padding will be applied to provide exactly `num_patches` patches unless `threshold` is set. |
| | When `threshold` is set, this value is treated as the maximum number of patches. |
| | Defaults to None, which does not limit number of the patches. |
| | overlap: the amount of overlap of neighboring patches in each dimension (a value between 0.0 and 1.0). |
| | If only one float number is given, it will be applied to all dimensions. Defaults to 0.0. |
| | sort_fn: when `num_patches` is provided, it determines if keep patches with highest values (`"max"`), |
| | lowest values (`"min"`), in random ("random"), or in their default order (`None`). Default to None. |
| | threshold: a value to keep only the patches whose sum of intensities are less than the threshold. |
| | Defaults to no filtering. |
| | pad_mode: the mode for padding the input image by `patch_size` to include patches that cross boundaries. |
| | Available modes: (Numpy) {``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | (PyTorch) {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. |
| | Defaults to `None`, which means no padding will be applied. |
| | See also: https://numpy.org/doc/stable/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | requires pytorch >= 1.10 for best compatibility. |
| | pad_kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | Returns: |
| | MetaTensor: the extracted patches as a single tensor (with patch dimension as the first dimension), |
| | with following metadata: |
| | |
| | - `PatchKeys.LOCATION`: the starting location of the patch in the image, |
| | - `PatchKeys.COUNT`: total number of patches in the image, |
| | - "spatial_shape": spatial size of the extracted patch, and |
| | - "offset": the amount of offset for the patches in the image (starting position of the first patch) |
| | |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | patch_size: Sequence[int], |
| | min_offset: Sequence[int] | int | None = None, |
| | max_offset: Sequence[int] | int | None = None, |
| | num_patches: int | None = None, |
| | overlap: Sequence[float] | float = 0.0, |
| | sort_fn: str | None = None, |
| | threshold: float | None = None, |
| | pad_mode: str | None = None, |
| | **pad_kwargs, |
| | ): |
| | super().__init__( |
| | patch_size=patch_size, |
| | offset=(), |
| | num_patches=num_patches, |
| | overlap=overlap, |
| | sort_fn=sort_fn, |
| | threshold=threshold, |
| | pad_mode=pad_mode, |
| | **pad_kwargs, |
| | ) |
| | self.min_offset = min_offset |
| | self.max_offset = max_offset |
| | self.num_patches = num_patches |
| | self.sort_fn = sort_fn |
| |
|
| | def randomize(self, array): |
| | if self.min_offset is None: |
| | min_offset = (0,) * len(self.patch_size) |
| | else: |
| | min_offset = ensure_tuple_rep(self.min_offset, len(self.patch_size)) |
| | if self.max_offset is None: |
| | max_offset = tuple(s % p for s, p in zip(array.shape[1:], self.patch_size)) |
| | else: |
| | max_offset = ensure_tuple_rep(self.max_offset, len(self.patch_size)) |
| |
|
| | self.offset = tuple(self.R.randint(low=low, high=high + 1) for low, high in zip(min_offset, max_offset)) |
| |
|
| | def filter_count(self, image_np: NdarrayOrTensor, locations: np.ndarray) -> tuple[NdarrayOrTensor, np.ndarray]: |
| | if self.sort_fn == GridPatchSort.RANDOM: |
| | idx = self.R.permutation(image_np.shape[0]) |
| | idx = idx[: self.num_patches] |
| | idx_np = convert_data_type(idx, np.ndarray)[0] |
| | image_np = image_np[idx] |
| | locations = locations[idx_np] |
| | return image_np, locations |
| | elif self.sort_fn not in (None, GridPatchSort.MIN, GridPatchSort.MAX): |
| | raise ValueError(f'`sort_fn` should be either "min", "max", "random" or None! {self.sort_fn} provided!') |
| | return super().filter_count(image_np, locations) |
| |
|
| | def __call__(self, array: NdarrayOrTensor, randomize: bool = True): |
| | if randomize: |
| | self.randomize(array) |
| | return super().__call__(array) |
| |
|
| |
|
| | class RandSimulateLowResolution(RandomizableTransform): |
| | """ |
| | Random simulation of low resolution corresponding to nnU-Net's SimulateLowResolutionTransform |
| | (https://github.com/MIC-DKFZ/batchgenerators/blob/7651ece69faf55263dd582a9f5cbd149ed9c3ad0/batchgenerators/transforms/resample_transforms.py#L23) |
| | First, the array/tensor is resampled at lower resolution as determined by the zoom_factor which is uniformly sampled |
| | from the `zoom_range`. Then, the array/tensor is resampled at the original resolution. |
| | """ |
| |
|
| | backend = Affine.backend |
| |
|
| | def __init__( |
| | self, |
| | prob: float = 0.1, |
| | downsample_mode: InterpolateMode | str = InterpolateMode.NEAREST, |
| | upsample_mode: InterpolateMode | str = InterpolateMode.TRILINEAR, |
| | zoom_range: Sequence[float] = (0.5, 1.0), |
| | align_corners=False, |
| | device: torch.device | None = None, |
| | ) -> None: |
| | """ |
| | Args: |
| | prob: probability of performing this augmentation |
| | downsample_mode: interpolation mode for downsampling operation |
| | upsample_mode: interpolation mode for upsampling operation |
| | zoom_range: range from which the random zoom factor for the downsampling and upsampling operation is |
| | sampled. It determines the shape of the downsampled tensor. |
| | align_corners: This only has an effect when downsample_mode or upsample_mode is 'linear', 'bilinear', |
| | 'bicubic' or 'trilinear'. Default: False |
| | See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html |
| | device: device on which the tensor will be allocated. |
| | |
| | """ |
| | RandomizableTransform.__init__(self, prob) |
| |
|
| | self.downsample_mode = downsample_mode |
| | self.upsample_mode = upsample_mode |
| | self.zoom_range = zoom_range |
| | self.align_corners = align_corners |
| | self.device = device |
| | self.zoom_factor = 1.0 |
| |
|
| | def randomize(self, data: Any | None = None) -> None: |
| | super().randomize(None) |
| | self.zoom_factor = self.R.uniform(self.zoom_range[0], self.zoom_range[1]) |
| | if not self._do_transform: |
| | return None |
| |
|
| | def __call__(self, img: torch.Tensor, randomize: bool = True) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: shape must be (num_channels, H, W[, D]), |
| | randomize: whether to execute `randomize()` function first, defaults to True. |
| | """ |
| | if randomize: |
| | self.randomize() |
| |
|
| | if self._do_transform: |
| | input_shape = img.shape[1:] |
| | target_shape = np.round(np.array(input_shape) * self.zoom_factor).astype(np.int_) |
| |
|
| | resize_tfm_downsample = Resize( |
| | spatial_size=target_shape, size_mode="all", mode=self.downsample_mode, anti_aliasing=False |
| | ) |
| |
|
| | resize_tfm_upsample = Resize( |
| | spatial_size=input_shape, |
| | size_mode="all", |
| | mode=self.upsample_mode, |
| | anti_aliasing=False, |
| | align_corners=self.align_corners, |
| | ) |
| | |
| | |
| | original_tack_meta_value = get_track_meta() |
| | set_track_meta(False) |
| |
|
| | img_downsampled = resize_tfm_downsample(img) |
| | img_upsampled = resize_tfm_upsample(img_downsampled) |
| |
|
| | |
| | set_track_meta(original_tack_meta_value) |
| |
|
| | |
| | img_upsampled = MetaTensor(img_upsampled) |
| | img_upsampled.copy_meta_from(img) |
| |
|
| | return img_upsampled |
| |
|
| | else: |
| | return img |
| |
|