| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | A collection of "vanilla" transforms for crop and pad operations. |
| | """ |
| |
|
| | from __future__ import annotations |
| |
|
| | import warnings |
| | from collections.abc import Callable, Sequence |
| | from itertools import chain |
| | from math import ceil |
| | from typing import Any |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai.config import IndexSelection |
| | from monai.config.type_definitions import NdarrayOrTensor |
| | from monai.data.meta_obj import get_track_meta |
| | from monai.data.meta_tensor import MetaTensor |
| | from monai.data.utils import get_random_patch, get_valid_patch_size |
| | from monai.transforms.croppad.functional import crop_func, pad_func |
| | from monai.transforms.inverse import InvertibleTransform, TraceableTransform |
| | from monai.transforms.traits import MultiSampleTrait |
| | from monai.transforms.transform import LazyTransform, Randomizable, Transform |
| | from monai.transforms.utils import ( |
| | compute_divisible_spatial_size, |
| | generate_label_classes_crop_centers, |
| | generate_pos_neg_label_crop_centers, |
| | generate_spatial_bounding_box, |
| | is_positive, |
| | map_binary_to_indices, |
| | map_classes_to_indices, |
| | weighted_patch_samples, |
| | ) |
| | from monai.utils import ImageMetaKey as Key |
| | from monai.utils import ( |
| | LazyAttr, |
| | Method, |
| | PytorchPadMode, |
| | TraceKeys, |
| | TransformBackends, |
| | convert_data_type, |
| | convert_to_tensor, |
| | deprecated_arg_default, |
| | ensure_tuple, |
| | ensure_tuple_rep, |
| | fall_back_tuple, |
| | look_up_option, |
| | pytorch_after, |
| | ) |
| |
|
| | __all__ = [ |
| | "Pad", |
| | "SpatialPad", |
| | "BorderPad", |
| | "DivisiblePad", |
| | "Crop", |
| | "SpatialCrop", |
| | "CenterSpatialCrop", |
| | "CenterScaleCrop", |
| | "RandSpatialCrop", |
| | "RandScaleCrop", |
| | "RandSpatialCropSamples", |
| | "CropForeground", |
| | "RandWeightedCrop", |
| | "RandCropByPosNegLabel", |
| | "RandCropByLabelClasses", |
| | "ResizeWithPadOrCrop", |
| | "BoundingRect", |
| | ] |
| |
|
| |
|
| | class Pad(InvertibleTransform, LazyTransform): |
| | """ |
| | Perform padding for a given an amount of padding in each dimension. |
| | |
| | `torch.nn.functional.pad` is used unless the mode or kwargs are not available in torch, |
| | in which case `np.pad` will be used. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | to_pad: the amount to pad in each dimension (including the channel) [(low_H, high_H), (low_W, high_W), ...]. |
| | if None, must provide in the `__call__` at runtime. |
| | mode: available modes: (Numpy) {``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | (PyTorch) {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | requires pytorch >= 1.10 for best compatibility. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | to_pad: tuple[tuple[int, int]] | None = None, |
| | mode: str = PytorchPadMode.CONSTANT, |
| | lazy: bool = False, |
| | **kwargs, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy) |
| | self.to_pad = to_pad |
| | self.mode = mode |
| | self.kwargs = kwargs |
| |
|
| | def compute_pad_width(self, spatial_shape: Sequence[int]) -> tuple[tuple[int, int]]: |
| | """ |
| | dynamically compute the pad width according to the spatial shape. |
| | the output is the amount of padding for all dimensions including the channel. |
| | |
| | Args: |
| | spatial_shape: spatial shape of the original image. |
| | |
| | """ |
| | raise NotImplementedError(f"subclass {self.__class__.__name__} must implement this method.") |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | to_pad: tuple[tuple[int, int]] | None = None, |
| | mode: str | None = None, |
| | lazy: bool | None = None, |
| | **kwargs, |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: data to be transformed, assuming `img` is channel-first and padding doesn't apply to the channel dim. |
| | to_pad: the amount to be padded in each dimension [(low_H, high_H), (low_W, high_W), ...]. |
| | default to `self.to_pad`. |
| | mode: available modes: (Numpy) {``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | (PyTorch) {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | lazy: a flag to override the lazy behaviour for this call, if set. Defaults to None. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| | to_pad_ = self.to_pad if to_pad is None else to_pad |
| | if to_pad_ is None: |
| | spatial_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | to_pad_ = self.compute_pad_width(spatial_shape) |
| | mode_ = self.mode if mode is None else mode |
| | kwargs_ = dict(self.kwargs) |
| | kwargs_.update(kwargs) |
| |
|
| | img_t = convert_to_tensor(data=img, track_meta=get_track_meta()) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return pad_func(img_t, to_pad_, self.get_transform_info(), mode_, lazy_, **kwargs_) |
| |
|
| | def inverse(self, data: MetaTensor) -> MetaTensor: |
| | transform = self.pop_transform(data) |
| | padded = transform[TraceKeys.EXTRA_INFO]["padded"] |
| | if padded[0][0] > 0 or padded[0][1] > 0: |
| | s = padded[0][0] |
| | e = min(max(padded[0][1], s + 1), len(data)) |
| | data = data[s : len(data) - e] |
| | roi_start = [i[0] for i in padded[1:]] |
| | roi_end = [i - j[1] for i, j in zip(data.shape[1:], padded[1:])] |
| | cropper = SpatialCrop(roi_start=roi_start, roi_end=roi_end) |
| | with cropper.trace_transform(False): |
| | return cropper(data) |
| |
|
| |
|
| | class SpatialPad(Pad): |
| | """ |
| | Performs padding to the data, symmetric for all sides or all on one side for each dimension. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_size: the spatial size of output data after padding, if a dimension of the input |
| | data size is larger than the pad size, will not pad that dimension. |
| | If its components have non-positive values, the corresponding size of input image will be used |
| | (no padding). for example: if the spatial size of input data is [30, 30, 30] and |
| | `spatial_size=[32, 25, -1]`, the spatial size of output data will be [32, 30, 30]. |
| | method: {``"symmetric"``, ``"end"``} |
| | Pad image symmetrically on every side or only pad at the end sides. Defaults to ``"symmetric"``. |
| | mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | spatial_size: Sequence[int] | int | tuple[tuple[int, ...] | int, ...], |
| | method: str = Method.SYMMETRIC, |
| | mode: str = PytorchPadMode.CONSTANT, |
| | lazy: bool = False, |
| | **kwargs, |
| | ) -> None: |
| | self.spatial_size = spatial_size |
| | self.method: Method = look_up_option(method, Method) |
| | super().__init__(mode=mode, lazy=lazy, **kwargs) |
| |
|
| | def compute_pad_width(self, spatial_shape: Sequence[int]) -> tuple[tuple[int, int]]: |
| | """ |
| | dynamically compute the pad width according to the spatial shape. |
| | |
| | Args: |
| | spatial_shape: spatial shape of the original image. |
| | |
| | """ |
| | spatial_size = fall_back_tuple(self.spatial_size, spatial_shape) |
| | if self.method == Method.SYMMETRIC: |
| | pad_width = [] |
| | for i, sp_i in enumerate(spatial_size): |
| | width = max(sp_i - spatial_shape[i], 0) |
| | pad_width.append((int(width // 2), int(width - (width // 2)))) |
| | else: |
| | pad_width = [(0, int(max(sp_i - spatial_shape[i], 0))) for i, sp_i in enumerate(spatial_size)] |
| | return tuple([(0, 0)] + pad_width) |
| |
|
| |
|
| | class BorderPad(Pad): |
| | """ |
| | Pad the input data by adding specified borders to every dimension. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_border: specified size for every spatial border. Any -ve values will be set to 0. It can be 3 shapes: |
| | |
| | - single int number, pad all the borders with the same size. |
| | - length equals the length of image shape, pad every spatial dimension separately. |
| | for example, image shape(CHW) is [1, 4, 4], spatial_border is [2, 1], |
| | pad every border of H dim with 2, pad every border of W dim with 1, result shape is [1, 8, 6]. |
| | - length equals 2 x (length of image shape), pad every border of every dimension separately. |
| | for example, image shape(CHW) is [1, 4, 4], spatial_border is [1, 2, 3, 4], pad top of H dim with 1, |
| | pad bottom of H dim with 2, pad left of W dim with 3, pad right of W dim with 4. |
| | the result shape is [1, 7, 11]. |
| | mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, spatial_border: Sequence[int] | int, mode: str = PytorchPadMode.CONSTANT, lazy: bool = False, **kwargs |
| | ) -> None: |
| | self.spatial_border = spatial_border |
| | super().__init__(mode=mode, lazy=lazy, **kwargs) |
| |
|
| | def compute_pad_width(self, spatial_shape: Sequence[int]) -> tuple[tuple[int, int]]: |
| | spatial_border = ensure_tuple(self.spatial_border) |
| | if not all(isinstance(b, int) for b in spatial_border): |
| | raise ValueError(f"self.spatial_border must contain only ints, got {spatial_border}.") |
| | spatial_border = tuple(max(0, b) for b in spatial_border) |
| |
|
| | if len(spatial_border) == 1: |
| | data_pad_width = [(int(spatial_border[0]), int(spatial_border[0])) for _ in spatial_shape] |
| | elif len(spatial_border) == len(spatial_shape): |
| | data_pad_width = [(int(sp), int(sp)) for sp in spatial_border[: len(spatial_shape)]] |
| | elif len(spatial_border) == len(spatial_shape) * 2: |
| | data_pad_width = [ |
| | (int(spatial_border[2 * i]), int(spatial_border[2 * i + 1])) for i in range(len(spatial_shape)) |
| | ] |
| | else: |
| | raise ValueError( |
| | f"Unsupported spatial_border length: {len(spatial_border)}, available options are " |
| | f"[1, len(spatial_shape)={len(spatial_shape)}, 2*len(spatial_shape)={2*len(spatial_shape)}]." |
| | ) |
| | return tuple([(0, 0)] + data_pad_width) |
| |
|
| |
|
| | class DivisiblePad(Pad): |
| | """ |
| | Pad the input data, so that the spatial sizes are divisible by `k`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | backend = SpatialPad.backend |
| |
|
| | def __init__( |
| | self, |
| | k: Sequence[int] | int, |
| | mode: str = PytorchPadMode.CONSTANT, |
| | method: str = Method.SYMMETRIC, |
| | lazy: bool = False, |
| | **kwargs, |
| | ) -> None: |
| | """ |
| | Args: |
| | k: the target k for each spatial dimension. |
| | if `k` is negative or 0, the original size is preserved. |
| | if `k` is an int, the same `k` be applied to all the input spatial dimensions. |
| | mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | method: {``"symmetric"``, ``"end"``} |
| | Pad image symmetrically on every side or only pad at the end sides. Defaults to ``"symmetric"``. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | See also :py:class:`monai.transforms.SpatialPad` |
| | """ |
| | self.k = k |
| | self.method: Method = Method(method) |
| | super().__init__(mode=mode, lazy=lazy, **kwargs) |
| |
|
| | def compute_pad_width(self, spatial_shape: Sequence[int]) -> tuple[tuple[int, int]]: |
| | new_size = compute_divisible_spatial_size(spatial_shape=spatial_shape, k=self.k) |
| | spatial_pad = SpatialPad(spatial_size=new_size, method=self.method) |
| | return spatial_pad.compute_pad_width(spatial_shape) |
| |
|
| |
|
| | class Crop(InvertibleTransform, LazyTransform): |
| | """ |
| | Perform crop operations on the input image. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH] |
| |
|
| | def __init__(self, lazy: bool = False): |
| | LazyTransform.__init__(self, lazy) |
| |
|
| | @staticmethod |
| | def compute_slices( |
| | roi_center: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_size: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_start: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_end: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_slices: Sequence[slice] | None = None, |
| | ) -> tuple[slice]: |
| | """ |
| | Compute the crop slices based on specified `center & size` or `start & end` or `slices`. |
| | |
| | Args: |
| | roi_center: voxel coordinates for center of the crop ROI. |
| | roi_size: size of the crop ROI, if a dimension of ROI size is larger than image size, |
| | will not crop that dimension of the image. |
| | roi_start: voxel coordinates for start of the crop ROI. |
| | roi_end: voxel coordinates for end of the crop ROI, if a coordinate is out of image, |
| | use the end coordinate of image. |
| | roi_slices: list of slices for each of the spatial dimensions. |
| | |
| | """ |
| | roi_start_t: torch.Tensor |
| |
|
| | if roi_slices: |
| | if not all(s.step is None or s.step == 1 for s in roi_slices): |
| | raise ValueError(f"only slice steps of 1/None are currently supported, got {roi_slices}.") |
| | return ensure_tuple(roi_slices) |
| | else: |
| | if roi_center is not None and roi_size is not None: |
| | roi_center_t = convert_to_tensor(data=roi_center, dtype=torch.int16, wrap_sequence=True, device="cpu") |
| | roi_size_t = convert_to_tensor(data=roi_size, dtype=torch.int16, wrap_sequence=True, device="cpu") |
| | _zeros = torch.zeros_like(roi_center_t) |
| | half = ( |
| | torch.divide(roi_size_t, 2, rounding_mode="floor") |
| | if pytorch_after(1, 8) |
| | else torch.floor_divide(roi_size_t, 2) |
| | ) |
| | roi_start_t = torch.maximum(roi_center_t - half, _zeros) |
| | roi_end_t = torch.maximum(roi_start_t + roi_size_t, roi_start_t) |
| | else: |
| | if roi_start is None or roi_end is None: |
| | raise ValueError("please specify either roi_center, roi_size or roi_start, roi_end.") |
| | roi_start_t = convert_to_tensor(data=roi_start, dtype=torch.int16, wrap_sequence=True) |
| | roi_start_t = torch.maximum(roi_start_t, torch.zeros_like(roi_start_t)) |
| | roi_end_t = convert_to_tensor(data=roi_end, dtype=torch.int16, wrap_sequence=True) |
| | roi_end_t = torch.maximum(roi_end_t, roi_start_t) |
| | |
| | if roi_start_t.numel() == 1: |
| | return ensure_tuple([slice(int(roi_start_t.item()), int(roi_end_t.item()))]) |
| | return ensure_tuple([slice(int(s), int(e)) for s, e in zip(roi_start_t.tolist(), roi_end_t.tolist())]) |
| |
|
| | def __call__( |
| | self, img: torch.Tensor, slices: tuple[slice, ...], lazy: bool | None = None |
| | ) -> torch.Tensor: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | slicing doesn't apply to the channel dim. |
| | |
| | """ |
| | slices_ = list(slices) |
| | sd = len(img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:]) |
| | if len(slices_) < sd: |
| | slices_ += [slice(None)] * (sd - len(slices_)) |
| | |
| | slices_ = list([slice(None)] + slices_[:sd]) |
| |
|
| | img_t: MetaTensor = convert_to_tensor(data=img, track_meta=get_track_meta()) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return crop_func(img_t, tuple(slices_), lazy_, self.get_transform_info()) |
| |
|
| | def inverse(self, img: MetaTensor) -> MetaTensor: |
| | transform = self.pop_transform(img) |
| | cropped = transform[TraceKeys.EXTRA_INFO]["cropped"] |
| | |
| | inverse_transform = BorderPad(cropped) |
| | |
| | with inverse_transform.trace_transform(False): |
| | return inverse_transform(img) |
| |
|
| |
|
| | class SpatialCrop(Crop): |
| | """ |
| | General purpose cropper to produce sub-volume region of interest (ROI). |
| | If a dimension of the expected ROI size is larger than the input image size, will not crop that dimension. |
| | So the cropped result may be smaller than the expected ROI, and the cropped results of several images may |
| | not have exactly the same shape. |
| | It can support to crop ND spatial (channel-first) data. |
| | |
| | The cropped region can be parameterised in various ways: |
| | - a list of slices for each spatial dimension (allows for use of negative indexing and `None`) |
| | - a spatial center and size |
| | - the start and end coordinates of the ROI |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | roi_center: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_size: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_start: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_end: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_slices: Sequence[slice] | None = None, |
| | lazy: bool = False, |
| | ) -> None: |
| | """ |
| | Args: |
| | roi_center: voxel coordinates for center of the crop ROI. |
| | roi_size: size of the crop ROI, if a dimension of ROI size is larger than image size, |
| | will not crop that dimension of the image. |
| | roi_start: voxel coordinates for start of the crop ROI. |
| | roi_end: voxel coordinates for end of the crop ROI, if a coordinate is out of image, |
| | use the end coordinate of image. |
| | roi_slices: list of slices for each of the spatial dimensions. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | |
| | """ |
| | super().__init__(lazy) |
| | self.slices = self.compute_slices( |
| | roi_center=roi_center, roi_size=roi_size, roi_start=roi_start, roi_end=roi_end, roi_slices=roi_slices |
| | ) |
| |
|
| | def __call__(self, img: torch.Tensor, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | slicing doesn't apply to the channel dim. |
| | |
| | """ |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return super().__call__(img=img, slices=ensure_tuple(self.slices), lazy=lazy_) |
| |
|
| |
|
| | class CenterSpatialCrop(Crop): |
| | """ |
| | Crop at the center of image with specified ROI size. |
| | If a dimension of the expected ROI size is larger than the input image size, will not crop that dimension. |
| | So the cropped result may be smaller than the expected ROI, and the cropped results of several images may |
| | not have exactly the same shape. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | roi_size: the spatial size of the crop region e.g. [224,224,128] |
| | if a dimension of ROI size is larger than image size, will not crop that dimension of the image. |
| | If its components have non-positive values, the corresponding size of input image will be used. |
| | for example: if the spatial size of input data is [40, 40, 40] and `roi_size=[32, 64, -1]`, |
| | the spatial size of output data will be [32, 40, 40]. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | def __init__(self, roi_size: Sequence[int] | int, lazy: bool = False) -> None: |
| | super().__init__(lazy=lazy) |
| | self.roi_size = roi_size |
| |
|
| | def compute_slices(self, spatial_size: Sequence[int]) -> tuple[slice]: |
| | roi_size = fall_back_tuple(self.roi_size, spatial_size) |
| | roi_center = [i // 2 for i in spatial_size] |
| | return super().compute_slices(roi_center=roi_center, roi_size=roi_size) |
| |
|
| | def __call__(self, img: torch.Tensor, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | slicing doesn't apply to the channel dim. |
| | |
| | """ |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return super().__call__( |
| | img=img, |
| | slices=self.compute_slices(img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:]), |
| | lazy=lazy_, |
| | ) |
| |
|
| |
|
| | class CenterScaleCrop(Crop): |
| | """ |
| | Crop at the center of image with specified scale of ROI size. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | roi_scale: specifies the expected scale of image size to crop. e.g. [0.3, 0.4, 0.5] or a number for all dims. |
| | If its components have non-positive values, will use `1.0` instead, which means the input image size. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | def __init__(self, roi_scale: Sequence[float] | float, lazy: bool = False): |
| | super().__init__(lazy=lazy) |
| | self.roi_scale = roi_scale |
| |
|
| | def __call__(self, img: torch.Tensor, lazy: bool | None = None) -> torch.Tensor: |
| | img_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | ndim = len(img_size) |
| | roi_size = [ceil(r * s) for r, s in zip(ensure_tuple_rep(self.roi_scale, ndim), img_size)] |
| | lazy_ = self.lazy if lazy is None else lazy |
| | cropper = CenterSpatialCrop(roi_size=roi_size, lazy=lazy_) |
| | return super().__call__(img=img, slices=cropper.compute_slices(img_size), lazy=lazy_) |
| |
|
| |
|
| | class RandSpatialCrop(Randomizable, Crop): |
| | """ |
| | Crop image with random size or specific size ROI. It can crop at a random position as center |
| | or at the image center. And allows to set the minimum and maximum size to limit the randomly generated ROI. |
| | |
| | Note: even `random_size=False`, if a dimension of the expected ROI size is larger than the input image size, |
| | will not crop that dimension. So the cropped result may be smaller than the expected ROI, and the cropped results |
| | of several images may not have exactly the same shape. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | roi_size: if `random_size` is True, it specifies the minimum crop region. |
| | if `random_size` is False, it specifies the expected ROI size to crop. e.g. [224, 224, 128] |
| | if a dimension of ROI size is larger than image size, will not crop that dimension of the image. |
| | If its components have non-positive values, the corresponding size of input image will be used. |
| | for example: if the spatial size of input data is [40, 40, 40] and `roi_size=[32, 64, -1]`, |
| | the spatial size of output data will be [32, 40, 40]. |
| | max_roi_size: if `random_size` is True and `roi_size` specifies the min crop region size, `max_roi_size` |
| | can specify the max crop region size. if None, defaults to the input image size. |
| | if its components have non-positive values, the corresponding size of input image will be used. |
| | random_center: crop at random position as center or the image center. |
| | random_size: crop with random size or specific size ROI. |
| | if True, the actual size is sampled from `randint(roi_size, max_roi_size + 1)`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | roi_size: Sequence[int] | int, |
| | max_roi_size: Sequence[int] | int | None = None, |
| | random_center: bool = True, |
| | random_size: bool = False, |
| | lazy: bool = False, |
| | ) -> None: |
| | super().__init__(lazy) |
| | self.roi_size = roi_size |
| | self.max_roi_size = max_roi_size |
| | self.random_center = random_center |
| | self.random_size = random_size |
| | self._size: Sequence[int] | None = None |
| | self._slices: tuple[slice, ...] |
| |
|
| | def randomize(self, img_size: Sequence[int]) -> None: |
| | self._size = fall_back_tuple(self.roi_size, img_size) |
| | if self.random_size: |
| | max_size = img_size if self.max_roi_size is None else fall_back_tuple(self.max_roi_size, img_size) |
| | if any(i > j for i, j in zip(self._size, max_size)): |
| | raise ValueError(f"min ROI size: {self._size} is larger than max ROI size: {max_size}.") |
| | self._size = tuple(self.R.randint(low=self._size[i], high=max_size[i] + 1) for i in range(len(img_size))) |
| | if self.random_center: |
| | valid_size = get_valid_patch_size(img_size, self._size) |
| | self._slices = get_random_patch(img_size, valid_size, self.R) |
| |
|
| | def __call__(self, img: torch.Tensor, randomize: bool = True, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | slicing doesn't apply to the channel dim. |
| | |
| | """ |
| | img_size = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | if randomize: |
| | self.randomize(img_size) |
| | if self._size is None: |
| | raise RuntimeError("self._size not specified.") |
| | lazy_ = self.lazy if lazy is None else lazy |
| | if self.random_center: |
| | return super().__call__(img=img, slices=self._slices, lazy=lazy_) |
| | cropper = CenterSpatialCrop(self._size, lazy=lazy_) |
| | return super().__call__(img=img, slices=cropper.compute_slices(img_size), lazy=lazy_) |
| |
|
| |
|
| | class RandScaleCrop(RandSpatialCrop): |
| | """ |
| | Subclass of :py:class:`monai.transforms.RandSpatialCrop`. Crop image with |
| | random size or specific size ROI. It can crop at a random position as |
| | center or at the image center. And allows to set the minimum and maximum |
| | scale of image size to limit the randomly generated ROI. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | roi_scale: if `random_size` is True, it specifies the minimum crop size: `roi_scale * image spatial size`. |
| | if `random_size` is False, it specifies the expected scale of image size to crop. e.g. [0.3, 0.4, 0.5]. |
| | If its components have non-positive values, will use `1.0` instead, which means the input image size. |
| | max_roi_scale: if `random_size` is True and `roi_scale` specifies the min crop region size, `max_roi_scale` |
| | can specify the max crop region size: `max_roi_scale * image spatial size`. |
| | if None, defaults to the input image size. if its components have non-positive values, |
| | will use `1.0` instead, which means the input image size. |
| | random_center: crop at random position as center or the image center. |
| | random_size: crop with random size or specified size ROI by `roi_scale * image spatial size`. |
| | if True, the actual size is sampled from |
| | `randint(roi_scale * image spatial size, max_roi_scale * image spatial size + 1)`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | roi_scale: Sequence[float] | float, |
| | max_roi_scale: Sequence[float] | float | None = None, |
| | random_center: bool = True, |
| | random_size: bool = False, |
| | lazy: bool = False, |
| | ) -> None: |
| | super().__init__( |
| | roi_size=-1, max_roi_size=None, random_center=random_center, random_size=random_size, lazy=lazy |
| | ) |
| | self.roi_scale = roi_scale |
| | self.max_roi_scale = max_roi_scale |
| |
|
| | def get_max_roi_size(self, img_size): |
| | ndim = len(img_size) |
| | self.roi_size = [ceil(r * s) for r, s in zip(ensure_tuple_rep(self.roi_scale, ndim), img_size)] |
| | if self.max_roi_scale is not None: |
| | self.max_roi_size = [ceil(r * s) for r, s in zip(ensure_tuple_rep(self.max_roi_scale, ndim), img_size)] |
| | else: |
| | self.max_roi_size = None |
| |
|
| | def randomize(self, img_size: Sequence[int]) -> None: |
| | self.get_max_roi_size(img_size) |
| | super().randomize(img_size) |
| |
|
| | def __call__(self, img: torch.Tensor, randomize: bool = True, lazy: bool | None = None) -> torch.Tensor: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | slicing doesn't apply to the channel dim. |
| | |
| | """ |
| | self.get_max_roi_size(img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:]) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | return super().__call__(img=img, randomize=randomize, lazy=lazy_) |
| |
|
| |
|
| | class RandSpatialCropSamples(Randomizable, TraceableTransform, LazyTransform, MultiSampleTrait): |
| | """ |
| | Crop image with random size or specific size ROI to generate a list of N samples. |
| | It can crop at a random position as center or at the image center. And allows to set |
| | the minimum size to limit the randomly generated ROI. |
| | It will return a list of cropped images. |
| | |
| | Note: even `random_size=False`, if a dimension of the expected ROI size is larger than the input image size, |
| | will not crop that dimension. So the cropped result may be smaller than the expected ROI, and the cropped |
| | results of several images may not have exactly the same shape. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | roi_size: if `random_size` is True, it specifies the minimum crop region. |
| | if `random_size` is False, it specifies the expected ROI size to crop. e.g. [224, 224, 128] |
| | if a dimension of ROI size is larger than image size, will not crop that dimension of the image. |
| | If its components have non-positive values, the corresponding size of input image will be used. |
| | for example: if the spatial size of input data is [40, 40, 40] and `roi_size=[32, 64, -1]`, |
| | the spatial size of output data will be [32, 40, 40]. |
| | num_samples: number of samples (crop regions) to take in the returned list. |
| | max_roi_size: if `random_size` is True and `roi_size` specifies the min crop region size, `max_roi_size` |
| | can specify the max crop region size. if None, defaults to the input image size. |
| | if its components have non-positive values, the corresponding size of input image will be used. |
| | random_center: crop at random position as center or the image center. |
| | random_size: crop with random size or specific size ROI. |
| | The actual size is sampled from `randint(roi_size, img_size)`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | |
| | Raises: |
| | ValueError: When ``num_samples`` is nonpositive. |
| | |
| | """ |
| |
|
| | backend = RandSpatialCrop.backend |
| |
|
| | def __init__( |
| | self, |
| | roi_size: Sequence[int] | int, |
| | num_samples: int, |
| | max_roi_size: Sequence[int] | int | None = None, |
| | random_center: bool = True, |
| | random_size: bool = False, |
| | lazy: bool = False, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy) |
| | if num_samples < 1: |
| | raise ValueError(f"num_samples must be positive, got {num_samples}.") |
| | self.num_samples = num_samples |
| | self.cropper = RandSpatialCrop(roi_size, max_roi_size, random_center, random_size, lazy) |
| |
|
| | def set_random_state( |
| | self, seed: int | None = None, state: np.random.RandomState | None = None |
| | ) -> RandSpatialCropSamples: |
| | super().set_random_state(seed, state) |
| | self.cropper.set_random_state(seed, state) |
| | return self |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, value: bool) -> None: |
| | self._lazy = value |
| | self.cropper.lazy = value |
| |
|
| | def randomize(self, data: Any | None = None) -> None: |
| | pass |
| |
|
| | def __call__(self, img: torch.Tensor, lazy: bool | None = None) -> list[torch.Tensor]: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | cropping doesn't change the channel dim. |
| | """ |
| | ret = [] |
| | lazy_ = self.lazy if lazy is None else lazy |
| | for i in range(self.num_samples): |
| | cropped = self.cropper(img, lazy=lazy_) |
| | if get_track_meta(): |
| | cropped.meta[Key.PATCH_INDEX] = i |
| | self.push_transform(cropped, replace=True, lazy=lazy_) |
| | ret.append(cropped) |
| | return ret |
| |
|
| |
|
| | class CropForeground(Crop): |
| | """ |
| | Crop an image using a bounding box. The bounding box is generated by selecting foreground using select_fn |
| | at channels channel_indices. margin is added in each spatial dimension of the bounding box. |
| | The typical usage is to help training and evaluation if the valid part is small in the whole medical image. |
| | Users can define arbitrary function to select expected foreground from the whole image or specified channels. |
| | And it can also add margin to every dim of the bounding box of foreground object. |
| | For example: |
| | |
| | .. code-block:: python |
| | |
| | image = np.array( |
| | [[[0, 0, 0, 0, 0], |
| | [0, 1, 2, 1, 0], |
| | [0, 1, 3, 2, 0], |
| | [0, 1, 2, 1, 0], |
| | [0, 0, 0, 0, 0]]]) # 1x5x5, single channel 5x5 image |
| | |
| | |
| | def threshold_at_one(x): |
| | # threshold at 1 |
| | return x > 1 |
| | |
| | |
| | cropper = CropForeground(select_fn=threshold_at_one, margin=0) |
| | print(cropper(image)) |
| | [[[2, 1], |
| | [3, 2], |
| | [2, 1]]] |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | """ |
| |
|
| | @deprecated_arg_default("allow_smaller", old_default=True, new_default=False, since="1.2", replaced="1.5") |
| | def __init__( |
| | self, |
| | select_fn: Callable = is_positive, |
| | channel_indices: IndexSelection | None = None, |
| | margin: Sequence[int] | int = 0, |
| | allow_smaller: bool = True, |
| | return_coords: bool = False, |
| | k_divisible: Sequence[int] | int = 1, |
| | mode: str = PytorchPadMode.CONSTANT, |
| | lazy: bool = False, |
| | **pad_kwargs, |
| | ) -> None: |
| | """ |
| | Args: |
| | select_fn: function to select expected foreground, default is to select values > 0. |
| | channel_indices: if defined, select foreground only on the specified channels |
| | of image. if None, select foreground on the whole image. |
| | margin: add margin value to spatial dims of the bounding box, if only 1 value provided, use it for all dims. |
| | allow_smaller: when computing box size with `margin`, whether to allow the image edges to be smaller than the |
| | final box edges. If `False`, part of a padded output box might be outside of the original image, if `True`, |
| | the image edges will be used as the box edges. Default to `True`. |
| | return_coords: whether return the coordinates of spatial bounding box for foreground. |
| | k_divisible: make each spatial dimension to be divisible by k, default to 1. |
| | if `k_divisible` is an int, the same `k` be applied to all the input spatial dimensions. |
| | mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | pad_kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| | LazyTransform.__init__(self, lazy) |
| | self.select_fn = select_fn |
| | self.channel_indices = ensure_tuple(channel_indices) if channel_indices is not None else None |
| | self.margin = margin |
| | self.allow_smaller = allow_smaller |
| | self.return_coords = return_coords |
| | self.k_divisible = k_divisible |
| | self.padder = Pad(mode=mode, lazy=lazy, **pad_kwargs) |
| |
|
| | @Crop.lazy.setter |
| | def lazy(self, _val: bool): |
| | self._lazy = _val |
| | self.padder.lazy = _val |
| |
|
| | @property |
| | def requires_current_data(self): |
| | return False |
| |
|
| | def compute_bounding_box(self, img: torch.Tensor) -> tuple[np.ndarray, np.ndarray]: |
| | """ |
| | Compute the start points and end points of bounding box to crop. |
| | And adjust bounding box coords to be divisible by `k`. |
| | |
| | """ |
| | box_start, box_end = generate_spatial_bounding_box( |
| | img, self.select_fn, self.channel_indices, self.margin, self.allow_smaller |
| | ) |
| | box_start_, *_ = convert_data_type(box_start, output_type=np.ndarray, dtype=np.int16, wrap_sequence=True) |
| | box_end_, *_ = convert_data_type(box_end, output_type=np.ndarray, dtype=np.int16, wrap_sequence=True) |
| | orig_spatial_size = box_end_ - box_start_ |
| | |
| | spatial_size = np.asarray(compute_divisible_spatial_size(orig_spatial_size.tolist(), k=self.k_divisible)) |
| | |
| | box_start_ = box_start_ - np.floor_divide(np.asarray(spatial_size) - orig_spatial_size, 2) |
| | box_end_ = box_start_ + spatial_size |
| | return box_start_, box_end_ |
| |
|
| | def crop_pad( |
| | self, |
| | img: torch.Tensor, |
| | box_start: np.ndarray, |
| | box_end: np.ndarray, |
| | mode: str | None = None, |
| | lazy: bool = False, |
| | **pad_kwargs, |
| | ) -> torch.Tensor: |
| | """ |
| | Crop and pad based on the bounding box. |
| | |
| | """ |
| | slices = self.compute_slices(roi_start=box_start, roi_end=box_end) |
| | cropped = super().__call__(img=img, slices=slices, lazy=lazy) |
| | pad_to_start = np.maximum(-box_start, 0) |
| | pad_to_end = np.maximum( |
| | box_end - np.asarray(img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:]), 0 |
| | ) |
| | pad = list(chain(*zip(pad_to_start.tolist(), pad_to_end.tolist()))) |
| | pad_width = BorderPad(spatial_border=pad).compute_pad_width( |
| | cropped.peek_pending_shape() if isinstance(cropped, MetaTensor) else cropped.shape[1:] |
| | ) |
| | ret = self.padder.__call__(img=cropped, to_pad=pad_width, mode=mode, lazy=lazy, **pad_kwargs) |
| | |
| | |
| | if get_track_meta() and isinstance(ret, MetaTensor): |
| | if not lazy: |
| | ret.applied_operations[-1][TraceKeys.EXTRA_INFO]["pad_info"] = ret.applied_operations.pop() |
| | else: |
| | pad_info = ret.pending_operations.pop() |
| | crop_info = ret.pending_operations.pop() |
| | extra = crop_info[TraceKeys.EXTRA_INFO] |
| | extra["pad_info"] = pad_info |
| | self.push_transform( |
| | ret, |
| | orig_size=crop_info.get(TraceKeys.ORIG_SIZE), |
| | sp_size=pad_info[LazyAttr.SHAPE], |
| | affine=crop_info[LazyAttr.AFFINE] @ pad_info[LazyAttr.AFFINE], |
| | lazy=lazy, |
| | extra_info=extra, |
| | ) |
| | return ret |
| |
|
| | def __call__( |
| | self, img: torch.Tensor, mode: str | None = None, lazy: bool | None = None, **pad_kwargs |
| | ) -> torch.Tensor: |
| | """ |
| | Apply the transform to `img`, assuming `img` is channel-first and |
| | slicing doesn't change the channel dim. |
| | """ |
| | box_start, box_end = self.compute_bounding_box(img) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | cropped = self.crop_pad(img, box_start, box_end, mode, lazy=lazy_, **pad_kwargs) |
| |
|
| | if self.return_coords: |
| | return cropped, box_start, box_end |
| | return cropped |
| |
|
| | def inverse(self, img: MetaTensor) -> MetaTensor: |
| | transform = self.get_most_recent_transform(img) |
| | |
| | pad_info = transform[TraceKeys.EXTRA_INFO].pop("pad_info") |
| | img.applied_operations.append(pad_info) |
| | |
| | inv = self.padder.inverse(img) |
| | |
| | return super().inverse(inv) |
| |
|
| |
|
| | class RandWeightedCrop(Randomizable, TraceableTransform, LazyTransform, MultiSampleTrait): |
| | """ |
| | Samples a list of `num_samples` image patches according to the provided `weight_map`. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_size: the spatial size of the image patch e.g. [224, 224, 128]. |
| | If its components have non-positive values, the corresponding size of `img` will be used. |
| | num_samples: number of samples (image patches) to take in the returned list. |
| | weight_map: weight map used to generate patch samples. The weights must be non-negative. |
| | Each element denotes a sampling weight of the spatial location. 0 indicates no sampling. |
| | It should be a single-channel array in shape, for example, `(1, spatial_dim_0, spatial_dim_1, ...)`. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | backend = SpatialCrop.backend |
| |
|
| | def __init__( |
| | self, |
| | spatial_size: Sequence[int] | int, |
| | num_samples: int = 1, |
| | weight_map: NdarrayOrTensor | None = None, |
| | lazy: bool = False, |
| | ): |
| | LazyTransform.__init__(self, lazy) |
| | self.spatial_size = ensure_tuple(spatial_size) |
| | self.num_samples = int(num_samples) |
| | self.weight_map = weight_map |
| | self.centers: list[np.ndarray] = [] |
| |
|
| | def randomize(self, weight_map: NdarrayOrTensor) -> None: |
| | self.centers = weighted_patch_samples( |
| | spatial_size=self.spatial_size, w=weight_map[0], n_samples=self.num_samples, r_state=self.R |
| | ) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, _val: bool): |
| | self._lazy = _val |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | weight_map: NdarrayOrTensor | None = None, |
| | randomize: bool = True, |
| | lazy: bool | None = None, |
| | ) -> list[torch.Tensor]: |
| | """ |
| | Args: |
| | img: input image to sample patches from. assuming `img` is a channel-first array. |
| | weight_map: weight map used to generate patch samples. The weights must be non-negative. |
| | Each element denotes a sampling weight of the spatial location. 0 indicates no sampling. |
| | It should be a single-channel array in shape, for example, `(1, spatial_dim_0, spatial_dim_1, ...)` |
| | randomize: whether to execute random operations, default to `True`. |
| | lazy: a flag to override the lazy behaviour for this call, if set. Defaults to None. |
| | |
| | Returns: |
| | A list of image patches |
| | """ |
| | img_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| |
|
| | if randomize: |
| | if weight_map is None: |
| | weight_map = self.weight_map |
| | if weight_map is None: |
| | raise ValueError("weight map must be provided for weighted patch sampling.") |
| | w_shape = weight_map.peek_pending_shape() if isinstance(weight_map, MetaTensor) else weight_map.shape[1:] |
| | if img_shape != w_shape: |
| | warnings.warn(f"image and weight map spatial shape mismatch: {img_shape} vs {w_shape}.") |
| | self.randomize(weight_map) |
| |
|
| | _spatial_size = fall_back_tuple(self.spatial_size, img_shape) |
| | results: list[torch.Tensor] = [] |
| | lazy_ = self.lazy if lazy is None else lazy |
| | for i, center in enumerate(self.centers): |
| | cropper = SpatialCrop(roi_center=center, roi_size=_spatial_size, lazy=lazy_) |
| | cropped = cropper(img) |
| | if get_track_meta(): |
| | ret_: MetaTensor = cropped |
| | ret_.meta[Key.PATCH_INDEX] = i |
| | ret_.meta["crop_center"] = center |
| | self.push_transform(ret_, replace=True, lazy=lazy_) |
| | results.append(cropped) |
| | return results |
| |
|
| |
|
| | class RandCropByPosNegLabel(Randomizable, TraceableTransform, LazyTransform, MultiSampleTrait): |
| | """ |
| | Crop random fixed sized regions with the center being a foreground or background voxel |
| | based on the Pos Neg Ratio. |
| | And will return a list of arrays for all the cropped images. |
| | For example, crop two (3 x 3) arrays from (5 x 5) array with pos/neg=1:: |
| | |
| | [[[0, 0, 0, 0, 0], |
| | [0, 1, 2, 1, 0], [[0, 1, 2], [[2, 1, 0], |
| | [0, 1, 3, 0, 0], --> [0, 1, 3], [3, 0, 0], |
| | [0, 0, 0, 0, 0], [0, 0, 0]] [0, 0, 0]] |
| | [0, 0, 0, 0, 0]]] |
| | |
| | If a dimension of the expected spatial size is larger than the input image size, |
| | will not crop that dimension. So the cropped result may be smaller than expected size, and the cropped |
| | results of several images may not have exactly same shape. |
| | And if the crop ROI is partly out of the image, will automatically adjust the crop center to ensure the |
| | valid crop ROI. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_size: the spatial size of the crop region e.g. [224, 224, 128]. |
| | if a dimension of ROI size is larger than image size, will not crop that dimension of the image. |
| | if its components have non-positive values, the corresponding size of `label` will be used. |
| | for example: if the spatial size of input data is [40, 40, 40] and `spatial_size=[32, 64, -1]`, |
| | the spatial size of output data will be [32, 40, 40]. |
| | label: the label image that is used for finding foreground/background, if None, must set at |
| | `self.__call__`. Non-zero indicates foreground, zero indicates background. |
| | pos: used with `neg` together to calculate the ratio ``pos / (pos + neg)`` for the probability |
| | to pick a foreground voxel as a center rather than a background voxel. |
| | neg: used with `pos` together to calculate the ratio ``pos / (pos + neg)`` for the probability |
| | to pick a foreground voxel as a center rather than a background voxel. |
| | num_samples: number of samples (crop regions) to take in each list. |
| | image: optional image data to help select valid area, can be same as `img` or another image array. |
| | if not None, use ``label == 0 & image > image_threshold`` to select the negative |
| | sample (background) center. So the crop center will only come from the valid image areas. |
| | image_threshold: if enabled `image`, use ``image > image_threshold`` to determine |
| | the valid image content areas. |
| | fg_indices: if provided pre-computed foreground indices of `label`, will ignore above `image` and |
| | `image_threshold`, and randomly select crop centers based on them, need to provide `fg_indices` |
| | and `bg_indices` together, expect to be 1 dim array of spatial indices after flattening. |
| | a typical usage is to call `FgBgToIndices` transform first and cache the results. |
| | bg_indices: if provided pre-computed background indices of `label`, will ignore above `image` and |
| | `image_threshold`, and randomly select crop centers based on them, need to provide `fg_indices` |
| | and `bg_indices` together, expect to be 1 dim array of spatial indices after flattening. |
| | a typical usage is to call `FgBgToIndices` transform first and cache the results. |
| | allow_smaller: if `False`, an exception will be raised if the image is smaller than |
| | the requested ROI in any dimension. If `True`, any smaller dimensions will be set to |
| | match the cropped size (i.e., no cropping in that dimension). |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | |
| | Raises: |
| | ValueError: When ``pos`` or ``neg`` are negative. |
| | ValueError: When ``pos=0`` and ``neg=0``. Incompatible values. |
| | |
| | """ |
| |
|
| | backend = SpatialCrop.backend |
| |
|
| | def __init__( |
| | self, |
| | spatial_size: Sequence[int] | int, |
| | label: torch.Tensor | None = None, |
| | pos: float = 1.0, |
| | neg: float = 1.0, |
| | num_samples: int = 1, |
| | image: torch.Tensor | None = None, |
| | image_threshold: float = 0.0, |
| | fg_indices: NdarrayOrTensor | None = None, |
| | bg_indices: NdarrayOrTensor | None = None, |
| | allow_smaller: bool = False, |
| | lazy: bool = False, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy) |
| | self.spatial_size = spatial_size |
| | self.label = label |
| | if pos < 0 or neg < 0: |
| | raise ValueError(f"pos and neg must be nonnegative, got pos={pos} neg={neg}.") |
| | if pos + neg == 0: |
| | raise ValueError("Incompatible values: pos=0 and neg=0.") |
| | self.pos_ratio = pos / (pos + neg) |
| | self.num_samples = num_samples |
| | self.image = image |
| | self.image_threshold = image_threshold |
| | self.centers: tuple[tuple] | None = None |
| | self.fg_indices = fg_indices |
| | self.bg_indices = bg_indices |
| | self.allow_smaller = allow_smaller |
| |
|
| | def randomize( |
| | self, |
| | label: torch.Tensor | None = None, |
| | fg_indices: NdarrayOrTensor | None = None, |
| | bg_indices: NdarrayOrTensor | None = None, |
| | image: torch.Tensor | None = None, |
| | ) -> None: |
| | fg_indices_ = self.fg_indices if fg_indices is None else fg_indices |
| | bg_indices_ = self.bg_indices if bg_indices is None else bg_indices |
| | if fg_indices_ is None or bg_indices_ is None: |
| | if label is None: |
| | raise ValueError("label must be provided.") |
| | fg_indices_, bg_indices_ = map_binary_to_indices(label, image, self.image_threshold) |
| | _shape = None |
| | if label is not None: |
| | _shape = label.peek_pending_shape() if isinstance(label, MetaTensor) else label.shape[1:] |
| | elif image is not None: |
| | _shape = image.peek_pending_shape() if isinstance(image, MetaTensor) else image.shape[1:] |
| | if _shape is None: |
| | raise ValueError("label or image must be provided to get the spatial shape.") |
| | self.centers = generate_pos_neg_label_crop_centers( |
| | self.spatial_size, |
| | self.num_samples, |
| | self.pos_ratio, |
| | _shape, |
| | fg_indices_, |
| | bg_indices_, |
| | self.R, |
| | self.allow_smaller, |
| | ) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, _val: bool): |
| | self._lazy = _val |
| |
|
| | @property |
| | def requires_current_data(self): |
| | return False |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | label: torch.Tensor | None = None, |
| | image: torch.Tensor | None = None, |
| | fg_indices: NdarrayOrTensor | None = None, |
| | bg_indices: NdarrayOrTensor | None = None, |
| | randomize: bool = True, |
| | lazy: bool | None = None, |
| | ) -> list[torch.Tensor]: |
| | """ |
| | Args: |
| | img: input data to crop samples from based on the pos/neg ratio of `label` and `image`. |
| | Assumes `img` is a channel-first array. |
| | label: the label image that is used for finding foreground/background, if None, use `self.label`. |
| | image: optional image data to help select valid area, can be same as `img` or another image array. |
| | use ``label == 0 & image > image_threshold`` to select the negative sample(background) center. |
| | so the crop center will only exist on valid image area. if None, use `self.image`. |
| | fg_indices: foreground indices to randomly select crop centers, |
| | need to provide `fg_indices` and `bg_indices` together. |
| | bg_indices: background indices to randomly select crop centers, |
| | need to provide `fg_indices` and `bg_indices` together. |
| | randomize: whether to execute the random operations, default to `True`. |
| | lazy: a flag to override the lazy behaviour for this call, if set. Defaults to None. |
| | |
| | """ |
| | if image is None: |
| | image = self.image |
| | if randomize: |
| | if label is None: |
| | label = self.label |
| | self.randomize(label, fg_indices, bg_indices, image) |
| | results: list[torch.Tensor] = [] |
| | if self.centers is not None: |
| | img_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | roi_size = fall_back_tuple(self.spatial_size, default=img_shape) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | for i, center in enumerate(self.centers): |
| | cropper = SpatialCrop(roi_center=center, roi_size=roi_size, lazy=lazy_) |
| | cropped = cropper(img) |
| | if get_track_meta(): |
| | ret_: MetaTensor = cropped |
| | ret_.meta[Key.PATCH_INDEX] = i |
| | ret_.meta["crop_center"] = center |
| | self.push_transform(ret_, replace=True, lazy=lazy_) |
| | results.append(cropped) |
| | return results |
| |
|
| |
|
| | class RandCropByLabelClasses(Randomizable, TraceableTransform, LazyTransform, MultiSampleTrait): |
| | """ |
| | Crop random fixed sized regions with the center being a class based on the specified ratios of every class. |
| | The label data can be One-Hot format array or Argmax data. And will return a list of arrays for all the |
| | cropped images. For example, crop two (3 x 3) arrays from (5 x 5) array with `ratios=[1, 2, 3, 1]`:: |
| | |
| | image = np.array([ |
| | [[0.0, 0.3, 0.4, 0.2, 0.0], |
| | [0.0, 0.1, 0.2, 0.1, 0.4], |
| | [0.0, 0.3, 0.5, 0.2, 0.0], |
| | [0.1, 0.2, 0.1, 0.1, 0.0], |
| | [0.0, 0.1, 0.2, 0.1, 0.0]] |
| | ]) |
| | label = np.array([ |
| | [[0, 0, 0, 0, 0], |
| | [0, 1, 2, 1, 0], |
| | [0, 1, 3, 0, 0], |
| | [0, 0, 0, 0, 0], |
| | [0, 0, 0, 0, 0]] |
| | ]) |
| | cropper = RandCropByLabelClasses( |
| | spatial_size=[3, 3], |
| | ratios=[1, 2, 3, 1], |
| | num_classes=4, |
| | num_samples=2, |
| | ) |
| | label_samples = cropper(img=label, label=label, image=image) |
| | |
| | The 2 randomly cropped samples of `label` can be: |
| | [[0, 1, 2], [[0, 0, 0], |
| | [0, 1, 3], [1, 2, 1], |
| | [0, 0, 0]] [1, 3, 0]] |
| | |
| | If a dimension of the expected spatial size is larger than the input image size, |
| | will not crop that dimension. So the cropped result may be smaller than expected size, and the cropped |
| | results of several images may not have exactly same shape. |
| | And if the crop ROI is partly out of the image, will automatically adjust the crop center to ensure the |
| | valid crop ROI. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_size: the spatial size of the crop region e.g. [224, 224, 128]. |
| | if a dimension of ROI size is larger than image size, will not crop that dimension of the image. |
| | if its components have non-positive values, the corresponding size of `label` will be used. |
| | for example: if the spatial size of input data is [40, 40, 40] and `spatial_size=[32, 64, -1]`, |
| | the spatial size of output data will be [32, 40, 40]. |
| | ratios: specified ratios of every class in the label to generate crop centers, including background class. |
| | if None, every class will have the same ratio to generate crop centers. |
| | label: the label image that is used for finding every class, if None, must set at `self.__call__`. |
| | num_classes: number of classes for argmax label, not necessary for One-Hot label. |
| | num_samples: number of samples (crop regions) to take in each list. |
| | image: if image is not None, only return the indices of every class that are within the valid |
| | region of the image (``image > image_threshold``). |
| | image_threshold: if enabled `image`, use ``image > image_threshold`` to |
| | determine the valid image content area and select class indices only in this area. |
| | indices: if provided pre-computed indices of every class, will ignore above `image` and |
| | `image_threshold`, and randomly select crop centers based on them, expect to be 1 dim array |
| | of spatial indices after flattening. a typical usage is to call `ClassesToIndices` transform first |
| | and cache the results for better performance. |
| | allow_smaller: if `False`, an exception will be raised if the image is smaller than |
| | the requested ROI in any dimension. If `True`, any smaller dimensions will remain |
| | unchanged. |
| | warn: if `True` prints a warning if a class is not present in the label. |
| | max_samples_per_class: maximum length of indices to sample in each class to reduce memory consumption. |
| | Default is None, no subsampling. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | """ |
| |
|
| | backend = SpatialCrop.backend |
| |
|
| | def __init__( |
| | self, |
| | spatial_size: Sequence[int] | int, |
| | ratios: list[float | int] | None = None, |
| | label: torch.Tensor | None = None, |
| | num_classes: int | None = None, |
| | num_samples: int = 1, |
| | image: torch.Tensor | None = None, |
| | image_threshold: float = 0.0, |
| | indices: list[NdarrayOrTensor] | None = None, |
| | allow_smaller: bool = False, |
| | warn: bool = True, |
| | max_samples_per_class: int | None = None, |
| | lazy: bool = False, |
| | ) -> None: |
| | LazyTransform.__init__(self, lazy) |
| | self.spatial_size = spatial_size |
| | self.ratios = ratios |
| | self.label = label |
| | self.num_classes = num_classes |
| | self.num_samples = num_samples |
| | self.image = image |
| | self.image_threshold = image_threshold |
| | self.centers: tuple[tuple] | None = None |
| | self.indices = indices |
| | self.allow_smaller = allow_smaller |
| | self.warn = warn |
| | self.max_samples_per_class = max_samples_per_class |
| |
|
| | def randomize( |
| | self, |
| | label: torch.Tensor | None = None, |
| | indices: list[NdarrayOrTensor] | None = None, |
| | image: torch.Tensor | None = None, |
| | ) -> None: |
| | indices_ = self.indices if indices is None else indices |
| | if indices_ is None: |
| | if label is None: |
| | raise ValueError("label must not be None.") |
| | indices_ = map_classes_to_indices( |
| | label, self.num_classes, image, self.image_threshold, self.max_samples_per_class |
| | ) |
| | _shape = None |
| | if label is not None: |
| | _shape = label.peek_pending_shape() if isinstance(label, MetaTensor) else label.shape[1:] |
| | elif image is not None: |
| | _shape = image.peek_pending_shape() if isinstance(image, MetaTensor) else image.shape[1:] |
| | if _shape is None: |
| | raise ValueError("label or image must be provided to infer the output spatial shape.") |
| | self.centers = generate_label_classes_crop_centers( |
| | self.spatial_size, self.num_samples, _shape, indices_, self.ratios, self.R, self.allow_smaller, self.warn |
| | ) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, _val: bool): |
| | self._lazy = _val |
| |
|
| | @property |
| | def requires_current_data(self): |
| | return False |
| |
|
| | def __call__( |
| | self, |
| | img: torch.Tensor, |
| | label: torch.Tensor | None = None, |
| | image: torch.Tensor | None = None, |
| | indices: list[NdarrayOrTensor] | None = None, |
| | randomize: bool = True, |
| | lazy: bool | None = None, |
| | ) -> list[torch.Tensor]: |
| | """ |
| | Args: |
| | img: input data to crop samples from based on the ratios of every class, assumes `img` is a |
| | channel-first array. |
| | label: the label image that is used for finding indices of every class, if None, use `self.label`. |
| | image: optional image data to help select valid area, can be same as `img` or another image array. |
| | use ``image > image_threshold`` to select the centers only in valid region. if None, use `self.image`. |
| | indices: list of indices for every class in the image, used to randomly select crop centers. |
| | randomize: whether to execute the random operations, default to `True`. |
| | lazy: a flag to override the lazy behaviour for this call, if set. Defaults to None. |
| | """ |
| | if image is None: |
| | image = self.image |
| | if randomize: |
| | if label is None: |
| | label = self.label |
| | self.randomize(label, indices, image) |
| | results: list[torch.Tensor] = [] |
| | if self.centers is not None: |
| | img_shape = img.peek_pending_shape() if isinstance(img, MetaTensor) else img.shape[1:] |
| | roi_size = fall_back_tuple(self.spatial_size, default=img_shape) |
| | lazy_ = self.lazy if lazy is None else lazy |
| | for i, center in enumerate(self.centers): |
| | cropper = SpatialCrop(roi_center=tuple(center), roi_size=roi_size, lazy=lazy_) |
| | cropped = cropper(img) |
| | if get_track_meta(): |
| | ret_: MetaTensor = cropped |
| | ret_.meta[Key.PATCH_INDEX] = i |
| | ret_.meta["crop_center"] = center |
| | self.push_transform(ret_, replace=True, lazy=lazy_) |
| | results.append(cropped) |
| |
|
| | return results |
| |
|
| |
|
| | class ResizeWithPadOrCrop(InvertibleTransform, LazyTransform): |
| | """ |
| | Resize an image to a target spatial size by either centrally cropping the image or |
| | padding it evenly with a user-specified mode. |
| | When the dimension is smaller than the target size, do symmetric padding along that dim. |
| | When the dimension is larger than the target size, do central cropping along that dim. |
| | |
| | This transform is capable of lazy execution. See the :ref:`Lazy Resampling topic<lazy_resampling>` |
| | for more information. |
| | |
| | Args: |
| | spatial_size: the spatial size of output data after padding or crop. |
| | If has non-positive values, the corresponding size of input image will be used (no padding). |
| | method: {``"symmetric"``, ``"end"``} |
| | Pad image symmetrically on every side or only pad at the end sides. Defaults to ``"symmetric"``. |
| | mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | pad_kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | lazy: a flag to indicate whether this transform should execute lazily or not. Defaults to False. |
| | |
| | """ |
| |
|
| | backend = list(set(SpatialPad.backend) & set(CenterSpatialCrop.backend)) |
| |
|
| | def __init__( |
| | self, |
| | spatial_size: Sequence[int] | int, |
| | method: str = Method.SYMMETRIC, |
| | mode: str = PytorchPadMode.CONSTANT, |
| | lazy: bool = False, |
| | **pad_kwargs, |
| | ): |
| | LazyTransform.__init__(self, lazy) |
| | self.padder = SpatialPad(spatial_size=spatial_size, method=method, mode=mode, lazy=lazy, **pad_kwargs) |
| | self.cropper = CenterSpatialCrop(roi_size=spatial_size, lazy=lazy) |
| |
|
| | @LazyTransform.lazy.setter |
| | def lazy(self, val: bool): |
| | self.padder.lazy = val |
| | self.cropper.lazy = val |
| | self._lazy = val |
| |
|
| | def __call__( |
| | self, img: torch.Tensor, mode: str | None = None, lazy: bool | None = None, **pad_kwargs |
| | ) -> torch.Tensor: |
| | """ |
| | Args: |
| | img: data to pad or crop, assuming `img` is channel-first and |
| | padding or cropping doesn't apply to the channel dim. |
| | mode: available modes for numpy array:{``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, |
| | ``"mean"``, ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | available modes for PyTorch Tensor: {``"constant"``, ``"reflect"``, ``"replicate"``, ``"circular"``}. |
| | One of the listed string values or a user supplied function. Defaults to ``"constant"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | https://pytorch.org/docs/stable/generated/torch.nn.functional.pad.html |
| | lazy: a flag to override the lazy behaviour for this call, if set. Defaults to None. |
| | pad_kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | |
| | """ |
| | lazy_ = self.lazy if lazy is None else lazy |
| | ret = self.padder(self.cropper(img, lazy_), mode=mode, lazy=lazy_, **pad_kwargs) |
| | |
| | if get_track_meta(): |
| | ret_: MetaTensor = ret |
| | if not lazy_: |
| | pad_info = ret_.applied_operations.pop() |
| | crop_info = ret_.applied_operations.pop() |
| | orig_size = crop_info.get(TraceKeys.ORIG_SIZE) |
| | self.push_transform( |
| | ret_, orig_size=orig_size, extra_info={"pad_info": pad_info, "crop_info": crop_info}, lazy=lazy_ |
| | ) |
| | else: |
| | pad_info = ret_.pending_operations.pop() |
| | crop_info = ret_.pending_operations.pop() |
| | orig_size = crop_info.get(TraceKeys.ORIG_SIZE) |
| | self.push_transform( |
| | ret_, |
| | orig_size=orig_size, |
| | sp_size=pad_info[LazyAttr.SHAPE], |
| | affine=crop_info[LazyAttr.AFFINE] @ pad_info[LazyAttr.AFFINE], |
| | extra_info={"pad_info": pad_info, "crop_info": crop_info}, |
| | lazy=lazy_, |
| | ) |
| |
|
| | return ret |
| |
|
| | def inverse(self, img: MetaTensor) -> MetaTensor: |
| | transform = self.pop_transform(img) |
| | return self.inverse_transform(img, transform) |
| |
|
| | def inverse_transform(self, img: MetaTensor, transform) -> MetaTensor: |
| | |
| | crop_info = transform[TraceKeys.EXTRA_INFO].pop("crop_info") |
| | pad_info = transform[TraceKeys.EXTRA_INFO].pop("pad_info") |
| | img.applied_operations.append(crop_info) |
| | img.applied_operations.append(pad_info) |
| | |
| | inv = self.padder.inverse(img) |
| | |
| | return self.cropper.inverse(inv) |
| |
|
| |
|
| | class BoundingRect(Transform): |
| | """ |
| | Compute coordinates of axis-aligned bounding rectangles from input image `img`. |
| | The output format of the coordinates is (shape is [channel, 2 * spatial dims]): |
| | |
| | [[1st_spatial_dim_start, 1st_spatial_dim_end, |
| | 2nd_spatial_dim_start, 2nd_spatial_dim_end, |
| | ..., |
| | Nth_spatial_dim_start, Nth_spatial_dim_end], |
| | |
| | ... |
| | |
| | [1st_spatial_dim_start, 1st_spatial_dim_end, |
| | 2nd_spatial_dim_start, 2nd_spatial_dim_end, |
| | ..., |
| | Nth_spatial_dim_start, Nth_spatial_dim_end]] |
| | |
| | The bounding boxes edges are aligned with the input image edges. |
| | This function returns [0, 0, ...] if there's no positive intensity. |
| | |
| | Args: |
| | select_fn: function to select expected foreground, default is to select values > 0. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, select_fn: Callable = is_positive) -> None: |
| | self.select_fn = select_fn |
| |
|
| | def __call__(self, img: NdarrayOrTensor) -> np.ndarray: |
| | """ |
| | See also: :py:class:`monai.transforms.utils.generate_spatial_bounding_box`. |
| | """ |
| | bbox = [] |
| |
|
| | for channel in range(img.shape[0]): |
| | start_, end_ = generate_spatial_bounding_box(img, select_fn=self.select_fn, channel_indices=channel) |
| | bbox.append([i for k in zip(start_, end_) for i in k]) |
| |
|
| | return np.stack(bbox, axis=0) |
| |
|