| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | A collection of generic interfaces for MONAI transforms. |
| | """ |
| |
|
| | from __future__ import annotations |
| |
|
| | import logging |
| | from abc import ABC, abstractmethod |
| | from collections.abc import Callable, Generator, Hashable, Iterable, Mapping |
| | from typing import Any, TypeVar |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai import config, transforms |
| | from monai.config import KeysCollection |
| | from monai.data.meta_tensor import MetaTensor |
| | from monai.transforms.traits import LazyTrait, RandomizableTrait, ThreadUnsafe |
| | from monai.utils import MAX_SEED, ensure_tuple, first |
| | from monai.utils.enums import TransformBackends |
| | from monai.utils.misc import MONAIEnvVars |
| |
|
| | __all__ = [ |
| | "ThreadUnsafe", |
| | "apply_transform", |
| | "Randomizable", |
| | "LazyTransform", |
| | "RandomizableTransform", |
| | "Transform", |
| | "MapTransform", |
| | ] |
| |
|
| | ReturnType = TypeVar("ReturnType") |
| |
|
| |
|
| | def _apply_transform( |
| | transform: Callable[..., ReturnType], |
| | data: Any, |
| | unpack_parameters: bool = False, |
| | lazy: bool | None = False, |
| | overrides: dict | None = None, |
| | logger_name: bool | str = False, |
| | ) -> ReturnType: |
| | """ |
| | Perform a transform 'transform' on 'data', according to the other parameters specified. |
| | |
| | If `data` is a tuple and `unpack_parameters` is True, each parameter of `data` is unpacked |
| | as arguments to `transform`. Otherwise `data` is considered as single argument to `transform`. |
| | |
| | If 'lazy' is True, this method first checks whether it can execute this method lazily. If it |
| | can't, it will ensure that all pending lazy transforms on 'data' are applied before applying |
| | this 'transform' to it. If 'lazy' is True, and 'overrides' are provided, those overrides will |
| | be applied to the pending operations on 'data'. See ``Compose`` for more details on lazy |
| | resampling, which is an experimental feature for 1.2. |
| | |
| | Please note, this class is function is designed to be called by ``apply_transform``. |
| | In general, you should not need to make specific use of it unless you are implementing |
| | pipeline execution mechanisms. |
| | |
| | Args: |
| | transform: a callable to be used to transform `data`. |
| | data: the tensorlike or dictionary of tensorlikes to be executed on |
| | unpack_parameters: whether to unpack parameters for `transform`. Defaults to False. |
| | lazy: whether to enable lazy evaluation for lazy transforms. If False, transforms will be |
| | carried out on a transform by transform basis. If True, all lazy transforms will |
| | be executed by accumulating changes and resampling as few times as possible. |
| | See the :ref:`Lazy Resampling topic<lazy_resampling> for more information about lazy resampling. |
| | overrides: this optional parameter allows you to specify a dictionary of parameters that should be overridden |
| | when executing a pipeline. These each parameter that is compatible with a given transform is then applied |
| | to that transform before it is executed. Note that overrides are currently only applied when |
| | :ref:`Lazy Resampling<lazy_resampling>` is enabled for the pipeline or a given transform. If lazy is False |
| | they are ignored. Currently supported args are: |
| | {``"mode"``, ``"padding_mode"``, ``"dtype"``, ``"align_corners"``, ``"resample_mode"``, ``device``}. |
| | logger_name: this optional parameter allows you to specify a logger by name for logging of pipeline execution. |
| | Setting this to False disables logging. Setting it to True enables logging to the default loggers. |
| | Setting a string overrides the logger name to which logging is performed. |
| | |
| | Returns: |
| | ReturnType: The return type of `transform`. |
| | """ |
| | from monai.transforms.lazy.functional import apply_pending_transforms_in_order |
| |
|
| | data = apply_pending_transforms_in_order(transform, data, lazy, overrides, logger_name) |
| |
|
| | if isinstance(data, tuple) and unpack_parameters: |
| | return transform(*data, lazy=lazy) if isinstance(transform, LazyTrait) else transform(*data) |
| |
|
| | return transform(data, lazy=lazy) if isinstance(transform, LazyTrait) else transform(data) |
| |
|
| |
|
| | def apply_transform( |
| | transform: Callable[..., ReturnType], |
| | data: Any, |
| | map_items: bool = True, |
| | unpack_items: bool = False, |
| | log_stats: bool | str = False, |
| | lazy: bool | None = None, |
| | overrides: dict | None = None, |
| | ) -> list[ReturnType] | ReturnType: |
| | """ |
| | Transform `data` with `transform`. |
| | |
| | If `data` is a list or tuple and `map_data` is True, each item of `data` will be transformed |
| | and this method returns a list of outcomes. |
| | otherwise transform will be applied once with `data` as the argument. |
| | |
| | Args: |
| | transform: a callable to be used to transform `data`. |
| | data: an object to be transformed. |
| | map_items: whether to apply transform to each item in `data`, |
| | if `data` is a list or tuple. Defaults to True. |
| | unpack_items: whether to unpack parameters using `*`. Defaults to False. |
| | log_stats: log errors when they occur in the processing pipeline. By default, this is set to False, which |
| | disables the logger for processing pipeline errors. Setting it to None or True will enable logging to the |
| | default logger name. Setting it to a string specifies the logger to which errors should be logged. |
| | lazy: whether to execute in lazy mode or not. See the :ref:`Lazy Resampling topic<lazy_resampling> for more |
| | information about lazy resampling. Defaults to None. |
| | overrides: optional overrides to apply to transform parameters. This parameter is ignored unless transforms |
| | are being executed lazily. See the :ref:`Lazy Resampling topic<lazy_resampling> for more details and |
| | examples of its usage. |
| | |
| | Raises: |
| | Exception: When ``transform`` raises an exception. |
| | |
| | Returns: |
| | Union[List[ReturnType], ReturnType]: The return type of `transform` or a list thereof. |
| | """ |
| | try: |
| | if isinstance(data, (list, tuple)) and map_items: |
| | return [_apply_transform(transform, item, unpack_items, lazy, overrides, log_stats) for item in data] |
| | return _apply_transform(transform, data, unpack_items, lazy, overrides, log_stats) |
| | except Exception as e: |
| | |
| | |
| | if MONAIEnvVars.debug(): |
| | raise |
| | if log_stats is not False and not isinstance(transform, transforms.compose.Compose): |
| | |
| | if isinstance(log_stats, str): |
| | datastats = transforms.utility.array.DataStats(data_shape=False, value_range=False, name=log_stats) |
| | else: |
| | datastats = transforms.utility.array.DataStats(data_shape=False, value_range=False) |
| | logger = logging.getLogger(datastats._logger_name) |
| | logger.error(f"\n=== Transform input info -- {type(transform).__name__} ===") |
| | if isinstance(data, (list, tuple)): |
| | data = data[0] |
| |
|
| | def _log_stats(data, prefix: str | None = "Data"): |
| | if isinstance(data, (np.ndarray, torch.Tensor)): |
| | |
| | datastats(img=data, data_shape=True, value_range=True, prefix=prefix) |
| | else: |
| | |
| | datastats(img=data, data_value=True, prefix=prefix) |
| |
|
| | if isinstance(data, dict): |
| | for k, v in data.items(): |
| | _log_stats(data=v, prefix=k) |
| | else: |
| | _log_stats(data=data) |
| | raise RuntimeError(f"applying transform {transform}") from e |
| |
|
| |
|
| | class Randomizable(ThreadUnsafe, RandomizableTrait): |
| | """ |
| | An interface for handling random state locally, currently based on a class |
| | variable `R`, which is an instance of `np.random.RandomState`. This |
| | provides the flexibility of component-specific determinism without |
| | affecting the global states. It is recommended to use this API with |
| | :py:class:`monai.data.DataLoader` for deterministic behaviour of the |
| | preprocessing pipelines. This API is not thread-safe. Additionally, |
| | deepcopying instance of this class often causes insufficient randomness as |
| | the random states will be duplicated. |
| | """ |
| |
|
| | R: np.random.RandomState = np.random.RandomState() |
| |
|
| | def set_random_state(self, seed: int | None = None, state: np.random.RandomState | None = None) -> Randomizable: |
| | """ |
| | Set the random state locally, to control the randomness, the derived |
| | classes should use :py:attr:`self.R` instead of `np.random` to introduce random |
| | factors. |
| | |
| | Args: |
| | seed: set the random state with an integer seed. |
| | state: set the random state with a `np.random.RandomState` object. |
| | |
| | Raises: |
| | TypeError: When ``state`` is not an ``Optional[np.random.RandomState]``. |
| | |
| | Returns: |
| | a Randomizable instance. |
| | |
| | """ |
| | if seed is not None: |
| | _seed = id(seed) if not isinstance(seed, (int, np.integer)) else seed |
| | _seed = _seed % MAX_SEED |
| | self.R = np.random.RandomState(_seed) |
| | return self |
| |
|
| | if state is not None: |
| | if not isinstance(state, np.random.RandomState): |
| | raise TypeError(f"state must be None or a np.random.RandomState but is {type(state).__name__}.") |
| | self.R = state |
| | return self |
| |
|
| | self.R = np.random.RandomState() |
| | return self |
| |
|
| | def randomize(self, data: Any) -> None: |
| | """ |
| | Within this method, :py:attr:`self.R` should be used, instead of `np.random`, to introduce random factors. |
| | |
| | all :py:attr:`self.R` calls happen here so that we have a better chance to |
| | identify errors of sync the random state. |
| | |
| | This method can generate the random factors based on properties of the input data. |
| | |
| | Raises: |
| | NotImplementedError: When the subclass does not override this method. |
| | """ |
| | raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
| |
|
| |
|
| | class Transform(ABC): |
| | """ |
| | An abstract class of a ``Transform``. |
| | A transform is callable that processes ``data``. |
| | |
| | It could be stateful and may modify ``data`` in place, |
| | the implementation should be aware of: |
| | |
| | #. thread safety when mutating its own states. |
| | When used from a multi-process context, transform's instance variables are read-only. |
| | thread-unsafe transforms should inherit :py:class:`monai.transforms.ThreadUnsafe`. |
| | #. ``data`` content unused by this transform may still be used in the |
| | subsequent transforms in a composed transform. |
| | #. storing too much information in ``data`` may cause some memory issue or IPC sync issue, |
| | especially in the multi-processing environment of PyTorch DataLoader. |
| | |
| | See Also |
| | |
| | :py:class:`monai.transforms.Compose` |
| | """ |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | backend: list[TransformBackends] = [] |
| |
|
| | @abstractmethod |
| | def __call__(self, data: Any): |
| | """ |
| | ``data`` is an element which often comes from an iteration over an |
| | iterable, such as :py:class:`torch.utils.data.Dataset`. This method should |
| | return an updated version of ``data``. |
| | To simplify the input validations, most of the transforms assume that |
| | |
| | - ``data`` is a Numpy ndarray, PyTorch Tensor or string, |
| | - the data shape can be: |
| | |
| | #. string data without shape, `LoadImage` transform expects file paths, |
| | #. most of the pre-/post-processing transforms expect: ``(num_channels, spatial_dim_1[, spatial_dim_2, ...])``, |
| | except for example: `AddChannel` expects (spatial_dim_1[, spatial_dim_2, ...]) |
| | |
| | - the channel dimension is often not omitted even if number of channels is one. |
| | |
| | This method can optionally take additional arguments to help execute transformation operation. |
| | |
| | Raises: |
| | NotImplementedError: When the subclass does not override this method. |
| | |
| | """ |
| | raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
| |
|
| |
|
| | class LazyTransform(Transform, LazyTrait): |
| | """ |
| | An implementation of functionality for lazy transforms that can be subclassed by array and |
| | dictionary transforms to simplify implementation of new lazy transforms. |
| | """ |
| |
|
| | def __init__(self, lazy: bool | None = False): |
| | if lazy is not None: |
| | if not isinstance(lazy, bool): |
| | raise TypeError(f"lazy must be a bool but is of type {type(lazy)}") |
| | self._lazy = lazy |
| |
|
| | @property |
| | def lazy(self): |
| | return self._lazy |
| |
|
| | @lazy.setter |
| | def lazy(self, lazy: bool | None): |
| | if lazy is not None: |
| | if not isinstance(lazy, bool): |
| | raise TypeError(f"lazy must be a bool but is of type {type(lazy)}") |
| | self._lazy = lazy |
| |
|
| | @property |
| | def requires_current_data(self): |
| | return False |
| |
|
| |
|
| | class RandomizableTransform(Randomizable, Transform): |
| | """ |
| | An interface for handling random state locally, currently based on a class variable `R`, |
| | which is an instance of `np.random.RandomState`. |
| | This class introduces a randomized flag `_do_transform`, is mainly for randomized data augmentation transforms. |
| | For example: |
| | |
| | .. code-block:: python |
| | |
| | from monai.transforms import RandomizableTransform |
| | |
| | class RandShiftIntensity100(RandomizableTransform): |
| | def randomize(self): |
| | super().randomize(None) |
| | self._offset = self.R.uniform(low=0, high=100) |
| | |
| | def __call__(self, img): |
| | self.randomize() |
| | if not self._do_transform: |
| | return img |
| | return img + self._offset |
| | |
| | transform = RandShiftIntensity() |
| | transform.set_random_state(seed=0) |
| | print(transform(10)) |
| | |
| | """ |
| |
|
| | def __init__(self, prob: float = 1.0, do_transform: bool = True): |
| | self._do_transform = do_transform |
| | self.prob = min(max(prob, 0.0), 1.0) |
| |
|
| | def randomize(self, data: Any) -> None: |
| | """ |
| | Within this method, :py:attr:`self.R` should be used, instead of `np.random`, to introduce random factors. |
| | |
| | all :py:attr:`self.R` calls happen here so that we have a better chance to |
| | identify errors of sync the random state. |
| | |
| | This method can generate the random factors based on properties of the input data. |
| | """ |
| | self._do_transform = self.R.rand() < self.prob |
| |
|
| |
|
| | class MapTransform(Transform): |
| | """ |
| | A subclass of :py:class:`monai.transforms.Transform` with an assumption |
| | that the ``data`` input of ``self.__call__`` is a MutableMapping such as ``dict``. |
| | |
| | The ``keys`` parameter will be used to get and set the actual data |
| | item to transform. That is, the callable of this transform should |
| | follow the pattern: |
| | |
| | .. code-block:: python |
| | |
| | def __call__(self, data): |
| | for key in self.keys: |
| | if key in data: |
| | # update output data with some_transform_function(data[key]). |
| | else: |
| | # raise exception unless allow_missing_keys==True. |
| | return data |
| | |
| | Raises: |
| | ValueError: When ``keys`` is an empty iterable. |
| | TypeError: When ``keys`` type is not in ``Union[Hashable, Iterable[Hashable]]``. |
| | |
| | """ |
| |
|
| | def __new__(cls, *args, **kwargs): |
| | if config.USE_META_DICT: |
| | |
| | cls.__call__ = transforms.attach_hook(cls.__call__, MapTransform.call_update, "post") |
| |
|
| | if hasattr(cls, "inverse"): |
| | |
| | cls.inverse: Any = transforms.attach_hook(cls.inverse, transforms.InvertibleTransform.inverse_update) |
| | return Transform.__new__(cls) |
| |
|
| | def __init__(self, keys: KeysCollection, allow_missing_keys: bool = False) -> None: |
| | super().__init__() |
| | self.keys: tuple[Hashable, ...] = ensure_tuple(keys) |
| | self.allow_missing_keys = allow_missing_keys |
| | if not self.keys: |
| | raise ValueError("keys must be non empty.") |
| | for key in self.keys: |
| | if not isinstance(key, Hashable): |
| | raise TypeError(f"keys must be one of (Hashable, Iterable[Hashable]) but is {type(keys).__name__}.") |
| |
|
| | def call_update(self, data): |
| | """ |
| | This function is to be called after every `self.__call__(data)`, |
| | update `data[key_transforms]` and `data[key_meta_dict]` using the content from MetaTensor `data[key]`, |
| | for MetaTensor backward compatibility 0.9.0. |
| | """ |
| | if not isinstance(data, (list, tuple, Mapping)): |
| | return data |
| | is_dict = False |
| | if isinstance(data, Mapping): |
| | data, is_dict = [data], True |
| | if not data or not isinstance(data[0], Mapping): |
| | return data[0] if is_dict else data |
| | list_d = [dict(x) for x in data] |
| | for idx, dict_i in enumerate(list_d): |
| | for k in dict_i: |
| | if not isinstance(dict_i[k], MetaTensor): |
| | continue |
| | list_d[idx] = transforms.sync_meta_info(k, dict_i, t=not isinstance(self, transforms.InvertD)) |
| | return list_d[0] if is_dict else list_d |
| |
|
| | @abstractmethod |
| | def __call__(self, data): |
| | """ |
| | ``data`` often comes from an iteration over an iterable, |
| | such as :py:class:`torch.utils.data.Dataset`. |
| | |
| | To simplify the input validations, this method assumes: |
| | |
| | - ``data`` is a Python dictionary, |
| | - ``data[key]`` is a Numpy ndarray, PyTorch Tensor or string, where ``key`` is an element |
| | of ``self.keys``, the data shape can be: |
| | |
| | #. string data without shape, `LoadImaged` transform expects file paths, |
| | #. most of the pre-/post-processing transforms expect: ``(num_channels, spatial_dim_1[, spatial_dim_2, ...])``, |
| | except for example: `AddChanneld` expects (spatial_dim_1[, spatial_dim_2, ...]) |
| | |
| | - the channel dimension is often not omitted even if number of channels is one. |
| | |
| | Raises: |
| | NotImplementedError: When the subclass does not override this method. |
| | |
| | returns: |
| | An updated dictionary version of ``data`` by applying the transform. |
| | |
| | """ |
| | raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
| |
|
| | def key_iterator(self, data: Mapping[Hashable, Any], *extra_iterables: Iterable | None) -> Generator: |
| | """ |
| | Iterate across keys and optionally extra iterables. If key is missing, exception is raised if |
| | `allow_missing_keys==False` (default). If `allow_missing_keys==True`, key is skipped. |
| | |
| | Args: |
| | data: data that the transform will be applied to |
| | extra_iterables: anything else to be iterated through |
| | """ |
| | |
| | ex_iters = extra_iterables or [[None] * len(self.keys)] |
| |
|
| | |
| | _ex_iters: list[Any] |
| | for key, *_ex_iters in zip(self.keys, *ex_iters): |
| | |
| | if key in data: |
| | yield (key,) + tuple(_ex_iters) if extra_iterables else key |
| | elif not self.allow_missing_keys: |
| | raise KeyError( |
| | f"Key `{key}` of transform `{self.__class__.__name__}` was missing in the data" |
| | " and allow_missing_keys==False." |
| | ) |
| |
|
| | def first_key(self, data: dict[Hashable, Any]): |
| | """ |
| | Get the first available key of `self.keys` in the input `data` dictionary. |
| | If no available key, return an empty tuple `()`. |
| | |
| | Args: |
| | data: data that the transform will be applied to. |
| | |
| | """ |
| | return first(self.key_iterator(data), ()) |
| |
|