| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | import warnings |
| | from collections.abc import Hashable, Mapping |
| | from contextlib import contextmanager |
| | from typing import Any |
| |
|
| | import torch |
| |
|
| | from monai import transforms |
| | from monai.data.meta_obj import MetaObj, get_track_meta |
| | from monai.data.meta_tensor import MetaTensor |
| | from monai.data.utils import to_affine_nd |
| | from monai.transforms.traits import InvertibleTrait |
| | from monai.transforms.transform import Transform |
| | from monai.utils import ( |
| | LazyAttr, |
| | MetaKeys, |
| | TraceKeys, |
| | TraceStatusKeys, |
| | convert_to_dst_type, |
| | convert_to_numpy, |
| | convert_to_tensor, |
| | ) |
| | from monai.utils.misc import MONAIEnvVars |
| |
|
| | __all__ = ["TraceableTransform", "InvertibleTransform"] |
| |
|
| |
|
| | class TraceableTransform(Transform): |
| | """ |
| | Maintains a stack of applied transforms to data. |
| | |
| | Data can be one of two types: |
| | 1. A `MetaTensor` (this is the preferred data type). |
| | 2. A dictionary of data containing arrays/tensors and auxiliary metadata. In |
| | this case, a key must be supplied (this dictionary-based approach is deprecated). |
| | |
| | If `data` is of type `MetaTensor`, then the applied transform will be added to ``data.applied_operations``. |
| | |
| | If `data` is a dictionary, then one of two things can happen: |
| | 1. If data[key] is a `MetaTensor`, the applied transform will be added to ``data[key].applied_operations``. |
| | 2. Else, the applied transform will be appended to an adjacent list using |
| | `trace_key`. If, for example, the key is `image`, then the transform |
| | will be appended to `image_transforms` (this dictionary-based approach is deprecated). |
| | |
| | Hopefully it is clear that there are three total possibilities: |
| | 1. data is `MetaTensor` |
| | 2. data is dictionary, data[key] is `MetaTensor` |
| | 3. data is dictionary, data[key] is not `MetaTensor` (this is a deprecated approach). |
| | |
| | The ``__call__`` method of this transform class must be implemented so |
| | that the transformation information is stored during the data transformation. |
| | |
| | The information in the stack of applied transforms must be compatible with the |
| | default collate, by only storing strings, numbers and arrays. |
| | |
| | `tracing` could be enabled by `self.set_tracing` or setting |
| | `MONAI_TRACE_TRANSFORM` when initializing the class. |
| | """ |
| |
|
| | tracing = MONAIEnvVars.trace_transform() != "0" |
| |
|
| | def set_tracing(self, tracing: bool) -> None: |
| | """Set whether to trace transforms.""" |
| | self.tracing = tracing |
| |
|
| | @staticmethod |
| | def trace_key(key: Hashable = None): |
| | """The key to store the stack of applied transforms.""" |
| | if key is None: |
| | return f"{TraceKeys.KEY_SUFFIX}" |
| | return f"{key}{TraceKeys.KEY_SUFFIX}" |
| |
|
| | @staticmethod |
| | def transform_info_keys(): |
| | """The keys to store necessary info of an applied transform.""" |
| | return (TraceKeys.CLASS_NAME, TraceKeys.ID, TraceKeys.TRACING, TraceKeys.DO_TRANSFORM) |
| |
|
| | def get_transform_info(self) -> dict: |
| | """ |
| | Return a dictionary with the relevant information pertaining to an applied transform. |
| | """ |
| | vals = ( |
| | self.__class__.__name__, |
| | id(self), |
| | self.tracing, |
| | self._do_transform if hasattr(self, "_do_transform") else True, |
| | ) |
| | return dict(zip(self.transform_info_keys(), vals)) |
| |
|
| | def push_transform(self, data, *args, **kwargs): |
| | """ |
| | Push to a stack of applied transforms of ``data``. |
| | |
| | Args: |
| | data: dictionary of data or `MetaTensor`. |
| | args: additional positional arguments to track_transform_meta. |
| | kwargs: additional keyword arguments to track_transform_meta, |
| | set ``replace=True`` (default False) to rewrite the last transform infor in |
| | applied_operation/pending_operation based on ``self.get_transform_info()``. |
| | """ |
| | lazy_eval = kwargs.get("lazy", False) |
| | transform_info = self.get_transform_info() |
| | do_transform = transform_info.get(TraceKeys.DO_TRANSFORM, True) |
| | kwargs = kwargs or {} |
| | replace = kwargs.pop("replace", False) |
| | if replace and get_track_meta() and isinstance(data, MetaTensor): |
| | if not lazy_eval: |
| | xform = self.pop_transform(data, check=False) if do_transform else {} |
| | meta_obj = self.push_transform(data, orig_size=xform.get(TraceKeys.ORIG_SIZE), extra_info=xform) |
| | return data.copy_meta_from(meta_obj) |
| | if do_transform: |
| | xform = data.pending_operations.pop() |
| | extra = xform.copy() |
| | xform.update(transform_info) |
| | else: |
| | xform, extra = transform_info, {} |
| | meta_obj = self.push_transform(data, transform_info=xform, lazy=True, extra_info=extra) |
| | return data.copy_meta_from(meta_obj) |
| | kwargs["lazy"] = lazy_eval |
| | if "transform_info" in kwargs and isinstance(kwargs["transform_info"], dict): |
| | kwargs["transform_info"].update(transform_info) |
| | else: |
| | kwargs["transform_info"] = transform_info |
| | meta_obj = TraceableTransform.track_transform_meta(data, *args, **kwargs) |
| | return data.copy_meta_from(meta_obj) if isinstance(data, MetaTensor) else data |
| |
|
| | @classmethod |
| | def track_transform_meta( |
| | cls, |
| | data, |
| | key: Hashable = None, |
| | sp_size=None, |
| | affine=None, |
| | extra_info: dict | None = None, |
| | orig_size: tuple | None = None, |
| | transform_info=None, |
| | lazy=False, |
| | ): |
| | """ |
| | Update a stack of applied/pending transforms metadata of ``data``. |
| | |
| | Args: |
| | data: dictionary of data or `MetaTensor`. |
| | key: if data is a dictionary, data[key] will be modified. |
| | sp_size: the expected output spatial size when the transform is applied. |
| | it can be tensor or numpy, but will be converted to a list of integers. |
| | affine: the affine representation of the (spatial) transform in the image space. |
| | When the transform is applied, meta_tensor.affine will be updated to ``meta_tensor.affine @ affine``. |
| | extra_info: if desired, any extra information pertaining to the applied |
| | transform can be stored in this dictionary. These are often needed for |
| | computing the inverse transformation. |
| | orig_size: sometimes during the inverse it is useful to know what the size |
| | of the original image was, in which case it can be supplied here. |
| | transform_info: info from self.get_transform_info(). |
| | lazy: whether to push the transform to pending_operations or applied_operations. |
| | |
| | Returns: |
| | |
| | For backward compatibility, if ``data`` is a dictionary, it returns the dictionary with |
| | updated ``data[key]``. Otherwise, this function returns a MetaObj with updated transform metadata. |
| | """ |
| | data_t = data[key] if key is not None else data |
| | out_obj = MetaObj() |
| | |
| | if isinstance(data_t, MetaTensor): |
| | out_obj.copy_meta_from(data_t, keys=out_obj.__dict__.keys()) |
| |
|
| | if lazy and (not get_track_meta()): |
| | warnings.warn("metadata is not tracked, please call 'set_track_meta(True)' if doing lazy evaluation.") |
| |
|
| | if not lazy and affine is not None and isinstance(data_t, MetaTensor): |
| | |
| | orig_affine = data_t.peek_pending_affine() |
| | orig_affine = convert_to_dst_type(orig_affine, affine, dtype=torch.float64)[0] |
| | try: |
| | affine = orig_affine @ to_affine_nd(len(orig_affine) - 1, affine, dtype=torch.float64) |
| | except RuntimeError as e: |
| | if orig_affine.ndim > 2: |
| | if data_t.is_batch: |
| | msg = "Transform applied to batched tensor, should be applied to instances only" |
| | else: |
| | msg = "Mismatch affine matrix, ensured that the batch dimension is not included in the calculation." |
| | raise RuntimeError(msg) from e |
| | else: |
| | raise |
| | out_obj.meta[MetaKeys.AFFINE] = convert_to_tensor(affine, device=torch.device("cpu"), dtype=torch.float64) |
| |
|
| | if not (get_track_meta() and transform_info and transform_info.get(TraceKeys.TRACING)): |
| | if isinstance(data, Mapping): |
| | if not isinstance(data, dict): |
| | data = dict(data) |
| | data[key] = data_t.copy_meta_from(out_obj) if isinstance(data_t, MetaTensor) else data_t |
| | return data |
| | return out_obj |
| |
|
| | info = transform_info.copy() |
| | |
| | if orig_size is not None: |
| | info[TraceKeys.ORIG_SIZE] = orig_size |
| | elif isinstance(data_t, MetaTensor): |
| | info[TraceKeys.ORIG_SIZE] = data_t.peek_pending_shape() |
| | elif hasattr(data_t, "shape"): |
| | info[TraceKeys.ORIG_SIZE] = data_t.shape[1:] |
| |
|
| | |
| | info[TraceKeys.LAZY] = lazy |
| |
|
| | |
| | if extra_info is not None: |
| | extra_info.pop(LazyAttr.SHAPE, None) |
| | extra_info.pop(LazyAttr.AFFINE, None) |
| | info[TraceKeys.EXTRA_INFO] = extra_info |
| |
|
| | |
| | if lazy: |
| | if sp_size is None: |
| | if LazyAttr.SHAPE not in info: |
| | info[LazyAttr.SHAPE] = info.get(TraceKeys.ORIG_SIZE, []) |
| | else: |
| | info[LazyAttr.SHAPE] = sp_size |
| | info[LazyAttr.SHAPE] = tuple(convert_to_numpy(info[LazyAttr.SHAPE], wrap_sequence=True).tolist()) |
| | if affine is None: |
| | if LazyAttr.AFFINE not in info: |
| | info[LazyAttr.AFFINE] = MetaTensor.get_default_affine() |
| | else: |
| | info[LazyAttr.AFFINE] = affine |
| | info[LazyAttr.AFFINE] = convert_to_tensor(info[LazyAttr.AFFINE], device=torch.device("cpu")) |
| | out_obj.push_pending_operation(info) |
| | else: |
| | if out_obj.pending_operations: |
| | transform_name = info.get(TraceKeys.CLASS_NAME, "") if isinstance(info, dict) else "" |
| | msg = ( |
| | f"Transform {transform_name} has been applied to a MetaTensor with pending operations: " |
| | f"{[x.get(TraceKeys.CLASS_NAME) for x in out_obj.pending_operations]}" |
| | ) |
| | if key is not None: |
| | msg += f" for key {key}" |
| |
|
| | pend = out_obj.pending_operations[-1] |
| | statuses = pend.get(TraceKeys.STATUSES, dict()) |
| | messages = statuses.get(TraceStatusKeys.PENDING_DURING_APPLY, list()) |
| | messages.append(msg) |
| | statuses[TraceStatusKeys.PENDING_DURING_APPLY] = messages |
| | info[TraceKeys.STATUSES] = statuses |
| | out_obj.push_applied_operation(info) |
| | if isinstance(data, Mapping): |
| | if not isinstance(data, dict): |
| | data = dict(data) |
| | if isinstance(data_t, MetaTensor): |
| | data[key] = data_t.copy_meta_from(out_obj) |
| | else: |
| | x_k = TraceableTransform.trace_key(key) |
| | if x_k not in data: |
| | data[x_k] = [] |
| | data[x_k].append(info) |
| | return data |
| | return out_obj |
| |
|
| | def check_transforms_match(self, transform: Mapping) -> None: |
| | """Check transforms are of same instance.""" |
| | xform_id = transform.get(TraceKeys.ID, "") |
| | if xform_id == id(self): |
| | return |
| | |
| | if xform_id == TraceKeys.NONE: |
| | return |
| | xform_name = transform.get(TraceKeys.CLASS_NAME, "") |
| | warning_msg = transform.get(TraceKeys.EXTRA_INFO, {}).get("warn") |
| | if warning_msg: |
| | warnings.warn(warning_msg) |
| | |
| | if torch.multiprocessing.get_start_method() in ("spawn", None) and xform_name == self.__class__.__name__: |
| | return |
| | raise RuntimeError( |
| | f"Error {self.__class__.__name__} getting the most recently " |
| | f"applied invertible transform {xform_name} {xform_id} != {id(self)}." |
| | ) |
| |
|
| | def get_most_recent_transform(self, data, key: Hashable = None, check: bool = True, pop: bool = False): |
| | """ |
| | Get most recent transform for the stack. |
| | |
| | Args: |
| | data: dictionary of data or `MetaTensor`. |
| | key: if data is a dictionary, data[key] will be modified. |
| | check: if true, check that `self` is the same type as the most recently-applied transform. |
| | pop: if true, remove the transform as it is returned. |
| | |
| | Returns: |
| | Dictionary of most recently applied transform |
| | |
| | Raises: |
| | - RuntimeError: data is neither `MetaTensor` nor dictionary |
| | """ |
| | if not self.tracing: |
| | raise RuntimeError("Transform Tracing must be enabled to get the most recent transform.") |
| | if isinstance(data, MetaTensor): |
| | all_transforms = data.applied_operations |
| | elif isinstance(data, Mapping): |
| | if key in data and isinstance(data[key], MetaTensor): |
| | all_transforms = data[key].applied_operations |
| | else: |
| | all_transforms = data.get(self.trace_key(key), MetaTensor.get_default_applied_operations()) |
| | else: |
| | raise ValueError(f"`data` should be either `MetaTensor` or dictionary, got {type(data)}.") |
| | if check: |
| | self.check_transforms_match(all_transforms[-1]) |
| | return all_transforms.pop() if pop else all_transforms[-1] |
| |
|
| | def pop_transform(self, data, key: Hashable = None, check: bool = True): |
| | """ |
| | Return and pop the most recent transform. |
| | |
| | Args: |
| | data: dictionary of data or `MetaTensor` |
| | key: if data is a dictionary, data[key] will be modified |
| | check: if true, check that `self` is the same type as the most recently-applied transform. |
| | |
| | Returns: |
| | Dictionary of most recently applied transform |
| | |
| | Raises: |
| | - RuntimeError: data is neither `MetaTensor` nor dictionary |
| | """ |
| | return self.get_most_recent_transform(data, key, check, pop=True) |
| |
|
| | @contextmanager |
| | def trace_transform(self, to_trace: bool): |
| | """Temporarily set the tracing status of a transform with a context manager.""" |
| | prev = self.tracing |
| | self.tracing = to_trace |
| | yield |
| | self.tracing = prev |
| |
|
| |
|
| | class InvertibleTransform(TraceableTransform, InvertibleTrait): |
| | """Classes for invertible transforms. |
| | |
| | This class exists so that an ``invert`` method can be implemented. This allows, for |
| | example, images to be cropped, rotated, padded, etc., during training and inference, |
| | and after be returned to their original size before saving to file for comparison in |
| | an external viewer. |
| | |
| | When the ``inverse`` method is called: |
| | |
| | - the inverse is called on each key individually, which allows for |
| | different parameters being passed to each label (e.g., different |
| | interpolation for image and label). |
| | |
| | - the inverse transforms are applied in a last-in-first-out order. As |
| | the inverse is applied, its entry is removed from the list detailing |
| | the applied transformations. That is to say that during the forward |
| | pass, the list of applied transforms grows, and then during the |
| | inverse it shrinks back down to an empty list. |
| | |
| | We currently check that the ``id()`` of the transform is the same in the forward and |
| | inverse directions. This is a useful check to ensure that the inverses are being |
| | processed in the correct order. |
| | |
| | Note to developers: When converting a transform to an invertible transform, you need to: |
| | |
| | #. Inherit from this class. |
| | #. In ``__call__``, add a call to ``push_transform``. |
| | #. Any extra information that might be needed for the inverse can be included with the |
| | dictionary ``extra_info``. This dictionary should have the same keys regardless of |
| | whether ``do_transform`` was `True` or `False` and can only contain objects that are |
| | accepted in pytorch data loader's collate function (e.g., `None` is not allowed). |
| | #. Implement an ``inverse`` method. Make sure that after performing the inverse, |
| | ``pop_transform`` is called. |
| | |
| | """ |
| |
|
| | def inverse_update(self, data): |
| | """ |
| | This function is to be called before every `self.inverse(data)`, |
| | update each MetaTensor `data[key]` using `data[key_transforms]` and `data[key_meta_dict]`, |
| | for MetaTensor backward compatibility 0.9.0. |
| | """ |
| | if not isinstance(data, dict) or not isinstance(self, transforms.MapTransform): |
| | return data |
| | d = dict(data) |
| | for k in self.key_iterator(data): |
| | transform_key = transforms.TraceableTransform.trace_key(k) |
| | if transform_key not in data or not data[transform_key]: |
| | continue |
| | d = transforms.sync_meta_info(k, data, t=False) |
| | return d |
| |
|
| | def inverse(self, data: Any) -> Any: |
| | """ |
| | Inverse of ``__call__``. |
| | |
| | Raises: |
| | NotImplementedError: When the subclass does not override this method. |
| | |
| | """ |
| | raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
| |
|