| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | from collections.abc import Callable, Mapping, Sequence |
| | from typing import Any |
| |
|
| | import numpy as np |
| | import torch |
| | import torch.nn as nn |
| |
|
| | from monai.data.meta_tensor import MetaTensor |
| | from monai.networks.utils import eval_mode |
| | from monai.transforms import Compose, GaussianSmooth, Lambda, ScaleIntensity, SpatialCrop |
| | from monai.utils import ensure_tuple_rep |
| |
|
| |
|
| | class OcclusionSensitivity: |
| | """ |
| | This class computes the occlusion sensitivity for a model's prediction of a given image. By occlusion sensitivity, |
| | we mean how the probability of a given prediction changes as the occluded section of an image changes. This can be |
| | useful to understand why a network is making certain decisions. |
| | |
| | As important parts of the image are occluded, the probability of classifying the image correctly will decrease. |
| | Hence, more negative values imply the corresponding occluded volume was more important in the decision process. |
| | |
| | Two ``torch.Tensor`` will be returned by the ``__call__`` method: an occlusion map and an image of the most probable |
| | class. Both images will be cropped if a bounding box used, but voxel sizes will always match the input. |
| | |
| | The occlusion map shows the inference probabilities when the corresponding part of the image is occluded. Hence, |
| | more -ve values imply that region was important in the decision process. The map will have shape ``BCHW(D)N``, |
| | where ``N`` is the number of classes to be inferred by the network. Hence, the occlusion for class ``i`` can |
| | be seen with ``map[...,i]``. |
| | |
| | The most probable class is an image of the probable class when the corresponding part of the image is occluded |
| | (equivalent to ``occ_map.argmax(dim=-1)``). |
| | |
| | See: R. R. Selvaraju et al. Grad-CAM: Visual Explanations from Deep Networks via |
| | Gradient-based Localization. https://doi.org/10.1109/ICCV.2017.74. |
| | |
| | Examples: |
| | |
| | .. code-block:: python |
| | |
| | # densenet 2d |
| | from monai.networks.nets import DenseNet121 |
| | from monai.visualize import OcclusionSensitivity |
| | import torch |
| | |
| | model_2d = DenseNet121(spatial_dims=2, in_channels=1, out_channels=3) |
| | occ_sens = OcclusionSensitivity(nn_module=model_2d) |
| | occ_map, most_probable_class = occ_sens(x=torch.rand((1, 1, 48, 64)), b_box=[2, 40, 1, 62]) |
| | |
| | # densenet 3d |
| | from monai.networks.nets import DenseNet |
| | from monai.visualize import OcclusionSensitivity |
| | |
| | model_3d = DenseNet(spatial_dims=3, in_channels=1, out_channels=3, init_features=2, growth_rate=2, block_config=(6,)) |
| | occ_sens = OcclusionSensitivity(nn_module=model_3d, n_batch=10) |
| | occ_map, most_probable_class = occ_sens(torch.rand(1, 1, 6, 6, 6), b_box=[1, 3, -1, -1, -1, -1]) |
| | |
| | See Also: |
| | |
| | - :py:class:`monai.visualize.occlusion_sensitivity.OcclusionSensitivity.` |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | nn_module: nn.Module, |
| | mask_size: int | Sequence = 16, |
| | n_batch: int = 16, |
| | verbose: bool = True, |
| | mode: str | float | Callable = "gaussian", |
| | overlap: float = 0.25, |
| | activate: bool | Callable = True, |
| | ) -> None: |
| | """ |
| | Occlusion sensitivity constructor. |
| | |
| | Args: |
| | nn_module: Classification model to use for inference |
| | mask_size: Size of box to be occluded, centred on the central voxel. If a single number |
| | is given, this is used for all dimensions. If a sequence is given, this is used for each dimension |
| | individually. |
| | n_batch: Number of images in a batch for inference. |
| | verbose: Use progress bar (if ``tqdm`` available). |
| | mode: what should the occluded region be replaced with? If a float is given, that value will be used |
| | throughout the occlusion. Else, ``gaussian``, ``mean_img`` and ``mean_patch`` can be supplied: |
| | |
| | * ``gaussian``: occluded region is multiplied by 1 - gaussian kernel. In this fashion, the occlusion |
| | will be 0 at the center and will be unchanged towards the edges, varying smoothly between. When |
| | gaussian is used, a weighted average will be used to combine overlapping regions. This will be |
| | done using the gaussian (not 1-gaussian) as occluded regions count more. |
| | * ``mean_patch``: occluded region will be replaced with the mean of occluded region. |
| | * ``mean_img``: occluded region will be replaced with the mean of the whole image. |
| | |
| | overlap: overlap between inferred regions. Should be in range 0<=x<1. |
| | activate: if ``True``, do softmax activation if num_channels > 1 else do ``sigmoid``. If ``False``, don't do any |
| | activation. If ``callable``, use callable on inferred outputs. |
| | |
| | """ |
| | self.nn_module = nn_module |
| | self.mask_size = mask_size |
| | self.n_batch = n_batch |
| | self.verbose = verbose |
| | self.overlap = overlap |
| | self.activate = activate |
| | |
| | if isinstance(mode, str) and mode not in ("gaussian", "mean_patch", "mean_img"): |
| | raise NotImplementedError |
| | self.mode = mode |
| |
|
| | @staticmethod |
| | def constant_occlusion(x: torch.Tensor, val: float, mask_size: Sequence) -> tuple[float, torch.Tensor]: |
| | """Occlude with a constant occlusion. Multiplicative is zero, additive is constant value.""" |
| | ones = torch.ones((*x.shape[:2], *mask_size), device=x.device, dtype=x.dtype) |
| | return 0, ones * val |
| |
|
| | @staticmethod |
| | def gaussian_occlusion(x: torch.Tensor, mask_size: Sequence, sigma: float = 0.25) -> tuple[torch.Tensor, float]: |
| | """ |
| | For Gaussian occlusion, Multiplicative is 1-Gaussian, additive is zero. |
| | Default sigma of 0.25 empirically shown to give reasonable kernel, see here: |
| | https://github.com/Project-MONAI/MONAI/pull/5230#discussion_r984520714. |
| | """ |
| | kernel = torch.zeros((x.shape[1], *mask_size), device=x.device, dtype=x.dtype) |
| | spatial_shape = kernel.shape[1:] |
| | |
| | center = [slice(None)] + [slice(s // 2, s // 2 + 1) for s in spatial_shape] |
| | |
| | kernel[center] = 1.0 |
| | |
| | |
| | gaussian = Compose( |
| | [GaussianSmooth(sigma=[b * sigma for b in spatial_shape]), Lambda(lambda x: -x), ScaleIntensity()] |
| | ) |
| | |
| | mul: torch.Tensor = gaussian(kernel)[None] |
| |
|
| | return mul, 0 |
| |
|
| | @staticmethod |
| | def predictor( |
| | cropped_grid: torch.Tensor, |
| | nn_module: nn.Module, |
| | x: torch.Tensor, |
| | mul: torch.Tensor | float, |
| | add: torch.Tensor | float, |
| | mask_size: Sequence, |
| | occ_mode: str, |
| | activate: bool | Callable, |
| | module_kwargs: Mapping[str, Any], |
| | ) -> torch.Tensor: |
| | """ |
| | Predictor function to be passed to the sliding window inferer. Takes a cropped meshgrid, |
| | referring to the coordinates in the input image. We use the index of the top-left corner |
| | in combination ``mask_size`` to figure out which region of the image is to be occluded. The |
| | occlusion is performed on the original image, ``x``, using ``cropped_region * mul + add``. ``mul`` |
| | and ``add`` are sometimes pre-computed (e.g., a constant Gaussian blur), or they are |
| | sometimes calculated on the fly (e.g., the mean of the occluded patch). For this reason |
| | ``occ_mode`` is given. Lastly, ``activate`` is used to activate after each call of the model. |
| | |
| | Args: |
| | cropped_grid: subsection of the meshgrid, where each voxel refers to the coordinate of |
| | the input image. The meshgrid is created by the ``OcclusionSensitivity`` class, and |
| | the generation of the subset is determined by ``sliding_window_inference``. |
| | nn_module: module to call on data. |
| | x: the image that was originally passed into ``OcclusionSensitivity.__call__``. |
| | mul: occluded region will be multiplied by this. Can be ``torch.Tensor`` or ``float``. |
| | add: after multiplication, this is added to the occluded region. Can be ``torch.Tensor`` or ``float``. |
| | mask_size: Size of box to be occluded, centred on the central voxel. Should be |
| | a sequence, one value for each spatial dimension. |
| | occ_mode: might be used to calculate ``mul`` and ``add`` on the fly. |
| | activate: if ``True``, do softmax activation if num_channels > 1 else do ``sigmoid``. If ``False``, don't do any |
| | activation. If ``callable``, use callable on inferred outputs. |
| | module_kwargs: kwargs to be passed onto module when inferring |
| | """ |
| | n_batch = cropped_grid.shape[0] |
| | sd = cropped_grid.ndim - 2 |
| | |
| | im = torch.repeat_interleave(x, n_batch, 0) |
| | |
| | corner_coord_slices = [slice(None)] * 2 + [slice(1)] * sd |
| | top_corners = cropped_grid[corner_coord_slices] |
| |
|
| | |
| | for b, t in enumerate(top_corners): |
| | |
| | slices = [slice(b, b + 1), slice(None)] + [slice(int(j), int(j) + m) for j, m in zip(t, mask_size)] |
| | to_occlude = im[slices] |
| | if occ_mode == "mean_patch": |
| | add, mul = OcclusionSensitivity.constant_occlusion(x, to_occlude.mean().item(), mask_size) |
| |
|
| | if callable(occ_mode): |
| | to_occlude = occ_mode(x, to_occlude) |
| | else: |
| | to_occlude = to_occlude * mul + add |
| | if add is None or mul is None: |
| | raise RuntimeError("Shouldn't be here, something's gone wrong...") |
| | im[slices] = to_occlude |
| | |
| | out: torch.Tensor = nn_module(im, **module_kwargs) |
| |
|
| | |
| | if callable(activate): |
| | out = activate(out) |
| | |
| | elif activate: |
| | out = out.sigmoid() if x.shape[1] == 1 else out.softmax(1) |
| |
|
| | |
| | |
| | for m in mask_size: |
| | out = torch.repeat_interleave(out.unsqueeze(-1), m, dim=-1) |
| |
|
| | return out |
| |
|
| | @staticmethod |
| | def crop_meshgrid( |
| | grid: MetaTensor, b_box: Sequence, mask_size: Sequence |
| | ) -> tuple[MetaTensor, SpatialCrop, Sequence]: |
| | """Crop the meshgrid so we only perform occlusion sensitivity on a subsection of the image.""" |
| | |
| | mask_edge = [(m - 1) // 2 for m in mask_size] |
| | bbox_min = [max(b - m, 0) for b, m in zip(b_box[::2], mask_edge)] |
| | bbox_max = [] |
| | for b, m, s in zip(b_box[1::2], mask_edge, grid.shape[2:]): |
| | |
| | if b == -1: |
| | bbox_max.append(s) |
| | |
| | else: |
| | bbox_max.append(min(b + m, s)) |
| | |
| | |
| | |
| | slices = [slice(s, e) for s, e in zip(bbox_min, bbox_max)] |
| | cropper = SpatialCrop(roi_slices=slices) |
| | cropped: MetaTensor = cropper(grid[0])[None] |
| | mask_size = list(mask_size) |
| | for i, s in enumerate(cropped.shape[2:]): |
| | mask_size[i] = min(s, mask_size[i]) |
| | return cropped, cropper, mask_size |
| |
|
| | def __call__( |
| | self, x: torch.Tensor, b_box: Sequence | None = None, **kwargs: Any |
| | ) -> tuple[torch.Tensor, torch.Tensor]: |
| | """ |
| | Args: |
| | x: Image to use for inference. Should be a tensor consisting of 1 batch. |
| | b_box: Bounding box on which to perform the analysis. The output image will be limited to this size. |
| | There should be a minimum and maximum for all spatial dimensions: ``[min1, max1, min2, max2,...]``. |
| | * By default, the whole image will be used. Decreasing the size will speed the analysis up, which might |
| | be useful for larger images. |
| | * Min and max are inclusive, so ``[0, 63, ...]`` will have size ``(64, ...)``. |
| | * Use -ve to use ``min=0`` and ``max=im.shape[x]-1`` for xth dimension. |
| | * N.B.: we add half of the mask size to the bounding box to ensure that the region of interest has a |
| | sufficiently large area surrounding it. |
| | kwargs: any extra arguments to be passed on to the module as part of its `__call__`. |
| | |
| | Returns: |
| | * Occlusion map: |
| | * Shows the inference probabilities when the corresponding part of the image is occluded. |
| | Hence, more -ve values imply that region was important in the decision process. |
| | * The map will have shape ``BCHW(D)N``, where N is the number of classes to be inferred by the |
| | network. Hence, the occlusion for class ``i`` can be seen with ``map[...,i]``. |
| | * If `per_channel==False`, output ``C`` will equal 1: ``B1HW(D)N`` |
| | * Most probable class: |
| | * The most probable class when the corresponding part of the image is occluded (``argmax(dim=-1)``). |
| | Both images will be cropped if a bounding box used, but voxel sizes will always match the input. |
| | """ |
| | if x.shape[0] > 1: |
| | raise ValueError("Expected batch size of 1.") |
| |
|
| | sd = x.ndim - 2 |
| | mask_size: Sequence = ensure_tuple_rep(self.mask_size, sd) |
| |
|
| | |
| | grid: MetaTensor = MetaTensor( |
| | np.stack(np.meshgrid(*[np.arange(0, i) for i in x.shape[2:]], indexing="ij"))[None], |
| | device=x.device, |
| | dtype=x.dtype, |
| | ) |
| | |
| | if b_box is not None: |
| | grid, cropper, mask_size = self.crop_meshgrid(grid, b_box, mask_size) |
| |
|
| | |
| | if any(m > g for g, m in zip(grid.shape[2:], mask_size)): |
| | raise ValueError(f"Image (spatial shape) {grid.shape[2:]} should be bigger than mask {mask_size}.") |
| |
|
| | |
| | add: float | torch.Tensor | None |
| | mul: float | torch.Tensor | None |
| | |
| | if isinstance(self.mode, float): |
| | mul, add = self.constant_occlusion(x, self.mode, mask_size) |
| | |
| | elif self.mode == "mean_img": |
| | mul, add = self.constant_occlusion(x, x.mean().item(), mask_size) |
| | |
| | elif self.mode == "gaussian": |
| | mul, add = self.gaussian_occlusion(x, mask_size) |
| | |
| | else: |
| | add, mul = None, None |
| |
|
| | with eval_mode(self.nn_module): |
| | |
| | from monai.inferers import sliding_window_inference |
| |
|
| | sensitivity_im: MetaTensor = sliding_window_inference( |
| | grid, |
| | roi_size=mask_size, |
| | sw_batch_size=self.n_batch, |
| | predictor=OcclusionSensitivity.predictor, |
| | overlap=self.overlap, |
| | mode="gaussian" if self.mode == "gaussian" else "constant", |
| | progress=self.verbose, |
| | nn_module=self.nn_module, |
| | x=x, |
| | add=add, |
| | mul=mul, |
| | mask_size=mask_size, |
| | occ_mode=self.mode, |
| | activate=self.activate, |
| | module_kwargs=kwargs, |
| | ) |
| |
|
| | if b_box is not None: |
| | |
| | sensitivity_im = cropper.inverse(sensitivity_im[0])[None] |
| | |
| | bbox_min = [max(b, 0) for b in b_box[::2]] |
| | bbox_max = [b if b > 0 else s for b, s in zip(b_box[1::2], x.shape[2:])] |
| | cropper = SpatialCrop(roi_start=bbox_min, roi_end=bbox_max) |
| | sensitivity_im = cropper(sensitivity_im[0])[None] |
| |
|
| | |
| | most_probable_class = sensitivity_im.argmax(dim=1, keepdim=True) |
| | return sensitivity_im, most_probable_class |
| |
|