| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | from collections.abc import Callable, Sequence |
| |
|
| | import torch |
| |
|
| | from monai.data import decollate_batch, list_data_collate |
| | from monai.engines import SupervisedEvaluator, SupervisedTrainer |
| | from monai.engines.utils import IterationEvents |
| | from monai.transforms import Compose |
| | from monai.utils.enums import CommonKeys |
| |
|
| |
|
| | class Interaction: |
| | """ |
| | Ignite process_function used to introduce interactions (simulation of clicks) for Deepgrow Training/Evaluation. |
| | For more details please refer to: https://pytorch.org/ignite/generated/ignite.engine.engine.Engine.html. |
| | This implementation is based on: |
| | |
| | Sakinis et al., Interactive segmentation of medical images through |
| | fully convolutional neural networks. (2019) https://arxiv.org/abs/1903.08205 |
| | |
| | Args: |
| | transforms: execute additional transformation during every iteration (before train). |
| | Typically, several Tensor based transforms composed by `Compose`. |
| | max_interactions: maximum number of interactions per iteration |
| | train: training or evaluation |
| | key_probability: field name to fill probability for every interaction |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | transforms: Sequence[Callable] | Callable, |
| | max_interactions: int, |
| | train: bool, |
| | key_probability: str = "probability", |
| | ) -> None: |
| | if not isinstance(transforms, Compose): |
| | transforms = Compose(transforms) |
| |
|
| | self.transforms: Compose = transforms |
| | self.max_interactions = max_interactions |
| | self.train = train |
| | self.key_probability = key_probability |
| |
|
| | def __call__(self, engine: SupervisedTrainer | SupervisedEvaluator, batchdata: dict[str, torch.Tensor]) -> dict: |
| | if batchdata is None: |
| | raise ValueError("Must provide batch data for current iteration.") |
| |
|
| | for j in range(self.max_interactions): |
| | inputs, _ = engine.prepare_batch(batchdata) |
| | inputs = inputs.to(engine.state.device) |
| |
|
| | engine.fire_event(IterationEvents.INNER_ITERATION_STARTED) |
| |
|
| | engine.network.eval() |
| | with torch.no_grad(): |
| | if engine.amp: |
| | with torch.cuda.amp.autocast(): |
| | predictions = engine.inferer(inputs, engine.network) |
| | else: |
| | predictions = engine.inferer(inputs, engine.network) |
| |
|
| | engine.fire_event(IterationEvents.INNER_ITERATION_COMPLETED) |
| |
|
| | batchdata.update({CommonKeys.PRED: predictions}) |
| |
|
| | |
| | batchdata_list = decollate_batch(batchdata, detach=True) |
| | for i in range(len(batchdata_list)): |
| | batchdata_list[i][self.key_probability] = ( |
| | (1.0 - ((1.0 / self.max_interactions) * j)) if self.train else 1.0 |
| | ) |
| | batchdata_list[i] = self.transforms(batchdata_list[i]) |
| |
|
| | |
| | batchdata = list_data_collate(batchdata_list) |
| |
|
| | return engine._iteration(engine, batchdata) |
| |
|