| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | from abc import ABC, abstractmethod |
| | from collections.abc import Callable, Sequence |
| | from typing import TYPE_CHECKING, Any, cast |
| |
|
| | import torch |
| |
|
| | from monai.config import IgniteInfo |
| | from monai.transforms import apply_transform |
| | from monai.utils import ensure_tuple, min_version, optional_import |
| | from monai.utils.enums import CommonKeys, GanKeys |
| |
|
| | if TYPE_CHECKING: |
| | from ignite.engine import EventEnum |
| | else: |
| | EventEnum, _ = optional_import( |
| | "ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "EventEnum", as_type="base" |
| | ) |
| |
|
| | __all__ = [ |
| | "IterationEvents", |
| | "get_devices_spec", |
| | "default_prepare_batch", |
| | "PrepareBatch", |
| | "PrepareBatchDefault", |
| | "PrepareBatchExtraInput", |
| | "default_make_latent", |
| | "engine_apply_transform", |
| | "default_metric_cmp_fn", |
| | ] |
| |
|
| |
|
| | class IterationEvents(EventEnum): |
| | """ |
| | Additional Events engine can register and trigger in the iteration process. |
| | Refer to the example in ignite: https://pytorch.org/ignite/generated/ignite.engine.events.EventEnum.html. |
| | These Events can be triggered during training iteration: |
| | `FORWARD_COMPLETED` is the Event when `network(image, label)` completed. |
| | `LOSS_COMPLETED` is the Event when `loss(pred, label)` completed. |
| | `BACKWARD_COMPLETED` is the Event when `loss.backward()` completed. |
| | `MODEL_COMPLETED` is the Event when all the model related operations completed. |
| | `INNER_ITERATION_STARTED` is the Event when the iteration has an inner loop and the loop is started. |
| | `INNER_ITERATION_COMPLETED` is the Event when the iteration has an inner loop and the loop is completed. |
| | """ |
| |
|
| | FORWARD_COMPLETED = "forward_completed" |
| | LOSS_COMPLETED = "loss_completed" |
| | BACKWARD_COMPLETED = "backward_completed" |
| | MODEL_COMPLETED = "model_completed" |
| | INNER_ITERATION_STARTED = "inner_iteration_started" |
| | INNER_ITERATION_COMPLETED = "inner_iteration_completed" |
| |
|
| |
|
| | def get_devices_spec(devices: Sequence[torch.device | str] | None = None) -> list[torch.device]: |
| | """ |
| | Get a valid specification for one or more devices. If `devices` is None get devices for all CUDA devices available. |
| | If `devices` is and zero-length structure a single CPU compute device is returned. In any other cases `devices` is |
| | returned unchanged. |
| | |
| | Args: |
| | devices: list of devices to request, None for all GPU devices, [] for CPU. |
| | |
| | Raises: |
| | RuntimeError: When all GPUs are selected (``devices=None``) but no GPUs are available. |
| | |
| | Returns: |
| | list of torch.device: list of devices. |
| | |
| | """ |
| | if devices is None: |
| | devices = [torch.device(f"cuda:{d:d}") for d in range(torch.cuda.device_count())] |
| |
|
| | if not devices: |
| | raise RuntimeError("No GPU devices available.") |
| |
|
| | elif len(devices) == 0: |
| | devices = [torch.device("cpu")] |
| |
|
| | else: |
| | devices = list(devices) |
| |
|
| | devices = [torch.device(d) if isinstance(d, str) else d for d in devices] |
| | return devices |
| |
|
| |
|
| | def default_prepare_batch( |
| | batchdata: dict[str, torch.Tensor] | torch.Tensor | Sequence[torch.Tensor], |
| | device: str | torch.device | None = None, |
| | non_blocking: bool = False, |
| | **kwargs: Any, |
| | ) -> tuple[torch.Tensor, torch.Tensor | None] | torch.Tensor: |
| | """ |
| | Default function to prepare the data for current iteration. |
| | |
| | The input `batchdata` is either a single tensor, a pair of tensors, or a dictionary of data. In the first case the |
| | return value is the tensor and None, in the second case the return value is the two tensors, and in the dictionary |
| | case the return value depends on what keys are present. if `CommonKeys.IMAGE` and `CommonKeys.LABEL` are present |
| | then the tensors they key to are returned, if only `CommonKeys.IMAGE` is present that tensor and None is returned. |
| | If `CommonKeys.REALS` is present this is returned with None. All returned tensors are moved to the given device |
| | using the given non-blocking argument before being returned. |
| | |
| | This function implements the expected API for a `prepare_batch` callable in Ignite: |
| | https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html |
| | |
| | Args: |
| | batchdata: input batch data which is either a single tensor, a pair, or a dictionary |
| | device: device to move every returned tensor to |
| | non_blocking: equivalent argument for `Tensor.to` |
| | kwargs: further arguments for `Tensor.to` |
| | |
| | Returns: |
| | image, label(optional). |
| | """ |
| | if not isinstance(batchdata, dict): |
| | if isinstance(batchdata, torch.Tensor): |
| | return batchdata.to(device=device, non_blocking=non_blocking, **kwargs), None |
| | elif len(batchdata) == 2: |
| | image, label = batchdata |
| | return ( |
| | image.to(device=device, non_blocking=non_blocking, **kwargs), |
| | label.to(device=device, non_blocking=non_blocking, **kwargs), |
| | ) |
| |
|
| | raise AssertionError("Default prepare_batch expects a single tensor, a tensor pair, or dictionary input data.") |
| |
|
| | if isinstance(batchdata.get(CommonKeys.LABEL), torch.Tensor): |
| | return ( |
| | batchdata[CommonKeys.IMAGE].to(device=device, non_blocking=non_blocking, **kwargs), |
| | batchdata[CommonKeys.LABEL].to(device=device, non_blocking=non_blocking, **kwargs), |
| | ) |
| |
|
| | if GanKeys.REALS in batchdata: |
| | return batchdata[GanKeys.REALS].to(device=device, non_blocking=non_blocking, **kwargs) |
| |
|
| | return batchdata[CommonKeys.IMAGE].to(device=device, non_blocking=non_blocking, **kwargs), None |
| |
|
| |
|
| | class PrepareBatch(ABC): |
| | """ |
| | Interface of customized prepare_batch in the trainer or evaluator workflows. |
| | It takes the data of current batch, target device and non_blocking flag as input. |
| | Args `batchdata`, `device`, `non_blocking` refer to the ignite API: |
| | https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html. |
| | `kwargs` supports other args for `Tensor.to()` API. |
| | """ |
| |
|
| | @abstractmethod |
| | def __call__( |
| | self, |
| | batchdata: dict[str, torch.Tensor], |
| | device: str | torch.device | None = None, |
| | non_blocking: bool = False, |
| | **kwargs: Any, |
| | ) -> Any: |
| | raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
| |
|
| |
|
| | class PrepareBatchDefault(PrepareBatch): |
| | """ |
| | This wraps `default_prepare_batch` to return `image` and `label` only, so is consistent with its API. |
| | """ |
| |
|
| | def __call__( |
| | self, |
| | batchdata: dict[str, torch.Tensor] | torch.Tensor | Sequence[torch.Tensor], |
| | device: str | torch.device | None = None, |
| | non_blocking: bool = False, |
| | **kwargs: Any, |
| | ) -> tuple[torch.Tensor, torch.Tensor | None] | torch.Tensor: |
| | """ |
| | Args `batchdata`, `device`, `non_blocking` refer to the ignite API: |
| | https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html. |
| | `kwargs` supports other args for `Tensor.to()` API. |
| | |
| | """ |
| | return default_prepare_batch(batchdata, device, non_blocking, **kwargs) |
| |
|
| |
|
| | class PrepareBatchExtraInput(PrepareBatch): |
| | """ |
| | Customized prepare batch callable for trainers or evaluators which support extra input data for the network. |
| | Extra items are specified by the `extra_keys` parameter and are extracted from the input dictionary (ie. the batch). |
| | This uses `default_prepare_batch` but requires dictionary inputs. |
| | |
| | Args: |
| | extra_keys: If a string or sequence of strings is provided, values from the input dictionary are extracted from |
| | those keys and passed to the network as extra positional arguments. If a dictionary is provided, every pair |
| | `(k, v)` in that dictionary will become a new keyword argument assigning to `k` the value in the input |
| | dictionary keyed to `v`. |
| | """ |
| |
|
| | def __init__(self, extra_keys: str | Sequence[str] | dict[str, str]) -> None: |
| | self.extra_keys = extra_keys |
| |
|
| | def __call__( |
| | self, |
| | batchdata: dict[str, torch.Tensor], |
| | device: str | torch.device | None = None, |
| | non_blocking: bool = False, |
| | **kwargs: Any, |
| | ) -> tuple[torch.Tensor, torch.Tensor, tuple, dict]: |
| | """ |
| | Args `batchdata`, `device`, `non_blocking` refer to the ignite API: |
| | https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html. |
| | `kwargs` supports other args for `Tensor.to()` API. |
| | """ |
| | image, label = default_prepare_batch(batchdata, device, non_blocking, **kwargs) |
| | args_ = list() |
| | kwargs_ = dict() |
| |
|
| | def _get_data(key: str) -> torch.Tensor: |
| | data = batchdata[key] |
| |
|
| | if isinstance(data, torch.Tensor): |
| | data = data.to(device=device, non_blocking=non_blocking, **kwargs) |
| |
|
| | return data |
| |
|
| | if isinstance(self.extra_keys, (str, list, tuple)): |
| | for k in ensure_tuple(self.extra_keys): |
| | args_.append(_get_data(k)) |
| | elif isinstance(self.extra_keys, dict): |
| | for k, v in self.extra_keys.items(): |
| | kwargs_.update({k: _get_data(v)}) |
| |
|
| | return cast(torch.Tensor, image), cast(torch.Tensor, label), tuple(args_), kwargs_ |
| |
|
| |
|
| | def default_make_latent( |
| | num_latents: int, |
| | latent_size: int, |
| | device: str | torch.device | None = None, |
| | non_blocking: bool = False, |
| | **kwargs: Any, |
| | ) -> torch.Tensor: |
| | return torch.randn(num_latents, latent_size).to(device=device, non_blocking=non_blocking, **kwargs) |
| |
|
| |
|
| | def engine_apply_transform(batch: Any, output: Any, transform: Callable[..., dict]) -> tuple[Any, Any]: |
| | """ |
| | Apply transform on `batch` and `output`. |
| | If `batch` and `output` are dictionaries, temporarily combine them for the transform, |
| | otherwise, apply the transform for `output` data only. |
| | |
| | """ |
| | if isinstance(batch, dict) and isinstance(output, dict): |
| | data = dict(batch) |
| | data.update(output) |
| | transformed_data = apply_transform(transform, data) |
| |
|
| | if not isinstance(transformed_data, dict): |
| | raise AssertionError("With a dict supplied to apply_transform a single dict return is expected.") |
| |
|
| | for k, v in transformed_data.items(): |
| | |
| | |
| | if k in output or k not in batch: |
| | output[k] = v |
| | else: |
| | batch[k] = v |
| | else: |
| | output = apply_transform(transform, output) |
| |
|
| | return batch, output |
| |
|
| |
|
| | def default_metric_cmp_fn(current_metric: float, prev_best: float) -> bool: |
| | """ |
| | The default function to compare metric values between current metric and previous best metric. |
| | |
| | Args: |
| | current_metric: metric value of current round computation. |
| | prev_best: the best metric value of previous rounds to compare with. |
| | |
| | """ |
| | return current_metric > prev_best |
| |
|