| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import os |
| from collections import OrderedDict |
| from collections.abc import Callable, Sequence |
| from typing import TYPE_CHECKING, Any |
|
|
| import numpy as np |
| import torch |
|
|
| from monai.config import IgniteInfo, KeysCollection, PathLike |
| from monai.utils import ensure_tuple, look_up_option, min_version, optional_import |
|
|
| idist, _ = optional_import("ignite", IgniteInfo.OPT_IMPORT_VERSION, min_version, "distributed") |
| if TYPE_CHECKING: |
| from ignite.engine import Engine |
| else: |
| Engine, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Engine") |
|
|
| __all__ = ["stopping_fn_from_metric", "stopping_fn_from_loss", "write_metrics_reports", "from_engine"] |
|
|
|
|
| def stopping_fn_from_metric(metric_name: str) -> Callable[[Engine], Any]: |
| """ |
| Returns a stopping function for ignite.handlers.EarlyStopping using the given metric name. |
| """ |
|
|
| def stopping_fn(engine: Engine) -> Any: |
| return engine.state.metrics[metric_name] |
|
|
| return stopping_fn |
|
|
|
|
| def stopping_fn_from_loss() -> Callable[[Engine], Any]: |
| """ |
| Returns a stopping function for ignite.handlers.EarlyStopping using the loss value. |
| """ |
|
|
| def stopping_fn(engine: Engine) -> Any: |
| return -engine.state.output |
|
|
| return stopping_fn |
|
|
|
|
| def write_metrics_reports( |
| save_dir: PathLike, |
| images: Sequence[str] | None, |
| metrics: dict[str, torch.Tensor | np.ndarray] | None, |
| metric_details: dict[str, torch.Tensor | np.ndarray] | None, |
| summary_ops: str | Sequence[str] | None, |
| deli: str = ",", |
| output_type: str = "csv", |
| class_labels: list[str] | None = None, |
| ) -> None: |
| """ |
| Utility function to write the metrics into files, contains 3 parts: |
| 1. if `metrics` dict is not None, write overall metrics into file, every line is a metric name and value pair. |
| 2. if `metric_details` dict is not None, write raw metric data of every image into file, every line for 1 image. |
| 3. if `summary_ops` is not None, compute summary based on operations on `metric_details` and write to file. |
| |
| Args: |
| save_dir: directory to save all the metrics reports. |
| images: name or path of every input image corresponding to the metric_details data. |
| if None, will use index number as the filename of every input image. |
| metrics: a dictionary of (metric name, metric value) pairs. |
| metric_details: a dictionary of (metric name, metric raw values) pairs, usually, it comes from metrics |
| computation, for example, the raw value can be the mean_dice of every channel of every input image. |
| summary_ops: expected computation operations to generate the summary report. |
| it can be: None, "*" or list of strings, default to None. |
| None - don't generate summary report for every expected metric_details. |
| "*" - generate summary report for every metric_details with all the supported operations. |
| list of strings - generate summary report for every metric_details with specified operations, they |
| should be within list: ["mean", "median", "max", "min", "<int>percentile", "std", "notnans"]. |
| the number in "<int>percentile" should be [0, 100], like: "15percentile". default: "90percentile". |
| for more details, please check: https://numpy.org/doc/stable/reference/generated/numpy.nanpercentile.html. |
| note that: for the overall summary, it computes `nanmean` of all classes for each image first, |
| then compute summary. example of the generated summary report:: |
| |
| class mean median max 5percentile 95percentile notnans |
| class0 6.0000 6.0000 7.0000 5.1000 6.9000 2.0000 |
| class1 6.0000 6.0000 6.0000 6.0000 6.0000 1.0000 |
| mean 6.2500 6.2500 7.0000 5.5750 6.9250 2.0000 |
| |
| deli: the delimiter character in the saved file, default to "," as the default output type is `csv`. |
| to be consistent with: https://docs.python.org/3/library/csv.html#csv.Dialect.delimiter. |
| output_type: expected output file type, supported types: ["csv"], default to "csv". |
| class_labels: list of class names used to name the classes in the output report, if None, |
| "class0", ..., "classn" are used, default to None. |
| |
| """ |
| if output_type.lower() != "csv": |
| raise ValueError(f"unsupported output type: {output_type}.") |
|
|
| if not os.path.exists(save_dir): |
| os.makedirs(save_dir) |
|
|
| if metrics is not None and len(metrics) > 0: |
| with open(os.path.join(save_dir, "metrics.csv"), "w") as f: |
| for k, v in metrics.items(): |
| f.write(f"{k}{deli}{str(v)}\n") |
| if metric_details is not None and len(metric_details) > 0: |
| for k, v in metric_details.items(): |
| if isinstance(v, torch.Tensor): |
| v = v.cpu().numpy() |
| if v.ndim == 0: |
| |
| v = v.reshape((1, 1)) |
| elif v.ndim == 1: |
| |
| v = v.reshape((-1, 1)) |
|
|
| |
| if class_labels is None: |
| class_labels = ["class" + str(i) for i in range(v.shape[1])] |
| else: |
| class_labels = [str(i) for i in class_labels] |
|
|
| class_labels += ["mean"] |
| v = np.concatenate([v, np.nanmean(v, axis=1, keepdims=True)], axis=1) |
|
|
| with open(os.path.join(save_dir, f"{k}_raw.csv"), "w") as f: |
| f.write(f"filename{deli}{deli.join(class_labels)}\n") |
| for i, b in enumerate(v): |
| f.write( |
| f"{images[i] if images is not None else str(i)}{deli}" |
| f"{deli.join([f'{c:.4f}' if isinstance(c, (int, float)) else str(c) for c in b])}\n" |
| ) |
|
|
| if summary_ops is not None: |
| supported_ops = OrderedDict( |
| { |
| "mean": np.nanmean, |
| "median": np.nanmedian, |
| "max": np.nanmax, |
| "min": np.nanmin, |
| "90percentile": lambda x: np.nanpercentile(x[0], x[1]), |
| "std": np.nanstd, |
| "notnans": lambda x: (~np.isnan(x)).sum(), |
| } |
| ) |
| ops = ensure_tuple(summary_ops) |
| if "*" in ops: |
| ops = tuple(supported_ops.keys()) |
|
|
| def _compute_op(op: str, d: np.ndarray) -> Any: |
| if not op.endswith("percentile"): |
| c_op = look_up_option(op, supported_ops) |
| return c_op(d) |
|
|
| threshold = int(op.split("percentile")[0]) |
| return supported_ops["90percentile"]((d, threshold)) |
|
|
| with open(os.path.join(save_dir, f"{k}_summary.csv"), "w") as f: |
| f.write(f"class{deli}{deli.join(ops)}\n") |
| for i, c in enumerate(np.transpose(v)): |
| f.write(f"{class_labels[i]}{deli}{deli.join([f'{_compute_op(k, c):.4f}' for k in ops])}\n") |
|
|
|
|
| def from_engine(keys: KeysCollection, first: bool = False) -> Callable: |
| """ |
| Utility function to simplify the `batch_transform` or `output_transform` args of ignite components |
| when handling dictionary or list of dictionaries(for example: `engine.state.batch` or `engine.state.output`). |
| Users only need to set the expected keys, then it will return a callable function to extract data from |
| dictionary and construct a tuple respectively. |
| |
| If data is a list of dictionaries after decollating, extract expected keys and construct lists respectively, |
| for example, if data is `[{"A": 1, "B": 2}, {"A": 3, "B": 4}]`, from_engine(["A", "B"]): `([1, 3], [2, 4])`. |
| |
| It can help avoid a complicated `lambda` function and make the arg of metrics more straight-forward. |
| For example, set the first key as the prediction and the second key as label to get the expected data |
| from `engine.state.output` for a metric:: |
| |
| from monai.handlers import MeanDice, from_engine |
| |
| metric = MeanDice( |
| include_background=False, |
| output_transform=from_engine(["pred", "label"]) |
| ) |
| |
| Args: |
| keys: specified keys to extract data from dictionary or decollated list of dictionaries. |
| first: whether only extract specified keys from the first item if input data is a list of dictionaries, |
| it's used to extract the scalar data which doesn't have batch dim and was replicated into every |
| dictionary when decollating, like `loss`, etc. |
| |
| |
| """ |
| _keys = ensure_tuple(keys) |
|
|
| def _wrapper(data): |
| if isinstance(data, dict): |
| return tuple(data[k] for k in _keys) |
| if isinstance(data, list) and isinstance(data[0], dict): |
| |
| |
| ret = [data[0][k] if first else [i[k] for i in data] for k in _keys] |
| return tuple(ret) if len(ret) > 1 else ret[0] |
|
|
| return _wrapper |
|
|
|
|
| def ignore_data(x: Any) -> None: |
| """ |
| Always return `None` for any input data. |
| A typical usage is to avoid logging the engine output of every iteration during evaluation. |
| |
| """ |
| return None |
|
|