| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | import inspect |
| | import itertools |
| | import math |
| | import os |
| | import pprint |
| | import random |
| | import shutil |
| | import subprocess |
| | import tempfile |
| | import types |
| | import warnings |
| | from ast import literal_eval |
| | from collections.abc import Callable, Iterable, Sequence |
| | from distutils.util import strtobool |
| | from math import log10 |
| | from pathlib import Path |
| | from typing import TYPE_CHECKING, Any, TypeVar, cast, overload |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai.config.type_definitions import NdarrayOrTensor, NdarrayTensor, PathLike |
| | from monai.utils.module import optional_import, version_leq |
| |
|
| | if TYPE_CHECKING: |
| | from yaml import SafeLoader |
| | else: |
| | SafeLoader, _ = optional_import("yaml", name="SafeLoader", as_type="base") |
| |
|
| | __all__ = [ |
| | "zip_with", |
| | "star_zip_with", |
| | "first", |
| | "issequenceiterable", |
| | "is_immutable", |
| | "ensure_tuple", |
| | "ensure_tuple_size", |
| | "ensure_tuple_rep", |
| | "to_tuple_of_dictionaries", |
| | "fall_back_tuple", |
| | "is_scalar_tensor", |
| | "is_scalar", |
| | "progress_bar", |
| | "get_seed", |
| | "set_determinism", |
| | "list_to_dict", |
| | "MAX_SEED", |
| | "copy_to_device", |
| | "str2bool", |
| | "str2list", |
| | "MONAIEnvVars", |
| | "ImageMetaKey", |
| | "is_module_ver_at_least", |
| | "has_option", |
| | "sample_slices", |
| | "check_parent_dir", |
| | "save_obj", |
| | "label_union", |
| | "path_to_uri", |
| | "pprint_edges", |
| | "check_key_duplicates", |
| | "CheckKeyDuplicatesYamlLoader", |
| | "ConvertUnits", |
| | "check_kwargs_exist_in_class_init", |
| | "run_cmd", |
| | ] |
| |
|
| | _seed = None |
| | _flag_deterministic = torch.backends.cudnn.deterministic |
| | _flag_cudnn_benchmark = torch.backends.cudnn.benchmark |
| | NP_MAX = np.iinfo(np.uint32).max |
| | MAX_SEED = NP_MAX + 1 |
| |
|
| |
|
| | def zip_with(op, *vals, mapfunc=map): |
| | """ |
| | Map `op`, using `mapfunc`, to each tuple derived from zipping the iterables in `vals`. |
| | """ |
| | return mapfunc(op, zip(*vals)) |
| |
|
| |
|
| | def star_zip_with(op, *vals): |
| | """ |
| | Use starmap as the mapping function in zipWith. |
| | """ |
| | return zip_with(op, *vals, mapfunc=itertools.starmap) |
| |
|
| |
|
| | T = TypeVar("T") |
| |
|
| |
|
| | @overload |
| | def first(iterable: Iterable[T], default: T) -> T: ... |
| |
|
| |
|
| | @overload |
| | def first(iterable: Iterable[T]) -> T | None: ... |
| |
|
| |
|
| | def first(iterable: Iterable[T], default: T | None = None) -> T | None: |
| | """ |
| | Returns the first item in the given iterable or `default` if empty, meaningful mostly with 'for' expressions. |
| | """ |
| | for i in iterable: |
| | return i |
| | return default |
| |
|
| |
|
| | def issequenceiterable(obj: Any) -> bool: |
| | """ |
| | Determine if the object is an iterable sequence and is not a string. |
| | """ |
| | try: |
| | if hasattr(obj, "ndim") and obj.ndim == 0: |
| | return False |
| | except Exception: |
| | return False |
| | return isinstance(obj, Iterable) and not isinstance(obj, (str, bytes)) |
| |
|
| |
|
| | def is_immutable(obj: Any) -> bool: |
| | """ |
| | Determine if the object is an immutable object. |
| | |
| | see also https://github.com/python/cpython/blob/3.11/Lib/copy.py#L109 |
| | """ |
| | return isinstance(obj, (type(None), int, float, bool, complex, str, tuple, bytes, type, range, slice)) |
| |
|
| |
|
| | def ensure_tuple(vals: Any, wrap_array: bool = False) -> tuple: |
| | """ |
| | Returns a tuple of `vals`. |
| | |
| | Args: |
| | vals: input data to convert to a tuple. |
| | wrap_array: if `True`, treat the input numerical array (ndarray/tensor) as one item of the tuple. |
| | if `False`, try to convert the array with `tuple(vals)`, default to `False`. |
| | |
| | """ |
| | if wrap_array and isinstance(vals, (np.ndarray, torch.Tensor)): |
| | return (vals,) |
| | return tuple(vals) if issequenceiterable(vals) else (vals,) |
| |
|
| |
|
| | def ensure_tuple_size(vals: Any, dim: int, pad_val: Any = 0, pad_from_start: bool = False) -> tuple: |
| | """ |
| | Returns a copy of `tup` with `dim` values by either shortened or padded with `pad_val` as necessary. |
| | """ |
| | tup = ensure_tuple(vals) |
| | pad_dim = dim - len(tup) |
| | if pad_dim <= 0: |
| | return tup[:dim] |
| | if pad_from_start: |
| | return (pad_val,) * pad_dim + tup |
| | return tup + (pad_val,) * pad_dim |
| |
|
| |
|
| | def ensure_tuple_rep(tup: Any, dim: int) -> tuple[Any, ...]: |
| | """ |
| | Returns a copy of `tup` with `dim` values by either shortened or duplicated input. |
| | |
| | Raises: |
| | ValueError: When ``tup`` is a sequence and ``tup`` length is not ``dim``. |
| | |
| | Examples:: |
| | |
| | >>> ensure_tuple_rep(1, 3) |
| | (1, 1, 1) |
| | >>> ensure_tuple_rep(None, 3) |
| | (None, None, None) |
| | >>> ensure_tuple_rep('test', 3) |
| | ('test', 'test', 'test') |
| | >>> ensure_tuple_rep([1, 2, 3], 3) |
| | (1, 2, 3) |
| | >>> ensure_tuple_rep(range(3), 3) |
| | (0, 1, 2) |
| | >>> ensure_tuple_rep([1, 2], 3) |
| | ValueError: Sequence must have length 3, got length 2. |
| | |
| | """ |
| | if isinstance(tup, torch.Tensor): |
| | tup = tup.detach().cpu().numpy() |
| | if isinstance(tup, np.ndarray): |
| | tup = tup.tolist() |
| | if not issequenceiterable(tup): |
| | return (tup,) * dim |
| | if len(tup) == dim: |
| | return tuple(tup) |
| |
|
| | raise ValueError(f"Sequence must have length {dim}, got {len(tup)}.") |
| |
|
| |
|
| | def to_tuple_of_dictionaries(dictionary_of_tuples: dict, keys: Any) -> tuple[dict[Any, Any], ...]: |
| | """ |
| | Given a dictionary whose values contain scalars or tuples (with the same length as ``keys``), |
| | Create a dictionary for each key containing the scalar values mapping to that key. |
| | |
| | Args: |
| | dictionary_of_tuples: a dictionary whose values are scalars or tuples whose length is |
| | the length of ``keys`` |
| | keys: a tuple of string values representing the keys in question |
| | |
| | Returns: |
| | a tuple of dictionaries that contain scalar values, one dictionary for each key |
| | |
| | Raises: |
| | ValueError: when values in the dictionary are tuples but not the same length as the length |
| | of ``keys`` |
| | |
| | Examples: |
| | >>> to_tuple_of_dictionaries({'a': 1 'b': (2, 3), 'c': (4, 4)}, ("x", "y")) |
| | ({'a':1, 'b':2, 'c':4}, {'a':1, 'b':3, 'c':4}) |
| | |
| | """ |
| |
|
| | keys = ensure_tuple(keys) |
| | if len(keys) == 0: |
| | return tuple({}) |
| |
|
| | dict_overrides = {k: ensure_tuple_rep(v, len(keys)) for k, v in dictionary_of_tuples.items()} |
| | return tuple({k: v[ik] for (k, v) in dict_overrides.items()} for ik in range(len(keys))) |
| |
|
| |
|
| | def fall_back_tuple( |
| | user_provided: Any, default: Sequence | NdarrayTensor, func: Callable = lambda x: x and x > 0 |
| | ) -> tuple[Any, ...]: |
| | """ |
| | Refine `user_provided` according to the `default`, and returns as a validated tuple. |
| | |
| | The validation is done for each element in `user_provided` using `func`. |
| | If `func(user_provided[idx])` returns False, the corresponding `default[idx]` will be used |
| | as the fallback. |
| | |
| | Typically used when `user_provided` is a tuple of window size provided by the user, |
| | `default` is defined by data, this function returns an updated `user_provided` with its non-positive |
| | components replaced by the corresponding components from `default`. |
| | |
| | Args: |
| | user_provided: item to be validated. |
| | default: a sequence used to provided the fallbacks. |
| | func: a Callable to validate every components of `user_provided`. |
| | |
| | Examples:: |
| | |
| | >>> fall_back_tuple((1, 2), (32, 32)) |
| | (1, 2) |
| | >>> fall_back_tuple(None, (32, 32)) |
| | (32, 32) |
| | >>> fall_back_tuple((-1, 10), (32, 32)) |
| | (32, 10) |
| | >>> fall_back_tuple((-1, None), (32, 32)) |
| | (32, 32) |
| | >>> fall_back_tuple((1, None), (32, 32)) |
| | (1, 32) |
| | >>> fall_back_tuple(0, (32, 32)) |
| | (32, 32) |
| | >>> fall_back_tuple(range(3), (32, 64, 48)) |
| | (32, 1, 2) |
| | >>> fall_back_tuple([0], (32, 32)) |
| | ValueError: Sequence must have length 2, got length 1. |
| | |
| | """ |
| | ndim = len(default) |
| | user = ensure_tuple_rep(user_provided, ndim) |
| | return tuple( |
| | user_c if func(user_c) else default_c for default_c, user_c in zip(default, user) |
| | ) |
| |
|
| |
|
| | def is_scalar_tensor(val: Any) -> bool: |
| | return isinstance(val, torch.Tensor) and val.ndim == 0 |
| |
|
| |
|
| | def is_scalar(val: Any) -> bool: |
| | if isinstance(val, torch.Tensor) and val.ndim == 0: |
| | return True |
| | return bool(np.isscalar(val)) |
| |
|
| |
|
| | def progress_bar(index: int, count: int, desc: str | None = None, bar_len: int = 30, newline: bool = False) -> None: |
| | """print a progress bar to track some time consuming task. |
| | |
| | Args: |
| | index: current status in progress. |
| | count: total steps of the progress. |
| | desc: description of the progress bar, if not None, show before the progress bar. |
| | bar_len: the total length of the bar on screen, default is 30 char. |
| | newline: whether to print in a new line for every index. |
| | """ |
| | end = "\r" if not newline else "\r\n" |
| | filled_len = int(bar_len * index // count) |
| | bar = f"{desc} " if desc is not None else "" |
| | bar += "[" + "=" * filled_len + " " * (bar_len - filled_len) + "]" |
| | print(f"{index}/{count} {bar}", end=end) |
| | if index == count: |
| | print("") |
| |
|
| |
|
| | def get_seed() -> int | None: |
| | return _seed |
| |
|
| |
|
| | def set_determinism( |
| | seed: int | None = NP_MAX, |
| | use_deterministic_algorithms: bool | None = None, |
| | additional_settings: Sequence[Callable[[int], Any]] | Callable[[int], Any] | None = None, |
| | ) -> None: |
| | """ |
| | Set random seed for modules to enable or disable deterministic training. |
| | |
| | Args: |
| | seed: the random seed to use, default is np.iinfo(np.int32).max. |
| | It is recommended to set a large seed, i.e. a number that has a good balance |
| | of 0 and 1 bits. Avoid having many 0 bits in the seed. |
| | if set to None, will disable deterministic training. |
| | use_deterministic_algorithms: Set whether PyTorch operations must use "deterministic" algorithms. |
| | additional_settings: additional settings that need to set random seed. |
| | |
| | Note: |
| | |
| | This function will not affect the randomizable objects in :py:class:`monai.transforms.Randomizable`, which |
| | have independent random states. For those objects, the ``set_random_state()`` method should be used to |
| | ensure the deterministic behavior (alternatively, :py:class:`monai.data.DataLoader` by default sets the seeds |
| | according to the global random state, please see also: :py:class:`monai.data.utils.worker_init_fn` and |
| | :py:class:`monai.data.utils.set_rnd`). |
| | """ |
| | if seed is None: |
| | |
| | seed_ = torch.default_generator.seed() % MAX_SEED |
| | torch.manual_seed(seed_) |
| | else: |
| | seed = int(seed) % MAX_SEED |
| | torch.manual_seed(seed) |
| |
|
| | global _seed |
| | _seed = seed |
| | random.seed(seed) |
| | np.random.seed(seed) |
| |
|
| | if additional_settings is not None: |
| | additional_settings = ensure_tuple(additional_settings) |
| | for func in additional_settings: |
| | func(seed) |
| |
|
| | if torch.backends.flags_frozen(): |
| | warnings.warn("PyTorch global flag support of backends is disabled, enable it to set global `cudnn` flags.") |
| | torch.backends.__allow_nonbracketed_mutation_flag = True |
| |
|
| | if seed is not None: |
| | torch.backends.cudnn.deterministic = True |
| | torch.backends.cudnn.benchmark = False |
| | else: |
| | torch.backends.cudnn.deterministic = _flag_deterministic |
| | torch.backends.cudnn.benchmark = _flag_cudnn_benchmark |
| | if use_deterministic_algorithms is not None: |
| | if hasattr(torch, "use_deterministic_algorithms"): |
| | torch.use_deterministic_algorithms(use_deterministic_algorithms) |
| | elif hasattr(torch, "set_deterministic"): |
| | torch.set_deterministic(use_deterministic_algorithms) |
| | else: |
| | warnings.warn("use_deterministic_algorithms=True, but PyTorch version is too old to set the mode.") |
| |
|
| |
|
| | def list_to_dict(items): |
| | """ |
| | To convert a list of "key=value" pairs into a dictionary. |
| | For examples: items: `["a=1", "b=2", "c=3"]`, return: {"a": "1", "b": "2", "c": "3"}. |
| | If no "=" in the pair, use None as the value, for example: ["a"], return: {"a": None}. |
| | Note that it will remove the blanks around keys and values. |
| | |
| | """ |
| |
|
| | def _parse_var(s): |
| | items = s.split("=", maxsplit=1) |
| | key = items[0].strip(" \n\r\t'") |
| | value = items[1].strip(" \n\r\t'") if len(items) > 1 else None |
| | return key, value |
| |
|
| | d = {} |
| | if items: |
| | for item in items: |
| | key, value = _parse_var(item) |
| |
|
| | try: |
| | if key in d: |
| | raise KeyError(f"encounter duplicated key {key}.") |
| | d[key] = literal_eval(value) |
| | except ValueError: |
| | try: |
| | d[key] = bool(strtobool(str(value))) |
| | except ValueError: |
| | d[key] = value |
| | return d |
| |
|
| |
|
| | def copy_to_device( |
| | obj: Any, device: str | torch.device | None, non_blocking: bool = True, verbose: bool = False |
| | ) -> Any: |
| | """ |
| | Copy object or tuple/list/dictionary of objects to ``device``. |
| | |
| | Args: |
| | obj: object or tuple/list/dictionary of objects to move to ``device``. |
| | device: move ``obj`` to this device. Can be a string (e.g., ``cpu``, ``cuda``, |
| | ``cuda:0``, etc.) or of type ``torch.device``. |
| | non_blocking: when `True`, moves data to device asynchronously if |
| | possible, e.g., moving CPU Tensors with pinned memory to CUDA devices. |
| | verbose: when `True`, will print a warning for any elements of incompatible type |
| | not copied to ``device``. |
| | Returns: |
| | Same as input, copied to ``device`` where possible. Original input will be |
| | unchanged. |
| | """ |
| |
|
| | if hasattr(obj, "to"): |
| | return obj.to(device, non_blocking=non_blocking) |
| | if isinstance(obj, tuple): |
| | return tuple(copy_to_device(o, device, non_blocking) for o in obj) |
| | if isinstance(obj, list): |
| | return [copy_to_device(o, device, non_blocking) for o in obj] |
| | if isinstance(obj, dict): |
| | return {k: copy_to_device(o, device, non_blocking) for k, o in obj.items()} |
| | if verbose: |
| | fn_name = cast(types.FrameType, inspect.currentframe()).f_code.co_name |
| | warnings.warn(f"{fn_name} called with incompatible type: " + f"{type(obj)}. Data will be returned unchanged.") |
| |
|
| | return obj |
| |
|
| |
|
| | def str2bool(value: str | bool, default: bool = False, raise_exc: bool = True) -> bool: |
| | """ |
| | Convert a string to a boolean. Case insensitive. |
| | True: yes, true, t, y, 1. False: no, false, f, n, 0. |
| | |
| | Args: |
| | value: string to be converted to a boolean. If value is a bool already, simply return it. |
| | raise_exc: if value not in tuples of expected true or false inputs, |
| | should we raise an exception? If not, return `default`. |
| | Raises |
| | ValueError: value not in tuples of expected true or false inputs and |
| | `raise_exc` is `True`. |
| | Useful with argparse, for example: |
| | parser.add_argument("--convert", default=False, type=str2bool) |
| | python mycode.py --convert=True |
| | """ |
| |
|
| | if isinstance(value, bool): |
| | return value |
| |
|
| | true_set = ("yes", "true", "t", "y", "1") |
| | false_set = ("no", "false", "f", "n", "0") |
| |
|
| | if isinstance(value, str): |
| | value = value.lower() |
| | if value in true_set: |
| | return True |
| | if value in false_set: |
| | return False |
| |
|
| | if raise_exc: |
| | raise ValueError(f"Got \"{value}\", expected a value from: {', '.join(true_set + false_set)}") |
| | return default |
| |
|
| |
|
| | def str2list(value: str | list | None, raise_exc: bool = True) -> list | None: |
| | """ |
| | Convert a string to a list. Useful with argparse commandline arguments: |
| | parser.add_argument("--blocks", default=[1,2,3], type=str2list) |
| | python mycode.py --blocks=1,2,2,4 |
| | |
| | Args: |
| | value: string (comma separated) to be converted to a list |
| | raise_exc: if not possible to convert to a list, raise an exception |
| | Raises |
| | ValueError: value not a string or list or not possible to convert |
| | """ |
| |
|
| | if value is None: |
| | return None |
| | elif isinstance(value, list): |
| | return value |
| | elif isinstance(value, str): |
| | v = value.split(",") |
| | for i in range(len(v)): |
| | try: |
| | a = literal_eval(v[i].strip()) |
| | v[i] = a |
| | except Exception: |
| | pass |
| | return v |
| | elif raise_exc: |
| | raise ValueError(f'Unable to convert "{value}", expected a comma-separated str, e.g. 1,2,3') |
| |
|
| | return None |
| |
|
| |
|
| | class MONAIEnvVars: |
| | """ |
| | Environment variables used by MONAI. |
| | """ |
| |
|
| | @staticmethod |
| | def data_dir() -> str | None: |
| | return os.environ.get("MONAI_DATA_DIRECTORY") |
| |
|
| | @staticmethod |
| | def debug() -> bool: |
| | val = os.environ.get("MONAI_DEBUG", False) |
| | return val if isinstance(val, bool) else str2bool(val) |
| |
|
| | @staticmethod |
| | def doc_images() -> str | None: |
| | return os.environ.get("MONAI_DOC_IMAGES") |
| |
|
| | @staticmethod |
| | def algo_hash() -> str | None: |
| | return os.environ.get("MONAI_ALGO_HASH", "e4cf5a1") |
| |
|
| | @staticmethod |
| | def trace_transform() -> str | None: |
| | return os.environ.get("MONAI_TRACE_TRANSFORM", "1") |
| |
|
| | @staticmethod |
| | def eval_expr() -> str | None: |
| | return os.environ.get("MONAI_EVAL_EXPR", "1") |
| |
|
| | @staticmethod |
| | def allow_missing_reference() -> str | None: |
| | return os.environ.get("MONAI_ALLOW_MISSING_REFERENCE", "1") |
| |
|
| | @staticmethod |
| | def extra_test_data() -> str | None: |
| | return os.environ.get("MONAI_EXTRA_TEST_DATA", "1") |
| |
|
| | @staticmethod |
| | def testing_algo_template() -> str | None: |
| | return os.environ.get("MONAI_TESTING_ALGO_TEMPLATE", None) |
| |
|
| |
|
| | class ImageMetaKey: |
| | """ |
| | Common key names in the metadata header of images |
| | """ |
| |
|
| | FILENAME_OR_OBJ = "filename_or_obj" |
| | PATCH_INDEX = "patch_index" |
| | SPATIAL_SHAPE = "spatial_shape" |
| |
|
| |
|
| | def has_option(obj: Callable, keywords: str | Sequence[str]) -> bool: |
| | """ |
| | Return a boolean indicating whether the given callable `obj` has the `keywords` in its signature. |
| | """ |
| | if not callable(obj): |
| | return False |
| | sig = inspect.signature(obj) |
| | return all(key in sig.parameters for key in ensure_tuple(keywords)) |
| |
|
| |
|
| | def is_module_ver_at_least(module, version): |
| | """Determine if a module's version is at least equal to the given value. |
| | |
| | Args: |
| | module: imported module's name, e.g., `np` or `torch`. |
| | version: required version, given as a tuple, e.g., `(1, 8, 0)`. |
| | Returns: |
| | `True` if module is the given version or newer. |
| | """ |
| | test_ver = ".".join(map(str, version)) |
| | return module.__version__ != test_ver and version_leq(test_ver, module.__version__) |
| |
|
| |
|
| | def sample_slices(data: NdarrayOrTensor, dim: int = 1, as_indices: bool = True, *slicevals: int) -> NdarrayOrTensor: |
| | """sample several slices of input numpy array or Tensor on specified `dim`. |
| | |
| | Args: |
| | data: input data to sample slices, can be numpy array or PyTorch Tensor. |
| | dim: expected dimension index to sample slices, default to `1`. |
| | as_indices: if `True`, `slicevals` arg will be treated as the expected indices of slice, like: `1, 3, 5` |
| | means `data[..., [1, 3, 5], ...]`, if `False`, `slicevals` arg will be treated as args for `slice` func, |
| | like: `1, None` means `data[..., [1:], ...]`, `1, 5` means `data[..., [1: 5], ...]`. |
| | slicevals: indices of slices or start and end indices of expected slices, depends on `as_indices` flag. |
| | |
| | """ |
| | slices = [slice(None)] * len(data.shape) |
| | slices[dim] = slicevals if as_indices else slice(*slicevals) |
| |
|
| | return data[tuple(slices)] |
| |
|
| |
|
| | def check_parent_dir(path: PathLike, create_dir: bool = True) -> None: |
| | """ |
| | Utility to check whether the parent directory of the `path` exists. |
| | |
| | Args: |
| | path: input path to check the parent directory. |
| | create_dir: if True, when the parent directory doesn't exist, create the directory, |
| | otherwise, raise exception. |
| | |
| | """ |
| | path = Path(path) |
| | path_dir = path.parent |
| | if not path_dir.exists(): |
| | if create_dir: |
| | path_dir.mkdir(parents=True) |
| | else: |
| | raise ValueError(f"the directory of specified path does not exist: `{path_dir}`.") |
| |
|
| |
|
| | def save_obj( |
| | obj: object, |
| | path: PathLike, |
| | create_dir: bool = True, |
| | atomic: bool = True, |
| | func: Callable | None = None, |
| | **kwargs: Any, |
| | ) -> None: |
| | """ |
| | Save an object to file with specified path. |
| | Support to serialize to a temporary file first, then move to final destination, |
| | so that files are guaranteed to not be damaged if exception occurs. |
| | |
| | Args: |
| | obj: input object data to save. |
| | path: target file path to save the input object. |
| | create_dir: whether to create dictionary of the path if not existing, default to `True`. |
| | atomic: if `True`, state is serialized to a temporary file first, then move to final destination. |
| | so that files are guaranteed to not be damaged if exception occurs. default to `True`. |
| | func: the function to save file, if None, default to `torch.save`. |
| | kwargs: other args for the save `func` except for the checkpoint and filename. |
| | default `func` is `torch.save()`, details of other args: |
| | https://pytorch.org/docs/stable/generated/torch.save.html. |
| | |
| | """ |
| | path = Path(path) |
| | check_parent_dir(path=path, create_dir=create_dir) |
| | if path.exists(): |
| | |
| | os.remove(path) |
| |
|
| | if func is None: |
| | func = torch.save |
| |
|
| | if not atomic: |
| | func(obj=obj, f=path, **kwargs) |
| | return |
| | try: |
| | |
| | with tempfile.TemporaryDirectory() as tempdir: |
| | temp_path: Path = Path(tempdir) / path.name |
| | func(obj=obj, f=temp_path, **kwargs) |
| | if temp_path.is_file(): |
| | shutil.move(str(temp_path), path) |
| | except PermissionError: |
| | pass |
| |
|
| |
|
| | def label_union(x: list | np.ndarray) -> list: |
| | """ |
| | Compute the union of class IDs in label and generate a list to include all class IDs |
| | Args: |
| | x: a list of numbers (for example, class_IDs) |
| | |
| | Returns |
| | a list showing the union (the union the class IDs) |
| | """ |
| | return list(set.union(set(np.array(x).tolist()))) |
| |
|
| |
|
| | def prob2class(x: torch.Tensor, sigmoid: bool = False, threshold: float = 0.5, **kwargs: Any) -> torch.Tensor: |
| | """ |
| | Compute the lab from the probability of predicted feature maps |
| | |
| | Args: |
| | sigmoid: If the sigmoid function should be used. |
| | threshold: threshold value to activate the sigmoid function. |
| | """ |
| | return torch.argmax(x, **kwargs) if not sigmoid else (x > threshold).int() |
| |
|
| |
|
| | def path_to_uri(path: PathLike) -> str: |
| | """ |
| | Convert a file path to URI. if not absolute path, will convert to absolute path first. |
| | |
| | Args: |
| | path: input file path to convert, can be a string or `Path` object. |
| | |
| | """ |
| | return Path(path).absolute().as_uri() |
| |
|
| |
|
| | def pprint_edges(val: Any, n_lines: int = 20) -> str: |
| | """ |
| | Pretty print the head and tail ``n_lines`` of ``val``, and omit the middle part if the part has more than 3 lines. |
| | |
| | Returns: the formatted string. |
| | """ |
| | val_str = pprint.pformat(val).splitlines(True) |
| | n_lines = max(n_lines, 1) |
| | if len(val_str) > n_lines * 2 + 3: |
| | hidden_n = len(val_str) - n_lines * 2 |
| | val_str = val_str[:n_lines] + [f"\n ... omitted {hidden_n} line(s)\n\n"] + val_str[-n_lines:] |
| | return "".join(val_str) |
| |
|
| |
|
| | def check_key_duplicates(ordered_pairs: Sequence[tuple[Any, Any]]) -> dict[Any, Any]: |
| | """ |
| | Checks if there is a duplicated key in the sequence of `ordered_pairs`. |
| | If there is - it will log a warning or raise ValueError |
| | (if configured by environmental var `MONAI_FAIL_ON_DUPLICATE_CONFIG==1`) |
| | |
| | Otherwise, it returns the dict made from this sequence. |
| | |
| | Satisfies a format for an `object_pairs_hook` in `json.load` |
| | |
| | Args: |
| | ordered_pairs: sequence of (key, value) |
| | """ |
| | keys = set() |
| | for k, _ in ordered_pairs: |
| | if k in keys: |
| | if os.environ.get("MONAI_FAIL_ON_DUPLICATE_CONFIG", "0") == "1": |
| | raise ValueError(f"Duplicate key: `{k}`") |
| | else: |
| | warnings.warn(f"Duplicate key: `{k}`") |
| | else: |
| | keys.add(k) |
| | return dict(ordered_pairs) |
| |
|
| |
|
| | class CheckKeyDuplicatesYamlLoader(SafeLoader): |
| |
|
| | def construct_mapping(self, node, deep=False): |
| | mapping = set() |
| | for key_node, _ in node.value: |
| | key = self.construct_object(key_node, deep=deep) |
| | if key in mapping: |
| | if os.environ.get("MONAI_FAIL_ON_DUPLICATE_CONFIG", "0") == "1": |
| | raise ValueError(f"Duplicate key: `{key}`") |
| | else: |
| | warnings.warn(f"Duplicate key: `{key}`") |
| | mapping.add(key) |
| | return super().construct_mapping(node, deep) |
| |
|
| |
|
| | class ConvertUnits: |
| | """ |
| | Convert the values from input unit to the target unit |
| | |
| | Args: |
| | input_unit: the unit of the input quantity |
| | target_unit: the unit of the target quantity |
| | |
| | """ |
| |
|
| | imperial_unit_of_length = {"inch": 0.0254, "foot": 0.3048, "yard": 0.9144, "mile": 1609.344} |
| |
|
| | unit_prefix = { |
| | "peta": 15, |
| | "tera": 12, |
| | "giga": 9, |
| | "mega": 6, |
| | "kilo": 3, |
| | "hecto": 2, |
| | "deca": 1, |
| | "deci": -1, |
| | "centi": -2, |
| | "milli": -3, |
| | "micro": -6, |
| | "nano": -9, |
| | "pico": -12, |
| | "femto": -15, |
| | } |
| | base_units = ["meter", "byte", "bit"] |
| |
|
| | def __init__(self, input_unit: str, target_unit: str) -> None: |
| | self.input_unit, input_base = self._get_valid_unit_and_base(input_unit) |
| | self.target_unit, target_base = self._get_valid_unit_and_base(target_unit) |
| | if input_base == target_base: |
| | self.unit_base = input_base |
| | else: |
| | raise ValueError( |
| | "Both input and target units should be from the same quantity. " |
| | f"Input quantity is {input_base} while target quantity is {target_base}" |
| | ) |
| | self._calculate_conversion_factor() |
| |
|
| | def _get_valid_unit_and_base(self, unit): |
| | unit = str(unit).lower() |
| | if unit in self.imperial_unit_of_length: |
| | return unit, "meter" |
| | for base_unit in self.base_units: |
| | if unit.endswith(base_unit): |
| | return unit, base_unit |
| | raise ValueError(f"Currently, it only supports length conversion but `{unit}` is given.") |
| |
|
| | def _get_unit_power(self, unit): |
| | """Calculate the power of the unit factor with respect to the base unit""" |
| | if unit in self.imperial_unit_of_length: |
| | return log10(self.imperial_unit_of_length[unit]) |
| |
|
| | prefix = unit[: len(self.unit_base)] |
| | if prefix == "": |
| | return 1.0 |
| | return self.unit_prefix[prefix] |
| |
|
| | def _calculate_conversion_factor(self): |
| | """Calculate unit conversion factor with respect to the input unit""" |
| | if self.input_unit == self.target_unit: |
| | return 1.0 |
| | input_power = self._get_unit_power(self.input_unit) |
| | target_power = self._get_unit_power(self.target_unit) |
| | self.conversion_factor = 10 ** (input_power - target_power) |
| |
|
| | def __call__(self, value: int | float) -> Any: |
| | return float(value) * self.conversion_factor |
| |
|
| |
|
| | def check_kwargs_exist_in_class_init(cls, kwargs): |
| | """ |
| | Check if the all keys in kwargs exist in the __init__ method of the class. |
| | |
| | Args: |
| | cls: the class to check. |
| | kwargs: kwargs to examine. |
| | |
| | Returns: |
| | a boolean indicating if all keys exist. |
| | a set of extra keys that are not used in the __init__. |
| | """ |
| | init_signature = inspect.signature(cls.__init__) |
| | init_params = set(init_signature.parameters) - {"self"} |
| | input_kwargs = set(kwargs) |
| | extra_kwargs = input_kwargs - init_params |
| |
|
| | return extra_kwargs == set(), extra_kwargs |
| |
|
| |
|
| | def run_cmd(cmd_list: list[str], **kwargs: Any) -> subprocess.CompletedProcess: |
| | """ |
| | Run a command by using ``subprocess.run`` with capture_output=True and stderr=subprocess.STDOUT |
| | so that the raise exception will have that information. The argument `capture_output` can be set explicitly |
| | if desired, but will be overriden with the debug status from the variable. |
| | |
| | Args: |
| | cmd_list: a list of strings describing the command to run. |
| | kwargs: keyword arguments supported by the ``subprocess.run`` method. |
| | |
| | Returns: |
| | a CompletedProcess instance after the command completes. |
| | """ |
| | debug = MONAIEnvVars.debug() |
| | kwargs["capture_output"] = kwargs.get("capture_output", debug) |
| |
|
| | if kwargs.pop("run_cmd_verbose", False): |
| | import monai |
| |
|
| | monai.apps.utils.get_logger("run_cmd").info(f"{cmd_list}") |
| | try: |
| | return subprocess.run(cmd_list, **kwargs) |
| | except subprocess.CalledProcessError as e: |
| | if not debug: |
| | raise |
| | output = str(e.stdout.decode(errors="replace")) |
| | errors = str(e.stderr.decode(errors="replace")) |
| | raise RuntimeError(f"subprocess call error {e.returncode}: {errors}, {output}.") from e |
| |
|
| |
|
| | def is_sqrt(num: Sequence[int] | int) -> bool: |
| | """ |
| | Determine if the input is a square number or a squence of square numbers. |
| | """ |
| | num = ensure_tuple(num) |
| | sqrt_num = [int(math.sqrt(_num)) for _num in num] |
| | ret = [_i * _j for _i, _j in zip(sqrt_num, sqrt_num)] |
| | return ensure_tuple(ret) == num |
| |
|
| |
|
| | def unsqueeze_right(arr: NdarrayOrTensor, ndim: int) -> NdarrayOrTensor: |
| | """Append 1-sized dimensions to `arr` to create a result with `ndim` dimensions.""" |
| | return arr[(...,) + (None,) * (ndim - arr.ndim)] |
| |
|
| |
|
| | def unsqueeze_left(arr: NdarrayOrTensor, ndim: int) -> NdarrayOrTensor: |
| | """Prepend 1-sized dimensions to `arr` to create a result with `ndim` dimensions.""" |
| | return arr[(None,) * (ndim - arr.ndim)] |
| |
|