| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | Utilities and types for defining networks, these depend on PyTorch. |
| | """ |
| |
|
| | from __future__ import annotations |
| |
|
| | import io |
| | import re |
| | import warnings |
| | from collections import OrderedDict |
| | from collections.abc import Callable, Mapping, Sequence |
| | from contextlib import contextmanager |
| | from copy import deepcopy |
| | from typing import Any |
| |
|
| | import numpy as np |
| | import torch |
| | import torch.nn as nn |
| |
|
| | from monai.apps.utils import get_logger |
| | from monai.config import PathLike |
| | from monai.utils.misc import ensure_tuple, save_obj, set_determinism |
| | from monai.utils.module import look_up_option, optional_import, pytorch_after |
| | from monai.utils.type_conversion import convert_to_dst_type, convert_to_tensor |
| |
|
| | onnx, _ = optional_import("onnx") |
| | onnxreference, _ = optional_import("onnx.reference") |
| | onnxruntime, _ = optional_import("onnxruntime") |
| |
|
| | __all__ = [ |
| | "one_hot", |
| | "predict_segmentation", |
| | "normalize_transform", |
| | "to_norm_affine", |
| | "normal_init", |
| | "icnr_init", |
| | "pixelshuffle", |
| | "eval_mode", |
| | "train_mode", |
| | "get_state_dict", |
| | "copy_model_state", |
| | "save_state", |
| | "convert_to_onnx", |
| | "convert_to_torchscript", |
| | "convert_to_trt", |
| | "meshgrid_ij", |
| | "meshgrid_xy", |
| | "replace_modules", |
| | "replace_modules_temp", |
| | "look_up_named_module", |
| | "set_named_module", |
| | "has_nvfuser_instance_norm", |
| | ] |
| |
|
| | logger = get_logger(module_name=__name__) |
| |
|
| | _has_nvfuser = None |
| |
|
| |
|
| | def has_nvfuser_instance_norm(): |
| | """whether the current environment has InstanceNorm3dNVFuser |
| | https://github.com/NVIDIA/apex/blob/23.05-devel/apex/normalization/instance_norm.py#L15-L16 |
| | """ |
| | global _has_nvfuser |
| | if _has_nvfuser is not None: |
| | return _has_nvfuser |
| |
|
| | _, _has_nvfuser = optional_import("apex.normalization", name="InstanceNorm3dNVFuser") |
| | if not _has_nvfuser: |
| | return False |
| | try: |
| | import importlib |
| |
|
| | importlib.import_module("instance_norm_nvfuser_cuda") |
| | except ImportError: |
| | _has_nvfuser = False |
| | return _has_nvfuser |
| |
|
| |
|
| | def look_up_named_module(name: str, mod, print_all_options=False): |
| | """ |
| | get the named module in `mod` by the attribute name, |
| | for example ``look_up_named_module(net, "features.3.1.attn")`` |
| | |
| | Args: |
| | name: a string representing the module attribute. |
| | mod: a pytorch module to be searched (in ``mod.named_modules()``). |
| | print_all_options: whether to print all named modules when `name` is not found in `mod`. Defaults to False. |
| | |
| | Returns: |
| | the corresponding pytorch module's subcomponent such as ``net.features[3][1].attn`` |
| | """ |
| | name_str = look_up_option( |
| | name, {n[0] for n in mod.named_modules()}, default=None, print_all_options=print_all_options |
| | ) |
| | if name_str is None: |
| | return None |
| | if name_str == "": |
| | return mod |
| | for n in name_str.split("."): |
| | if n.isdigit(): |
| | mod = mod[int(n)] |
| | else: |
| | n = look_up_option(n, {item[0] for item in mod.named_modules()}, default=None, print_all_options=False) |
| | if n is None: |
| | return None |
| | mod = getattr(mod, n) |
| | return mod |
| |
|
| |
|
| | def set_named_module(mod, name: str, new_layer): |
| | """ |
| | look up `name` in `mod` and replace the layer with `new_layer`, return the updated `mod`. |
| | |
| | Args: |
| | mod: a pytorch module to be updated. |
| | name: a string representing the target module attribute. |
| | new_layer: a new module replacing the corresponding layer at ``mod.name``. |
| | |
| | Returns: |
| | an updated ``mod`` |
| | |
| | See also: :py:func:`monai.networks.utils.look_up_named_module`. |
| | """ |
| | mods_attr = name.rsplit(".", 1) |
| | submods, attr = mods_attr if len(mods_attr) == 2 else ("", name) |
| | if not attr: |
| | return new_layer |
| | _mod = look_up_named_module(submods, mod) |
| | setattr(_mod, attr, new_layer) |
| | return mod |
| |
|
| |
|
| | def one_hot(labels: torch.Tensor, num_classes: int, dtype: torch.dtype = torch.float, dim: int = 1) -> torch.Tensor: |
| | """ |
| | For every value v in `labels`, the value in the output will be either 1 or 0. Each vector along the `dim`-th |
| | dimension has the "one-hot" format, i.e., it has a total length of `num_classes`, |
| | with a one and `num_class-1` zeros. |
| | Note that this will include the background label, thus a binary mask should be treated as having two classes. |
| | |
| | Args: |
| | labels: input tensor of integers to be converted into the 'one-hot' format. Internally `labels` will be |
| | converted into integers `labels.long()`. |
| | num_classes: number of output channels, the corresponding length of `labels[dim]` will be converted to |
| | `num_classes` from `1`. |
| | dtype: the data type of the output one_hot label. |
| | dim: the dimension to be converted to `num_classes` channels from `1` channel, should be non-negative number. |
| | |
| | Example: |
| | |
| | For a tensor `labels` of dimensions [B]1[spatial_dims], return a tensor of dimensions `[B]N[spatial_dims]` |
| | when `num_classes=N` number of classes and `dim=1`. |
| | |
| | .. code-block:: python |
| | |
| | from monai.networks.utils import one_hot |
| | import torch |
| | |
| | a = torch.randint(0, 2, size=(1, 2, 2, 2)) |
| | out = one_hot(a, num_classes=2, dim=0) |
| | print(out.shape) # torch.Size([2, 2, 2, 2]) |
| | |
| | a = torch.randint(0, 2, size=(2, 1, 2, 2, 2)) |
| | out = one_hot(a, num_classes=2, dim=1) |
| | print(out.shape) # torch.Size([2, 2, 2, 2, 2]) |
| | |
| | """ |
| |
|
| | |
| | if labels.ndim < dim + 1: |
| | shape = list(labels.shape) + [1] * (dim + 1 - len(labels.shape)) |
| | labels = torch.reshape(labels, shape) |
| |
|
| | sh = list(labels.shape) |
| |
|
| | if sh[dim] != 1: |
| | raise AssertionError("labels should have a channel with length equal to one.") |
| |
|
| | sh[dim] = num_classes |
| |
|
| | o = torch.zeros(size=sh, dtype=dtype, device=labels.device) |
| | labels = o.scatter_(dim=dim, index=labels.long(), value=1) |
| |
|
| | return labels |
| |
|
| |
|
| | def predict_segmentation(logits: torch.Tensor, mutually_exclusive: bool = False, threshold: float = 0.0) -> Any: |
| | """ |
| | Given the logits from a network, computing the segmentation by thresholding all values above 0 |
| | if multi-labels task, computing the `argmax` along the channel axis if multi-classes task, |
| | logits has shape `BCHW[D]`. |
| | |
| | Args: |
| | logits: raw data of model output. |
| | mutually_exclusive: if True, `logits` will be converted into a binary matrix using |
| | a combination of argmax, which is suitable for multi-classes task. Defaults to False. |
| | threshold: thresholding the prediction values if multi-labels task. |
| | """ |
| | if not mutually_exclusive: |
| | return (logits >= threshold).int() |
| | if logits.shape[1] == 1: |
| | warnings.warn("single channel prediction, `mutually_exclusive=True` ignored, use threshold instead.") |
| | return (logits >= threshold).int() |
| | return logits.argmax(1, keepdim=True) |
| |
|
| |
|
| | def normalize_transform( |
| | shape, |
| | device: torch.device | str | None = None, |
| | dtype: torch.dtype | None = None, |
| | align_corners: bool = False, |
| | zero_centered: bool = False, |
| | ) -> torch.Tensor: |
| | """ |
| | Compute an affine matrix according to the input shape. |
| | The transform normalizes the homogeneous image coordinates to the |
| | range of `[-1, 1]`. Currently the following source coordinates are supported: |
| | |
| | - `align_corners=False`, `zero_centered=False`, normalizing from ``[-0.5, d-0.5]``. |
| | - `align_corners=True`, `zero_centered=False`, normalizing from ``[0, d-1]``. |
| | - `align_corners=False`, `zero_centered=True`, normalizing from ``[-(d-1)/2, (d-1)/2]``. |
| | - `align_corners=True`, `zero_centered=True`, normalizing from ``[-d/2, d/2]``. |
| | |
| | Args: |
| | shape: input spatial shape, a sequence of integers. |
| | device: device on which the returned affine will be allocated. |
| | dtype: data type of the returned affine |
| | align_corners: if True, consider -1 and 1 to refer to the centers of the |
| | corner pixels rather than the image corners. |
| | See also: https://pytorch.org/docs/stable/nn.functional.html#torch.nn.functional.grid_sample |
| | zero_centered: whether the coordinates are normalized from a zero-centered range, default to `False`. |
| | Setting this flag and `align_corners` will jointly specify the normalization source range. |
| | """ |
| | shape = convert_to_tensor(shape, torch.float64, device=device, wrap_sequence=True, track_meta=False) |
| | norm = shape.clone().detach().to(dtype=torch.float64, device=device) |
| | if align_corners: |
| | norm[norm <= 1.0] = 2.0 |
| | norm = 2.0 / (norm if zero_centered else norm - 1.0) |
| | norm = torch.diag(torch.cat((norm, torch.ones((1,), dtype=torch.float64, device=device)))) |
| | if not zero_centered: |
| | norm[:-1, -1] = -1.0 |
| | else: |
| | norm[norm <= 0.0] = 2.0 |
| | norm = 2.0 / (norm - 1.0 if zero_centered else norm) |
| | norm = torch.diag(torch.cat((norm, torch.ones((1,), dtype=torch.float64, device=device)))) |
| | if not zero_centered: |
| | norm[:-1, -1] = 1.0 / shape - 1.0 |
| | norm = norm.unsqueeze(0).to(dtype=dtype) |
| | norm.requires_grad = False |
| | return norm |
| |
|
| |
|
| | def to_norm_affine( |
| | affine: torch.Tensor, |
| | src_size: Sequence[int], |
| | dst_size: Sequence[int], |
| | align_corners: bool = False, |
| | zero_centered: bool = False, |
| | ) -> torch.Tensor: |
| | """ |
| | Given ``affine`` defined for coordinates in the pixel space, compute the corresponding affine |
| | for the normalized coordinates. |
| | |
| | Args: |
| | affine: Nxdxd batched square matrix |
| | src_size: source image spatial shape |
| | dst_size: target image spatial shape |
| | align_corners: if True, consider -1 and 1 to refer to the centers of the |
| | corner pixels rather than the image corners. |
| | See also: https://pytorch.org/docs/stable/nn.functional.html#torch.nn.functional.grid_sample |
| | zero_centered: whether the coordinates are normalized from a zero-centered range, default to `False`. |
| | See also: :py:func:`monai.networks.utils.normalize_transform`. |
| | |
| | Raises: |
| | TypeError: When ``affine`` is not a ``torch.Tensor``. |
| | ValueError: When ``affine`` is not Nxdxd. |
| | ValueError: When ``src_size`` or ``dst_size`` dimensions differ from ``affine``. |
| | |
| | """ |
| | if not isinstance(affine, torch.Tensor): |
| | raise TypeError(f"affine must be a torch.Tensor but is {type(affine).__name__}.") |
| | if affine.ndimension() != 3 or affine.shape[1] != affine.shape[2]: |
| | raise ValueError(f"affine must be Nxdxd, got {tuple(affine.shape)}.") |
| | sr = affine.shape[1] - 1 |
| | if sr != len(src_size) or sr != len(dst_size): |
| | raise ValueError(f"affine suggests {sr}D, got src={len(src_size)}D, dst={len(dst_size)}D.") |
| |
|
| | src_xform = normalize_transform(src_size, affine.device, affine.dtype, align_corners, zero_centered) |
| | dst_xform = normalize_transform(dst_size, "cpu", affine.dtype, align_corners, zero_centered) |
| | return src_xform @ affine @ convert_to_dst_type(np.linalg.inv(dst_xform.numpy()), dst=affine)[0] |
| |
|
| |
|
| | def normal_init( |
| | m, std: float = 0.02, normal_func: Callable[[torch.Tensor, float, float], Any] = torch.nn.init.normal_ |
| | ) -> None: |
| | """ |
| | Initialize the weight and bias tensors of `m' and its submodules to values from a normal distribution with a |
| | stddev of `std'. Weight tensors of convolution and linear modules are initialized with a mean of 0, batch |
| | norm modules with a mean of 1. The callable `normal_func', used to assign values, should have the same arguments |
| | as its default normal_(). This can be used with `nn.Module.apply` to visit submodules of a network. |
| | """ |
| | cname = m.__class__.__name__ |
| |
|
| | if getattr(m, "weight", None) is not None and (cname.find("Conv") != -1 or cname.find("Linear") != -1): |
| | normal_func(m.weight.data, 0.0, std) |
| | if getattr(m, "bias", None) is not None: |
| | nn.init.constant_(m.bias.data, 0.0) |
| |
|
| | elif cname.find("BatchNorm") != -1: |
| | normal_func(m.weight.data, 1.0, std) |
| | nn.init.constant_(m.bias.data, 0) |
| |
|
| |
|
| | def icnr_init(conv, upsample_factor, init=nn.init.kaiming_normal_): |
| | """ |
| | ICNR initialization for 2D/3D kernels adapted from Aitken et al.,2017 , "Checkerboard artifact free |
| | sub-pixel convolution". |
| | """ |
| | out_channels, in_channels, *dims = conv.weight.shape |
| | scale_factor = upsample_factor ** len(dims) |
| |
|
| | oc2 = int(out_channels / scale_factor) |
| |
|
| | kernel = torch.zeros([oc2, in_channels] + dims) |
| | kernel = init(kernel) |
| | kernel = kernel.transpose(0, 1) |
| | kernel = kernel.reshape(oc2, in_channels, -1) |
| | kernel = kernel.repeat(1, 1, scale_factor) |
| | kernel = kernel.reshape([in_channels, out_channels] + dims) |
| | kernel = kernel.transpose(0, 1) |
| | conv.weight.data.copy_(kernel) |
| |
|
| |
|
| | def pixelshuffle(x: torch.Tensor, spatial_dims: int, scale_factor: int) -> torch.Tensor: |
| | """ |
| | Apply pixel shuffle to the tensor `x` with spatial dimensions `spatial_dims` and scaling factor `scale_factor`. |
| | |
| | See: Shi et al., 2016, "Real-Time Single Image and Video Super-Resolution |
| | Using a nEfficient Sub-Pixel Convolutional Neural Network." |
| | |
| | See: Aitken et al., 2017, "Checkerboard artifact free sub-pixel convolution". |
| | |
| | Args: |
| | x: Input tensor |
| | spatial_dims: number of spatial dimensions, typically 2 or 3 for 2D or 3D |
| | scale_factor: factor to rescale the spatial dimensions by, must be >=1 |
| | |
| | Returns: |
| | Reshuffled version of `x`. |
| | |
| | Raises: |
| | ValueError: When input channels of `x` are not divisible by (scale_factor ** spatial_dims) |
| | """ |
| | dim, factor = spatial_dims, scale_factor |
| | input_size = list(x.size()) |
| | batch_size, channels = input_size[:2] |
| | scale_divisor = factor**dim |
| |
|
| | if channels % scale_divisor != 0: |
| | raise ValueError( |
| | f"Number of input channels ({channels}) must be evenly " |
| | f"divisible by scale_factor ** dimensions ({factor}**{dim}={scale_divisor})." |
| | ) |
| |
|
| | org_channels = int(channels // scale_divisor) |
| | output_size = [batch_size, org_channels] + [d * factor for d in input_size[2:]] |
| |
|
| | indices = list(range(2, 2 + 2 * dim)) |
| | indices = indices[dim:] + indices[:dim] |
| | permute_indices = [0, 1] |
| | for idx in range(dim): |
| | permute_indices.extend(indices[idx::dim]) |
| |
|
| | x = x.reshape([batch_size, org_channels] + [factor] * dim + input_size[2:]) |
| | x = x.permute(permute_indices).reshape(output_size) |
| | return x |
| |
|
| |
|
| | @contextmanager |
| | def eval_mode(*nets: nn.Module): |
| | """ |
| | Set network(s) to eval mode and then return to original state at the end. |
| | |
| | Args: |
| | nets: Input network(s) |
| | |
| | Examples |
| | |
| | .. code-block:: python |
| | |
| | t=torch.rand(1,1,16,16) |
| | p=torch.nn.Conv2d(1,1,3) |
| | print(p.training) # True |
| | with eval_mode(p): |
| | print(p.training) # False |
| | print(p(t).sum().backward()) # will correctly raise an exception as gradients are calculated |
| | """ |
| |
|
| | |
| | |
| | training = [n for n in nets if hasattr(n, "training") and n.training] |
| |
|
| | try: |
| | |
| | with torch.no_grad(): |
| | yield [n.eval() if hasattr(n, "eval") else n for n in nets] |
| | finally: |
| | |
| | for n in training: |
| | if hasattr(n, "train"): |
| | n.train() |
| |
|
| |
|
| | @contextmanager |
| | def train_mode(*nets: nn.Module): |
| | """ |
| | Set network(s) to train mode and then return to original state at the end. |
| | |
| | Args: |
| | nets: Input network(s) |
| | |
| | Examples |
| | |
| | .. code-block:: python |
| | |
| | t=torch.rand(1,1,16,16) |
| | p=torch.nn.Conv2d(1,1,3) |
| | p.eval() |
| | print(p.training) # False |
| | with train_mode(p): |
| | print(p.training) # True |
| | print(p(t).sum().backward()) # No exception |
| | """ |
| |
|
| | |
| | |
| | eval_list = [n for n in nets if hasattr(n, "training") and (not n.training)] |
| |
|
| | try: |
| | |
| | with torch.set_grad_enabled(True): |
| | yield [n.train() if hasattr(n, "train") else n for n in nets] |
| | finally: |
| | |
| | for n in eval_list: |
| | if hasattr(n, "eval"): |
| | n.eval() |
| |
|
| |
|
| | def get_state_dict(obj: torch.nn.Module | Mapping): |
| | """ |
| | Get the state dict of input object if has `state_dict`, otherwise, return object directly. |
| | For data parallel model, automatically convert it to regular model first. |
| | |
| | Args: |
| | obj: input object to check and get the state_dict. |
| | |
| | """ |
| | if isinstance(obj, (nn.DataParallel, nn.parallel.DistributedDataParallel)): |
| | obj = obj.module |
| | return obj.state_dict() if hasattr(obj, "state_dict") else obj |
| |
|
| |
|
| | def copy_model_state( |
| | dst: torch.nn.Module | Mapping, |
| | src: torch.nn.Module | Mapping, |
| | dst_prefix="", |
| | mapping=None, |
| | exclude_vars=None, |
| | inplace=True, |
| | filter_func=None, |
| | ): |
| | """ |
| | Compute a module state_dict, of which the keys are the same as `dst`. The values of `dst` are overwritten |
| | by the ones from `src` whenever their keys match. The method provides additional `dst_prefix` for |
| | the `dst` key when matching them. `mapping` can be a `{"src_key": "dst_key"}` dict, indicating |
| | `dst[dst_prefix + dst_key] = src[src_key]`. |
| | This function is mainly to return a model state dict |
| | for loading the `src` model state into the `dst` model, `src` and `dst` can have different dict keys, but |
| | their corresponding values normally have the same shape. |
| | |
| | Args: |
| | dst: a pytorch module or state dict to be updated. |
| | src: a pytorch module or state dict used to get the values used for the update. |
| | dst_prefix: `dst` key prefix, so that `dst[dst_prefix + src_key]` |
| | will be assigned to the value of `src[src_key]`. |
| | mapping: a `{"src_key": "dst_key"}` dict, indicating that `dst[dst_prefix + dst_key]` |
| | to be assigned to the value of `src[src_key]`. |
| | exclude_vars: a regular expression to match the `dst` variable names, |
| | so that their values are not overwritten by `src`. |
| | inplace: whether to set the `dst` module with the updated `state_dict` via `load_state_dict`. |
| | This option is only available when `dst` is a `torch.nn.Module`. |
| | filter_func: a filter function used to filter the weights to be loaded. |
| | See 'filter_swinunetr' in "monai.networks.nets.swin_unetr.py". |
| | |
| | Examples: |
| | .. code-block:: python |
| | |
| | from monai.networks.nets import BasicUNet |
| | from monai.networks.utils import copy_model_state |
| | |
| | model_a = BasicUNet(in_channels=1, out_channels=4) |
| | model_b = BasicUNet(in_channels=1, out_channels=2) |
| | model_a_b, changed, unchanged = copy_model_state( |
| | model_a, model_b, exclude_vars="conv_0.conv_0", inplace=False) |
| | # dst model updated: 76 of 82 variables. |
| | model_a.load_state_dict(model_a_b) |
| | # <All keys matched successfully> |
| | |
| | Returns: an OrderedDict of the updated `dst` state, the changed, and unchanged keys. |
| | |
| | """ |
| | src_dict = get_state_dict(src) |
| | dst_dict = OrderedDict(get_state_dict(dst)) |
| |
|
| | to_skip = {s_key for s_key in src_dict if exclude_vars and re.compile(exclude_vars).search(s_key)} |
| |
|
| | |
| | all_keys, updated_keys = list(dst_dict), list() |
| | for s, val in src_dict.items(): |
| | dst_key = f"{dst_prefix}{s}" |
| | if dst_key in dst_dict and dst_key not in to_skip and dst_dict[dst_key].shape == val.shape: |
| | dst_dict[dst_key] = val |
| | updated_keys.append(dst_key) |
| | for s in mapping if mapping else {}: |
| | dst_key = f"{dst_prefix}{mapping[s]}" |
| | if dst_key in dst_dict and dst_key not in to_skip: |
| | if dst_dict[dst_key].shape != src_dict[s].shape: |
| | warnings.warn(f"Param. shape changed from {dst_dict[dst_key].shape} to {src_dict[s].shape}.") |
| | dst_dict[dst_key] = src_dict[s] |
| | updated_keys.append(dst_key) |
| | if filter_func is not None: |
| | for key, value in src_dict.items(): |
| | new_pair = filter_func(key, value) |
| | if new_pair is not None and new_pair[0] not in to_skip: |
| | dst_dict[new_pair[0]] = new_pair[1] |
| | updated_keys.append(new_pair[0]) |
| |
|
| | updated_keys = sorted(set(updated_keys)) |
| | unchanged_keys = sorted(set(all_keys).difference(updated_keys)) |
| | logger.info(f"'dst' model updated: {len(updated_keys)} of {len(dst_dict)} variables.") |
| | if inplace and isinstance(dst, torch.nn.Module): |
| | if isinstance(dst, (nn.DataParallel, nn.parallel.DistributedDataParallel)): |
| | dst = dst.module |
| | dst.load_state_dict(dst_dict) |
| | return dst_dict, updated_keys, unchanged_keys |
| |
|
| |
|
| | def save_state(src: torch.nn.Module | dict, path: PathLike, **kwargs): |
| | """ |
| | Save the state dict of input source data with PyTorch `save`. |
| | It can save `nn.Module`, `state_dict`, a dictionary of `nn.Module` or `state_dict`. |
| | And automatically convert the data parallel module to regular module. |
| | For example:: |
| | |
| | save_state(net, path) |
| | save_state(net.state_dict(), path) |
| | save_state({"net": net, "opt": opt}, path) |
| | net_dp = torch.nn.DataParallel(net) |
| | save_state(net_dp, path) |
| | |
| | Refer to: https://pytorch.org/ignite/v0.4.8/generated/ignite.handlers.DiskSaver.html. |
| | |
| | Args: |
| | src: input data to save, can be `nn.Module`, `state_dict`, a dictionary of `nn.Module` or `state_dict`. |
| | path: target file path to save the input object. |
| | kwargs: other args for the `save_obj` except for the `obj` and `path`. |
| | default `func` is `torch.save()`, details of the args: |
| | https://pytorch.org/docs/stable/generated/torch.save.html. |
| | |
| | """ |
| |
|
| | ckpt: dict = {} |
| | if isinstance(src, dict): |
| | for k, v in src.items(): |
| | ckpt[k] = get_state_dict(v) |
| | else: |
| | ckpt = get_state_dict(src) |
| |
|
| | save_obj(obj=ckpt, path=path, **kwargs) |
| |
|
| |
|
| | def convert_to_onnx( |
| | model: nn.Module, |
| | inputs: Sequence[Any], |
| | input_names: Sequence[str] | None = None, |
| | output_names: Sequence[str] | None = None, |
| | opset_version: int | None = None, |
| | dynamic_axes: Mapping[str, Mapping[int, str]] | Mapping[str, Sequence[int]] | None = None, |
| | filename: Any | None = None, |
| | verify: bool = False, |
| | device: torch.device | None = None, |
| | use_ort: bool = False, |
| | ort_provider: Sequence[str] | None = None, |
| | rtol: float = 1e-4, |
| | atol: float = 0.0, |
| | use_trace: bool = True, |
| | **kwargs, |
| | ): |
| | """ |
| | Utility to convert a model into ONNX model and optionally verify with ONNX or onnxruntime. |
| | See also: https://pytorch.org/docs/stable/onnx.html for how to convert a PyTorch model to ONNX. |
| | |
| | Args: |
| | model: source PyTorch model to save. |
| | inputs: input sample data used by pytorch.onnx.export. It is also used in ONNX model verification. |
| | input_names: optional input names of the ONNX model. |
| | output_names: optional output names of the ONNX model. |
| | opset_version: version of the (ai.onnx) opset to target. Must be >= 7 and not exceed |
| | the latest opset version supported by PyTorch, for more details: |
| | https://github.com/onnx/onnx/blob/main/docs/Operators.md and |
| | https://github.com/pytorch/pytorch/blob/master/torch/onnx/_constants.py |
| | dynamic_axes: specifies axes of tensors as dynamic (i.e. known only at run-time). If set to None, |
| | the exported model will have the shapes of all input and output tensors set to match given |
| | ones, for more details: https://pytorch.org/docs/stable/onnx.html#torch.onnx.export. |
| | filename: optional filename to save the ONNX model, if None, don't save the ONNX model. |
| | verify: whether to verify the ONNX model with ONNX or onnxruntime. |
| | device: target PyTorch device to verify the model, if None, use CUDA if available. |
| | use_ort: whether to use onnxruntime to verify the model. |
| | ort_provider": onnxruntime provider to use, default is ["CPUExecutionProvider"]. |
| | rtol: the relative tolerance when comparing the outputs of PyTorch model and TorchScript model. |
| | atol: the absolute tolerance when comparing the outputs of PyTorch model and TorchScript model. |
| | use_trace: whether to use `torch.jit.trace` to export the torchscript model. |
| | kwargs: other arguments except `obj` for `torch.jit.script()` to convert model, for more details: |
| | https://pytorch.org/docs/master/generated/torch.jit.script.html. |
| | |
| | """ |
| | model.eval() |
| | with torch.no_grad(): |
| | torch_versioned_kwargs = {} |
| | if use_trace: |
| | |
| | mode_to_export = model |
| | else: |
| | if not pytorch_after(1, 10): |
| | if "example_outputs" not in kwargs: |
| | |
| | raise TypeError( |
| | "example_outputs is required in scripting mode before PyTorch 1.10." |
| | "Please provide example outputs or use trace mode to export onnx model." |
| | ) |
| | torch_versioned_kwargs["example_outputs"] = kwargs["example_outputs"] |
| | del kwargs["example_outputs"] |
| | mode_to_export = torch.jit.script(model, **kwargs) |
| |
|
| | if filename is None: |
| | f = io.BytesIO() |
| | torch.onnx.export( |
| | mode_to_export, |
| | tuple(inputs), |
| | f=f, |
| | input_names=input_names, |
| | output_names=output_names, |
| | dynamic_axes=dynamic_axes, |
| | opset_version=opset_version, |
| | **torch_versioned_kwargs, |
| | ) |
| | onnx_model = onnx.load_model_from_string(f.getvalue()) |
| | else: |
| | torch.onnx.export( |
| | mode_to_export, |
| | tuple(inputs), |
| | f=filename, |
| | input_names=input_names, |
| | output_names=output_names, |
| | dynamic_axes=dynamic_axes, |
| | opset_version=opset_version, |
| | **torch_versioned_kwargs, |
| | ) |
| | onnx_model = onnx.load(filename) |
| |
|
| | if verify: |
| | if device is None: |
| | device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| |
|
| | inputs = [i.to(device) if isinstance(i, torch.Tensor) else i for i in inputs] |
| | model = model.to(device) |
| |
|
| | with torch.no_grad(): |
| | set_determinism(seed=0) |
| | torch_out = ensure_tuple(model(*inputs), True) |
| |
|
| | set_determinism(seed=0) |
| | model_input_names = [i.name for i in onnx_model.graph.input] |
| | input_dict = dict(zip(model_input_names, [i.cpu().numpy() for i in inputs])) |
| | if use_ort: |
| | ort_sess = onnxruntime.InferenceSession( |
| | onnx_model.SerializeToString(), providers=ort_provider if ort_provider else ["CPUExecutionProvider"] |
| | ) |
| | onnx_out = ort_sess.run(None, input_dict) |
| | else: |
| | sess = onnxreference.ReferenceEvaluator(onnx_model) |
| | onnx_out = sess.run(None, input_dict) |
| | set_determinism(seed=None) |
| | |
| | for r1, r2 in zip(torch_out, onnx_out): |
| | if isinstance(r1, torch.Tensor): |
| | assert_fn = torch.testing.assert_close if pytorch_after(1, 11) else torch.testing.assert_allclose |
| | assert_fn(r1.cpu(), convert_to_tensor(r2, dtype=r1.dtype), rtol=rtol, atol=atol) |
| |
|
| | return onnx_model |
| |
|
| |
|
| | def convert_to_torchscript( |
| | model: nn.Module, |
| | filename_or_obj: Any | None = None, |
| | extra_files: dict | None = None, |
| | verify: bool = False, |
| | inputs: Sequence[Any] | None = None, |
| | device: torch.device | None = None, |
| | rtol: float = 1e-4, |
| | atol: float = 0.0, |
| | use_trace: bool = False, |
| | **kwargs, |
| | ): |
| | """ |
| | Utility to convert a model into TorchScript model and save to file, |
| | with optional input / output data verification. |
| | |
| | Args: |
| | model: source PyTorch model to save. |
| | filename_or_obj: if not None, specify a file-like object (has to implement write and flush) |
| | or a string containing a file path name to save the TorchScript model. |
| | extra_files: map from filename to contents which will be stored as part of the save model file. |
| | for more details: https://pytorch.org/docs/stable/generated/torch.jit.save.html. |
| | verify: whether to verify the input and output of TorchScript model. |
| | if `filename_or_obj` is not None, load the saved TorchScript model and verify. |
| | inputs: input test data to verify model, should be a sequence of data, every item maps to a argument |
| | of `model()` function. |
| | device: target device to verify the model, if None, use CUDA if available. |
| | rtol: the relative tolerance when comparing the outputs of PyTorch model and TorchScript model. |
| | atol: the absolute tolerance when comparing the outputs of PyTorch model and TorchScript model. |
| | use_trace: whether to use `torch.jit.trace` to export the TorchScript model. |
| | kwargs: other arguments except `obj` for `torch.jit.script()` or `torch.jit.trace()` (if use_trace is True) |
| | to convert model, for more details: https://pytorch.org/docs/master/generated/torch.jit.script.html. |
| | |
| | """ |
| | model.eval() |
| | with torch.no_grad(): |
| | if use_trace: |
| | if inputs is None: |
| | raise ValueError("Missing input data for tracing convert.") |
| | script_module = torch.jit.trace(model, example_inputs=inputs, **kwargs) |
| | else: |
| | script_module = torch.jit.script(model, **kwargs) |
| | if filename_or_obj is not None: |
| | torch.jit.save(m=script_module, f=filename_or_obj, _extra_files=extra_files) |
| |
|
| | if verify: |
| | if device is None: |
| | device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| | if inputs is None: |
| | raise ValueError("Missing input data for verification.") |
| |
|
| | inputs = [i.to(device) if isinstance(i, torch.Tensor) else i for i in inputs] |
| | ts_model = torch.jit.load(filename_or_obj) if filename_or_obj is not None else script_module |
| | ts_model.eval().to(device) |
| | model = model.to(device) |
| |
|
| | with torch.no_grad(): |
| | set_determinism(seed=0) |
| | torch_out = ensure_tuple(model(*inputs)) |
| | set_determinism(seed=0) |
| | torchscript_out = ensure_tuple(ts_model(*inputs)) |
| | set_determinism(seed=None) |
| | |
| | for r1, r2 in zip(torch_out, torchscript_out): |
| | if isinstance(r1, torch.Tensor) or isinstance(r2, torch.Tensor): |
| | assert_fn = torch.testing.assert_close if pytorch_after(1, 11) else torch.testing.assert_allclose |
| | assert_fn(r1, r2, rtol=rtol, atol=atol) |
| |
|
| | return script_module |
| |
|
| |
|
| | def _onnx_trt_compile( |
| | onnx_model, |
| | min_shape: Sequence[int], |
| | opt_shape: Sequence[int], |
| | max_shape: Sequence[int], |
| | device: int, |
| | precision: str, |
| | input_names: Sequence[str] | None, |
| | output_names: Sequence[str] | None, |
| | ): |
| | """ |
| | This function takes an ONNX model as input, exports it to a TensorRT engine, wraps the TensorRT engine |
| | to a TensorRT engine-based TorchScript model and return the TorchScript model. |
| | |
| | Args: |
| | onnx_model: the source ONNX model to compile. |
| | min_shape: the minimum input shape of the converted TensorRT model. |
| | opt_shape: the optimization input shape of the model, on which the TensorRT optimizes. |
| | max_shape: the maximum input shape of the converted TensorRT model. |
| | device: the target GPU index to convert and verify the model. |
| | precision: the weight precision of the converted TensorRT engine-based TorchScript model. |
| | Should be 'fp32' or 'fp16'. |
| | input_names: optional input names of the ONNX model. Should be a sequence like |
| | `['input_0', 'input_1', ..., 'input_N']` where N equals to the number of the |
| | model inputs. |
| | output_names: optional output names of the ONNX model. Should be a sequence like |
| | `['output_0', 'output_1', ..., 'output_N']` where N equals to the number of |
| | the model outputs. |
| | |
| | """ |
| | trt, _ = optional_import("tensorrt", "8.5.3") |
| | torch_tensorrt, _ = optional_import("torch_tensorrt", "1.4.0") |
| |
|
| | input_shapes = (min_shape, opt_shape, max_shape) |
| | |
| | input_names = [] if not input_names else input_names |
| | output_names = [] if not output_names else output_names |
| |
|
| | |
| | torch_tensorrt.set_device(device) |
| | logger = trt.Logger(trt.Logger.WARNING) |
| | builder = trt.Builder(logger) |
| | network = builder.create_network(1 << int(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH)) |
| | profile = builder.create_optimization_profile() |
| | if input_names: |
| | profile.set_shape(input_names[0], *input_shapes) |
| |
|
| | |
| | parser = trt.OnnxParser(network, logger) |
| | success = parser.parse(onnx_model.SerializeToString()) |
| | if not success: |
| | parser_error_message = "" |
| | for idx in range(parser.num_errors): |
| | parser_error_message += parser.get_error(idx).desc() + "\n" |
| | raise Exception(f"TensorRT cannot parse the ONNX model, due to:\n{parser_error_message}") |
| |
|
| | |
| | config = builder.create_builder_config() |
| | config.add_optimization_profile(profile) |
| | if precision == "fp16": |
| | config.set_flag(trt.BuilderFlag.FP16) |
| | serialized_engine = builder.build_serialized_network(network, config) |
| | f = io.BytesIO() |
| | f.write(serialized_engine) |
| |
|
| | |
| | trt_model = torch_tensorrt.ts.embed_engine_in_new_module( |
| | f.getvalue(), |
| | device=torch.device(f"cuda:{device}"), |
| | input_binding_names=input_names, |
| | output_binding_names=output_names, |
| | ) |
| | return trt_model |
| |
|
| |
|
| | def convert_to_trt( |
| | model: nn.Module, |
| | precision: str, |
| | input_shape: Sequence[int], |
| | dynamic_batchsize: Sequence[int] | None = None, |
| | use_trace: bool = False, |
| | filename_or_obj: Any | None = None, |
| | verify: bool = False, |
| | device: int | None = None, |
| | use_onnx: bool | None = False, |
| | onnx_input_names: Sequence[str] | None = ("input_0",), |
| | onnx_output_names: Sequence[str] | None = ("output_0",), |
| | rtol: float = 1e-2, |
| | atol: float = 0.0, |
| | **kwargs, |
| | ): |
| | """ |
| | Utility to export a model into a TensorRT engine-based TorchScript model with optional input / output data verification. |
| | |
| | There are two ways to export a model: |
| | 1, Torch-TensorRT way: PyTorch module ---> TorchScript module ---> TensorRT engine-based TorchScript. |
| | 2, ONNX-TensorRT way: PyTorch module ---> TorchScript module ---> ONNX model ---> TensorRT engine ---> |
| | TensorRT engine-based TorchScript. |
| | |
| | When exporting through the first way, some models suffer from the slowdown problem, since Torch-TensorRT |
| | may only convert a little part of the PyTorch model to the TensorRT engine. However when exporting through |
| | the second way, some Python data structures like `dict` are not supported. And some TorchScript models are |
| | not supported by the ONNX if exported through `torch.jit.script`. |
| | |
| | Args: |
| | model: a source PyTorch model to convert. |
| | precision: the weight precision of the converted TensorRT engine based TorchScript models. Should be 'fp32' or 'fp16'. |
| | input_shape: the input shape that is used to convert the model. Should be a list like [N, C, H, W] or |
| | [N, C, H, W, D]. |
| | dynamic_batchsize: a sequence with three elements to define the batch size range of the input for the model to be |
| | converted. Should be a sequence like [MIN_BATCH, OPT_BATCH, MAX_BATCH]. After converted, the batchsize of model |
| | input should between `MIN_BATCH` and `MAX_BATCH` and the `OPT_BATCH` is the best performance batchsize that the |
| | TensorRT tries to fit. The `OPT_BATCH` should be the most frequently used input batchsize in the application, |
| | default to None. |
| | use_trace: whether using `torch.jit.trace` to convert the PyTorch model to a TorchScript model and then convert to |
| | a TensorRT engine based TorchScript model or an ONNX model (if `use_onnx` is True), default to False. |
| | filename_or_obj: if not None, specify a file-like object (has to implement write and flush) or a string containing a |
| | file path name to load the TensorRT engine based TorchScript model for verifying. |
| | verify: whether to verify the input and output of the TensorRT engine based TorchScript model. |
| | device: the target GPU index to convert and verify the model. If None, use #0 GPU. |
| | use_onnx: whether to use the ONNX-TensorRT way to export the TensorRT engine-based TorchScript model. |
| | onnx_input_names: optional input names of the ONNX model. This arg is only useful when `use_onnx` is True. Should be |
| | a sequence like `('input_0', 'input_1', ..., 'input_N')` where N equals to the number of the model inputs. If not |
| | given, will use `('input_0',)`, which supposes the model only has one input. |
| | onnx_output_names: optional output names of the ONNX model. This arg is only useful when `use_onnx` is True. Should be |
| | a sequence like `('output_0', 'output_1', ..., 'output_N')` where N equals to the number of the model outputs. If |
| | not given, will use `('output_0',)`, which supposes the model only has one output. |
| | rtol: the relative tolerance when comparing the outputs between the PyTorch model and TensorRT model. |
| | atol: the absolute tolerance when comparing the outputs between the PyTorch model and TensorRT model. |
| | kwargs: other arguments except `module`, `inputs`, `enabled_precisions` and `device` for `torch_tensorrt.compile()` |
| | to compile model, for more details: https://pytorch.org/TensorRT/py_api/torch_tensorrt.html#torch-tensorrt-py. |
| | """ |
| |
|
| | torch_tensorrt, _ = optional_import("torch_tensorrt", version="1.4.0") |
| |
|
| | if not torch.cuda.is_available(): |
| | raise Exception("Cannot find any GPU devices.") |
| |
|
| | if not input_shape: |
| | raise ValueError("Missing the input shape for model convert.") |
| |
|
| | if not dynamic_batchsize: |
| | warnings.warn(f"There is no dynamic batch range. The converted model only takes {input_shape} shape input.") |
| |
|
| | if (dynamic_batchsize is not None) and (len(dynamic_batchsize) != 3): |
| | warnings.warn(f"The dynamic batch range sequence should have 3 elements, but got {dynamic_batchsize} elements.") |
| |
|
| | device = device if device else 0 |
| | target_device = torch.device(f"cuda:{device}") if device else torch.device("cuda:0") |
| | convert_precision = torch.float32 if precision == "fp32" else torch.half |
| | inputs = [torch.rand(ensure_tuple(input_shape)).to(target_device)] |
| |
|
| | def scale_batch_size(input_shape: Sequence[int], scale_num: int): |
| | scale_shape = [*input_shape] |
| | scale_shape[0] *= scale_num |
| | return scale_shape |
| |
|
| | |
| | if dynamic_batchsize: |
| | min_input_shape = scale_batch_size(input_shape, dynamic_batchsize[0]) |
| | opt_input_shape = scale_batch_size(input_shape, dynamic_batchsize[1]) |
| | max_input_shape = scale_batch_size(input_shape, dynamic_batchsize[2]) |
| | else: |
| | min_input_shape = opt_input_shape = max_input_shape = input_shape |
| |
|
| | |
| | model = model.eval().to(target_device) |
| | ir_model = convert_to_torchscript(model, device=target_device, inputs=inputs, use_trace=use_trace) |
| | ir_model.eval() |
| |
|
| | if use_onnx: |
| | |
| | dynamic_axes = {k: {0: "batchsize"} for k in onnx_input_names} if onnx_input_names else {} |
| | dynamic_axes.update({k: {0: "batchsize"} for k in onnx_output_names} if onnx_output_names else {}) |
| | ir_model = convert_to_onnx( |
| | model, inputs, onnx_input_names, onnx_output_names, use_trace=use_trace, dynamic_axes=dynamic_axes |
| | ) |
| |
|
| | |
| | trt_model = _onnx_trt_compile( |
| | ir_model, |
| | min_shape=min_input_shape, |
| | opt_shape=opt_input_shape, |
| | max_shape=max_input_shape, |
| | device=device, |
| | precision=precision, |
| | input_names=onnx_input_names, |
| | output_names=onnx_output_names, |
| | ) |
| | else: |
| | |
| | ir_model.to(target_device) |
| | with torch.no_grad(): |
| | with torch.cuda.device(device=device): |
| | input_placeholder = [ |
| | torch_tensorrt.Input( |
| | min_shape=min_input_shape, opt_shape=opt_input_shape, max_shape=max_input_shape |
| | ) |
| | ] |
| | trt_model = torch_tensorrt.compile( |
| | ir_model, |
| | inputs=input_placeholder, |
| | enabled_precisions=convert_precision, |
| | device=target_device, |
| | ir="torchscript", |
| | **kwargs, |
| | ) |
| |
|
| | |
| | if verify: |
| | if inputs is None: |
| | raise ValueError("Missing input data for verification.") |
| |
|
| | trt_model = torch.jit.load(filename_or_obj) if filename_or_obj is not None else trt_model |
| |
|
| | with torch.no_grad(): |
| | set_determinism(seed=0) |
| | torch_out = ensure_tuple(model(*inputs)) |
| | set_determinism(seed=0) |
| | trt_out = ensure_tuple(trt_model(*inputs)) |
| | set_determinism(seed=None) |
| | |
| | for r1, r2 in zip(torch_out, trt_out): |
| | if isinstance(r1, torch.Tensor) or isinstance(r2, torch.Tensor): |
| | assert_fn = torch.testing.assert_close if pytorch_after(1, 11) else torch.testing.assert_allclose |
| | assert_fn(r1, r2, rtol=rtol, atol=atol) |
| |
|
| | return trt_model |
| |
|
| |
|
| | def meshgrid_ij(*tensors): |
| | if torch.meshgrid.__kwdefaults__ is not None and "indexing" in torch.meshgrid.__kwdefaults__: |
| | return torch.meshgrid(*tensors, indexing="ij") |
| |
|
| | return torch.meshgrid(*tensors) |
| |
|
| |
|
| | def meshgrid_xy(*tensors): |
| | if torch.meshgrid.__kwdefaults__ is not None and "indexing" in torch.meshgrid.__kwdefaults__: |
| | return torch.meshgrid(*tensors, indexing="xy") |
| |
|
| | return torch.meshgrid(tensors[1], tensors[0], *tensors[2:]) |
| |
|
| |
|
| | def _replace_modules( |
| | parent: torch.nn.Module, |
| | name: str, |
| | new_module: torch.nn.Module, |
| | out: list[tuple[str, torch.nn.Module]], |
| | strict_match: bool = True, |
| | match_device: bool = True, |
| | ) -> None: |
| | """ |
| | Helper function for :py:class:`monai.networks.utils.replace_modules`. |
| | """ |
| | if match_device: |
| | devices = list({i.device for i in parent.parameters()}) |
| | |
| | if len(devices) == 1: |
| | new_module.to(devices[0]) |
| | idx = name.find(".") |
| | |
| | if idx != -1: |
| | parent_name = name[:idx] |
| | parent = getattr(parent, parent_name) |
| | name = name[idx + 1 :] |
| | _out: list[tuple[str, torch.nn.Module]] = [] |
| | _replace_modules(parent, name, new_module, _out) |
| | |
| | out += [(f"{parent_name}.{r[0]}", r[1]) for r in _out] |
| | |
| | else: |
| | if strict_match: |
| | old_module = getattr(parent, name) |
| | setattr(parent, name, new_module) |
| | out += [(name, old_module)] |
| | else: |
| | for mod_name, _ in parent.named_modules(): |
| | if name in mod_name: |
| | _replace_modules(parent, mod_name, deepcopy(new_module), out, strict_match=True) |
| |
|
| |
|
| | def replace_modules( |
| | parent: torch.nn.Module, |
| | name: str, |
| | new_module: torch.nn.Module, |
| | strict_match: bool = True, |
| | match_device: bool = True, |
| | ) -> list[tuple[str, torch.nn.Module]]: |
| | """ |
| | Replace sub-module(s) in a parent module. |
| | |
| | The name of the module to be replace can be nested e.g., |
| | `features.denseblock1.denselayer1.layers.relu1`. If this is the case (there are "." |
| | in the module name), then this function will recursively call itself. |
| | |
| | Args: |
| | parent: module that contains the module to be replaced |
| | name: name of module to be replaced. Can include ".". |
| | new_module: `torch.nn.Module` to be placed at position `name` inside `parent`. This will |
| | be deep copied if `strict_match == False` multiple instances are independent. |
| | strict_match: if `True`, module name must `== name`. If false then |
| | `name in named_modules()` will be used. `True` can be used to change just |
| | one module, whereas `False` can be used to replace all modules with similar |
| | name (e.g., `relu`). |
| | match_device: if `True`, the device of the new module will match the model. Requires all |
| | of `parent` to be on the same device. |
| | |
| | Returns: |
| | List of tuples of replaced modules. Element 0 is module name, element 1 is the replaced module. |
| | |
| | Raises: |
| | AttributeError: if `strict_match` is `True` and `name` is not a named module in `parent`. |
| | """ |
| | out: list[tuple[str, torch.nn.Module]] = [] |
| | _replace_modules(parent, name, new_module, out, strict_match, match_device) |
| | return out |
| |
|
| |
|
| | @contextmanager |
| | def replace_modules_temp( |
| | parent: torch.nn.Module, |
| | name: str, |
| | new_module: torch.nn.Module, |
| | strict_match: bool = True, |
| | match_device: bool = True, |
| | ): |
| | """ |
| | Temporarily replace sub-module(s) in a parent module (context manager). |
| | |
| | See :py:class:`monai.networks.utils.replace_modules`. |
| | """ |
| | replaced: list[tuple[str, torch.nn.Module]] = [] |
| | try: |
| | |
| | _replace_modules(parent, name, new_module, replaced, strict_match, match_device) |
| | yield |
| | finally: |
| | |
| | for name, module in replaced: |
| | _replace_modules(parent, name, module, [], strict_match=True, match_device=match_device) |
| |
|
| |
|
| | def freeze_layers(model: nn.Module, freeze_vars=None, exclude_vars=None): |
| | """ |
| | A utilty function to help freeze specific layers. |
| | |
| | Args: |
| | model: a source PyTorch model to freeze layer. |
| | freeze_vars: a regular expression to match the `model` variable names, |
| | so that their `requires_grad` will set to `False`. |
| | exclude_vars: a regular expression to match the `model` variable names, |
| | except for matched variable names, other `requires_grad` will set to `False`. |
| | |
| | Raises: |
| | ValueError: when freeze_vars and exclude_vars are both specified. |
| | |
| | """ |
| | if freeze_vars is not None and exclude_vars is not None: |
| | raise ValueError("Incompatible values: freeze_vars and exclude_vars are both specified.") |
| | src_dict = get_state_dict(model) |
| |
|
| | frozen_keys = list() |
| | if freeze_vars is not None: |
| | to_freeze = {s_key for s_key in src_dict if freeze_vars and re.compile(freeze_vars).search(s_key)} |
| | for name, param in model.named_parameters(): |
| | if name in to_freeze: |
| | param.requires_grad = False |
| | frozen_keys.append(name) |
| | elif not param.requires_grad: |
| | param.requires_grad = True |
| | warnings.warn( |
| | f"The freeze_vars does not include {param}, but requires_grad is False, change it to True." |
| | ) |
| | if exclude_vars is not None: |
| | to_exclude = {s_key for s_key in src_dict if exclude_vars and re.compile(exclude_vars).search(s_key)} |
| | for name, param in model.named_parameters(): |
| | if name not in to_exclude: |
| | param.requires_grad = False |
| | frozen_keys.append(name) |
| | elif not param.requires_grad: |
| | param.requires_grad = True |
| | warnings.warn(f"The exclude_vars includes {param}, but requires_grad is False, change it to True.") |
| |
|
| | logger.info(f"{len(frozen_keys)} of {len(src_dict)} variables frozen.") |
| |
|