| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | from collections.abc import Sequence |
| |
|
| | import torch |
| | import torch.nn as nn |
| |
|
| | from monai.networks.layers.factories import Act, Norm, split_args |
| | from monai.networks.nets.regressor import Regressor |
| |
|
| | __all__ = ["Classifier", "Discriminator", "Critic"] |
| |
|
| |
|
| | class Classifier(Regressor): |
| | """ |
| | Defines a classification network from Regressor by specifying the output shape as a single dimensional tensor |
| | with size equal to the number of classes to predict. The final activation function can also be specified, eg. |
| | softmax or sigmoid. |
| | |
| | Args: |
| | in_shape: tuple of integers stating the dimension of the input tensor (minus batch dimension) |
| | classes: integer stating the dimension of the final output tensor |
| | channels: tuple of integers stating the output channels of each convolutional layer |
| | strides: tuple of integers stating the stride (downscale factor) of each convolutional layer |
| | kernel_size: integer or tuple of integers stating size of convolutional kernels |
| | num_res_units: integer stating number of convolutions in residual units, 0 means no residual units |
| | act: name or type defining activation layers |
| | norm: name or type defining normalization layers |
| | dropout: optional float value in range [0, 1] stating dropout probability for layers, None for no dropout |
| | bias: boolean stating if convolution layers should have a bias component |
| | last_act: name defining the last activation layer |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | in_shape: Sequence[int], |
| | classes: int, |
| | channels: Sequence[int], |
| | strides: Sequence[int], |
| | kernel_size: Sequence[int] | int = 3, |
| | num_res_units: int = 2, |
| | act=Act.PRELU, |
| | norm=Norm.INSTANCE, |
| | dropout: float | None = None, |
| | bias: bool = True, |
| | last_act: str | None = None, |
| | ) -> None: |
| | super().__init__(in_shape, (classes,), channels, strides, kernel_size, num_res_units, act, norm, dropout, bias) |
| |
|
| | if last_act is not None: |
| | last_act_name, last_act_args = split_args(last_act) |
| | last_act_type = Act[last_act_name] |
| |
|
| | self.final.add_module("lastact", last_act_type(**last_act_args)) |
| |
|
| |
|
| | class Discriminator(Classifier): |
| | """ |
| | Defines a discriminator network from Classifier with a single output value and sigmoid activation by default. This |
| | is meant for use with GANs or other applications requiring a generic discriminator network. |
| | |
| | Args: |
| | in_shape: tuple of integers stating the dimension of the input tensor (minus batch dimension) |
| | channels: tuple of integers stating the output channels of each convolutional layer |
| | strides: tuple of integers stating the stride (downscale factor) of each convolutional layer |
| | kernel_size: integer or tuple of integers stating size of convolutional kernels |
| | num_res_units: integer stating number of convolutions in residual units, 0 means no residual units |
| | act: name or type defining activation layers |
| | norm: name or type defining normalization layers |
| | dropout: optional float value in range [0, 1] stating dropout probability for layers, None for no dropout |
| | bias: boolean stating if convolution layers should have a bias component |
| | last_act: name defining the last activation layer |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | in_shape: Sequence[int], |
| | channels: Sequence[int], |
| | strides: Sequence[int], |
| | kernel_size: Sequence[int] | int = 3, |
| | num_res_units: int = 2, |
| | act=Act.PRELU, |
| | norm=Norm.INSTANCE, |
| | dropout: float | None = 0.25, |
| | bias: bool = True, |
| | last_act=Act.SIGMOID, |
| | ) -> None: |
| | super().__init__(in_shape, 1, channels, strides, kernel_size, num_res_units, act, norm, dropout, bias, last_act) |
| |
|
| |
|
| | class Critic(Classifier): |
| | """ |
| | Defines a critic network from Classifier with a single output value and no final activation. The final layer is |
| | `nn.Flatten` instead of `nn.Linear`, the final result is computed as the mean over the first dimension. This is |
| | meant to be used with Wasserstein GANs. |
| | |
| | Args: |
| | in_shape: tuple of integers stating the dimension of the input tensor (minus batch dimension) |
| | channels: tuple of integers stating the output channels of each convolutional layer |
| | strides: tuple of integers stating the stride (downscale factor) of each convolutional layer |
| | kernel_size: integer or tuple of integers stating size of convolutional kernels |
| | num_res_units: integer stating number of convolutions in residual units, 0 means no residual units |
| | act: name or type defining activation layers |
| | norm: name or type defining normalization layers |
| | dropout: optional float value in range [0, 1] stating dropout probability for layers, None for no dropout |
| | bias: boolean stating if convolution layers should have a bias component |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | in_shape: Sequence[int], |
| | channels: Sequence[int], |
| | strides: Sequence[int], |
| | kernel_size: Sequence[int] | int = 3, |
| | num_res_units: int = 2, |
| | act=Act.PRELU, |
| | norm=Norm.INSTANCE, |
| | dropout: float | None = 0.25, |
| | bias: bool = True, |
| | ) -> None: |
| | super().__init__(in_shape, 1, channels, strides, kernel_size, num_res_units, act, norm, dropout, bias, None) |
| |
|
| | def _get_final_layer(self, in_shape: Sequence[int]): |
| | return nn.Flatten() |
| |
|
| | def forward(self, x: torch.Tensor) -> torch.Tensor: |
| | x = self.net(x) |
| | x = self.final(x) |
| | x = x.mean(1) |
| | return x.view((x.shape[0], -1)) |
| |
|