| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | A collection of "vanilla" transforms for box operations |
| | https://github.com/Project-MONAI/MONAI/wiki/MONAI_Design |
| | """ |
| |
|
| | from __future__ import annotations |
| |
|
| | from typing import Any, Sequence |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai.config.type_definitions import DtypeLike, NdarrayOrTensor, NdarrayTensor |
| | from monai.data.box_utils import ( |
| | BoxMode, |
| | clip_boxes_to_image, |
| | convert_box_mode, |
| | convert_box_to_standard_mode, |
| | get_spatial_dims, |
| | spatial_crop_boxes, |
| | standardize_empty_box, |
| | ) |
| | from monai.transforms import Rotate90, SpatialCrop |
| | from monai.transforms.transform import Transform |
| | from monai.utils import ensure_tuple, ensure_tuple_rep, fall_back_tuple, look_up_option |
| | from monai.utils.enums import TransformBackends |
| |
|
| | from .box_ops import ( |
| | apply_affine_to_boxes, |
| | convert_box_to_mask, |
| | convert_mask_to_box, |
| | flip_boxes, |
| | resize_boxes, |
| | rot90_boxes, |
| | select_labels, |
| | zoom_boxes, |
| | ) |
| |
|
| | __all__ = [ |
| | "StandardizeEmptyBox", |
| | "ConvertBoxToStandardMode", |
| | "ConvertBoxMode", |
| | "AffineBox", |
| | "ZoomBox", |
| | "ResizeBox", |
| | "FlipBox", |
| | "ClipBoxToImage", |
| | "BoxToMask", |
| | "MaskToBox", |
| | "SpatialCropBox", |
| | "RotateBox90", |
| | ] |
| |
|
| |
|
| | class StandardizeEmptyBox(Transform): |
| | """ |
| | When boxes are empty, this transform standardize it to shape of (0,4) or (0,6). |
| | |
| | Args: |
| | spatial_dims: number of spatial dimensions of the bounding boxes. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, spatial_dims: int) -> None: |
| | self.spatial_dims = spatial_dims |
| |
|
| | def __call__(self, boxes: NdarrayOrTensor) -> NdarrayOrTensor: |
| | """ |
| | Args: |
| | boxes: source bounding boxes, Nx4 or Nx6 or 0xM torch tensor or ndarray. |
| | """ |
| | return standardize_empty_box(boxes, spatial_dims=self.spatial_dims) |
| |
|
| |
|
| | class ConvertBoxMode(Transform): |
| | """ |
| | This transform converts the boxes in src_mode to the dst_mode. |
| | |
| | Args: |
| | src_mode: source box mode. If it is not given, this func will assume it is ``StandardMode()``. |
| | dst_mode: target box mode. If it is not given, this func will assume it is ``StandardMode()``. |
| | |
| | Note: |
| | ``StandardMode`` = :class:`~monai.data.box_utils.CornerCornerModeTypeA`, |
| | also represented as "xyxy" for 2D and "xyzxyz" for 3D. |
| | |
| | src_mode and dst_mode can be: |
| | #. str: choose from :class:`~monai.utils.enums.BoxModeName`, for example, |
| | - "xyxy": boxes has format [xmin, ymin, xmax, ymax] |
| | - "xyzxyz": boxes has format [xmin, ymin, zmin, xmax, ymax, zmax] |
| | - "xxyy": boxes has format [xmin, xmax, ymin, ymax] |
| | - "xxyyzz": boxes has format [xmin, xmax, ymin, ymax, zmin, zmax] |
| | - "xyxyzz": boxes has format [xmin, ymin, xmax, ymax, zmin, zmax] |
| | - "xywh": boxes has format [xmin, ymin, xsize, ysize] |
| | - "xyzwhd": boxes has format [xmin, ymin, zmin, xsize, ysize, zsize] |
| | - "ccwh": boxes has format [xcenter, ycenter, xsize, ysize] |
| | - "cccwhd": boxes has format [xcenter, ycenter, zcenter, xsize, ysize, zsize] |
| | #. BoxMode class: choose from the subclasses of :class:`~monai.data.box_utils.BoxMode`, for example, |
| | - CornerCornerModeTypeA: equivalent to "xyxy" or "xyzxyz" |
| | - CornerCornerModeTypeB: equivalent to "xxyy" or "xxyyzz" |
| | - CornerCornerModeTypeC: equivalent to "xyxy" or "xyxyzz" |
| | - CornerSizeMode: equivalent to "xywh" or "xyzwhd" |
| | - CenterSizeMode: equivalent to "ccwh" or "cccwhd" |
| | #. BoxMode object: choose from the subclasses of :class:`~monai.data.box_utils.BoxMode`, for example, |
| | - CornerCornerModeTypeA(): equivalent to "xyxy" or "xyzxyz" |
| | - CornerCornerModeTypeB(): equivalent to "xxyy" or "xxyyzz" |
| | - CornerCornerModeTypeC(): equivalent to "xyxy" or "xyxyzz" |
| | - CornerSizeMode(): equivalent to "xywh" or "xyzwhd" |
| | - CenterSizeMode(): equivalent to "ccwh" or "cccwhd" |
| | #. None: will assume mode is ``StandardMode()`` |
| | |
| | Example: |
| | .. code-block:: python |
| | |
| | boxes = torch.ones(10,4) |
| | # convert boxes with format [xmin, ymin, xmax, ymax] to [xcenter, ycenter, xsize, ysize]. |
| | box_converter = ConvertBoxMode(src_mode="xyxy", dst_mode="ccwh") |
| | box_converter(boxes) |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | src_mode: str | BoxMode | type[BoxMode] | None = None, |
| | dst_mode: str | BoxMode | type[BoxMode] | None = None, |
| | ) -> None: |
| | self.src_mode = src_mode |
| | self.dst_mode = dst_mode |
| |
|
| | def __call__(self, boxes: NdarrayOrTensor) -> NdarrayOrTensor: |
| | """ |
| | Converts the boxes in src_mode to the dst_mode. |
| | |
| | Args: |
| | boxes: source bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | |
| | Returns: |
| | bounding boxes with target mode, with same data type as ``boxes``, does not share memory with ``boxes`` |
| | """ |
| | return convert_box_mode(boxes, src_mode=self.src_mode, dst_mode=self.dst_mode) |
| |
|
| |
|
| | class ConvertBoxToStandardMode(Transform): |
| | """ |
| | Convert given boxes to standard mode. |
| | Standard mode is "xyxy" or "xyzxyz", |
| | representing box format of [xmin, ymin, xmax, ymax] or [xmin, ymin, zmin, xmax, ymax, zmax]. |
| | |
| | Args: |
| | mode: source box mode. If it is not given, this func will assume it is ``StandardMode()``. |
| | It follows the same format with ``src_mode`` in :class:`~monai.apps.detection.transforms.array.ConvertBoxMode` . |
| | |
| | Example: |
| | .. code-block:: python |
| | |
| | boxes = torch.ones(10,6) |
| | # convert boxes with format [xmin, xmax, ymin, ymax, zmin, zmax] to [xmin, ymin, zmin, xmax, ymax, zmax] |
| | box_converter = ConvertBoxToStandardMode(mode="xxyyzz") |
| | box_converter(boxes) |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, mode: str | BoxMode | type[BoxMode] | None = None) -> None: |
| | self.mode = mode |
| |
|
| | def __call__(self, boxes: NdarrayOrTensor) -> NdarrayOrTensor: |
| | """ |
| | Convert given boxes to standard mode. |
| | Standard mode is "xyxy" or "xyzxyz", |
| | representing box format of [xmin, ymin, xmax, ymax] or [xmin, ymin, zmin, xmax, ymax, zmax]. |
| | |
| | Args: |
| | boxes: source bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | |
| | Returns: |
| | bounding boxes with standard mode, with same data type as ``boxes``, does not share memory with ``boxes`` |
| | """ |
| | return convert_box_to_standard_mode(boxes, mode=self.mode) |
| |
|
| |
|
| | class AffineBox(Transform): |
| | """ |
| | Applies affine matrix to the boxes |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __call__(self, boxes: NdarrayOrTensor, affine: NdarrayOrTensor | None) -> NdarrayOrTensor: |
| | """ |
| | Args: |
| | boxes: source bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | affine: affine matrix to be applied to the box coordinate |
| | """ |
| | if affine is None: |
| | return boxes |
| |
|
| | return apply_affine_to_boxes(boxes, affine=affine) |
| |
|
| |
|
| | class ZoomBox(Transform): |
| | """ |
| | Zooms an ND Box with same padding or slicing setting with Zoom(). |
| | |
| | Args: |
| | zoom: The zoom factor along the spatial axes. |
| | If a float, zoom is the same for each spatial axis. |
| | If a sequence, zoom should contain one value for each spatial axis. |
| | keep_size: Should keep original size (padding/slicing if needed), default is True. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, zoom: Sequence[float] | float, keep_size: bool = False, **kwargs: Any) -> None: |
| | self.zoom = zoom |
| | self.keep_size = keep_size |
| | self.kwargs = kwargs |
| |
|
| | def __call__(self, boxes: NdarrayTensor, src_spatial_size: Sequence[int] | int | None = None) -> NdarrayTensor: |
| | """ |
| | Args: |
| | boxes: source bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | src_spatial_size: original image spatial size before zooming, used only when keep_size=True. |
| | """ |
| | spatial_dims: int = get_spatial_dims(boxes=boxes) |
| | self._zoom = ensure_tuple_rep(self.zoom, spatial_dims) |
| |
|
| | if not self.keep_size: |
| | return zoom_boxes(boxes, self._zoom) |
| |
|
| | if src_spatial_size is None: |
| | raise ValueError("keep_size=True, src_spatial_size must be provided.") |
| |
|
| | src_spatial_size = ensure_tuple_rep(src_spatial_size, spatial_dims) |
| | dst_spatial_size = [int(round(z * ss)) for z, ss in zip(self._zoom, src_spatial_size)] |
| | self._zoom = tuple(ds / float(ss) for ss, ds in zip(src_spatial_size, dst_spatial_size)) |
| | zoomed_boxes = zoom_boxes(boxes, self._zoom) |
| |
|
| | |
| | if not np.allclose(np.array(src_spatial_size), np.array(dst_spatial_size)): |
| | for axis, (od, zd) in enumerate(zip(src_spatial_size, dst_spatial_size)): |
| | diff = od - zd |
| | half = abs(diff) // 2 |
| | if diff > 0: |
| | zoomed_boxes[:, axis] = zoomed_boxes[:, axis] + half |
| | zoomed_boxes[:, axis + spatial_dims] = zoomed_boxes[:, axis + spatial_dims] + half |
| | elif diff < 0: |
| | zoomed_boxes[:, axis] = zoomed_boxes[:, axis] - half |
| | zoomed_boxes[:, axis + spatial_dims] = zoomed_boxes[:, axis + spatial_dims] - half |
| | return zoomed_boxes |
| |
|
| |
|
| | class ResizeBox(Transform): |
| | """ |
| | Resize the input boxes when the corresponding image is |
| | resized to given spatial size (with scaling, not cropping/padding). |
| | |
| | Args: |
| | spatial_size: expected shape of spatial dimensions after resize operation. |
| | if some components of the `spatial_size` are non-positive values, the transform will use the |
| | corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| | to `(32, 64)` if the second spatial dimension size of img is `64`. |
| | size_mode: should be "all" or "longest", if "all", will use `spatial_size` for all the spatial dims, |
| | if "longest", rescale the image so that only the longest side is equal to specified `spatial_size`, |
| | which must be an int number in this case, keeping the aspect ratio of the initial image, refer to: |
| | https://albumentations.ai/docs/api_reference/augmentations/geometric/resize/ |
| | #albumentations.augmentations.geometric.resize.LongestMaxSize. |
| | kwargs: other arguments for the `np.pad` or `torch.pad` function. |
| | note that `np.pad` treats channel dimension as the first dimension. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, spatial_size: Sequence[int] | int, size_mode: str = "all", **kwargs: Any) -> None: |
| | self.size_mode = look_up_option(size_mode, ["all", "longest"]) |
| | self.spatial_size = spatial_size |
| |
|
| | def __call__(self, boxes: NdarrayOrTensor, src_spatial_size: Sequence[int] | int) -> NdarrayOrTensor: |
| | """ |
| | Args: |
| | boxes: source bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | src_spatial_size: original image spatial size before resizing. |
| | |
| | Raises: |
| | ValueError: When ``self.spatial_size`` length is less than ``boxes`` spatial dimensions. |
| | """ |
| | input_ndim = get_spatial_dims(boxes=boxes) |
| | src_spatial_size_ = ensure_tuple_rep(src_spatial_size, input_ndim) |
| |
|
| | if self.size_mode == "all": |
| | |
| | output_ndim = len(ensure_tuple(self.spatial_size)) |
| | if output_ndim != input_ndim: |
| | raise ValueError( |
| | "len(spatial_size) must be greater or equal to img spatial dimensions, " |
| | f"got spatial_size={output_ndim} img={input_ndim}." |
| | ) |
| | spatial_size_ = fall_back_tuple(self.spatial_size, src_spatial_size_) |
| | else: |
| | if not isinstance(self.spatial_size, int): |
| | raise ValueError("spatial_size must be an int number if size_mode is 'longest'.") |
| | scale = self.spatial_size / max(src_spatial_size_) |
| | spatial_size_ = tuple(int(round(s * scale)) for s in src_spatial_size_) |
| |
|
| | return resize_boxes(boxes, src_spatial_size_, spatial_size_) |
| |
|
| |
|
| | class FlipBox(Transform): |
| | """ |
| | Reverses the box coordinates along the given spatial axis. Preserves shape. |
| | |
| | Args: |
| | spatial_axis: spatial axes along which to flip over. Default is None. |
| | The default `axis=None` will flip over all of the axes of the input array. |
| | If axis is negative it counts from the last to the first axis. |
| | If axis is a tuple of ints, flipping is performed on all of the axes |
| | specified in the tuple. |
| | |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, spatial_axis: Sequence[int] | int | None = None) -> None: |
| | self.spatial_axis = spatial_axis |
| |
|
| | def __call__(self, boxes: NdarrayOrTensor, spatial_size: Sequence[int] | int): |
| | """ |
| | Args: |
| | boxes: bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | spatial_size: image spatial size. |
| | """ |
| |
|
| | return flip_boxes(boxes, spatial_size=spatial_size, flip_axes=self.spatial_axis) |
| |
|
| |
|
| | class ClipBoxToImage(Transform): |
| | """ |
| | Clip the bounding boxes and the associated labels/scores to make sure they are within the image. |
| | There might be multiple arrays of labels/scores associated with one array of boxes. |
| | |
| | Args: |
| | remove_empty: whether to remove the boxes and corresponding labels that are actually empty |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__(self, remove_empty: bool = False) -> None: |
| | self.remove_empty = remove_empty |
| |
|
| | def __call__( |
| | self, |
| | boxes: NdarrayOrTensor, |
| | labels: Sequence[NdarrayOrTensor] | NdarrayOrTensor, |
| | spatial_size: Sequence[int] | int, |
| | ) -> tuple[NdarrayOrTensor, tuple | NdarrayOrTensor]: |
| | """ |
| | Args: |
| | boxes: bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | labels: Sequence of array. Each element represents classification labels or scores |
| | corresponding to ``boxes``, sized (N,). |
| | spatial_size: The spatial size of the image where the boxes are attached. len(spatial_size) should be in [2, 3]. |
| | |
| | Returns: |
| | - clipped boxes, does not share memory with original boxes |
| | - clipped labels, does not share memory with original labels |
| | |
| | Example: |
| | .. code-block:: python |
| | |
| | box_clipper = ClipBoxToImage(remove_empty=True) |
| | boxes = torch.ones(2, 6) |
| | class_labels = torch.Tensor([0, 1]) |
| | pred_scores = torch.Tensor([[0.4,0.3,0.3], [0.5,0.1,0.4]]) |
| | labels = (class_labels, pred_scores) |
| | spatial_size = [32, 32, 32] |
| | boxes_clip, labels_clip_tuple = box_clipper(boxes, labels, spatial_size) |
| | """ |
| | spatial_dims: int = get_spatial_dims(boxes=boxes) |
| | spatial_size = ensure_tuple_rep(spatial_size, spatial_dims) |
| |
|
| | boxes_clip, keep = clip_boxes_to_image(boxes, spatial_size, self.remove_empty) |
| | return boxes_clip, select_labels(labels, keep) |
| |
|
| |
|
| | class BoxToMask(Transform): |
| | """ |
| | Convert box to int16 mask image, which has the same size with the input image. |
| | |
| | Args: |
| | bg_label: background labels for the output mask image, make sure it is smaller than any foreground(fg) labels. |
| | ellipse_mask: bool. |
| | |
| | - If True, it assumes the object shape is close to ellipse or ellipsoid. |
| | - If False, it assumes the object shape is close to rectangle or cube and well occupies the bounding box. |
| | - If the users are going to apply random rotation as data augmentation, we suggest setting ellipse_mask=True |
| | See also Kalra et al. "Towards Rotation Invariance in Object Detection", ICCV 2021. |
| | """ |
| |
|
| | backend = [TransformBackends.NUMPY] |
| |
|
| | def __init__(self, bg_label: int = -1, ellipse_mask: bool = False) -> None: |
| | self.bg_label = bg_label |
| | self.ellipse_mask = ellipse_mask |
| |
|
| | def __call__( |
| | self, boxes: NdarrayOrTensor, labels: NdarrayOrTensor, spatial_size: Sequence[int] | int |
| | ) -> NdarrayOrTensor: |
| | """ |
| | Args: |
| | boxes: bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode``. |
| | labels: classification foreground(fg) labels corresponding to `boxes`, dtype should be int, sized (N,). |
| | spatial_size: image spatial size. |
| | |
| | Return: |
| | - int16 array, sized (num_box, H, W). Each channel represents a box. |
| | The foreground region in channel c has intensity of labels[c]. |
| | The background intensity is bg_label. |
| | """ |
| | return convert_box_to_mask(boxes, labels, spatial_size, self.bg_label, self.ellipse_mask) |
| |
|
| |
|
| | class MaskToBox(Transform): |
| | """ |
| | Convert int16 mask image to box, which has the same size with the input image. |
| | Pairs with :py:class:`monai.apps.detection.transforms.array.BoxToMask`. |
| | Please make sure the same ``min_fg_label`` is used when using the two transforms in pairs. |
| | |
| | Args: |
| | bg_label: background labels for the output mask image, make sure it is smaller than any foreground(fg) labels. |
| | box_dtype: output dtype for boxes |
| | label_dtype: output dtype for labels |
| | """ |
| |
|
| | backend = [TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | bg_label: int = -1, |
| | box_dtype: DtypeLike | torch.dtype = torch.float32, |
| | label_dtype: DtypeLike | torch.dtype = torch.long, |
| | ) -> None: |
| | self.bg_label = bg_label |
| | self.box_dtype = box_dtype |
| | self.label_dtype = label_dtype |
| |
|
| | def __call__(self, boxes_mask: NdarrayOrTensor) -> tuple[NdarrayOrTensor, NdarrayOrTensor]: |
| | """ |
| | Args: |
| | boxes_mask: int16 array, sized (num_box, H, W). Each channel represents a box. |
| | The foreground region in channel c has intensity of labels[c]. |
| | The background intensity is bg_label. |
| | |
| | Return: |
| | - bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode``. |
| | - classification foreground(fg) labels, dtype should be int, sized (N,). |
| | """ |
| | return convert_mask_to_box(boxes_mask, self.bg_label, self.box_dtype, self.label_dtype) |
| |
|
| |
|
| | class SpatialCropBox(SpatialCrop): |
| | """ |
| | General purpose box cropper when the corresponding image is cropped by SpatialCrop(*) with the same ROI. |
| | The difference is that we do not support negative indexing for roi_slices. |
| | |
| | If a dimension of the expected ROI size is bigger than the input image size, will not crop that dimension. |
| | So the cropped result may be smaller than the expected ROI, and the cropped results of several images may |
| | not have exactly the same shape. |
| | It can support to crop ND spatial boxes. |
| | |
| | The cropped region can be parameterised in various ways: |
| | - a list of slices for each spatial dimension (do not allow for use of negative indexing) |
| | - a spatial center and size |
| | - the start and end coordinates of the ROI |
| | |
| | Args: |
| | roi_center: voxel coordinates for center of the crop ROI. |
| | roi_size: size of the crop ROI, if a dimension of ROI size is bigger than image size, |
| | will not crop that dimension of the image. |
| | roi_start: voxel coordinates for start of the crop ROI. |
| | roi_end: voxel coordinates for end of the crop ROI, if a coordinate is out of image, |
| | use the end coordinate of image. |
| | roi_slices: list of slices for each of the spatial dimensions. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __init__( |
| | self, |
| | roi_center: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_size: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_start: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_end: Sequence[int] | NdarrayOrTensor | None = None, |
| | roi_slices: Sequence[slice] | None = None, |
| | ) -> None: |
| | super().__init__(roi_center, roi_size, roi_start, roi_end, roi_slices) |
| | for s in self.slices: |
| | if s.start < 0 or s.stop < 0 or (s.step is not None and s.step < 0): |
| | raise ValueError("Currently negative indexing is not supported for SpatialCropBox.") |
| |
|
| | def __call__( |
| | self, boxes: NdarrayTensor, labels: Sequence[NdarrayOrTensor] | NdarrayOrTensor |
| | ) -> tuple[NdarrayTensor, tuple | NdarrayOrTensor]: |
| | """ |
| | Args: |
| | boxes: bounding boxes, Nx4 or Nx6 torch tensor or ndarray. The box mode is assumed to be ``StandardMode`` |
| | labels: Sequence of array. Each element represents classification labels or scores |
| | |
| | Returns: |
| | - cropped boxes, does not share memory with original boxes |
| | - cropped labels, does not share memory with original labels |
| | |
| | Example: |
| | .. code-block:: python |
| | |
| | box_cropper = SpatialCropPadBox(roi_start=[0, 1, 4], roi_end=[21, 15, 8]) |
| | boxes = torch.ones(2, 6) |
| | class_labels = torch.Tensor([0, 1]) |
| | pred_scores = torch.Tensor([[0.4,0.3,0.3], [0.5,0.1,0.4]]) |
| | labels = (class_labels, pred_scores) |
| | boxes_crop, labels_crop_tuple = box_cropper(boxes, labels) |
| | """ |
| | spatial_dims = min(len(self.slices), get_spatial_dims(boxes=boxes)) |
| | boxes_crop, keep = spatial_crop_boxes( |
| | boxes, |
| | [self.slices[axis].start for axis in range(spatial_dims)], |
| | [self.slices[axis].stop for axis in range(spatial_dims)], |
| | ) |
| | return boxes_crop, select_labels(labels, keep) |
| |
|
| |
|
| | class RotateBox90(Rotate90): |
| | """ |
| | Rotate a boxes by 90 degrees in the plane specified by `axes`. |
| | See box_ops.rot90_boxes for additional details |
| | |
| | Args: |
| | k: number of times to rotate by 90 degrees. |
| | spatial_axes: 2 int numbers, defines the plane to rotate with 2 spatial axes. |
| | Default: (0, 1), this is the first two axis in spatial dimensions. |
| | If axis is negative it counts from the last to the first axis. |
| | """ |
| |
|
| | backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
| |
|
| | def __call__(self, boxes: NdarrayTensor, spatial_size: Sequence[int] | int) -> NdarrayTensor: |
| | """ |
| | Args: |
| | img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| | """ |
| | return rot90_boxes(boxes, spatial_size, self.k, self.spatial_axes) |
| |
|