| | from collections import OrderedDict |
| | from typing import Union, Tuple, List |
| |
|
| | import numpy as np |
| | import pandas as pd |
| | import torch |
| | from batchgenerators.augmentations.utils import resize_segmentation |
| | from scipy.ndimage.interpolation import map_coordinates |
| | from skimage.transform import resize |
| |
|
| | ANISO_THRESHOLD = 3 |
| |
|
| | def get_do_separate_z(spacing: Union[Tuple[float, ...], List[float], np.ndarray], anisotropy_threshold=ANISO_THRESHOLD): |
| | do_separate_z = (np.max(spacing) / np.min(spacing)) > anisotropy_threshold |
| | return do_separate_z |
| |
|
| |
|
| | def get_lowres_axis(new_spacing: Union[Tuple[float, ...], List[float], np.ndarray]): |
| | axis = np.where(max(new_spacing) / np.array(new_spacing) == 1)[0] |
| | return axis |
| |
|
| |
|
| | def compute_new_shape(old_shape: Union[Tuple[int, ...], List[int], np.ndarray], |
| | old_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
| | new_spacing: Union[Tuple[float, ...], List[float], np.ndarray]) -> np.ndarray: |
| | assert len(old_spacing) == len(old_shape) |
| | assert len(old_shape) == len(new_spacing) |
| | |
| | new_shape = np.array([int(round(i / j * k)) for i, j, k in zip(old_spacing, new_spacing, old_shape)]) |
| | return new_shape |
| |
|
| |
|
| | def resample_data_or_seg_to_spacing(data: np.ndarray, |
| | current_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
| | new_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
| | is_seg: bool = False, |
| | order: int = 3, order_z: int = 0, |
| | force_separate_z: Union[bool, None] = False, |
| | separate_z_anisotropy_threshold: float = ANISO_THRESHOLD): |
| | if force_separate_z is not None: |
| | do_separate_z = force_separate_z |
| | if force_separate_z: |
| | axis = get_lowres_axis(current_spacing) |
| | else: |
| | axis = None |
| | else: |
| | if get_do_separate_z(current_spacing, separate_z_anisotropy_threshold): |
| | do_separate_z = True |
| | axis = get_lowres_axis(current_spacing) |
| | elif get_do_separate_z(new_spacing, separate_z_anisotropy_threshold): |
| | do_separate_z = True |
| | axis = get_lowres_axis(new_spacing) |
| | else: |
| | do_separate_z = False |
| | axis = None |
| |
|
| | if axis is not None: |
| | if len(axis) == 3: |
| | |
| | do_separate_z = False |
| | elif len(axis) == 2: |
| | |
| | |
| | do_separate_z = False |
| | else: |
| | pass |
| |
|
| | if data is not None: |
| | assert len(data.shape) == 4, "data must be c x y z" |
| |
|
| | shape = np.array(data[0].shape) |
| | new_shape = compute_new_shape(shape[1:], current_spacing, new_spacing) |
| |
|
| | data_reshaped = resample_data_or_seg(data, new_shape, is_seg, axis, order, do_separate_z, order_z=order_z) |
| | return data_reshaped |
| |
|
| |
|
| | def resample_data_or_seg_to_shape(data: Union[torch.Tensor, np.ndarray], |
| | new_shape: Union[Tuple[int, ...], List[int], np.ndarray], |
| | current_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
| | new_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
| | is_seg: bool = False, |
| | order: int = 3, order_z: int = 0, |
| | force_separate_z: Union[bool, None] = False, |
| | separate_z_anisotropy_threshold: float = ANISO_THRESHOLD): |
| | """ |
| | needed for segmentation export. Stupid, I know. Maybe we can fix that with Leos new resampling functions |
| | """ |
| | if isinstance(data, torch.Tensor): |
| | data = data.cpu().numpy() |
| | if force_separate_z is not None: |
| | do_separate_z = force_separate_z |
| | if force_separate_z: |
| | axis = get_lowres_axis(current_spacing) |
| | else: |
| | axis = None |
| | else: |
| | if get_do_separate_z(current_spacing, separate_z_anisotropy_threshold): |
| | do_separate_z = True |
| | axis = get_lowres_axis(current_spacing) |
| | elif get_do_separate_z(new_spacing, separate_z_anisotropy_threshold): |
| | do_separate_z = True |
| | axis = get_lowres_axis(new_spacing) |
| | else: |
| | do_separate_z = False |
| | axis = None |
| |
|
| | if axis is not None: |
| | if len(axis) == 3: |
| | |
| | do_separate_z = False |
| | elif len(axis) == 2: |
| | |
| | |
| | do_separate_z = False |
| | else: |
| | pass |
| |
|
| | if data is not None: |
| | assert len(data.shape) == 4, "data must be c x y z" |
| |
|
| | data_reshaped = resample_data_or_seg(data, new_shape, is_seg, axis, order, do_separate_z, order_z=order_z) |
| | return data_reshaped |
| |
|
| |
|
| | def resample_data_or_seg(data: np.ndarray, new_shape: Union[Tuple[float, ...], List[float], np.ndarray], |
| | is_seg: bool = False, axis: Union[None, int] = None, order: int = 3, |
| | do_separate_z: bool = False, order_z: int = 0): |
| | """ |
| | separate_z=True will resample with order 0 along z |
| | :param data: |
| | :param new_shape: |
| | :param is_seg: |
| | :param axis: |
| | :param order: |
| | :param do_separate_z: |
| | :param order_z: only applies if do_separate_z is True |
| | :return: |
| | """ |
| | assert len(data.shape) == 4, "data must be (c, x, y, z)" |
| | assert len(new_shape) == len(data.shape) - 1 |
| |
|
| | if is_seg: |
| | resize_fn = resize_segmentation |
| | kwargs = OrderedDict() |
| | else: |
| | resize_fn = resize |
| | kwargs = {'mode': 'edge', 'anti_aliasing': False} |
| | dtype_data = data.dtype |
| | shape = np.array(data[0].shape) |
| | new_shape = np.array(new_shape) |
| | if np.any(shape != new_shape): |
| | data = data.astype(float) |
| | if do_separate_z: |
| | |
| | assert len(axis) == 1, "only one anisotropic axis supported" |
| | axis = axis[0] |
| | if axis == 0: |
| | new_shape_2d = new_shape[1:] |
| | elif axis == 1: |
| | new_shape_2d = new_shape[[0, 2]] |
| | else: |
| | new_shape_2d = new_shape[:-1] |
| |
|
| | reshaped_final_data = [] |
| | for c in range(data.shape[0]): |
| | reshaped_data = [] |
| | for slice_id in range(shape[axis]): |
| | if axis == 0: |
| | reshaped_data.append(resize_fn(data[c, slice_id], new_shape_2d, order, **kwargs)) |
| | elif axis == 1: |
| | reshaped_data.append(resize_fn(data[c, :, slice_id], new_shape_2d, order, **kwargs)) |
| | else: |
| | reshaped_data.append(resize_fn(data[c, :, :, slice_id], new_shape_2d, order, **kwargs)) |
| | reshaped_data = np.stack(reshaped_data, axis) |
| | if shape[axis] != new_shape[axis]: |
| |
|
| | |
| | rows, cols, dim = new_shape[0], new_shape[1], new_shape[2] |
| | orig_rows, orig_cols, orig_dim = reshaped_data.shape |
| |
|
| | row_scale = float(orig_rows) / rows |
| | col_scale = float(orig_cols) / cols |
| | dim_scale = float(orig_dim) / dim |
| |
|
| | map_rows, map_cols, map_dims = np.mgrid[:rows, :cols, :dim] |
| | map_rows = row_scale * (map_rows + 0.5) - 0.5 |
| | map_cols = col_scale * (map_cols + 0.5) - 0.5 |
| | map_dims = dim_scale * (map_dims + 0.5) - 0.5 |
| |
|
| | coord_map = np.array([map_rows, map_cols, map_dims]) |
| | if not is_seg or order_z == 0: |
| | reshaped_final_data.append(map_coordinates(reshaped_data, coord_map, order=order_z, |
| | mode='nearest')[None]) |
| | else: |
| | unique_labels = np.sort(pd.unique(reshaped_data.ravel())) |
| | reshaped = np.zeros(new_shape, dtype=dtype_data) |
| |
|
| | for i, cl in enumerate(unique_labels): |
| | reshaped_multihot = np.round( |
| | map_coordinates((reshaped_data == cl).astype(float), coord_map, order=order_z, |
| | mode='nearest')) |
| | reshaped[reshaped_multihot > 0.5] = cl |
| | reshaped_final_data.append(reshaped[None]) |
| | else: |
| | reshaped_final_data.append(reshaped_data[None]) |
| | reshaped_final_data = np.vstack(reshaped_final_data) |
| | else: |
| | |
| | reshaped = [] |
| | for c in range(data.shape[0]): |
| | reshaped.append(resize_fn(data[c], new_shape, order, **kwargs)[None]) |
| | reshaped_final_data = np.vstack(reshaped) |
| | return reshaped_final_data.astype(dtype_data) |
| | else: |
| | |
| | return data |
| |
|