| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | import json |
| | import logging |
| | import random |
| | import warnings |
| | from collections.abc import Hashable, Mapping, Sequence, Sized |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai.config import KeysCollection |
| | from monai.data import MetaTensor |
| | from monai.networks.layers import GaussianFilter |
| | from monai.transforms.transform import MapTransform, Randomizable, Transform |
| | from monai.utils import min_version, optional_import |
| |
|
| | measure, _ = optional_import("skimage.measure", "0.14.2", min_version) |
| |
|
| | logger = logging.getLogger(__name__) |
| |
|
| | distance_transform_cdt, _ = optional_import("scipy.ndimage.morphology", name="distance_transform_cdt") |
| |
|
| |
|
| | class DiscardAddGuidanced(MapTransform): |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | number_intensity_ch: int = 1, |
| | probability: float = 1.0, |
| | label_names: Sized | None = None, |
| | allow_missing_keys: bool = False, |
| | ): |
| | """ |
| | Discard positive and negative points according to discard probability |
| | |
| | Args: |
| | keys: The ``keys`` parameter will be used to get and set the actual data item to transform |
| | number_intensity_ch: number of intensity channels |
| | probability: probability of discarding clicks |
| | """ |
| | super().__init__(keys, allow_missing_keys) |
| |
|
| | self.number_intensity_ch = number_intensity_ch |
| | self.discard_probability = probability |
| | self.label_names = label_names or [] |
| |
|
| | def _apply(self, image): |
| | if self.discard_probability >= 1.0 or np.random.choice( |
| | [True, False], p=[self.discard_probability, 1 - self.discard_probability] |
| | ): |
| | signal = np.zeros( |
| | (len(self.label_names), image.shape[-3], image.shape[-2], image.shape[-1]), dtype=np.float32 |
| | ) |
| | if image.shape[0] == self.number_intensity_ch + len(self.label_names): |
| | image[self.number_intensity_ch :, ...] = signal |
| | else: |
| | image = np.concatenate([image, signal], axis=0) |
| | return image |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "image": |
| | tmp_image = self._apply(d[key]) |
| | if isinstance(d[key], MetaTensor): |
| | d[key].array = tmp_image |
| | else: |
| | d[key] = tmp_image |
| | else: |
| | print("This transform only applies to the image") |
| | return d |
| |
|
| |
|
| | class NormalizeLabelsInDatasetd(MapTransform): |
| |
|
| | def __init__( |
| | self, keys: KeysCollection, label_names: dict[str, int] | None = None, allow_missing_keys: bool = False |
| | ): |
| | """ |
| | Normalize label values according to label names dictionary |
| | |
| | Args: |
| | keys: The ``keys`` parameter will be used to get and set the actual data item to transform |
| | label_names: all label names |
| | """ |
| | super().__init__(keys, allow_missing_keys) |
| |
|
| | self.label_names = label_names or {} |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | |
| | new_label_names = {} |
| | label = np.zeros(d[key].shape) |
| | |
| | for idx, (key_label, val_label) in enumerate(self.label_names.items(), start=1): |
| | if key_label != "background": |
| | new_label_names[key_label] = idx |
| | label[d[key] == val_label] = idx |
| | if key_label == "background": |
| | new_label_names["background"] = 0 |
| |
|
| | d["label_names"] = new_label_names |
| | if isinstance(d[key], MetaTensor): |
| | d[key].array = label |
| | else: |
| | d[key] = label |
| | return d |
| |
|
| |
|
| | class SingleLabelSelectiond(MapTransform): |
| |
|
| | def __init__( |
| | self, keys: KeysCollection, label_names: Sequence[str] | None = None, allow_missing_keys: bool = False |
| | ): |
| | """ |
| | Selects one label at a time to train the DeepEdit |
| | |
| | Args: |
| | keys: The ``keys`` parameter will be used to get and set the actual data item to transform |
| | label_names: all label names |
| | """ |
| | super().__init__(keys, allow_missing_keys) |
| |
|
| | self.label_names: Sequence[str] = label_names or [] |
| | self.all_label_values = { |
| | "spleen": 1, |
| | "right kidney": 2, |
| | "left kidney": 3, |
| | "gallbladder": 4, |
| | "esophagus": 5, |
| | "liver": 6, |
| | "stomach": 7, |
| | "aorta": 8, |
| | "inferior vena cava": 9, |
| | "portal_vein": 10, |
| | "splenic_vein": 11, |
| | "pancreas": 12, |
| | "right adrenal gland": 13, |
| | "left adrenal gland": 14, |
| | } |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "label": |
| | |
| | t_label = np.random.choice(self.label_names) |
| | d["current_label"] = t_label |
| | d[key][d[key] != self.all_label_values[t_label]] = 0.0 |
| | |
| | max_label_val = self.label_names.index(t_label) + 1 |
| | d[key][d[key] > 0] = max_label_val |
| | print(f"Using label {t_label} with number: {d[key].max()}") |
| | else: |
| | warnings.warn("This transform only applies to the label") |
| | return d |
| |
|
| |
|
| | class AddGuidanceSignalDeepEditd(MapTransform): |
| | """ |
| | Add Guidance signal for input image. Multilabel DeepEdit |
| | |
| | Based on the "guidance" points, apply Gaussian to them and add them as new channel for input image. |
| | |
| | Args: |
| | guidance: key to store guidance. |
| | sigma: standard deviation for Gaussian kernel. |
| | number_intensity_ch: channel index. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | guidance: str = "guidance", |
| | sigma: int = 3, |
| | number_intensity_ch: int = 1, |
| | allow_missing_keys: bool = False, |
| | ): |
| | super().__init__(keys, allow_missing_keys) |
| | self.guidance = guidance |
| | self.sigma = sigma |
| | self.number_intensity_ch = number_intensity_ch |
| |
|
| | def _get_signal(self, image, guidance): |
| | dimensions = 3 if len(image.shape) > 3 else 2 |
| | guidance = guidance.tolist() if isinstance(guidance, np.ndarray) else guidance |
| | guidance = json.loads(guidance) if isinstance(guidance, str) else guidance |
| |
|
| | |
| | if len(guidance): |
| | if dimensions == 3: |
| | |
| | signal = np.zeros((1, image.shape[-3], image.shape[-2], image.shape[-1]), dtype=np.float32) |
| | else: |
| | signal = np.zeros((1, image.shape[-2], image.shape[-1]), dtype=np.float32) |
| |
|
| | sshape = signal.shape |
| | for point in guidance: |
| | if np.any(np.asarray(point) < 0): |
| | continue |
| |
|
| | if dimensions == 3: |
| | |
| | p1 = max(0, min(int(point[-3]), sshape[-3] - 1)) |
| | p2 = max(0, min(int(point[-2]), sshape[-2] - 1)) |
| | p3 = max(0, min(int(point[-1]), sshape[-1] - 1)) |
| | signal[:, p1, p2, p3] = 1.0 |
| | else: |
| | p1 = max(0, min(int(point[-2]), sshape[-2] - 1)) |
| | p2 = max(0, min(int(point[-1]), sshape[-1] - 1)) |
| | signal[:, p1, p2] = 1.0 |
| |
|
| | |
| | if np.max(signal[0]) > 0: |
| | signal_tensor = torch.tensor(signal[0]) |
| | pt_gaussian = GaussianFilter(len(signal_tensor.shape), sigma=self.sigma) |
| | signal_tensor = pt_gaussian(signal_tensor.unsqueeze(0).unsqueeze(0)) |
| | signal_tensor = signal_tensor.squeeze(0).squeeze(0) |
| | signal[0] = signal_tensor.detach().cpu().numpy() |
| | signal[0] = (signal[0] - np.min(signal[0])) / (np.max(signal[0]) - np.min(signal[0])) |
| | return signal |
| | else: |
| | if dimensions == 3: |
| | signal = np.zeros((1, image.shape[-3], image.shape[-2], image.shape[-1]), dtype=np.float32) |
| | else: |
| | signal = np.zeros((1, image.shape[-2], image.shape[-1]), dtype=np.float32) |
| | return signal |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "image": |
| | image = d[key] |
| | tmp_image = image[0 : 0 + self.number_intensity_ch, ...] |
| | guidance = d[self.guidance] |
| | for key_label in guidance.keys(): |
| | |
| | signal = self._get_signal(image, guidance[key_label]) |
| | tmp_image = np.concatenate([tmp_image, signal], axis=0) |
| | if isinstance(d[key], MetaTensor): |
| | d[key].array = tmp_image |
| | else: |
| | d[key] = tmp_image |
| | return d |
| | else: |
| | print("This transform only applies to image key") |
| | return d |
| |
|
| |
|
| | class FindAllValidSlicesDeepEditd(MapTransform): |
| | """ |
| | Find/List all valid slices in the labels. |
| | Label is assumed to be a 4D Volume with shape CHWD, where C=1. |
| | |
| | Args: |
| | sids: key to store slices indices having valid label map. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, sids: Hashable = "sids", allow_missing_keys: bool = False): |
| | super().__init__(keys, allow_missing_keys) |
| | self.sids = sids |
| |
|
| | def _apply(self, label, d): |
| | sids = {} |
| | for key_label in d["label_names"].keys(): |
| | l_ids = [] |
| | for sid in range(label.shape[-1]): |
| | if d["label_names"][key_label] in label[0][..., sid]: |
| | l_ids.append(sid) |
| | sids[key_label] = l_ids |
| | return sids |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "label": |
| | label = d[key] |
| | if label.shape[0] != 1: |
| | raise ValueError("Only supports single channel labels!") |
| |
|
| | if len(label.shape) != 4: |
| | raise ValueError("Only supports label with shape CHWD!") |
| |
|
| | sids = self._apply(label, d) |
| | if sids is not None and len(sids.keys()): |
| | d[self.sids] = sids |
| | return d |
| | else: |
| | print("This transform only applies to label key") |
| | return d |
| |
|
| |
|
| | class AddInitialSeedPointDeepEditd(Randomizable, MapTransform): |
| | """ |
| | Add random guidance as initial seed point for a given label. |
| | |
| | Note that the label is of size (C, D, H, W) or (C, H, W) |
| | |
| | The guidance is of size (2, N, # of dims) where N is number of guidance added. |
| | # of dims = 4 when C, D, H, W; # of dims = 3 when (C, H, W) |
| | |
| | Args: |
| | guidance: key to store guidance. |
| | sids: key that represents lists of valid slice indices for the given label. |
| | sid: key that represents the slice to add initial seed point. If not present, random sid will be chosen. |
| | connected_regions: maximum connected regions to use for adding initial points. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | guidance: str = "guidance", |
| | sids: str = "sids", |
| | sid: str = "sid", |
| | connected_regions: int = 5, |
| | allow_missing_keys: bool = False, |
| | ): |
| | super().__init__(keys, allow_missing_keys) |
| | self.sids_key = sids |
| | self.sid_key = sid |
| | self.sid: dict[str, int] = dict() |
| | self.guidance = guidance |
| | self.connected_regions = connected_regions |
| |
|
| | def _apply(self, label, sid, key_label): |
| | dimensions = 3 if len(label.shape) > 3 else 2 |
| | self.default_guidance = [-1] * (dimensions + 1) |
| |
|
| | dims = dimensions |
| | if sid is not None and dimensions == 3: |
| | dims = 2 |
| | label = label[0][..., sid][np.newaxis] |
| |
|
| | |
| | label = (label > 0.5).astype(np.float32) |
| | |
| | |
| | blobs_labels = measure.label(label.astype(int), background=0) if dims == 2 else label |
| | if np.max(blobs_labels) <= 0: |
| | raise AssertionError(f"SLICES NOT FOUND FOR LABEL: {key_label}") |
| |
|
| | pos_guidance = [] |
| | for ridx in range(1, 2 if dims == 3 else self.connected_regions + 1): |
| | if dims == 2: |
| | label = (blobs_labels == ridx).astype(np.float32) |
| | if np.sum(label) == 0: |
| | pos_guidance.append(self.default_guidance) |
| | continue |
| |
|
| | |
| | |
| | |
| | distance = distance_transform_cdt(label).flatten() |
| | probability = np.exp(distance) - 1.0 |
| |
|
| | idx = np.where(label.flatten() > 0)[0] |
| | seed = self.R.choice(idx, size=1, p=probability[idx] / np.sum(probability[idx])) |
| | dst = distance[seed] |
| |
|
| | g = np.asarray(np.unravel_index(seed, label.shape)).transpose().tolist()[0] |
| | g[0] = dst[0] |
| | if dimensions == 2 or dims == 3: |
| | pos_guidance.append(g) |
| | else: |
| | |
| | pos_guidance.append([g[0], g[-2], g[-1], sid]) |
| |
|
| | return np.asarray([pos_guidance]) |
| |
|
| | def _randomize(self, d, key_label): |
| | sids = d.get(self.sids_key).get(key_label) if d.get(self.sids_key) is not None else None |
| | sid = d.get(self.sid_key).get(key_label) if d.get(self.sid_key) is not None else None |
| | if sids is not None and sids: |
| | if sid is None or sid not in sids: |
| | sid = self.R.choice(sids, replace=False) |
| | else: |
| | logger.info(f"Not slice IDs for label: {key_label}") |
| | sid = None |
| | self.sid[key_label] = sid |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "label": |
| | label_guidances = {} |
| | for key_label in d["sids"].keys(): |
| | |
| | self._randomize(d, key_label) |
| | |
| | tmp_label = np.copy(d[key]) |
| | |
| | if key_label != "background": |
| | tmp_label[tmp_label != float(d["label_names"][key_label])] = 0 |
| | else: |
| | tmp_label[tmp_label != float(d["label_names"][key_label])] = 1 |
| | tmp_label = 1 - tmp_label |
| | label_guidances[key_label] = json.dumps( |
| | self._apply(tmp_label, self.sid.get(key_label), key_label).astype(int).tolist() |
| | ) |
| | d[self.guidance] = label_guidances |
| | return d |
| | else: |
| | print("This transform only applies to label key") |
| | return d |
| |
|
| |
|
| | class FindDiscrepancyRegionsDeepEditd(MapTransform): |
| | """ |
| | Find discrepancy between prediction and actual during click interactions during training. |
| | |
| | Args: |
| | pred: key to prediction source. |
| | discrepancy: key to store discrepancies found between label and prediction. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | pred: str = "pred", |
| | discrepancy: str = "discrepancy", |
| | allow_missing_keys: bool = False, |
| | ): |
| | super().__init__(keys, allow_missing_keys) |
| | self.pred = pred |
| | self.discrepancy = discrepancy |
| |
|
| | @staticmethod |
| | def disparity(label, pred): |
| | disparity = label - pred |
| | |
| | |
| | pos_disparity = (disparity > 0).astype(np.float32) |
| | neg_disparity = (disparity < 0).astype(np.float32) |
| | return [pos_disparity, neg_disparity] |
| |
|
| | def _apply(self, label, pred): |
| | return self.disparity(label, pred) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "label": |
| | all_discrepancies = {} |
| | for _, (key_label, val_label) in enumerate(d["label_names"].items()): |
| | if key_label != "background": |
| | |
| | label = np.copy(d[key]) |
| | label[label != val_label] = 0 |
| | |
| | label = (label > 0.5).astype(np.float32) |
| | |
| | pred = np.copy(d[self.pred]) |
| | pred[pred != val_label] = 0 |
| | |
| | pred = (pred > 0.5).astype(np.float32) |
| | else: |
| | |
| | label = np.copy(d[key]) |
| | label[label != val_label] = 1 |
| | label = 1 - label |
| | |
| | label = (label > 0.5).astype(np.float32) |
| | |
| | pred = np.copy(d[self.pred]) |
| | pred[pred != val_label] = 1 |
| | pred = 1 - pred |
| | |
| | pred = (pred > 0.5).astype(np.float32) |
| | all_discrepancies[key_label] = self._apply(label, pred) |
| | d[self.discrepancy] = all_discrepancies |
| | return d |
| | else: |
| | print("This transform only applies to 'label' key") |
| | return d |
| |
|
| |
|
| | class AddRandomGuidanceDeepEditd(Randomizable, MapTransform): |
| | """ |
| | Add random guidance based on discrepancies that were found between label and prediction. |
| | |
| | Args: |
| | guidance: key to guidance source, shape (2, N, # of dim) |
| | discrepancy: key to discrepancy map between label and prediction shape (2, C, H, W, D) or (2, C, H, W) |
| | probability: key to click/interaction probability, shape (1) |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | guidance: str = "guidance", |
| | discrepancy: str = "discrepancy", |
| | probability: str = "probability", |
| | allow_missing_keys: bool = False, |
| | ): |
| | super().__init__(keys, allow_missing_keys) |
| | self.guidance_key = guidance |
| | self.discrepancy = discrepancy |
| | self.probability = probability |
| | self._will_interact = None |
| | self.is_pos: bool | None = None |
| | self.is_other: bool | None = None |
| | self.default_guidance = None |
| | self.guidance: dict[str, list[list[int]]] = {} |
| |
|
| | def randomize(self, data=None): |
| | probability = data[self.probability] |
| | self._will_interact = self.R.choice([True, False], p=[probability, 1.0 - probability]) |
| |
|
| | def find_guidance(self, discrepancy): |
| | distance = distance_transform_cdt(discrepancy).flatten() |
| | probability = np.exp(distance.flatten()) - 1.0 |
| | idx = np.where(discrepancy.flatten() > 0)[0] |
| |
|
| | if np.sum(discrepancy > 0) > 0: |
| | seed = self.R.choice(idx, size=1, p=probability[idx] / np.sum(probability[idx])) |
| | dst = distance[seed] |
| |
|
| | g = np.asarray(np.unravel_index(seed, discrepancy.shape)).transpose().tolist()[0] |
| | g[0] = dst[0] |
| | return g |
| | return None |
| |
|
| | def add_guidance(self, guidance, discrepancy, label_names, labels): |
| | |
| | pos_discr = discrepancy[0] |
| |
|
| | |
| | other_discrepancy_areas = {} |
| | for _, (key_label, val_label) in enumerate(label_names.items()): |
| | if key_label != "background": |
| | tmp_label = np.copy(labels) |
| | tmp_label[tmp_label != val_label] = 0 |
| | tmp_label = (tmp_label > 0.5).astype(np.float32) |
| | other_discrepancy_areas[key_label] = np.sum(discrepancy[1] * tmp_label) |
| | else: |
| | tmp_label = np.copy(labels) |
| | tmp_label[tmp_label != val_label] = 1 |
| | tmp_label = 1 - tmp_label |
| | other_discrepancy_areas[key_label] = np.sum(discrepancy[1] * tmp_label) |
| |
|
| | |
| | if np.sum(pos_discr) > 0: |
| | guidance.append(self.find_guidance(pos_discr)) |
| | self.is_pos = True |
| |
|
| | |
| | for key_label in label_names.keys(): |
| | |
| | if other_discrepancy_areas[key_label] > 50: |
| | self.is_other = True |
| | if key_label != "background": |
| | tmp_label = np.copy(labels) |
| | tmp_label[tmp_label != label_names[key_label]] = 0 |
| | tmp_label = (tmp_label > 0.5).astype(np.float32) |
| | self.guidance[key_label].append(self.find_guidance(discrepancy[1] * tmp_label)) |
| | else: |
| | tmp_label = np.copy(labels) |
| | tmp_label[tmp_label != label_names[key_label]] = 1 |
| | tmp_label = 1 - tmp_label |
| | self.guidance[key_label].append(self.find_guidance(discrepancy[1] * tmp_label)) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | guidance = d[self.guidance_key] |
| | discrepancy = d[self.discrepancy] |
| | self.randomize(data) |
| | if self._will_interact: |
| | |
| | for key_label in d["label_names"].keys(): |
| | tmp_gui = guidance[key_label] |
| | tmp_gui = tmp_gui.tolist() if isinstance(tmp_gui, np.ndarray) else tmp_gui |
| | tmp_gui = json.loads(tmp_gui) if isinstance(tmp_gui, str) else tmp_gui |
| | self.guidance[key_label] = [j for j in tmp_gui if -1 not in j] |
| |
|
| | |
| | for key_label in d["label_names"].keys(): |
| | |
| | self.add_guidance(self.guidance[key_label], discrepancy[key_label], d["label_names"], d["label"]) |
| |
|
| | |
| | num_clicks = random.randint(1, 10) |
| | counter = 0 |
| | keep_guidance = [] |
| | while True: |
| | aux_label = random.choice(list(d["label_names"].keys())) |
| | if aux_label in keep_guidance: |
| | pass |
| | else: |
| | keep_guidance.append(aux_label) |
| | counter = counter + len(self.guidance[aux_label]) |
| | |
| | if counter >= num_clicks: |
| | for key_label in d["label_names"].keys(): |
| | if key_label not in keep_guidance: |
| | self.guidance[key_label] = [] |
| | logger.info(f"Number of simulated clicks: {counter}") |
| | break |
| |
|
| | |
| | if len(keep_guidance) == len(d["label_names"].keys()): |
| | logger.info(f"Number of simulated clicks: {counter}") |
| | break |
| | d[self.guidance_key] = self.guidance |
| | return d |
| |
|
| |
|
| | class AddGuidanceFromPointsDeepEditd(Transform): |
| | """ |
| | Add guidance based on user clicks. ONLY WORKS FOR 3D |
| | |
| | We assume the input is loaded by LoadImaged and has the shape of (H, W, D) originally. |
| | Clicks always specify the coordinates in (H, W, D) |
| | |
| | Args: |
| | ref_image: key to reference image to fetch current and original image details. |
| | guidance: output key to store guidance. |
| | meta_keys: explicitly indicate the key of the metadata dictionary of `ref_image`. |
| | for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
| | the metadata is a dictionary object which contains: filename, original_shape, etc. |
| | if None, will try to construct meta_keys by `{ref_image}_{meta_key_postfix}`. |
| | meta_key_postfix: if meta_key is None, use `{ref_image}_{meta_key_postfix}` to fetch the metadata according |
| | to the key data, default is `meta_dict`, the metadata is a dictionary object. |
| | For example, to handle key `image`, read/write affine matrices from the |
| | metadata `image_meta_dict` dictionary's `affine` field. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | ref_image: str, |
| | guidance: str = "guidance", |
| | label_names: dict | None = None, |
| | meta_keys: str | None = None, |
| | meta_key_postfix: str = "meta_dict", |
| | ): |
| | self.ref_image = ref_image |
| | self.guidance = guidance |
| | self.label_names = label_names or {} |
| | self.meta_keys = meta_keys |
| | self.meta_key_postfix = meta_key_postfix |
| |
|
| | @staticmethod |
| | def _apply(clicks, factor): |
| | if len(clicks): |
| | guidance = np.multiply(clicks, factor).astype(int).tolist() |
| | return guidance |
| | else: |
| | return [] |
| |
|
| | def __call__(self, data): |
| | d = dict(data) |
| | meta_dict_key = self.meta_keys or f"{self.ref_image}_{self.meta_key_postfix}" |
| | |
| | if isinstance(d[self.ref_image], MetaTensor): |
| | meta_dict = d[self.ref_image].meta |
| | elif meta_dict_key in d: |
| | meta_dict = d[meta_dict_key] |
| | else: |
| | raise ValueError( |
| | f"{meta_dict_key} is not found. Please check whether it is the correct the image meta key." |
| | ) |
| |
|
| | if "spatial_shape" not in meta_dict: |
| | raise RuntimeError('Missing "spatial_shape" in meta_dict!') |
| |
|
| | |
| | original_shape = meta_dict["spatial_shape"] |
| | current_shape = list(d[self.ref_image].shape)[1:] |
| |
|
| | |
| | factor = np.array(current_shape) / original_shape |
| |
|
| | |
| | all_guidances = {} |
| | for key_label in self.label_names.keys(): |
| | clicks = d.get(key_label, []) |
| | clicks = list(np.array(clicks).astype(int)) |
| | all_guidances[key_label] = self._apply(clicks, factor) |
| | d[self.guidance] = all_guidances |
| | return d |
| |
|
| |
|
| | class ResizeGuidanceMultipleLabelDeepEditd(Transform): |
| | """ |
| | Resize the guidance based on cropped vs resized image. |
| | |
| | """ |
| |
|
| | def __init__(self, guidance: str, ref_image: str) -> None: |
| | self.guidance = guidance |
| | self.ref_image = ref_image |
| |
|
| | def __call__(self, data): |
| | d = dict(data) |
| | |
| | current_shape = d[self.ref_image].shape[1:] |
| |
|
| | meta_dict_key = "image_meta_dict" |
| | |
| | if isinstance(d[self.ref_image], MetaTensor): |
| | meta_dict = d[self.ref_image].meta |
| | elif meta_dict_key in d: |
| | meta_dict = d[meta_dict_key] |
| | else: |
| | raise ValueError( |
| | f"{meta_dict_key} is not found. Please check whether it is the correct the image meta key." |
| | ) |
| |
|
| | original_shape = meta_dict["spatial_shape"] |
| |
|
| | factor = np.divide(current_shape, original_shape) |
| | all_guidances = {} |
| | for key_label in d[self.guidance].keys(): |
| | guidance = ( |
| | np.multiply(d[self.guidance][key_label], factor).astype(int).tolist() |
| | if len(d[self.guidance][key_label]) |
| | else [] |
| | ) |
| | all_guidances[key_label] = guidance |
| |
|
| | d[self.guidance] = all_guidances |
| | return d |
| |
|
| |
|
| | class SplitPredsLabeld(MapTransform): |
| | """ |
| | Split preds and labels for individual evaluation |
| | |
| | """ |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "pred": |
| | for idx, (key_label, _) in enumerate(d["label_names"].items()): |
| | if key_label != "background": |
| | d[f"pred_{key_label}"] = d[key][idx + 1, ...][None] |
| | d[f"label_{key_label}"] = d["label"][idx + 1, ...][None] |
| | elif key != "pred": |
| | logger.info("This is only for pred key") |
| | return d |
| |
|
| |
|
| | class AddInitialSeedPointMissingLabelsd(Randomizable, MapTransform): |
| | """ |
| | Add random guidance as initial seed point for a given label. |
| | Note that the label is of size (C, D, H, W) or (C, H, W) |
| | The guidance is of size (2, N, # of dims) where N is number of guidance added. |
| | # of dims = 4 when C, D, H, W; # of dims = 3 when (C, H, W) |
| | Args: |
| | guidance: key to store guidance. |
| | sids: key that represents lists of valid slice indices for the given label. |
| | sid: key that represents the slice to add initial seed point. If not present, random sid will be chosen. |
| | connected_regions: maximum connected regions to use for adding initial points. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | guidance: str = "guidance", |
| | sids: str = "sids", |
| | sid: str = "sid", |
| | connected_regions: int = 5, |
| | allow_missing_keys: bool = False, |
| | ): |
| | super().__init__(keys, allow_missing_keys) |
| | self.sids_key = sids |
| | self.sid_key = sid |
| | self.sid: dict[str, int] = dict() |
| | self.guidance = guidance |
| | self.connected_regions = connected_regions |
| |
|
| | def _apply(self, label, sid): |
| | dimensions = 3 if len(label.shape) > 3 else 2 |
| | self.default_guidance = [-1] * (dimensions + 1) |
| |
|
| | dims = dimensions |
| | if sid is not None and dimensions == 3: |
| | dims = 2 |
| | label = label[0][..., sid][np.newaxis] |
| |
|
| | |
| | label = (label > 0.5).astype(np.float32) |
| | |
| | |
| | blobs_labels = measure.label(label.astype(int), background=0) if dims == 2 else label |
| |
|
| | label_guidance = [] |
| | |
| | if np.max(blobs_labels) <= 0: |
| | label_guidance.append(self.default_guidance) |
| | else: |
| | for ridx in range(1, 2 if dims == 3 else self.connected_regions + 1): |
| | if dims == 2: |
| | label = (blobs_labels == ridx).astype(np.float32) |
| | if np.sum(label) == 0: |
| | label_guidance.append(self.default_guidance) |
| | continue |
| |
|
| | |
| | |
| | |
| | |
| | distance = distance_transform_cdt(label).flatten() |
| | probability = np.exp(distance) - 1.0 |
| |
|
| | idx = np.where(label.flatten() > 0)[0] |
| | seed = self.R.choice(idx, size=1, p=probability[idx] / np.sum(probability[idx])) |
| | dst = distance[seed] |
| |
|
| | g = np.asarray(np.unravel_index(seed, label.shape)).transpose().tolist()[0] |
| | g[0] = dst[0] |
| | if dimensions == 2 or dims == 3: |
| | label_guidance.append(g) |
| | else: |
| | |
| | label_guidance.append([g[0], g[-2], g[-1], sid]) |
| |
|
| | return np.asarray(label_guidance) |
| |
|
| | def _randomize(self, d, key_label): |
| | sids = d.get(self.sids_key).get(key_label) if d.get(self.sids_key) is not None else None |
| | sid = d.get(self.sid_key).get(key_label) if d.get(self.sid_key) is not None else None |
| | if sids is not None and sids: |
| | if sid is None or sid not in sids: |
| | sid = self.R.choice(sids, replace=False) |
| | else: |
| | logger.info(f"Not slice IDs for label: {key_label}") |
| | sid = None |
| | self.sid[key_label] = sid |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "label": |
| | label_guidances = {} |
| | for key_label in d["sids"].keys(): |
| | |
| | self._randomize(d, key_label) |
| | |
| | tmp_label = np.copy(d[key]) |
| | |
| | if key_label != "background": |
| | tmp_label[tmp_label != float(d["label_names"][key_label])] = 0 |
| | else: |
| | tmp_label[tmp_label != float(d["label_names"][key_label])] = 1 |
| | tmp_label = 1 - tmp_label |
| | label_guidances[key_label] = json.dumps( |
| | self._apply(tmp_label, self.sid.get(key_label)).astype(int).tolist() |
| | ) |
| | d[self.guidance] = label_guidances |
| | return d |
| | else: |
| | print("This transform only applies to label key") |
| | return d |
| |
|
| |
|
| | class FindAllValidSlicesMissingLabelsd(MapTransform): |
| | """ |
| | Find/List all valid slices in the labels. |
| | Label is assumed to be a 4D Volume with shape CHWD, where C=1. |
| | Args: |
| | sids: key to store slices indices having valid label map. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, sids: Hashable = "sids", allow_missing_keys: bool = False): |
| | super().__init__(keys, allow_missing_keys) |
| | self.sids = sids |
| |
|
| | def _apply(self, label, d): |
| | sids = {} |
| | for key_label in d["label_names"].keys(): |
| | l_ids = [] |
| | for sid in range(label.shape[-1]): |
| | if d["label_names"][key_label] in label[0][..., sid]: |
| | l_ids.append(sid) |
| | |
| | if l_ids == []: |
| | l_ids = [-1] * 10 |
| | sids[key_label] = l_ids |
| | return sids |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> dict[Hashable, np.ndarray]: |
| | d: dict = dict(data) |
| | for key in self.key_iterator(d): |
| | if key == "label": |
| | label = d[key] |
| | if label.shape[0] != 1: |
| | raise ValueError("Only supports single channel labels!") |
| |
|
| | if len(label.shape) != 4: |
| | raise ValueError("Only supports label with shape CHWD!") |
| |
|
| | sids = self._apply(label, d) |
| | if sids is not None and len(sids.keys()): |
| | d[self.sids] = sids |
| | return d |
| | else: |
| | print("This transform only applies to label key") |
| | return d |
| |
|