| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import copy |
| import os |
|
|
| import numpy as np |
|
|
| import monai |
| from monai.bundle import ConfigParser |
| from monai.utils import StrEnum, ensure_tuple, optional_import |
|
|
| tqdm, has_tqdm = optional_import("tqdm", name="tqdm") |
| nib, _ = optional_import("nibabel") |
|
|
| logger = monai.apps.utils.get_logger(__name__) |
|
|
| __all__ = ["analyze_data", "create_new_data_copy", "create_new_dataset_json", "NNUNETMode"] |
|
|
|
|
| class NNUNETMode(StrEnum): |
| N_2D = "2d" |
| N_3D_FULLRES = "3d_fullres" |
| N_3D_LOWRES = "3d_lowres" |
| N_3D_CASCADE_FULLRES = "3d_cascade_fullres" |
|
|
|
|
| def analyze_data(datalist_json: dict, data_dir: str) -> tuple[int, int]: |
| """ |
| Analyze (training) data |
| |
| Args: |
| datalist_json: original data list .json (required by most monai tutorials). |
| data_dir: raw data directory. |
| """ |
| img = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| os.path.join(data_dir, datalist_json["training"][0]["image"]) |
| ) |
| num_input_channels = img.size()[0] if img.dim() == 4 else 1 |
| logger.info(f"num_input_channels: {num_input_channels}") |
|
|
| num_foreground_classes = 0 |
| for _i in range(len(datalist_json["training"])): |
| seg = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| os.path.join(data_dir, datalist_json["training"][_i]["label"]) |
| ) |
| num_foreground_classes = max(num_foreground_classes, int(seg.max())) |
| logger.info(f"num_foreground_classes: {num_foreground_classes}") |
|
|
| return num_input_channels, num_foreground_classes |
|
|
|
|
| def create_new_data_copy( |
| test_key: str, datalist_json: dict, data_dir: str, num_input_channels: int, output_datafolder: str |
| ) -> None: |
| """ |
| Create and organize a new copy of data to meet the requirements of nnU-Net V2 |
| |
| Args: |
| test_key: key for test data in the data list .json. |
| datalist_json: original data list .json (required by most monai tutorials). |
| data_dir: raw data directory. |
| num_input_channels: number of input (image) channels. |
| output_datafolder: output folder. |
| """ |
| _index = 0 |
| new_datalist_json: dict = {"training": [], test_key: []} |
|
|
| for _key, _folder, _label_folder in list( |
| zip(["training", test_key], ["imagesTr", "imagesTs"], ["labelsTr", "labelsTs"]) |
| ): |
| if _key is None: |
| continue |
|
|
| logger.info(f"converting data section: {_key}...") |
| for _k in tqdm(range(len(datalist_json[_key]))) if has_tqdm else range(len(datalist_json[_key])): |
| orig_img_name = ( |
| datalist_json[_key][_k]["image"] |
| if isinstance(datalist_json[_key][_k], dict) |
| else datalist_json[_key][_k] |
| ) |
| img_name = f"case_{_index}" |
| _index += 1 |
|
|
| |
| nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| os.path.join(data_dir, orig_img_name) |
| ) |
| affine = nda.meta["original_affine"] |
| nda = nda.numpy() |
| for _l in range(num_input_channels): |
| outimg = nib.Nifti1Image(nda[_l, ...], affine) |
| index = "_" + str(_l + 10000)[-4:] |
| nib.save(outimg, os.path.join(output_datafolder, _folder, img_name + index + ".nii.gz")) |
|
|
| |
| if isinstance(datalist_json[_key][_k], dict) and "label" in datalist_json[_key][_k]: |
| nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| os.path.join(data_dir, datalist_json[_key][_k]["label"]) |
| ) |
| affine = nda.meta["original_affine"] |
| nda = nda.numpy().astype(np.uint8) |
| nda = nda[0, ...] if nda.ndim == 4 and nda.shape[0] == 1 else nda |
| nib.save( |
| nib.Nifti1Image(nda, affine), os.path.join(output_datafolder, _label_folder, img_name + ".nii.gz") |
| ) |
|
|
| if isinstance(datalist_json[_key][_k], dict): |
| _val = copy.deepcopy(datalist_json[_key][_k]) |
| _val["new_name"] = img_name |
| new_datalist_json[_key].append(_val) |
| else: |
| new_datalist_json[_key].append({"image": datalist_json[_key][_k], "new_name": img_name}) |
|
|
| ConfigParser.export_config_file( |
| config=new_datalist_json, |
| filepath=os.path.join(output_datafolder, "datalist.json"), |
| fmt="json", |
| sort_keys=True, |
| indent=4, |
| ensure_ascii=False, |
| ) |
|
|
| return |
|
|
|
|
| def create_new_dataset_json( |
| modality: str, num_foreground_classes: int, num_input_channels: int, num_training_data: int, output_filepath: str |
| ) -> None: |
| """ |
| Create a new copy of dataset .json to meet the requirements of nnU-Net V2 |
| |
| Args: |
| modality: image modality, could a string or a list of strings. |
| num_foreground_classes: number of foreground classes. |
| num_input_channels: number of input (image) channels. |
| num_training_data: number of training data. |
| output_filepath: output file path/name. |
| """ |
| new_json_data: dict = {} |
|
|
| |
| modality = ensure_tuple(modality) |
|
|
| new_json_data["channel_names"] = {} |
| for _j in range(num_input_channels): |
| new_json_data["channel_names"][str(_j)] = modality[_j] |
|
|
| new_json_data["labels"] = {} |
| new_json_data["labels"]["background"] = 0 |
| for _j in range(num_foreground_classes): |
| new_json_data["labels"][f"class{_j + 1}"] = _j + 1 |
|
|
| |
| new_json_data["numTraining"] = num_training_data |
| new_json_data["file_ending"] = ".nii.gz" |
|
|
| ConfigParser.export_config_file( |
| config=new_json_data, |
| |
| filepath=output_filepath, |
| fmt="json", |
| sort_keys=True, |
| indent=4, |
| ensure_ascii=False, |
| ) |
|
|
| return |
|
|