| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | import os |
| | import shutil |
| | import sys |
| | import warnings |
| | from collections.abc import Callable, Sequence |
| | from pathlib import Path |
| | from typing import Any |
| |
|
| | import numpy as np |
| |
|
| | from monai.apps.tcia import ( |
| | DCM_FILENAME_REGEX, |
| | download_tcia_series_instance, |
| | get_tcia_metadata, |
| | get_tcia_ref_uid, |
| | match_tcia_ref_uid_in_study, |
| | ) |
| | from monai.apps.utils import download_and_extract |
| | from monai.config.type_definitions import PathLike |
| | from monai.data import ( |
| | CacheDataset, |
| | PydicomReader, |
| | load_decathlon_datalist, |
| | load_decathlon_properties, |
| | partition_dataset, |
| | select_cross_validation_folds, |
| | ) |
| | from monai.transforms import LoadImaged, Randomizable |
| | from monai.utils import ensure_tuple |
| |
|
| | __all__ = ["MedNISTDataset", "DecathlonDataset", "CrossValidation", "TciaDataset"] |
| |
|
| |
|
| | class MedNISTDataset(Randomizable, CacheDataset): |
| | """ |
| | The Dataset to automatically download MedNIST data and generate items for training, validation or test. |
| | It's based on `CacheDataset` to accelerate the training process. |
| | |
| | Args: |
| | root_dir: target directory to download and load MedNIST dataset. |
| | section: expected data section, can be: `training`, `validation` or `test`. |
| | transform: transforms to execute operations on input data. |
| | download: whether to download and extract the MedNIST from resource link, default is False. |
| | if expected file already exists, skip downloading even set it to True. |
| | user can manually copy `MedNIST.tar.gz` file or `MedNIST` folder to root directory. |
| | seed: random seed to randomly split training, validation and test datasets, default is 0. |
| | val_frac: percentage of validation fraction in the whole dataset, default is 0.1. |
| | test_frac: percentage of test fraction in the whole dataset, default is 0.1. |
| | cache_num: number of items to be cached. Default is `sys.maxsize`. |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | cache_rate: percentage of cached data in total, default is 1.0 (cache all). |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | num_workers: the number of worker threads if computing cache in the initialization. |
| | If num_workers is None then the number returned by os.cpu_count() is used. |
| | If a value less than 1 is specified, 1 will be used instead. |
| | progress: whether to display a progress bar when downloading dataset and computing the transform cache content. |
| | copy_cache: whether to `deepcopy` the cache content before applying the random transforms, |
| | default to `True`. if the random transforms don't modify the cached content |
| | (for example, randomly crop from the cached image and deepcopy the crop region) |
| | or if every cache item is only used once in a `multi-processing` environment, |
| | may set `copy=False` for better performance. |
| | as_contiguous: whether to convert the cached NumPy array or PyTorch tensor to be contiguous. |
| | it may help improve the performance of following logic. |
| | runtime_cache: whether to compute cache at the runtime, default to `False` to prepare |
| | the cache content at initialization. See: :py:class:`monai.data.CacheDataset`. |
| | |
| | Raises: |
| | ValueError: When ``root_dir`` is not a directory. |
| | RuntimeError: When ``dataset_dir`` doesn't exist and downloading is not selected (``download=False``). |
| | |
| | """ |
| |
|
| | resource = "https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/MedNIST.tar.gz" |
| | md5 = "0bc7306e7427e00ad1c5526a6677552d" |
| | compressed_file_name = "MedNIST.tar.gz" |
| | dataset_folder_name = "MedNIST" |
| |
|
| | def __init__( |
| | self, |
| | root_dir: PathLike, |
| | section: str, |
| | transform: Sequence[Callable] | Callable = (), |
| | download: bool = False, |
| | seed: int = 0, |
| | val_frac: float = 0.1, |
| | test_frac: float = 0.1, |
| | cache_num: int = sys.maxsize, |
| | cache_rate: float = 1.0, |
| | num_workers: int | None = 1, |
| | progress: bool = True, |
| | copy_cache: bool = True, |
| | as_contiguous: bool = True, |
| | runtime_cache: bool = False, |
| | ) -> None: |
| | root_dir = Path(root_dir) |
| | if not root_dir.is_dir(): |
| | raise ValueError("Root directory root_dir must be a directory.") |
| | self.section = section |
| | self.val_frac = val_frac |
| | self.test_frac = test_frac |
| | self.set_random_state(seed=seed) |
| | tarfile_name = root_dir / self.compressed_file_name |
| | dataset_dir = root_dir / self.dataset_folder_name |
| | self.num_class = 0 |
| | if download: |
| | download_and_extract( |
| | url=self.resource, |
| | filepath=tarfile_name, |
| | output_dir=root_dir, |
| | hash_val=self.md5, |
| | hash_type="md5", |
| | progress=progress, |
| | ) |
| |
|
| | if not dataset_dir.is_dir(): |
| | raise RuntimeError( |
| | f"Cannot find dataset directory: {dataset_dir}, please use download=True to download it." |
| | ) |
| | data = self._generate_data_list(dataset_dir) |
| | if transform == (): |
| | transform = LoadImaged("image") |
| | CacheDataset.__init__( |
| | self, |
| | data=data, |
| | transform=transform, |
| | cache_num=cache_num, |
| | cache_rate=cache_rate, |
| | num_workers=num_workers, |
| | progress=progress, |
| | copy_cache=copy_cache, |
| | as_contiguous=as_contiguous, |
| | runtime_cache=runtime_cache, |
| | ) |
| |
|
| | def randomize(self, data: np.ndarray) -> None: |
| | self.R.shuffle(data) |
| |
|
| | def get_num_classes(self) -> int: |
| | """Get number of classes.""" |
| | return self.num_class |
| |
|
| | def _generate_data_list(self, dataset_dir: PathLike) -> list[dict]: |
| | """ |
| | Raises: |
| | ValueError: When ``section`` is not one of ["training", "validation", "test"]. |
| | |
| | """ |
| | dataset_dir = Path(dataset_dir) |
| | class_names = sorted(f"{x.name}" for x in dataset_dir.iterdir() if x.is_dir()) |
| | self.num_class = len(class_names) |
| | image_files = [[f"{x}" for x in (dataset_dir / class_names[i]).iterdir()] for i in range(self.num_class)] |
| | num_each = [len(image_files[i]) for i in range(self.num_class)] |
| | image_files_list = [] |
| | image_class = [] |
| | class_name = [] |
| | for i in range(self.num_class): |
| | image_files_list.extend(image_files[i]) |
| | image_class.extend([i] * num_each[i]) |
| | class_name.extend([class_names[i]] * num_each[i]) |
| |
|
| | length = len(image_files_list) |
| | indices = np.arange(length) |
| | self.randomize(indices) |
| |
|
| | test_length = int(length * self.test_frac) |
| | val_length = int(length * self.val_frac) |
| | if self.section == "test": |
| | section_indices = indices[:test_length] |
| | elif self.section == "validation": |
| | section_indices = indices[test_length : test_length + val_length] |
| | elif self.section == "training": |
| | section_indices = indices[test_length + val_length :] |
| | else: |
| | raise ValueError( |
| | f'Unsupported section: {self.section}, available options are ["training", "validation", "test"].' |
| | ) |
| | |
| | return [ |
| | {"image": image_files_list[i], "label": image_class[i], "class_name": class_name[i]} |
| | for i in section_indices |
| | ] |
| |
|
| |
|
| | class DecathlonDataset(Randomizable, CacheDataset): |
| | """ |
| | The Dataset to automatically download the data of Medical Segmentation Decathlon challenge |
| | (http://medicaldecathlon.com/) and generate items for training, validation or test. |
| | It will also load these properties from the JSON config file of dataset. user can call `get_properties()` |
| | to get specified properties or all the properties loaded. |
| | It's based on :py:class:`monai.data.CacheDataset` to accelerate the training process. |
| | |
| | Args: |
| | root_dir: user's local directory for caching and loading the MSD datasets. |
| | task: which task to download and execute: one of list ("Task01_BrainTumour", "Task02_Heart", |
| | "Task03_Liver", "Task04_Hippocampus", "Task05_Prostate", "Task06_Lung", "Task07_Pancreas", |
| | "Task08_HepaticVessel", "Task09_Spleen", "Task10_Colon"). |
| | section: expected data section, can be: `training`, `validation` or `test`. |
| | transform: transforms to execute operations on input data. |
| | for further usage, use `EnsureChannelFirstd` to convert the shape to [C, H, W, D]. |
| | download: whether to download and extract the Decathlon from resource link, default is False. |
| | if expected file already exists, skip downloading even set it to True. |
| | user can manually copy tar file or dataset folder to the root directory. |
| | val_frac: percentage of validation fraction in the whole dataset, default is 0.2. |
| | seed: random seed to randomly shuffle the datalist before splitting into training and validation, default is 0. |
| | note to set same seed for `training` and `validation` sections. |
| | cache_num: number of items to be cached. Default is `sys.maxsize`. |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | cache_rate: percentage of cached data in total, default is 1.0 (cache all). |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | num_workers: the number of worker threads if computing cache in the initialization. |
| | If num_workers is None then the number returned by os.cpu_count() is used. |
| | If a value less than 1 is specified, 1 will be used instead. |
| | progress: whether to display a progress bar when downloading dataset and computing the transform cache content. |
| | copy_cache: whether to `deepcopy` the cache content before applying the random transforms, |
| | default to `True`. if the random transforms don't modify the cached content |
| | (for example, randomly crop from the cached image and deepcopy the crop region) |
| | or if every cache item is only used once in a `multi-processing` environment, |
| | may set `copy=False` for better performance. |
| | as_contiguous: whether to convert the cached NumPy array or PyTorch tensor to be contiguous. |
| | it may help improve the performance of following logic. |
| | runtime_cache: whether to compute cache at the runtime, default to `False` to prepare |
| | the cache content at initialization. See: :py:class:`monai.data.CacheDataset`. |
| | |
| | Raises: |
| | ValueError: When ``root_dir`` is not a directory. |
| | ValueError: When ``task`` is not one of ["Task01_BrainTumour", "Task02_Heart", |
| | "Task03_Liver", "Task04_Hippocampus", "Task05_Prostate", "Task06_Lung", "Task07_Pancreas", |
| | "Task08_HepaticVessel", "Task09_Spleen", "Task10_Colon"]. |
| | RuntimeError: When ``dataset_dir`` doesn't exist and downloading is not selected (``download=False``). |
| | |
| | Example:: |
| | |
| | transform = Compose( |
| | [ |
| | LoadImaged(keys=["image", "label"]), |
| | EnsureChannelFirstd(keys=["image", "label"]), |
| | ScaleIntensityd(keys="image"), |
| | ToTensord(keys=["image", "label"]), |
| | ] |
| | ) |
| | |
| | val_data = DecathlonDataset( |
| | root_dir="./", task="Task09_Spleen", transform=transform, section="validation", seed=12345, download=True |
| | ) |
| | |
| | print(val_data[0]["image"], val_data[0]["label"]) |
| | |
| | """ |
| |
|
| | resource = { |
| | "Task01_BrainTumour": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task01_BrainTumour.tar", |
| | "Task02_Heart": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task02_Heart.tar", |
| | "Task03_Liver": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task03_Liver.tar", |
| | "Task04_Hippocampus": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task04_Hippocampus.tar", |
| | "Task05_Prostate": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task05_Prostate.tar", |
| | "Task06_Lung": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task06_Lung.tar", |
| | "Task07_Pancreas": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task07_Pancreas.tar", |
| | "Task08_HepaticVessel": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task08_HepaticVessel.tar", |
| | "Task09_Spleen": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task09_Spleen.tar", |
| | "Task10_Colon": "https://msd-for-monai.s3-us-west-2.amazonaws.com/Task10_Colon.tar", |
| | } |
| | md5 = { |
| | "Task01_BrainTumour": "240a19d752f0d9e9101544901065d872", |
| | "Task02_Heart": "06ee59366e1e5124267b774dbd654057", |
| | "Task03_Liver": "a90ec6c4aa7f6a3d087205e23d4e6397", |
| | "Task04_Hippocampus": "9d24dba78a72977dbd1d2e110310f31b", |
| | "Task05_Prostate": "35138f08b1efaef89d7424d2bcc928db", |
| | "Task06_Lung": "8afd997733c7fc0432f71255ba4e52dc", |
| | "Task07_Pancreas": "4f7080cfca169fa8066d17ce6eb061e4", |
| | "Task08_HepaticVessel": "641d79e80ec66453921d997fbf12a29c", |
| | "Task09_Spleen": "410d4a301da4e5b2f6f86ec3ddba524e", |
| | "Task10_Colon": "bad7a188931dc2f6acf72b08eb6202d0", |
| | } |
| |
|
| | def __init__( |
| | self, |
| | root_dir: PathLike, |
| | task: str, |
| | section: str, |
| | transform: Sequence[Callable] | Callable = (), |
| | download: bool = False, |
| | seed: int = 0, |
| | val_frac: float = 0.2, |
| | cache_num: int = sys.maxsize, |
| | cache_rate: float = 1.0, |
| | num_workers: int = 1, |
| | progress: bool = True, |
| | copy_cache: bool = True, |
| | as_contiguous: bool = True, |
| | runtime_cache: bool = False, |
| | ) -> None: |
| | root_dir = Path(root_dir) |
| | if not root_dir.is_dir(): |
| | raise ValueError("Root directory root_dir must be a directory.") |
| | self.section = section |
| | self.val_frac = val_frac |
| | self.set_random_state(seed=seed) |
| | if task not in self.resource: |
| | raise ValueError(f"Unsupported task: {task}, available options are: {list(self.resource.keys())}.") |
| | dataset_dir = root_dir / task |
| | tarfile_name = f"{dataset_dir}.tar" |
| | if download: |
| | download_and_extract( |
| | url=self.resource[task], |
| | filepath=tarfile_name, |
| | output_dir=root_dir, |
| | hash_val=self.md5[task], |
| | hash_type="md5", |
| | progress=progress, |
| | ) |
| |
|
| | if not dataset_dir.exists(): |
| | raise RuntimeError( |
| | f"Cannot find dataset directory: {dataset_dir}, please use download=True to download it." |
| | ) |
| | self.indices: np.ndarray = np.array([]) |
| | data = self._generate_data_list(dataset_dir) |
| | |
| | property_keys = [ |
| | "name", |
| | "description", |
| | "reference", |
| | "licence", |
| | "tensorImageSize", |
| | "modality", |
| | "labels", |
| | "numTraining", |
| | "numTest", |
| | ] |
| | self._properties = load_decathlon_properties(dataset_dir / "dataset.json", property_keys) |
| | if transform == (): |
| | transform = LoadImaged(["image", "label"]) |
| | CacheDataset.__init__( |
| | self, |
| | data=data, |
| | transform=transform, |
| | cache_num=cache_num, |
| | cache_rate=cache_rate, |
| | num_workers=num_workers, |
| | progress=progress, |
| | copy_cache=copy_cache, |
| | as_contiguous=as_contiguous, |
| | runtime_cache=runtime_cache, |
| | ) |
| |
|
| | def get_indices(self) -> np.ndarray: |
| | """ |
| | Get the indices of datalist used in this dataset. |
| | |
| | """ |
| | return self.indices |
| |
|
| | def randomize(self, data: np.ndarray) -> None: |
| | self.R.shuffle(data) |
| |
|
| | def get_properties(self, keys: Sequence[str] | str | None = None) -> dict: |
| | """ |
| | Get the loaded properties of dataset with specified keys. |
| | If no keys specified, return all the loaded properties. |
| | |
| | """ |
| | if keys is None: |
| | return self._properties |
| | if self._properties is not None: |
| | return {key: self._properties[key] for key in ensure_tuple(keys)} |
| | return {} |
| |
|
| | def _generate_data_list(self, dataset_dir: PathLike) -> list[dict]: |
| | |
| | dataset_dir = Path(dataset_dir) |
| | section = "training" if self.section in ["training", "validation"] else "test" |
| | datalist = load_decathlon_datalist(dataset_dir / "dataset.json", True, section) |
| | return self._split_datalist(datalist) |
| |
|
| | def _split_datalist(self, datalist: list[dict]) -> list[dict]: |
| | if self.section == "test": |
| | return datalist |
| | length = len(datalist) |
| | indices = np.arange(length) |
| | self.randomize(indices) |
| |
|
| | val_length = int(length * self.val_frac) |
| | if self.section == "training": |
| | self.indices = indices[val_length:] |
| | else: |
| | self.indices = indices[:val_length] |
| |
|
| | return [datalist[i] for i in self.indices] |
| |
|
| |
|
| | class TciaDataset(Randomizable, CacheDataset): |
| | """ |
| | The Dataset to automatically download the data from a public The Cancer Imaging Archive (TCIA) dataset |
| | and generate items for training, validation or test. |
| | |
| | The Highdicom library is used to load dicom data with modality "SEG", but only a part of collections are |
| | supported, such as: "C4KC-KiTS", "NSCLC-Radiomics", "NSCLC-Radiomics-Interobserver1", " QIN-PROSTATE-Repeatability" |
| | and "PROSTATEx". Therefore, if "seg" is included in `keys` of the `LoadImaged` transform and loading some |
| | other collections, errors may be raised. For supported collections, the original "SEG" information may not |
| | always be consistent for each dicom file. Therefore, to avoid creating different format of labels, please use |
| | the `label_dict` argument of `PydicomReader` when calling the `LoadImaged` transform. The prepared label dicts |
| | of collections that are mentioned above is also saved in: `monai.apps.tcia.TCIA_LABEL_DICT`. You can also refer |
| | to the second example bellow. |
| | |
| | |
| | This class is based on :py:class:`monai.data.CacheDataset` to accelerate the training process. |
| | |
| | Args: |
| | root_dir: user's local directory for caching and loading the TCIA dataset. |
| | collection: name of a TCIA collection. |
| | a TCIA dataset is defined as a collection. Please check the following list to browse |
| | the collection list (only public collections can be downloaded): |
| | https://www.cancerimagingarchive.net/collections/ |
| | section: expected data section, can be: `training`, `validation` or `test`. |
| | transform: transforms to execute operations on input data. |
| | for further usage, use `EnsureChannelFirstd` to convert the shape to [C, H, W, D]. |
| | If not specified, `LoadImaged(reader="PydicomReader", keys=["image"])` will be used as the default |
| | transform. In addition, we suggest to set the argument `labels` for `PydicomReader` if segmentations |
| | are needed to be loaded. The original labels for each dicom series may be different, using this argument |
| | is able to unify the format of labels. |
| | download: whether to download and extract the dataset, default is False. |
| | if expected file already exists, skip downloading even set it to True. |
| | user can manually copy tar file or dataset folder to the root directory. |
| | download_len: number of series that will be downloaded, the value should be larger than 0 or -1, where -1 means |
| | all series will be downloaded. Default is -1. |
| | seg_type: modality type of segmentation that is used to do the first step download. Default is "SEG". |
| | modality_tag: tag of modality. Default is (0x0008, 0x0060). |
| | ref_series_uid_tag: tag of referenced Series Instance UID. Default is (0x0020, 0x000e). |
| | ref_sop_uid_tag: tag of referenced SOP Instance UID. Default is (0x0008, 0x1155). |
| | specific_tags: tags that will be loaded for "SEG" series. This argument will be used in |
| | `monai.data.PydicomReader`. Default is [(0x0008, 0x1115), (0x0008,0x1140), (0x3006, 0x0010), |
| | (0x0020,0x000D), (0x0010,0x0010), (0x0010,0x0020), (0x0020,0x0011), (0x0020,0x0012)]. |
| | fname_regex: a regular expression to match the file names when the input is a folder. |
| | If provided, only the matched files will be included. For example, to include the file name |
| | "image_0001.dcm", the regular expression could be `".*image_(\\d+).dcm"`. |
| | Default to `"^(?!.*LICENSE).*"`, ignoring any file name containing `"LICENSE"`. |
| | val_frac: percentage of validation fraction in the whole dataset, default is 0.2. |
| | seed: random seed to randomly shuffle the datalist before splitting into training and validation, default is 0. |
| | note to set same seed for `training` and `validation` sections. |
| | cache_num: number of items to be cached. Default is `sys.maxsize`. |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | cache_rate: percentage of cached data in total, default is 0.0 (no cache). |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | num_workers: the number of worker threads if computing cache in the initialization. |
| | If num_workers is None then the number returned by os.cpu_count() is used. |
| | If a value less than 1 is specified, 1 will be used instead. |
| | progress: whether to display a progress bar when downloading dataset and computing the transform cache content. |
| | copy_cache: whether to `deepcopy` the cache content before applying the random transforms, |
| | default to `True`. if the random transforms don't modify the cached content |
| | (for example, randomly crop from the cached image and deepcopy the crop region) |
| | or if every cache item is only used once in a `multi-processing` environment, |
| | may set `copy=False` for better performance. |
| | as_contiguous: whether to convert the cached NumPy array or PyTorch tensor to be contiguous. |
| | it may help improve the performance of following logic. |
| | runtime_cache: whether to compute cache at the runtime, default to `False` to prepare |
| | the cache content at initialization. See: :py:class:`monai.data.CacheDataset`. |
| | |
| | Example:: |
| | |
| | # collection is "Pancreatic-CT-CBCT-SEG", seg_type is "RTSTRUCT" |
| | data = TciaDataset( |
| | root_dir="./", collection="Pancreatic-CT-CBCT-SEG", seg_type="RTSTRUCT", download=True |
| | ) |
| | |
| | # collection is "C4KC-KiTS", seg_type is "SEG", and load both images and segmentations |
| | from monai.apps.tcia import TCIA_LABEL_DICT |
| | transform = Compose( |
| | [ |
| | LoadImaged(reader="PydicomReader", keys=["image", "seg"], label_dict=TCIA_LABEL_DICT["C4KC-KiTS"]), |
| | EnsureChannelFirstd(keys=["image", "seg"]), |
| | ResampleToMatchd(keys="image", key_dst="seg"), |
| | ] |
| | ) |
| | data = TciaDataset( |
| | root_dir="./", collection="C4KC-KiTS", section="validation", seed=12345, download=True |
| | ) |
| | |
| | print(data[0]["seg"].shape) |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | root_dir: PathLike, |
| | collection: str, |
| | section: str, |
| | transform: Sequence[Callable] | Callable = (), |
| | download: bool = False, |
| | download_len: int = -1, |
| | seg_type: str = "SEG", |
| | modality_tag: tuple = (0x0008, 0x0060), |
| | ref_series_uid_tag: tuple = (0x0020, 0x000E), |
| | ref_sop_uid_tag: tuple = (0x0008, 0x1155), |
| | specific_tags: tuple = ( |
| | (0x0008, 0x1115), |
| | (0x0008, 0x1140), |
| | (0x3006, 0x0010), |
| | (0x0020, 0x000D), |
| | (0x0010, 0x0010), |
| | (0x0010, 0x0020), |
| | (0x0020, 0x0011), |
| | (0x0020, 0x0012), |
| | ), |
| | fname_regex: str = DCM_FILENAME_REGEX, |
| | seed: int = 0, |
| | val_frac: float = 0.2, |
| | cache_num: int = sys.maxsize, |
| | cache_rate: float = 0.0, |
| | num_workers: int = 1, |
| | progress: bool = True, |
| | copy_cache: bool = True, |
| | as_contiguous: bool = True, |
| | runtime_cache: bool = False, |
| | ) -> None: |
| | root_dir = Path(root_dir) |
| | if not root_dir.is_dir(): |
| | raise ValueError("Root directory root_dir must be a directory.") |
| |
|
| | self.section = section |
| | self.val_frac = val_frac |
| | self.seg_type = seg_type |
| | self.modality_tag = modality_tag |
| | self.ref_series_uid_tag = ref_series_uid_tag |
| | self.ref_sop_uid_tag = ref_sop_uid_tag |
| |
|
| | self.set_random_state(seed=seed) |
| | download_dir = os.path.join(root_dir, collection) |
| | load_tags = list(specific_tags) |
| | load_tags += [modality_tag] |
| | self.load_tags = load_tags |
| | if download: |
| | seg_series_list = get_tcia_metadata( |
| | query=f"getSeries?Collection={collection}&Modality={seg_type}", attribute="SeriesInstanceUID" |
| | ) |
| | if download_len > 0: |
| | seg_series_list = seg_series_list[:download_len] |
| | if len(seg_series_list) == 0: |
| | raise ValueError(f"Cannot find data with collection: {collection} seg_type: {seg_type}") |
| | for series_uid in seg_series_list: |
| | self._download_series_reference_data(series_uid, download_dir) |
| |
|
| | if not os.path.exists(download_dir): |
| | raise RuntimeError(f"Cannot find dataset directory: {download_dir}.") |
| | self.fname_regex = fname_regex |
| |
|
| | self.indices: np.ndarray = np.array([]) |
| | self.datalist = self._generate_data_list(download_dir) |
| |
|
| | if transform == (): |
| | transform = LoadImaged(keys=["image"], reader="PydicomReader", fname_regex=self.fname_regex) |
| | CacheDataset.__init__( |
| | self, |
| | data=self.datalist, |
| | transform=transform, |
| | cache_num=cache_num, |
| | cache_rate=cache_rate, |
| | num_workers=num_workers, |
| | progress=progress, |
| | copy_cache=copy_cache, |
| | as_contiguous=as_contiguous, |
| | runtime_cache=runtime_cache, |
| | ) |
| |
|
| | def get_indices(self) -> np.ndarray: |
| | """ |
| | Get the indices of datalist used in this dataset. |
| | |
| | """ |
| | return self.indices |
| |
|
| | def randomize(self, data: np.ndarray) -> None: |
| | self.R.shuffle(data) |
| |
|
| | def _download_series_reference_data(self, series_uid: str, download_dir: str) -> None: |
| | """ |
| | First of all, download a series from TCIA according to `series_uid`. |
| | Then find all referenced series and download. |
| | """ |
| | seg_first_dir = os.path.join(download_dir, "raw", series_uid) |
| | download_tcia_series_instance( |
| | series_uid=series_uid, download_dir=download_dir, output_dir=seg_first_dir, check_md5=False |
| | ) |
| | dicom_files = [f for f in sorted(os.listdir(seg_first_dir)) if f.endswith(".dcm")] |
| | |
| | dcm_path = os.path.join(seg_first_dir, dicom_files[0]) |
| | ds = PydicomReader(stop_before_pixels=True, specific_tags=self.load_tags).read(dcm_path) |
| | |
| | patient_id = ds.PatientID if ds.PatientID else ds.PatientName |
| | if not patient_id: |
| | warnings.warn(f"unable to find patient name of dicom file: {dcm_path}, use 'patient' instead.") |
| | patient_id = "patient" |
| | |
| | series_num = ds.SeriesNumber if ds.SeriesNumber else ds.AcquisitionNumber |
| | if not series_num: |
| | warnings.warn(f"unable to find series number of dicom file: {dcm_path}, use '0' instead.") |
| | series_num = 0 |
| |
|
| | series_num = str(series_num) |
| | seg_dir = os.path.join(download_dir, patient_id, series_num, self.seg_type.lower()) |
| | dcm_dir = os.path.join(download_dir, patient_id, series_num, "image") |
| |
|
| | |
| | ref_uid_list = [] |
| | for dcm_file in dicom_files: |
| | dcm_path = os.path.join(seg_first_dir, dcm_file) |
| | ds = PydicomReader(stop_before_pixels=True, specific_tags=self.load_tags).read(dcm_path) |
| | if ds[self.modality_tag].value == self.seg_type: |
| | ref_uid = get_tcia_ref_uid( |
| | ds, find_sop=False, ref_series_uid_tag=self.ref_series_uid_tag, ref_sop_uid_tag=self.ref_sop_uid_tag |
| | ) |
| | if ref_uid == "": |
| | ref_sop_uid = get_tcia_ref_uid( |
| | ds, |
| | find_sop=True, |
| | ref_series_uid_tag=self.ref_series_uid_tag, |
| | ref_sop_uid_tag=self.ref_sop_uid_tag, |
| | ) |
| | ref_uid = match_tcia_ref_uid_in_study(ds.StudyInstanceUID, ref_sop_uid) |
| | if ref_uid != "": |
| | ref_uid_list.append(ref_uid) |
| | if not ref_uid_list: |
| | warnings.warn(f"Cannot find the referenced Series Instance UID from series: {series_uid}.") |
| | else: |
| | download_tcia_series_instance( |
| | series_uid=ref_uid_list[0], download_dir=download_dir, output_dir=dcm_dir, check_md5=False |
| | ) |
| | if not os.path.exists(seg_dir): |
| | shutil.copytree(seg_first_dir, seg_dir) |
| |
|
| | def _generate_data_list(self, dataset_dir: PathLike) -> list[dict]: |
| | |
| | dataset_dir = Path(dataset_dir) |
| | datalist = [] |
| | patient_list = [f.name for f in os.scandir(dataset_dir) if f.is_dir() and f.name != "raw"] |
| | for patient_id in patient_list: |
| | series_list = [f.name for f in os.scandir(os.path.join(dataset_dir, patient_id)) if f.is_dir()] |
| | for series_num in series_list: |
| | seg_key = self.seg_type.lower() |
| | image_path = os.path.join(dataset_dir, patient_id, series_num, "image") |
| | mask_path = os.path.join(dataset_dir, patient_id, series_num, seg_key) |
| |
|
| | if os.path.exists(image_path): |
| | datalist.append({"image": image_path, seg_key: mask_path}) |
| | else: |
| | datalist.append({seg_key: mask_path}) |
| |
|
| | return self._split_datalist(datalist) |
| |
|
| | def _split_datalist(self, datalist: list[dict]) -> list[dict]: |
| | if self.section == "test": |
| | return datalist |
| | length = len(datalist) |
| | indices = np.arange(length) |
| | self.randomize(indices) |
| |
|
| | val_length = int(length * self.val_frac) |
| | if self.section == "training": |
| | self.indices = indices[val_length:] |
| | else: |
| | self.indices = indices[:val_length] |
| |
|
| | return [datalist[i] for i in self.indices] |
| |
|
| |
|
| | class CrossValidation: |
| | """ |
| | Cross validation dataset based on the general dataset which must have `_split_datalist` API. |
| | |
| | Args: |
| | dataset_cls: dataset class to be used to create the cross validation partitions. |
| | It must have `_split_datalist` API. |
| | nfolds: number of folds to split the data for cross validation. |
| | seed: random seed to randomly shuffle the datalist before splitting into N folds, default is 0. |
| | dataset_params: other additional parameters for the dataset_cls base class. |
| | |
| | Example of 5 folds cross validation training:: |
| | |
| | cvdataset = CrossValidation( |
| | dataset_cls=DecathlonDataset, |
| | nfolds=5, |
| | seed=12345, |
| | root_dir="./", |
| | task="Task09_Spleen", |
| | section="training", |
| | transform=train_transform, |
| | download=True, |
| | ) |
| | dataset_fold0_train = cvdataset.get_dataset(folds=[1, 2, 3, 4]) |
| | dataset_fold0_val = cvdataset.get_dataset(folds=0, transform=val_transform, download=False) |
| | # execute training for fold 0 ... |
| | |
| | dataset_fold1_train = cvdataset.get_dataset(folds=[0, 2, 3, 4]) |
| | dataset_fold1_val = cvdataset.get_dataset(folds=1, transform=val_transform, download=False) |
| | # execute training for fold 1 ... |
| | |
| | ... |
| | |
| | dataset_fold4_train = ... |
| | # execute training for fold 4 ... |
| | |
| | """ |
| |
|
| | def __init__(self, dataset_cls: object, nfolds: int = 5, seed: int = 0, **dataset_params: Any) -> None: |
| | if not hasattr(dataset_cls, "_split_datalist"): |
| | raise ValueError("dataset class must have _split_datalist API.") |
| | self.dataset_cls = dataset_cls |
| | self.nfolds = nfolds |
| | self.seed = seed |
| | self.dataset_params = dataset_params |
| |
|
| | def get_dataset(self, folds: Sequence[int] | int, **dataset_params: Any) -> object: |
| | """ |
| | Generate dataset based on the specified fold indices in the cross validation group. |
| | |
| | Args: |
| | folds: index of folds for training or validation, if a list of values, concatenate the data. |
| | dataset_params: other additional parameters for the dataset_cls base class, will override |
| | the same parameters in `self.dataset_params`. |
| | |
| | """ |
| | nfolds = self.nfolds |
| | seed = self.seed |
| | dataset_params_ = dict(self.dataset_params) |
| | dataset_params_.update(dataset_params) |
| |
|
| | class _NsplitsDataset(self.dataset_cls): |
| |
|
| | def _split_datalist(self, datalist: list[dict]) -> list[dict]: |
| | data = partition_dataset(data=datalist, num_partitions=nfolds, shuffle=True, seed=seed) |
| | return select_cross_validation_folds(partitions=data, folds=folds) |
| |
|
| | return _NsplitsDataset(**dataset_params_) |
| |
|