| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import math |
| | from typing import Dict, Sequence, Union |
| |
|
| | import torch |
| | from torch.utils.data import Dataset, IterableDataset |
| |
|
| | from monai.data.utils import iter_patch |
| | from monai.utils import NumpyPadMode, ensure_tuple |
| |
|
| |
|
| | class GridPatchDataset(IterableDataset): |
| | """ |
| | Yields patches from arrays read from an input dataset. The patches are chosen in a contiguous grid sampling scheme. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | dataset: Dataset, |
| | patch_size: Sequence[int], |
| | start_pos: Sequence[int] = (), |
| | mode: Union[NumpyPadMode, str] = NumpyPadMode.WRAP, |
| | **pad_opts: Dict, |
| | ) -> None: |
| | """ |
| | Initializes this dataset in terms of the input dataset and patch size. The `patch_size` is the size of the |
| | patch to sample from the input arrays. It is assumed the arrays first dimension is the channel dimension which |
| | will be yielded in its entirety so this should not be specified in `patch_size`. For example, for an input 3D |
| | array with 1 channel of size (1, 20, 20, 20) a regular grid sampling of eight patches (1, 10, 10, 10) would be |
| | specified by a `patch_size` of (10, 10, 10). |
| | |
| | Args: |
| | dataset: the dataset to read array data from |
| | patch_size: size of patches to generate slices for, 0/None selects whole dimension |
| | start_pos: starting position in the array, default is 0 for each dimension |
| | mode: {``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, ``"mean"``, |
| | ``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
| | One of the listed string values or a user supplied function. Defaults to ``"wrap"``. |
| | See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
| | pad_opts: padding options, see numpy.pad |
| | """ |
| |
|
| | self.dataset = dataset |
| | self.patch_size = (None,) + tuple(patch_size) |
| | self.start_pos = ensure_tuple(start_pos) |
| | self.mode: NumpyPadMode = NumpyPadMode(mode) |
| | self.pad_opts = pad_opts |
| |
|
| | def __iter__(self): |
| | worker_info = torch.utils.data.get_worker_info() |
| | iter_start = 0 |
| | iter_end = len(self.dataset) |
| |
|
| | if worker_info is not None: |
| | |
| | per_worker = int(math.ceil((iter_end - iter_start) / float(worker_info.num_workers))) |
| | worker_id = worker_info.id |
| | iter_start = iter_start + worker_id * per_worker |
| | iter_end = min(iter_start + per_worker, iter_end) |
| |
|
| | for index in range(iter_start, iter_end): |
| | arrays = self.dataset[index] |
| |
|
| | iters = [iter_patch(a, self.patch_size, self.start_pos, False, self.mode, **self.pad_opts) for a in arrays] |
| |
|
| | yield from zip(*iters) |
| |
|