| | import numpy as np |
| | import os |
| | from batchgenerators.utilities.file_and_folder_operations import isfile, subfiles |
| | import multiprocessing |
| |
|
| | def _convert_to_npy(npz_file: str, unpack_segmentation: bool = True, overwrite_existing: bool = False) -> None: |
| | |
| | a = np.load(npz_file) |
| | if overwrite_existing or not isfile(npz_file[:-3] + "npy"): |
| | np.save(npz_file[:-3] + "npy", a['data']) |
| | np.save(npz_file[:-4] + "_global.npy", a['data_global']) |
| | np.save(npz_file[:-4] + "_global_seg.npy", a['seg_global']) |
| |
|
| | if unpack_segmentation and (overwrite_existing or not isfile(npz_file[:-4] + "_seg.npy")): |
| | np.save(npz_file[:-4] + "_seg.npy", a['seg']) |
| |
|
| | def unpack_dataset(folder: str, unpack_segmentation: bool = True, overwrite_existing: bool = False, |
| | num_processes: int = 8): |
| | """ |
| | all npz files in this folder belong to the dataset, unpack them all |
| | """ |
| | with multiprocessing.get_context("spawn").Pool(num_processes) as p: |
| | npz_files = subfiles(folder, True, None, ".npz", True) |
| | p.starmap(_convert_to_npy, zip(npz_files, |
| | [unpack_segmentation] * len(npz_files), |
| | [overwrite_existing] * len(npz_files)) |
| | ) |
| |
|