| | import torch |
| | import numpy as np |
| | import SimpleITK |
| | import os |
| | import sys |
| | from monai.inferers import SlidingWindowInferer |
| |
|
| | class Customalgorithm(): |
| | def __init__(self): |
| | """ |
| | Do not modify the `self.input_dir` and `self.output_dir`. |
| | (Check https://grand-challenge.org/algorithms/interfaces/) |
| | """ |
| | self.input_dir = "/input/" |
| | self.output_dir = "/output/images/head-neck-segmentation/" |
| |
|
| | |
| | self.out_spacing = [3.0, 1.0, 1.0] |
| |
|
| | |
| |
|
| | self.device = torch.device("cuda") |
| |
|
| | self.patch_size = [64, 128, 128] |
| |
|
| | def filte_state_dict(self, sd): |
| | if "module" in sd : |
| | sd = sd["module"] |
| | new_sd = {} |
| | for k, v in sd.items(): |
| | k = str(k) |
| | new_k = k[7:] if k.startswith("module") else k |
| | new_sd[new_k] = v |
| | del sd |
| | return new_sd |
| |
|
| | def convert_mha_to_nii(self, mha_input_path, nii_out_path): |
| | img = SimpleITK.ReadImage(mha_input_path) |
| | print(img.GetSize()) |
| | SimpleITK.WriteImage(img, nii_out_path, True) |
| |
|
| | def convert_nii_to_mha(self, nii_input_path, mha_out_path): |
| | img = SimpleITK.ReadImage(nii_input_path) |
| | SimpleITK.WriteImage(img, mha_out_path, True) |
| |
|
| | def read(self, mha_path): |
| | img = SimpleITK.ReadImage(mha_path) |
| | spacing = img.GetSpacing() |
| | raw_size = SimpleITK.GetArrayFromImage(img).shape |
| | img = SimpleITK.GetArrayFromImage(img)[None,].astype(np.float32) |
| | properties = { |
| | "spacing": spacing, |
| | "raw_size": raw_size |
| | } |
| | return img, properties |
| |
|
| | def check_gpu(self): |
| | """ |
| | Check if GPU is available. Note that the Grand Challenge only has one available GPU. |
| | """ |
| | print('Checking GPU availability') |
| | is_available = torch.cuda.is_available() |
| | print('Available: ' + str(is_available)) |
| | print(f'Device count: {torch.cuda.device_count()}') |
| | if is_available: |
| | print(f'Current device: {torch.cuda.current_device()}') |
| | print('Device name: ' + torch.cuda.get_device_name(0)) |
| | print('Device memory: ' + |
| | str(torch.cuda.get_device_properties(0).total_memory)) |
| |
|
| | def load_inputs(self): |
| | """ |
| | Read input data (two modalities) from `self.input_dir` (/input/). |
| | Please do not modify the path for CT and contrast-CT images. |
| | """ |
| | ct_mha = os.listdir(os.path.join(self.input_dir, 'images/head-neck-ct/'))[0] |
| | ctc_mha = os.listdir(os.path.join(self.input_dir, 'images/head-neck-contrast-enhanced-ct/'))[0] |
| | uuid = os.path.splitext(ct_mha)[0] |
| |
|
| | img, properties = self.read(os.path.join(self.input_dir, 'images/head-neck-ct/', ct_mha)) |
| | img_c, _ = self.read(os.path.join(self.input_dir, 'images/head-neck-contrast-enhanced-ct/', ctc_mha)) |
| |
|
| | data = np.concatenate([img, img_c], axis=0) |
| | del img |
| | del img_c |
| | |
| | return uuid, data, properties |
| |
|
| | def crop(self, data, properties): |
| | from light_training.preprocessing.cropping.cropping import crop_to_nonzero |
| |
|
| | seg = np.zeros_like(data) |
| |
|
| | shape_before_cropping = data.shape[1:] |
| | |
| | properties['shape_before_cropping'] = shape_before_cropping |
| | |
| | data, seg, bbox = crop_to_nonzero(data, seg) |
| | del seg |
| |
|
| | properties['bbox_used_for_cropping'] = bbox |
| |
|
| | return data, properties |
| |
|
| | def resample(self, data, properties): |
| | from light_training.preprocessing.resampling.default_resampling import compute_new_shape, resample_data_or_seg_to_shape |
| | |
| | shape_before_resample = data.shape[1:] |
| | properties['shape_after_cropping_before_resample'] = shape_before_resample |
| | new_shape = compute_new_shape(data.shape[1:], original_spacing_trans, self.out_spacing) |
| |
|
| | assert len(data.shape) == 4 |
| |
|
| | data = resample_data_or_seg_to_shape(data, new_shape, |
| | original_spacing, |
| | self.out_spacing, |
| | order=3, |
| | order_z=0) |
| | properties['shape_after_resample'] = new_shape |
| |
|
| | return data, properties |
| | |
| | def preprocess(self, data, properties, crop_first=True): |
| | from light_training.process_framework.norm import norm_func |
| |
|
| | original_spacing = list(properties['spacing']) |
| | |
| | original_spacing_trans = original_spacing[::-1] |
| | properties["original_spacing_trans"] = original_spacing_trans |
| | properties["target_spacing_trans"] = self.out_spacing |
| |
|
| | if crop_first: |
| | data, properties = self.crop(data, properties) |
| |
|
| | data = norm_func(data) |
| | |
| | if not crop_first: |
| | data, properties = self.crop(data, properties) |
| |
|
| | |
| | data, properties = self.resample(data, properties) |
| |
|
| | data = data[None,] |
| |
|
| | data = torch.from_numpy(data) |
| |
|
| | return data, properties |
| |
|
| | def predict(self, data, properties, uid): |
| | torch.cuda.empty_cache() |
| |
|
| | from models.nnunet3d import NNUNetWrapper |
| | model = NNUNetWrapper(norm="ins") |
| |
|
| | new_sd = self.filte_state_dict(torch.load("./weight/unet3d_0_addaug_bs2_ep1000_ds_gpu4/final_model_0.8552.pt", map_location="cpu")) |
| | model.load_state_dict(new_sd) |
| |
|
| | del new_sd |
| | torch.cuda.empty_cache() |
| | |
| | |
| | model.eval() |
| | window_infer = SlidingWindowInferer(roi_size=self.patch_size, |
| | sw_batch_size=1, |
| | overlap=0.5, |
| | progress=True, |
| | mode="gaussian") |
| | |
| | predictor = Predictor(window_infer, mirror_axes=None) |
| | try: |
| | ensemble_output = predictor.maybe_mirror_and_predict(data, model, self.device) |
| |
|
| | except RuntimeError: |
| | ensemble_output = predictor.maybe_mirror_and_predict(data, model, torch.device("cpu")) |
| | torch.cuda.empty_cache() |
| | del model |
| | del data |
| |
|
| | print(f"prediction done") |
| | ensemble_output = predictor.predict_raw_probability(ensemble_output, properties) |
| | print(f"non linear....") |
| | |
| | ensemble_output = ensemble_output > 0 |
| |
|
| | print(f"restore crop...") |
| | ensemble_output = predictor.predict_noncrop_probability(ensemble_output, properties) |
| |
|
| | raw_spacing = properties["spacing"] |
| | case_name = uid |
| | print(f"uuid is {uid}") |
| | os.makedirs(os.path.dirname(self.output_dir), exist_ok=True) |
| |
|
| | print(f"saving....") |
| | predictor.save_to_nii_multi_organ(ensemble_output, |
| | raw_spacing, |
| | save_dir=self.output_dir, |
| | case_name=case_name, |
| | postprocess=False) |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | def post_process(self): |
| | self.check_gpu() |
| | print('Start processing') |
| | uuid, data, properties = self.load_inputs() |
| |
|
| | data, properties = self.preprocess(data, properties) |
| | print(properties) |
| | print('Start prediction') |
| | self.predict(data, properties, uuid) |
| | |
| | |
| |
|
| | def process(self): |
| | """ |
| | Read inputs from /input, process with your algorithm and write to /output |
| | """ |
| | self.post_process() |
| |
|
| |
|
| | if __name__ == "__main__": |
| | Customalgorithm().process() |
| |
|