| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | import copy |
| | import os |
| |
|
| | import numpy as np |
| |
|
| | import monai |
| | from monai.bundle import ConfigParser |
| | from monai.utils import StrEnum, ensure_tuple, optional_import |
| |
|
| | tqdm, has_tqdm = optional_import("tqdm", name="tqdm") |
| | nib, _ = optional_import("nibabel") |
| |
|
| | logger = monai.apps.utils.get_logger(__name__) |
| |
|
| | __all__ = ["analyze_data", "create_new_data_copy", "create_new_dataset_json", "NNUNETMode"] |
| |
|
| |
|
| | class NNUNETMode(StrEnum): |
| | N_2D = "2d" |
| | N_3D_FULLRES = "3d_fullres" |
| | N_3D_LOWRES = "3d_lowres" |
| | N_3D_CASCADE_FULLRES = "3d_cascade_fullres" |
| |
|
| |
|
| | def analyze_data(datalist_json: dict, data_dir: str) -> tuple[int, int]: |
| | """ |
| | Analyze (training) data |
| | |
| | Args: |
| | datalist_json: original data list .json (required by most monai tutorials). |
| | data_dir: raw data directory. |
| | """ |
| | img = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| | os.path.join(data_dir, datalist_json["training"][0]["image"]) |
| | ) |
| | num_input_channels = img.size()[0] if img.dim() == 4 else 1 |
| | logger.info(f"num_input_channels: {num_input_channels}") |
| |
|
| | num_foreground_classes = 0 |
| | for _i in range(len(datalist_json["training"])): |
| | seg = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| | os.path.join(data_dir, datalist_json["training"][_i]["label"]) |
| | ) |
| | num_foreground_classes = max(num_foreground_classes, int(seg.max())) |
| | logger.info(f"num_foreground_classes: {num_foreground_classes}") |
| |
|
| | return num_input_channels, num_foreground_classes |
| |
|
| |
|
| | def create_new_data_copy( |
| | test_key: str, datalist_json: dict, data_dir: str, num_input_channels: int, output_datafolder: str |
| | ) -> None: |
| | """ |
| | Create and organize a new copy of data to meet the requirements of nnU-Net V2 |
| | |
| | Args: |
| | test_key: key for test data in the data list .json. |
| | datalist_json: original data list .json (required by most monai tutorials). |
| | data_dir: raw data directory. |
| | num_input_channels: number of input (image) channels. |
| | output_datafolder: output folder. |
| | """ |
| | _index = 0 |
| | new_datalist_json: dict = {"training": [], test_key: []} |
| |
|
| | for _key, _folder, _label_folder in list( |
| | zip(["training", test_key], ["imagesTr", "imagesTs"], ["labelsTr", "labelsTs"]) |
| | ): |
| | if _key is None: |
| | continue |
| |
|
| | logger.info(f"converting data section: {_key}...") |
| | for _k in tqdm(range(len(datalist_json[_key]))) if has_tqdm else range(len(datalist_json[_key])): |
| | orig_img_name = ( |
| | datalist_json[_key][_k]["image"] |
| | if isinstance(datalist_json[_key][_k], dict) |
| | else datalist_json[_key][_k] |
| | ) |
| | img_name = f"case_{_index}" |
| | _index += 1 |
| |
|
| | |
| | nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| | os.path.join(data_dir, orig_img_name) |
| | ) |
| | affine = nda.meta["original_affine"] |
| | nda = nda.numpy() |
| | for _l in range(num_input_channels): |
| | outimg = nib.Nifti1Image(nda[_l, ...], affine) |
| | index = "_" + str(_l + 10000)[-4:] |
| | nib.save(outimg, os.path.join(output_datafolder, _folder, img_name + index + ".nii.gz")) |
| |
|
| | |
| | if isinstance(datalist_json[_key][_k], dict) and "label" in datalist_json[_key][_k]: |
| | nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
| | os.path.join(data_dir, datalist_json[_key][_k]["label"]) |
| | ) |
| | affine = nda.meta["original_affine"] |
| | nda = nda.numpy().astype(np.uint8) |
| | nda = nda[0, ...] if nda.ndim == 4 and nda.shape[0] == 1 else nda |
| | nib.save( |
| | nib.Nifti1Image(nda, affine), os.path.join(output_datafolder, _label_folder, img_name + ".nii.gz") |
| | ) |
| |
|
| | if isinstance(datalist_json[_key][_k], dict): |
| | _val = copy.deepcopy(datalist_json[_key][_k]) |
| | _val["new_name"] = img_name |
| | new_datalist_json[_key].append(_val) |
| | else: |
| | new_datalist_json[_key].append({"image": datalist_json[_key][_k], "new_name": img_name}) |
| |
|
| | ConfigParser.export_config_file( |
| | config=new_datalist_json, |
| | filepath=os.path.join(output_datafolder, "datalist.json"), |
| | fmt="json", |
| | sort_keys=True, |
| | indent=4, |
| | ensure_ascii=False, |
| | ) |
| |
|
| | return |
| |
|
| |
|
| | def create_new_dataset_json( |
| | modality: str, num_foreground_classes: int, num_input_channels: int, num_training_data: int, output_filepath: str |
| | ) -> None: |
| | """ |
| | Create a new copy of dataset .json to meet the requirements of nnU-Net V2 |
| | |
| | Args: |
| | modality: image modality, could a string or a list of strings. |
| | num_foreground_classes: number of foreground classes. |
| | num_input_channels: number of input (image) channels. |
| | num_training_data: number of training data. |
| | output_filepath: output file path/name. |
| | """ |
| | new_json_data: dict = {} |
| |
|
| | |
| | modality = ensure_tuple(modality) |
| |
|
| | new_json_data["channel_names"] = {} |
| | for _j in range(num_input_channels): |
| | new_json_data["channel_names"][str(_j)] = modality[_j] |
| |
|
| | new_json_data["labels"] = {} |
| | new_json_data["labels"]["background"] = 0 |
| | for _j in range(num_foreground_classes): |
| | new_json_data["labels"][f"class{_j + 1}"] = _j + 1 |
| |
|
| | |
| | new_json_data["numTraining"] = num_training_data |
| | new_json_data["file_ending"] = ".nii.gz" |
| |
|
| | ConfigParser.export_config_file( |
| | config=new_json_data, |
| | |
| | filepath=output_filepath, |
| | fmt="json", |
| | sort_keys=True, |
| | indent=4, |
| | ensure_ascii=False, |
| | ) |
| |
|
| | return |
| |
|