nnUNet_calvingfront_detection / nnunet /dataset_conversion /Task029_LiverTumorSegmentationChallenge.py
ho11laqe's picture
init
ecf08bc
raw history blame
No virus
4.54 kB
# Copyright 2020 Division of Medical Image Computing, German Cancer Research Center (DKFZ), Heidelberg, Germany
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from collections import OrderedDict
import SimpleITK as sitk
from batchgenerators.utilities.file_and_folder_operations import *
from multiprocessing import Pool
import numpy as np
from nnunet.configuration import default_num_threads
from scipy.ndimage import label
def export_segmentations(indir, outdir):
niftis = subfiles(indir, suffix='nii.gz', join=False)
for n in niftis:
identifier = str(n.split("_")[-1][:-7])
outfname = join(outdir, "test-segmentation-%s.nii" % identifier)
img = sitk.ReadImage(join(indir, n))
sitk.WriteImage(img, outfname)
def export_segmentations_postprocess(indir, outdir):
maybe_mkdir_p(outdir)
niftis = subfiles(indir, suffix='nii.gz', join=False)
for n in niftis:
print("\n", n)
identifier = str(n.split("_")[-1][:-7])
outfname = join(outdir, "test-segmentation-%s.nii" % identifier)
img = sitk.ReadImage(join(indir, n))
img_npy = sitk.GetArrayFromImage(img)
lmap, num_objects = label((img_npy > 0).astype(int))
sizes = []
for o in range(1, num_objects + 1):
sizes.append((lmap == o).sum())
mx = np.argmax(sizes) + 1
print(sizes)
img_npy[lmap != mx] = 0
img_new = sitk.GetImageFromArray(img_npy)
img_new.CopyInformation(img)
sitk.WriteImage(img_new, outfname)
if __name__ == "__main__":
train_dir = "/media/fabian/DeepLearningData/tmp/LITS-Challenge-Train-Data"
test_dir = "/media/fabian/My Book/datasets/LiTS/test_data"
output_folder = "/media/fabian/My Book/MedicalDecathlon/MedicalDecathlon_raw_splitted/Task029_LITS"
img_dir = join(output_folder, "imagesTr")
lab_dir = join(output_folder, "labelsTr")
img_dir_te = join(output_folder, "imagesTs")
maybe_mkdir_p(img_dir)
maybe_mkdir_p(lab_dir)
maybe_mkdir_p(img_dir_te)
def load_save_train(args):
data_file, seg_file = args
pat_id = data_file.split("/")[-1]
pat_id = "train_" + pat_id.split("-")[-1][:-4]
img_itk = sitk.ReadImage(data_file)
sitk.WriteImage(img_itk, join(img_dir, pat_id + "_0000.nii.gz"))
img_itk = sitk.ReadImage(seg_file)
sitk.WriteImage(img_itk, join(lab_dir, pat_id + ".nii.gz"))
return pat_id
def load_save_test(args):
data_file = args
pat_id = data_file.split("/")[-1]
pat_id = "test_" + pat_id.split("-")[-1][:-4]
img_itk = sitk.ReadImage(data_file)
sitk.WriteImage(img_itk, join(img_dir_te, pat_id + "_0000.nii.gz"))
return pat_id
nii_files_tr_data = subfiles(train_dir, True, "volume", "nii", True)
nii_files_tr_seg = subfiles(train_dir, True, "segmen", "nii", True)
nii_files_ts = subfiles(test_dir, True, "test-volume", "nii", True)
p = Pool(default_num_threads)
train_ids = p.map(load_save_train, zip(nii_files_tr_data, nii_files_tr_seg))
test_ids = p.map(load_save_test, nii_files_ts)
p.close()
p.join()
json_dict = OrderedDict()
json_dict['name'] = "LITS"
json_dict['description'] = "LITS"
json_dict['tensorImageSize'] = "4D"
json_dict['reference'] = "see challenge website"
json_dict['licence'] = "see challenge website"
json_dict['release'] = "0.0"
json_dict['modality'] = {
"0": "CT"
}
json_dict['labels'] = {
"0": "background",
"1": "liver",
"2": "tumor"
}
json_dict['numTraining'] = len(train_ids)
json_dict['numTest'] = len(test_ids)
json_dict['training'] = [{'image': "./imagesTr/%s.nii.gz" % i, "label": "./labelsTr/%s.nii.gz" % i} for i in train_ids]
json_dict['test'] = ["./imagesTs/%s.nii.gz" % i for i in test_ids]
with open(os.path.join(output_folder, "dataset.json"), 'w') as f:
json.dump(json_dict, f, indent=4, sort_keys=True)