nnUNet_calvingfront_detection / documentation /training_example_Hippocampus.md
ho11laqe's picture
init
ecf08bc
# Example: 3D U-Net training on the Hippocampus dataset
This is a step-by-step example on how to run a 3D full resolution Training with the Hippocampus dataset from the
Medical Segmentation Decathlon.
1) Install nnU-Net by following the instructions [here](../readme.md#installation). Make sure to set all relevant paths,
also see [here](setting_up_paths.md). This step is necessary so that nnU-Net knows where to store raw data,
preprocessed data and trained models.
2) Download the Hippocampus dataset of the Medical Segmentation Decathlon from
[here](https://drive.google.com/drive/folders/1HqEgzS8BV2c7xYNrZdEAnrHk7osJJ--2). Then extract the archive to a
destination of your choice.
3) Decathlon data come as 4D niftis. This is not compatible with nnU-Net (see dataset format specified
[here](dataset_conversion.md)). Convert the Hippocampus dataset into the correct format with
```bash
nnUNet_convert_decathlon_task -i /xxx/Task04_Hippocampus
```
Note that `Task04_Hippocampus` must be the folder that has the three 'imagesTr', 'labelsTr', 'imagesTs' subfolders!
The converted dataset can be found in $nnUNet_raw_data_base/nnUNet_raw_data ($nnUNet_raw_data_base is the folder for
raw data that you specified during installation)
4) You can now run nnU-Nets pipeline configuration (and the preprocessing) with the following line:
```bash
nnUNet_plan_and_preprocess -t 4
```
Where 4 refers to the task ID of the Hippocampus dataset.
5) Now you can already start network training. This is how you train a 3d full resoltion U-Net on the Hippocampus dataset:
```bash
nnUNet_train 3d_fullres nnUNetTrainerV2 4 0
```
nnU-Net per default requires all trainings as 5-fold cross validation. The command above will run only the training for the
first fold (fold 0). 4 is the task identifier of the hippocampus dataset. Training one fold should take about 9
hours on a modern GPU.
This tutorial is only intended to demonstrate how easy it is to get nnU-Net running. You do not need to finish the
network training - pretrained models for the hippocampus task are available (see [here](../readme.md#run-inference)).
The only prerequisite for running nnU-Net on your custom dataset is to bring it into a structured, nnU-Net compatible
format. nnU-Net will take care of the rest. See [here](dataset_conversion.md) for instructions on how to convert
datasets into nnU-Net compatible format.