The dataset viewer is not available for this dataset.
Cannot get the config names for the dataset.
Error code:   ConfigNamesError
Exception:    ReadTimeout
Message:      (ReadTimeoutError("HTTPSConnectionPool(host='huggingface.co', port=443): Read timed out. (read timeout=10)"), '(Request ID: 93697133-7b5b-4610-bb2a-63e68b72c60f)')
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response
                  config_names = get_dataset_config_names(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 347, in get_dataset_config_names
                  dataset_module = dataset_module_factory(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1914, in dataset_module_factory
                  raise e1 from None
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1889, in dataset_module_factory
                  return HubDatasetModuleFactoryWithoutScript(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1269, in get_module
                  patterns = get_data_patterns(base_path, download_config=self.download_config)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/data_files.py", line 501, in get_data_patterns
                  return _get_data_files_patterns(resolver)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/data_files.py", line 275, in _get_data_files_patterns
                  data_files = pattern_resolver(pattern)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/data_files.py", line 388, in resolve_pattern
                  for filepath, info in fs.glob(pattern, detail=True, **glob_kwargs).items()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py", line 416, in glob
                  path = self.resolve_path(path, revision=kwargs.get("revision")).unresolve()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py", line 183, in resolve_path
                  repo_and_revision_exist, err = self._repo_and_revision_exist(repo_type, repo_id, revision)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py", line 131, in _repo_and_revision_exist
                  self._api.repo_info(repo_id, revision=revision, repo_type=repo_type, timeout=HF_HUB_ETAG_TIMEOUT)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
                  return fn(*args, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/hf_api.py", line 2588, in repo_info
                  return method(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
                  return fn(*args, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/hf_api.py", line 2445, in dataset_info
                  r = get_session().get(path, headers=headers, timeout=timeout, params=params)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py", line 602, in get
                  return self.request("GET", url, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py", line 589, in request
                  resp = self.send(prep, **send_kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py", line 703, in send
                  r = adapter.send(request, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/utils/_http.py", line 66, in send
                  return super().send(request, *args, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/requests/adapters.py", line 635, in send
                  raise ReadTimeout(e, request=request)
              requests.exceptions.ReadTimeout: (ReadTimeoutError("HTTPSConnectionPool(host='huggingface.co', port=443): Read timed out. (read timeout=10)"), '(Request ID: 93697133-7b5b-4610-bb2a-63e68b72c60f)')

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Dataset Summary

The largest, fully-annotated abdominal CT dataset to date, including 9,262 CT volumes with annotations for 25 different anatomical structures.


Join the Touchstone Benchmarking Project

The Touchstone Project aims to compare diverse semantic segmentation and pre-training algorithms. We, the CCVL research group at Johns Hopkins University, invite creators of these algorithms to contribute to the initiative. With our support, contributors will train their methodologies on the largest fully-annotated abdominal CT datasets to date. Subsequently, we will evaluate the trained models using a large internal dataset at Johns Hopkins University. If you are the creator of a semantic segmentation or pre-training algorithm and wish to advance medical AI by participating in the Benchmark Project, please reach out to pedro.salvadorbassi2@unibo.it. We will provide you further details on the project and explain your opportunities to collaborate in our future publications!


Note for Touchstone Benchmarking Project

This dataset should be only used for the second round of the Touchstone Project, and not to update first-round checkpoints. The first round dataset (5,195 annotated CT volumes, 9 annotated structures) is available at: AbdomenAtlas1.0Mini and AbdomenAtlas1.0MiniBeta


Downloading Instructions

1- Install the Hugging Face library:

pip install huggingface_hub[hf_transfer]==0.24.0
HF_HUB_ENABLE_HF_TRANSFER=1
[Optional] Alternative without HF Trasnsfer (slower)
pip install huggingface_hub==0.24.0

2- Download the dataset:

mkdir AbdomenAtlas
cd AbdomenAtlas
huggingface-cli download BodyMaps/AbdomenAtlas1.1Mini --repo-type dataset --local-dir .
[Optional] Resume downloading

In case you had a previous interrupted download, just run the huggingface-cli download command above again.

huggingface-cli download BodyMaps/AbdomenAtlas1.1Mini --repo-type dataset --local-dir .

Paper

AbdomenAtlas-8K: Annotating 8,000 CT Volumes for Multi-Organ Segmentation in Three Weeks
Chongyu Qu1, Tiezheng Zhang1, Hualin Qiao2, Jie Liu3, Yucheng Tang4, Alan L. Yuille1, and Zongwei Zhou1,*
1 Johns Hopkins University,
2 Rutgers University,
3 City University of Hong Kong,
4 NVIDIA
NeurIPS 2023
paper | code | dataset | annotation | poster

How Well Do Supervised 3D Models Transfer to Medical Imaging Tasks?
Wenxuan Li, Alan Yuille, and Zongwei Zhou*
Johns Hopkins University
International Conference on Learning Representations (ICLR) 2024 (oral; top 1.2%)
paper | code

Citation

@article{qu2023abdomenatlas,
  title={Abdomenatlas-8k: Annotating 8,000 CT volumes for multi-organ segmentation in three weeks},
  author={Qu, Chongyu and Zhang, Tiezheng and Qiao, Hualin and Tang, Yucheng and Yuille, Alan L and Zhou, Zongwei},
  journal={Advances in Neural Information Processing Systems},
  volume={36},
  year={2023}
}

@inproceedings{li2024well,
  title={How Well Do Supervised Models Transfer to 3D Image Segmentation?},
  author={Li, Wenxuan and Yuille, Alan and Zhou, Zongwei},
  booktitle={The Twelfth International Conference on Learning Representations},
  year={2024}
}

Acknowledgements

This work was supported by the Lustgarten Foundation for Pancreatic Cancer Research and partially by the Patrick J. McGovern Foundation Award. We appreciate the effort of the MONAI Team to provide open-source code for the community.

License

AbdomenAtlas 1.1 is licensed under CC BY-NC-SA 4.0.

Uploading AbdomenAtlas to HuggingFace

The file AbdomenAtlasUploadMultipleFolders.ipynb has the code we used to upload AbdomenAtlas to Hugging Face. It may be ncessary to run the script multiple times, until it finishes without an uploading error. The uploading script requires PyTorch, huggingface_hub, and Jupyter Notebook.

Downloads last month
0