Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Deep Incubation

This repository contains the pre-trained models for Deep Incubation.

Title:  Deep Incubation: Training Large Models by Divide-and-Conquering
Authors:  Zanlin Ni, Yulin Wang, Jiangwei Yu, Haojun Jiang, Yue Cao, Gao Huang (Corresponding Author)
Institute: Tsinghua University and Beijing Academy of Artificial Intelligence (BAAI)
Publish:   arXiv preprint (arXiv 2212.04129)
Contact:  nzl22 at mails dot tsinghua dot edu dot cn

Models

model image size #param. top-1 acc. checkpoint
ViT-B 224x224 87M 82.4% πŸ€— HF link
ViT-B 384x384 87M 84.2% πŸ€— HF link
ViT-L 224x224 304M 83.9% πŸ€— HF link
ViT-L 384x384 304M 85.3% πŸ€— HF link
ViT-H 224x224 632M 84.3% πŸ€— HF link
ViT-H 392x392 632M 85.6% πŸ€— HF link

Data Preparation

  • The ImageNet dataset should be prepared as follows:
data
β”œβ”€β”€ train
β”‚   β”œβ”€β”€ folder 1 (class 1)
β”‚   β”œβ”€β”€ folder 2 (class 1)
β”‚   β”œβ”€β”€ ...
β”œβ”€β”€ val
β”‚   β”œβ”€β”€ folder 1 (class 1)
β”‚   β”œβ”€β”€ folder 2 (class 1)
β”‚   β”œβ”€β”€ ...

Citation

If you find our work helpful, please star🌟 this repo and citeπŸ“‘ our paper. Thanks for your support!

@article{Ni2022Incub,
  title={Deep Incubation: Training Large Models by Divide-and-Conquering},
  author={Ni, Zanlin and Wang, Yulin and Yu, Jiangwei and Jiang, Haojun and Cao, Yue and Huang, Gao},
  journal={arXiv preprint arXiv:2212.04129},
  year={2022}
}

Acknowledgements

Our implementation is mainly based on deit. We thank to their clean codebase.

Contact

If you have any questions or concerns, please send mail to nzl22@mails.tsinghua.edu.cn.

Downloads last month
0
Unable to determine this model's library. Check the docs .