YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

DCNv4

News

  • Jan 15, 2024: 🚀 Compared with InternImage, the new FlashInternImage powered with DCNv4 has faster inference speed, faster convergence, and better performance!!!
  • Jan 15, 2024: 🚀 "DCNv4" is released!

Introduction

We introduce Deformable Convolution v4 (DCNv4), a highly efficient and effective operator designed for a broad spectrum of vision applications. DCNv4 addresses the limitations of its predecessor, DCNv3, with two key enhancements: 1. removing softmax normalization in spatial aggregation to enhance its dynamic property and expressive power and 2. optimizing memory access to minimize redundant operations for speedup. These improvements result in a significantly faster convergence compared to DCNv3 and a substantial increase in processing speed, with DCNv4 achieving more than three times the forward speed. DCNv4 demonstrates exceptional performance across various tasks, including image classification, instance and semantic segmentation, and notably, image generation. When integrated into generative models like U-Net in the latent diffusion model, DCNv4 outperforms its baseline, underscoring its possibility to enhance generative models. In practical applications, replacing DCNv3 with DCNv4 in the InternImage model to create FlashInternImage results in up to 80% speed increase and further performance improvement without further modifications. The advancements in speed and efficiency of DCNv4, combined with its robust performance across diverse vision tasks, show its potential as a foundational building block for future vision models.

Released Models

ImageNet Image Classification
name pretrain resolution acc@1 #param download
FlashInternImage-T ImageNet-1K 224x224 83.6 30M ckpt | cfg
FlashInternImage-S ImageNet-1K 224x224 84.4 50M ckpt | cfg
FlashInternImage-B ImageNet-1K 224x224 84.9 97M ckpt | cfg
FlashInternImage-L ImageNet-22K 384x384 88.1 223M ckpt | cfg
COCO Object Detection and Instance Segmentation
backbone method schd box mAP mask mAP Config Download
FlashInternImage-T Mask-RCNN 1x 48.0 43.1 config ckpt | log
FlashInternImage-T Mask-RCNN 3x 49.5 44.0 config ckpt | log
FlashInternImage-S Mask-RCNN 1x 49.2 44.0 config ckpt | log
FlashInternImage-S Mask-RCNN 3x 50.5 44.9 config ckpt | log
FlashInternImage-B Mask-RCNN 1x 50.1 44.5 config ckpt | log
FlashInternImage-B Mask-RCNN 3x 50.6 45.4 config ckpt | log
backbone method schd box mAP mask mAP Config Download
FlashInternImage-L Cascade Mask R-CNN 1x 55.6 48.2 config ckpt | log
FlashInternImage-L Cascade Mask R-CNN 3x 56.7 48.9 config ckpt
backbone method lr type pretrain schd box mAP Config Download
FlashInternImage-T DINO layer-wise lr ImageNet-1K 1x 54.7 config ckpt | log
FlashInternImage-S DINO layer-wise lr ImageNet-1K 1x 55.3 config ckpt | log
FlashInternImage-B DINO layer-wise lr ImageNet-1K 1x 56.0 config ckpt | log
FlashInternImage-L DINO 0.1x backbone lr ImageNet-22K 1x 58.8 config ckpt | log
ADE20K Semantic Segmentation
backbone method resolution mIoU (ss/ms) Config Download
FlashInternImage-T UperNet 512x512 49.3 / 50.3 config ckpt | log
FlashInternImage-S UperNet 512x512 50.6 / 51.6 config ckpt | log
FlashInternImage-B UperNet 512x512 52.0 / 52.6 config ckpt | log
FlashInternImage-L UperNet 640x640 55.6 / 56.0 config ckpt | log
backbone method resolution mIoU (ss) Config Download
FlashInternImage-T Mask2Former 512x512 51.2 config ckpt | log
FlashInternImage-S Mask2Former 640x640 52.6 config ckpt | log
FlashInternImage-B Mask2Former 640x640 53.4 config ckpt | log
FlashInternImage-L Mask2Former 640x640 56.7 config ckpt | log

Citations

If this work is helpful for your research, please consider citing the following BibTeX entry.


@article{xiong2024efficient,
      title={Efficient Deformable ConvNets: Rethinking Dynamic and Sparse Operator for Vision Applications}, 
      author={Yuwen Xiong and Zhiqi Li and Yuntao Chen and Feng Wang and Xizhou Zhu and Jiapeng Luo and Wenhai Wang and Tong Lu and Hongsheng Li and Yu Qiao and Lewei Lu and Jie Zhou and Jifeng Dai},
      journal={arXiv preprint arXiv:2401.06197},
      year={2024}
}

@article{wang2022internimage,
  title={InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions},
  author={Wang, Wenhai and Dai, Jifeng and Chen, Zhe and Huang, Zhenhang and Li, Zhiqi and Zhu, Xizhou and Hu, Xiaowei and Lu, Tong and Lu, Lewei and Li, Hongsheng and others},
  journal={arXiv preprint arXiv:2211.05778},
  year={2022}
}

@inproceedings{zhu2022uni,
  title={Uni-perceiver: Pre-training unified architecture for generic perception for zero-shot and few-shot tasks},
  author={Zhu, Xizhou and Zhu, Jinguo and Li, Hao and Wu, Xiaoshi and Li, Hongsheng and Wang, Xiaohua and Dai, Jifeng},
  booktitle={CVPR},
  pages={16804--16815},
  year={2022}
}

@article{zhu2022uni,
  title={Uni-perceiver-moe: Learning sparse generalist models with conditional moes},
  author={Zhu, Jinguo and Zhu, Xizhou and Wang, Wenhai and Wang, Xiaohua and Li, Hongsheng and Wang, Xiaogang and Dai, Jifeng},
  journal={arXiv preprint arXiv:2206.04674},
  year={2022}
}

@article{li2022uni,
  title={Uni-Perceiver v2: A Generalist Model for Large-Scale Vision and Vision-Language Tasks},
  author={Li, Hao and Zhu, Jinguo and Jiang, Xiaohu and Zhu, Xizhou and Li, Hongsheng and Yuan, Chun and Wang, Xiaohua and Qiao, Yu and Wang, Xiaogang and Wang, Wenhai and others},
  journal={arXiv preprint arXiv:2211.09808},
  year={2022}
}

@article{yang2022bevformer,
  title={BEVFormer v2: Adapting Modern Image Backbones to Bird's-Eye-View Recognition via Perspective Supervision},
  author={Yang, Chenyu and Chen, Yuntao and Tian, Hao and Tao, Chenxin and Zhu, Xizhou and Zhang, Zhaoxiang and Huang, Gao and Li, Hongyang and Qiao, Yu and Lu, Lewei and others},
  journal={arXiv preprint arXiv:2211.10439},
  year={2022}
}

@article{su2022towards,
  title={Towards All-in-one Pre-training via Maximizing Multi-modal Mutual Information},
  author={Su, Weijie and Zhu, Xizhou and Tao, Chenxin and Lu, Lewei and Li, Bin and Huang, Gao and Qiao, Yu and Wang, Xiaogang and Zhou, Jie and Dai, Jifeng},
  journal={arXiv preprint arXiv:2211.09807},
  year={2022}
}

@inproceedings{li2022bevformer,
  title={Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers},
  author={Li, Zhiqi and Wang, Wenhai and Li, Hongyang and Xie, Enze and Sima, Chonghao and Lu, Tong and Qiao, Yu and Dai, Jifeng},
  booktitle={ECCV},
  pages={1--18},
  year={2022},
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Collection including OpenGVLab/DCNv4