Edit model card

ViTMatte model

ViTMatte model trained on Composition-1k. It was introduced in the paper ViTMatte: Boosting Image Matting with Pretrained Plain Vision Transformers by Yao et al. and first released in this repository.

Disclaimer: The team releasing ViTMatte did not write a model card for this model so this model card has been written by the Hugging Face team.

Model description

ViTMatte is a simple approach to image matting, the task of accurately estimating the foreground object in an image. The model consists of a Vision Transformer (ViT) with a lightweight head on top.

drawing

ViTMatte high-level overview. Taken from the original paper.

Intended uses & limitations

You can use the raw model for image matting. See the model hub to look for other fine-tuned versions that may interest you.

How to use

We refer to the docs.

BibTeX entry and citation info

@misc{yao2023vitmatte,
      title={ViTMatte: Boosting Image Matting with Pretrained Plain Vision Transformers}, 
      author={Jingfeng Yao and Xinggang Wang and Shusheng Yang and Baoyuan Wang},
      year={2023},
      eprint={2305.15272},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}
Downloads last month
137,585
Safetensors
Model size
25.8M params
Tensor type
F32
·
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for hustvl/vitmatte-small-composition-1k

Quantizations
1 model

Space using hustvl/vitmatte-small-composition-1k 1