Create README.md

#1
by merve HF staff - opened
Files changed (1) hide show
  1. README.md +93 -0
README.md ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: image-classification
4
+ datasets:
5
+ - imagenet-1k
6
+ ---
7
+ # Hiera Tiny MAE trained on IN1K fine-tuned on IN1K
8
+
9
+
10
+ **Hiera** is a _hierarchical_ vision transformer that is fast, powerful, and, above all, _simple_. It outperforms the state-of-the-art across a wide array of image and video tasks _while being much faster_.
11
+
12
+ <p align="center">
13
+ <img src="https://github.com/facebookresearch/hiera/raw/main/examples/img/inference_speed.png" width="75%">
14
+ </p>
15
+
16
+ ## How does it work?
17
+ ![A diagram of Hiera's architecture.](https://github.com/facebookresearch/hiera/raw/main/examples/img/hiera_arch.png)
18
+
19
+ Vision transformers like [ViT](https://arxiv.org/abs/2010.11929) use the same spatial resolution and number of features throughout the whole network. But this is inefficient: the early layers don't need that many features, and the later layers don't need that much spatial resolution. Prior hierarchical models like [ResNet](https://arxiv.org/abs/1512.03385) accounted for this by using fewer features at the start and less spatial resolution at the end.
20
+
21
+ Several domain specific vision transformers have been introduced that employ this hierarchical design, such as [Swin](https://arxiv.org/abs/2103.14030) or [MViT](https://arxiv.org/abs/2104.11227). But in the pursuit of state-of-the-art results using fully supervised training on ImageNet-1K, these models have become more and more complicated as they add specialized modules to make up for spatial biases that ViTs lack. While these changes produce effective models with attractive FLOP counts, under the hood the added complexity makes these models _slower_ overall.
22
+
23
+ We show that a lot of this bulk is actually _unnecessary_. Instead of manually adding spatial bases through architectural changes, we opt to _teach_ the model these biases instead. By training with [MAE](https://arxiv.org/abs/2111.06377), we can simplify or remove _all_ of these bulky modules in existing transformers and _increase accuracy_ in the process. The result is Hiera, an extremely efficient and simple architecture that outperforms the state-of-the-art in several image and video recognition tasks.
24
+
25
+ ## Installation
26
+
27
+ Hiera requires a reasonably recent version of [torch](https://pytorch.org/get-started/locally/).
28
+ After that, you can install hiera through [pip](https://pypi.org/project/hiera-transformer/):
29
+ ```bash
30
+ pip install hiera-transformer
31
+ ```
32
+ This repo _should_ support the latest timm version, but timm is a constantly updating package. Create an issue if you have problems with a newer version of timm.
33
+ With the `hiera-transformer` and `huggingface-hub` packages installed, you can simply run, e.g.,
34
+ ```py
35
+ from hiera import Hiera
36
+ model = Hiera.from_pretrained("facebook/hiera_base_224.mae_in1k_ft_in1k") # mae pt then in1k ft'd model
37
+ model = Hiera.from_pretrained("facebook/hiera_base_224.mae_in1k") # just mae pt, no ft
38
+ ```
39
+ to load a model. Use `<model_name>.<checkpoint_name>` from model zoo below.
40
+
41
+ If you want to save a model, use `model.config` as the config, e.g.,
42
+ ```py
43
+ model.save_pretrained("hiera-base-224", config=model.config)
44
+ ```
45
+
46
+
47
+
48
+ ### Inference
49
+
50
+ ```py
51
+ import hiera
52
+ model = hiera.hiera_base_224(pretrained=True, checkpoint="mae_in1k_ft_in1k")
53
+ ```
54
+ Then you can run inference like any other model:
55
+ ```py
56
+ output = model(x)
57
+ ```
58
+ Video inference works the same way, just use a `16x224` model instead.
59
+
60
+ **Note**: for efficiency, Hiera re-orders its tokens at the start of the network (see the `Roll` and `Unroll` modules in `hiera_utils.py`). Thus, tokens _aren't in spatial order_ by default. If you'd like to use intermediate feature maps for a downstream task, pass the `return_intermediates` flag when running the model:
61
+ ```py
62
+ output, intermediates = model(x, return_intermediates=True)
63
+ ```
64
+
65
+ #### MAE Inference
66
+ By default, the models do not include the MAE decoder. If you would like to use the decoder or compute MAE loss, you can instantiate an mae version by running:
67
+ ```py
68
+ import hiera
69
+ model = hiera.mae_hiera_base_224(pretrained=True, checkpoint="mae_in1k")
70
+ ```
71
+ Then when you run inference on the model, it will return a 4-tuple of `(loss, predictions, labels, mask)` where predictions and labels are for the _deleted tokens_ only. The returned mask will be `True` if the token is visible and `False` if it's deleted. You can change the masking ratio by passing it during inference:
72
+ ```py
73
+ loss, preds, labels, mask = model(x, mask_ratio=0.6)
74
+ ```
75
+ The default mask ratio is `0.6` for images, but you should pass in `0.9` for video. See the paper for details.
76
+
77
+ **Note:** We use _normalized pixel targets_ for MAE pretraining, meaning the patches are each individually normalized before the model model has to predict them. Thus, you have to unnormalize them using the ground truth before visualizing them. See `get_pixel_label_2d` in `hiera_mae.py` for details.
78
+
79
+
80
+
81
+ ## Citation
82
+ If you use Hiera or this code in your work, please cite:
83
+ ```
84
+ @article{ryali2023hiera,
85
+ title={Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles},
86
+ author={Ryali, Chaitanya and Hu, Yuan-Ting and Bolya, Daniel and Wei, Chen and Fan, Haoqi and Huang, Po-Yao and Aggarwal, Vaibhav and Chowdhury, Arkabandhu and Poursaeed, Omid and Hoffman, Judy and Malik, Jitendra and Li, Yanghao and Feichtenhofer, Christoph},
87
+ journal={ICML},
88
+ year={2023}
89
+ }
90
+ ```
91
+
92
+ [arxiv-link]: https://arxiv.org/abs/2306.00989/
93
+ [icml-link]: https://icml.cc/Conferences/2023