Spaces:
Running
Running
File size: 1,555 Bytes
a145ae6 bc158f1 adbf9af bc158f1 9aeca25 c3921ec 58e8463 d16dc7c 58e8463 bc158f1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 |
---
title: README
emoji: π
colorFrom: gray
colorTo: blue
sdk: static
pinned: false
---
# All Things ViTs: Understanding and Interpreting Attention in Vision (CVPR'23 tutorial)
*By: [Hila Chefer](https://hila-chefer.github.io) and [Sayak Paul](https://sayak.dev)*
*Website: [all-things-vits.github.io/atv/](https://all-things-vits.github.io/atv/)*
*Code samples: [all-things-vits/code-samples](https://github.com/all-things-vits/code-samples)*
*Abstract: The attention mechanism has revolutionized deep learning research across many disciplines starting from NLP and expanding to vision, speech, and more. Different from other mechanisms, the elegant and general attention mechanism is easily adaptable and eliminates modality-specific inductive biases. As attention becomes increasingly popular, it is crucial to develop tools to allow researchers to understand and explain the inner workings of the mechanism to facilitate better and more responsible use of it. This tutorial focuses on understanding and interpreting attention in the vision and the multi-modal setting. We present state-of-the-art research on representation probing, interpretability, and attention-based semantic guidance, alongside hands-on demos to facilitate interactivity. Additionally, we discuss open questions arising from recent works and future research directions.*
<p align="center">
<img src="https://i.imgur.com/BcqCbcC.jpg" width=750/>
</p>
This organization hosts all the interactive demos to be presented at the tutorial. Below, you can find some of them.
|