xinyu1205 commited on
Commit
9b15849
1 Parent(s): 5e172ed

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -0
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ pipeline_tag: image-to-text
6
+ tags:
7
+ - image tagging, image captioning
8
+ ---
9
+
10
+ # Recognize Anything & Tag2Text
11
+
12
+ Model card for <a href="https://recognize-anything.github.io/">Recognize Anything: A Strong Image Tagging Model </a> and <a href="https://tag2text.github.io/">Tag2Text: Guiding Vision-Language Model via Image Tagging</a>.
13
+
14
+ **Recognition and localization are two foundation computer vision tasks.**
15
+ - **The Segment Anything Model (SAM)** excels in **localization capabilities**, while it falls short when it comes to **recognition tasks**.
16
+ - **The Recognize Anything Model (RAM) and Tag2Text** exhibits **exceptional recognition abilities**, in terms of **both accuracy and scope**.
17
+
18
+
19
+ | ![RAM.jpg](https://github.com/xinyu1205/recognize-anything/blob/main/images/localization_and_recognition.jpg) |
20
+ |:--:|
21
+ | <b> Pull figure from recognize-anything official repo | Image source: https://recognize-anything.github.io/ </b>|
22
+
23
+ ## TL;DR
24
+
25
+ Authors from the [paper](https://arxiv.org/abs/2306.03514) write in the abstract:
26
+
27
+ *We present the Recognize Anything Model~(RAM): a strong foundation model for image tagging. RAM makes a substantial step for large models in computer vision, demonstrating the zero-shot ability to recognize any common category with high accuracy. By leveraging large-scale image-text pairs for training instead of manual annotations, RAM introduces a new paradigm for image tagging. We evaluate the tagging capability of RAM on numerous benchmarks and observe an impressive zero-shot performance, which significantly outperforms CLIP and BLIP. Remarkably, RAM even surpasses fully supervised models and exhibits a competitive performance compared with the Google tagging API.*
28
+
29
+
30
+ ## BibTex and citation info
31
+
32
+ ```
33
+ @article{zhang2023recognize,
34
+ title={Recognize Anything: A Strong Image Tagging Model},
35
+ author={Zhang, Youcai and Huang, Xinyu and Ma, Jinyu and Li, Zhaoyang and Luo, Zhaochuan and Xie, Yanchun and Qin, Yuzhuo and Luo, Tong and Li, Yaqian and Liu, Shilong and others},
36
+ journal={arXiv preprint arXiv:2306.03514},
37
+ year={2023}
38
+ }
39
+
40
+ @article{huang2023tag2text,
41
+ title={Tag2Text: Guiding Vision-Language Model via Image Tagging},
42
+ author={Huang, Xinyu and Zhang, Youcai and Ma, Jinyu and Tian, Weiwei and Feng, Rui and Zhang, Yuejie and Li, Yaqian and Guo, Yandong and Zhang, Lei},
43
+ journal={arXiv preprint arXiv:2303.05657},
44
+ year={2023}
45
+ }
46
+ ```
47
+
48
+
49
+
50
+
51
+