xinyu1205 commited on
Commit
771b5f9
1 Parent(s): c9cb5fa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -20,7 +20,7 @@ RAM++ outperforms existing SOTA image fundamental reocngition models in terms of
20
 
21
  Authors from the [paper](https://arxiv.org/abs/2306.03514) write in the abstract:
22
 
23
- *We present the Recognize Anything Model~(RAM): a strong foundation model for image tagging. RAM makes a substantial step for large models in computer vision, demonstrating the zero-shot ability to recognize any common category with high accuracy. By leveraging large-scale image-text pairs for training instead of manual annotations, RAM introduces a new paradigm for image tagging. We evaluate the tagging capability of RAM on numerous benchmarks and observe an impressive zero-shot performance, which significantly outperforms CLIP and BLIP. Remarkably, RAM even surpasses fully supervised models and exhibits a competitive performance compared with the Google tagging API.*
24
 
25
 
26
  ## BibTex and citation info
@@ -28,7 +28,6 @@ Authors from the [paper](https://arxiv.org/abs/2306.03514) write in the abstract
28
  ```
29
 
30
 
31
-
32
  @article{zhang2023recognize,
33
  title={Recognize Anything: A Strong Image Tagging Model},
34
  author={Zhang, Youcai and Huang, Xinyu and Ma, Jinyu and Li, Zhaoyang and Luo, Zhaochuan and Xie, Yanchun and Qin, Yuzhuo and Luo, Tong and Li, Yaqian and Liu, Shilong and others},
 
20
 
21
  Authors from the [paper](https://arxiv.org/abs/2306.03514) write in the abstract:
22
 
23
+ *We introduce the Recognize Anything Plus Model~(RAM++), a fundamental image recognition model with strong open-set recognition capabilities, by injecting semantic concepts into image tagging training framework. Previous approaches are either image tagging models constrained by limited semantics, or vision-language models with shallow interaction for suboptimal performance in multi-tag recognition. In contrast, RAM++ integrates image-text alignment and image-tagging within a unified fine-grained interaction framework based on image-tags-text triplets. This design enables RAM++ not only excel in identifying predefined categories, but also significantly augment the recognition ability in open-set categories. Moreover, RAM++ employs large language models~(LLMs) to generate diverse visual tag descriptions, pioneering the integration of LLM's knowledge into image tagging training. This approach empowers RAM++ to integrate visual description concepts for open-set recognition during inference. Evaluations on comprehensive image recognition benchmarks demonstrate RAM++ exceeds existing state-of-the-art (SOTA) fundamental image recognition models on most aspects. *
24
 
25
 
26
  ## BibTex and citation info
 
28
  ```
29
 
30
 
 
31
  @article{zhang2023recognize,
32
  title={Recognize Anything: A Strong Image Tagging Model},
33
  author={Zhang, Youcai and Huang, Xinyu and Ma, Jinyu and Li, Zhaoyang and Luo, Zhaochuan and Xie, Yanchun and Qin, Yuzhuo and Luo, Tong and Li, Yaqian and Liu, Shilong and others},