Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -54,6 +54,7 @@ license_link: LICENSE
54
  - [Training](#training)
55
  - [Limitations](#limitations)
56
  - [Why Yi-VL?](#why-yi-vl)
 
57
  - [Benchmarks](#benchmarks)
58
  - [Showcases](#showcases)
59
  - [How to use Yi-VL?](#how-to-use-yi-vl)
@@ -159,7 +160,23 @@ This is the initial release of the Yi-VL, which comes with some known limitation
159
  - Other limitations of the Yi LLM.
160
 
161
  # Why Yi-VL?
162
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
163
  ## Benchmarks
164
 
165
  Yi-VL outperforms all existing open-source models in [MMMU](https://mmmu-benchmark.github.io) and [CMMMU](https://cmmmu-benchmark.github.io/), two advanced benchmarks that include massive multi-discipline multimodal questions (based on data available up to January 2024).
 
54
  - [Training](#training)
55
  - [Limitations](#limitations)
56
  - [Why Yi-VL?](#why-yi-vl)
57
+ - [Tech report](#tech-report)
58
  - [Benchmarks](#benchmarks)
59
  - [Showcases](#showcases)
60
  - [How to use Yi-VL?](#how-to-use-yi-vl)
 
160
  - Other limitations of the Yi LLM.
161
 
162
  # Why Yi-VL?
163
+
164
+ ## Tech report
165
+
166
+ For detailed capabilities of the Yi series model, see [Yi: Open Foundation Models by 01.AI](https://arxiv.org/abs/2403.04652).
167
+
168
+ ### Citation
169
+ ```
170
+ @misc{ai2024yi,
171
+ title={Yi: Open Foundation Models by 01.AI},
172
+ author={01. AI and : and Alex Young and Bei Chen and Chao Li and Chengen Huang and Ge Zhang and Guanwei Zhang and Heng Li and Jiangcheng Zhu and Jianqun Chen and Jing Chang and Kaidong Yu and Peng Liu and Qiang Liu and Shawn Yue and Senbin Yang and Shiming Yang and Tao Yu and Wen Xie and Wenhao Huang and Xiaohui Hu and Xiaoyi Ren and Xinyao Niu and Pengcheng Nie and Yuchi Xu and Yudong Liu and Yue Wang and Yuxuan Cai and Zhenyu Gu and Zhiyuan Liu and Zonghong Dai},
173
+ year={2024},
174
+ eprint={2403.04652},
175
+ archivePrefix={arXiv},
176
+ primaryClass={cs.CL}
177
+ }
178
+ ```
179
+
180
  ## Benchmarks
181
 
182
  Yi-VL outperforms all existing open-source models in [MMMU](https://mmmu-benchmark.github.io) and [CMMMU](https://cmmmu-benchmark.github.io/), two advanced benchmarks that include massive multi-discipline multimodal questions (based on data available up to January 2024).