Link model to paper
Browse filesThis PR improves the discoverability of the model by linking it to https://huggingface.co/papers/2407.07895
README.md
CHANGED
@@ -46,4 +46,18 @@ Use the code below to evaluate the model.
|
|
46 |
Please first edit /path/to/ckpt to the path of checkpoint, /path/to/images to the path of "interleave_data" in scripts/interleave/eval_all.sh and then run
|
47 |
```bash
|
48 |
bash scripts/interleave/eval_all.sh
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
```
|
|
|
46 |
Please first edit /path/to/ckpt to the path of checkpoint, /path/to/images to the path of "interleave_data" in scripts/interleave/eval_all.sh and then run
|
47 |
```bash
|
48 |
bash scripts/interleave/eval_all.sh
|
49 |
+
```
|
50 |
+
|
51 |
+
## Bibtex citation
|
52 |
+
|
53 |
+
```bibtex
|
54 |
+
@misc{li2024llavanextinterleavetacklingmultiimagevideo,
|
55 |
+
title={LLaVA-NeXT-Interleave: Tackling Multi-image, Video, and 3D in Large Multimodal Models},
|
56 |
+
author={Feng Li and Renrui Zhang and Hao Zhang and Yuanhan Zhang and Bo Li and Wei Li and Zejun Ma and Chunyuan Li},
|
57 |
+
year={2024},
|
58 |
+
eprint={2407.07895},
|
59 |
+
archivePrefix={arXiv},
|
60 |
+
primaryClass={cs.CV},
|
61 |
+
url={https://arxiv.org/abs/2407.07895},
|
62 |
+
}
|
63 |
```
|