wchai commited on
Commit
85310e6
1 Parent(s): 0207a43

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -2
README.md CHANGED
@@ -73,7 +73,7 @@ model-index:
73
  ## Resources
74
 
75
  - [Website](https://rese1f.github.io/aurora-web/)
76
- - [arXiv: Paper]()
77
  - [GitHub: Code](https://github.com/rese1f/aurora)
78
  - [Huggingface: AuroraCap Model](https://huggingface.co/collections/Reself/auroracap-66d117ffe13bedda96702013)
79
  - [Huggingface: VDC Benchmark](https://huggingface.co/datasets/Reself/Video-Detailed-Caption)
@@ -99,4 +99,13 @@ Q: Why do we provide both official LLaVA-format and Xtuner format weights for Au
99
 
100
  A: While Xtuner supports saving checkpoints in multiple formats, it currently only allows continued training with the Xtuner format. Therefore, we currently provide the model in the Xtuner format for both continued training and inference. In the future, we will provide the model in the official LLaVA format for both training and inference, enabling quicker SGLang deployment and integration with the transformers.
101
 
102
- ## Citation
 
 
 
 
 
 
 
 
 
 
73
  ## Resources
74
 
75
  - [Website](https://rese1f.github.io/aurora-web/)
76
+ - [arXiv: Paper](https://arxiv.org/abs/2410.03051)
77
  - [GitHub: Code](https://github.com/rese1f/aurora)
78
  - [Huggingface: AuroraCap Model](https://huggingface.co/collections/Reself/auroracap-66d117ffe13bedda96702013)
79
  - [Huggingface: VDC Benchmark](https://huggingface.co/datasets/Reself/Video-Detailed-Caption)
 
99
 
100
  A: While Xtuner supports saving checkpoints in multiple formats, it currently only allows continued training with the Xtuner format. Therefore, we currently provide the model in the Xtuner format for both continued training and inference. In the future, we will provide the model in the official LLaVA format for both training and inference, enabling quicker SGLang deployment and integration with the transformers.
101
 
102
+ ## Citation
103
+
104
+ ```
105
+ @article{chai2024auroracap,
106
+ title={AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark },
107
+ author={Wenhao Chai, Enxin Song, Yilun Du, Chenlin Meng, Vashisht Madhavan, Omer Bar-Tal, Jeng-Neng Hwang, Saining Xie, Christopher D. Manning},
108
+ journal={arXiv preprint arXiv:2410.03051},
109
+ year={2024}
110
+ }
111
+ ```