ruohongz commited on
Commit
6189f87
1 Parent(s): 3f5f6aa

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -0
README.md ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ inference: false
3
+ license: apache-2.0
4
+ ---
5
+
6
+ <br>
7
+ <br>
8
+
9
+ # LLaVA-Hound Model Card
10
+
11
+ ## Model details
12
+
13
+ **Model type:**
14
+ LLaVA-Hound is an open-source video large multimodal model, fine-tuned from video instruction following data based on large language model.
15
+
16
+ This model is the fine-tuned on image instruction and video caption.
17
+
18
+ Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-7b-v1.5)
19
+
20
+ **Model date:**
21
+ Trained on March 15, 2024.
22
+
23
+ **Paper or resources for more information:**
24
+ https://github.com/RifleZhang/LLaVA-Hound-DPO
25
+
26
+ ## License
27
+ [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-7b-v1.5) license.
28
+
29
+ **Where to send questions or comments about the model:**
30
+ https://github.com/RifleZhang/LLaVA-Hound-DPO/issues
31
+
32
+ ## Intended use
33
+ **Primary intended uses:**
34
+ Video (and image) instruction-following.
35
+
36
+ **Primary intended users:**
37
+ Researchers in artificial intelligence, large multimodal model, etc.
38
+
39
+ ## Training dataset
40
+ ShareGPTVideo dataset.
41
+
42
+ ## Evaluation
43
+ Follow https://github.com/RifleZhang/LLaVA-Hound-DPO/blob/main/README.md