KunpengSong commited on
Commit
01d622b
1 Parent(s): 6d3e5a7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -22
README.md CHANGED
@@ -7,39 +7,23 @@ inference: false
7
 
8
  # This page is still under construction.
9
 
10
- # LLaVA Model Card
11
 
12
  ## Model details
13
 
14
  **Model type:**
15
- LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data.
16
- It is an auto-regressive language model, based on the transformer architecture.
17
-
18
- **Model date:**
19
- LLaVA-v1.5-7B was trained in September 2023.
20
 
21
  **Paper or resources for more information:**
22
- https://llava-vl.github.io/
23
-
24
- ## License
25
- Llama 2 is licensed under the LLAMA 2 Community License,
26
- Copyright (c) Meta Platforms, Inc. All Rights Reserved.
27
 
28
  **Where to send questions or comments about the model:**
29
- https://github.com/haotian-liu/LLaVA/issues
30
 
31
  ## Intended use
32
  **Primary intended uses:**
33
- The primary use of LLaVA is research on large multimodal models and chatbots.
34
 
35
  **Primary intended users:**
36
  The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
37
-
38
- ## Training dataset
39
- - 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
40
- - 158K GPT-generated multimodal instruction-following data.
41
- - 450K academic-task-oriented VQA data mixture.
42
- - 40K ShareGPT data.
43
-
44
- ## Evaluation dataset
45
- A collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs.
 
7
 
8
  # This page is still under construction.
9
 
10
+ # MoMA Model Card
11
 
12
  ## Model details
13
 
14
  **Model type:**
15
+ MoMA is an open-source image personalization model. It has new attention layers and a multi-modal large language model fine-tuned from LLaVA-7B.
 
 
 
 
16
 
17
  **Paper or resources for more information:**
18
+ + Github: https://github.com/KunpengSong/MoMA
19
+ + Paper: https://arxiv.org/abs/2404.05674
 
 
 
20
 
21
  **Where to send questions or comments about the model:**
22
+ https://github.com/KunpengSong/MoMA
23
 
24
  ## Intended use
25
  **Primary intended uses:**
26
+ The primary use of LLaVA is research on personalized image generation tasks.
27
 
28
  **Primary intended users:**
29
  The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.