|
---
|
|
license: llama2
|
|
---
|
|
|
|
# v-MLLM Model Card
|
|
|
|
## Model details
|
|
|
|
**Model type:**
|
|
v-MLLM is an open-source MLLM trained on Visual-Modality Instruction (VIM) corpus, it can robustly follow the text-modality instructions and visual-modality instructions.
|
|
|
|
**Model date:**
|
|
v-MLLM-13B was trained in January 2024.
|
|
|
|
**Github for more information:**
|
|
https://github.com/VIM-Bench/VIM_TOOL
|
|
|
|
## License
|
|
v-MLLM is licensed under the LLAMA 2 Community License,
|
|
Copyright (c) Meta Platforms, Inc. All Rights Reserved.
|
|
|
|
## Intended use
|
|
**Primary intended uses:**
|
|
The primary use of v-MLLM is for research on multimodal large language models.
|
|
|
|
**Primary intended users:**
|
|
The primary intended users of the model are researchers in computer vision, natural language processing, machine learning, and artificial intelligence.
|
|
|
|
## Training dataset
|
|
- 846k VIM corpus based on LVIS-Instruct4V corpus. |