|
---
|
|
datasets:
|
|
- liuhaotian/LLaVA-Instruct-150K
|
|
- liuhaotian/LLaVA-CC3M-Pretrain-595K
|
|
language:
|
|
- en
|
|
metrics:
|
|
- accuracy
|
|
pipeline_tag: visual-question-answering
|
|
---
|
|
# DinoV2-SigLIP-Phi3(LoRA) VLM |
|
|
|
* **Vision Encoder** - DinoV2 + SigLIP @384px resolution. [Why 2 vision encoders?](https://arxiv.org/abs/2401.06209) |
|
* **Connector** - MLP (Dino and SigLIP features are concatenated and then projected to Phi3 representation space) |
|
* **Language Model** - Phi3 + LoRA |
|
* **Pre-train (Align) Dataset** - LLaVA-CC3M-Pretrain-595K |
|
* **Fine-tune (Instruction) Dataset** - LLAVA-v1.5-Instruct + LRV-Instruct |
|
|
|
Scripts to build and train the models are available at [NMS05/DinoV2-SigLIP-Phi3-LoRA-VLM](https://github.com/NMS05/DinoV2-SigLIP-Phi3-LoRA-VLM). |