Edit model card

Model Card for Model ID

This is a multimodal implementation of Phi2 model inspired by LlaVA-Phi.

Model Details

  1. LLM Backbone: Phi2
  2. Vision Tower: clip-vit-large-patch14-336
  3. Pretraining Dataset: LAION-CC-SBU dataset with BLIP captions(200k samples)
  4. Finetuning Dataset: Instruct 150k dataset based on COCO
  5. Finetuned Model: marianna13/llava-phi-2-3b

Model Sources

Downloads last month
4
Safetensors
Model size
2.79B params
Tensor type
F32
·

Datasets used to train sid819/Llava-Phi2