Overview

LLaVA (Large Language and Vision Assistant) is an open-source chatbot trained to handle multimodal instruction-following tasks. It is a fine-tuned Vicuna-7B model, designed to process both text and image inputs. This auto-regressive language model leverages the transformer architecture to improve interactions in vision-language tasks, making it useful for research in computer vision, natural language processing, machine learning, and artificial intelligence.

LLaVA-v1.6-Vicuna-7B is the latest iteration, trained in December 2023, and optimized for improved instruction-following performance in multimodal settings.

Variants

No Variant Cortex CLI command
1 llava-v1.6-vicuna-7b-f16 cortex run llava-v1.6:gguf-f16
2 llava-v1.6-vicuna-7b-q4_km cortex run llava-v1.6:gguf-q4-km

Use it with Jan (UI)

  1. Install Jan using Quickstart
  2. Use in Jan model Hub:
    cortexso/llava-v1.6
    

Use it with Cortex (CLI)

  1. Install Cortex using Quickstart
  2. Run the model with command:
    cortex run llava-v1.6
    

Credits

Downloads last month
129
GGUF
Model size
6.74B params
Architecture
llama

4-bit

16-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.