Edit model card

Model Summary

Video-CCAM-14B is a lightweight Video-MLLM built on Phi-3-medium-4k-instruct and SigLIP SO400M.

Usage

Inference using Huggingface transformers on NVIDIA GPUs. Requirements tested on python 3.10:

torch==2.1.0
torchvision==0.16.0
transformers==4.40.2
peft==0.10.0

Inference & Evaluation

Please refer to Video-CCAM on inference and evaluation.

Video-MME: 53.2/57.4 (96 frames)

MVBench: 61.43 (16 frames)

Acknowledgement

  • xtuner: Video-CCAM-14B is trained using the xtuner framework. Thanks for their excellent works!
  • Phi-3-medium-4k-instruct: Powerful language models developed by Microsoft.
  • SigLIP SO400M: Outstanding vision encoder developed by Google.

License

The model is licensed under the MIT license.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Collection including JaronTHU/Video-CCAM-14B