Visual Question Answering
Transformers
TensorBoard
Safetensors
internvl_chat
feature-extraction
custom_code

V100 GPU Support?

#5
by KULAJIANG - opened

Could you please support V100’s inference? Thank you very much.

OpenGVLab org

Hello, thanks for your attention. We are preparing quantitative models that can run on a 32G V100.

Thank you very much for your reply. Looking forward to further developments!

Look forward to further developments!

https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5-Int8
Thanks. The key problem for V100 is the flash attention. I believe that quantitative models can not solve it without changing the attention code.

KULAJIANG changed discussion status to closed

Sign up or log in to comment