Visual Question Answering
Transformers
TensorBoard
Safetensors
internvl_chat
feature-extraction
custom_code

Where can I deploy this model for inference?

#7
by catworld1212 - opened

Hi, I'm impressed with the work on InternVL and I'm interested in deploying its inference as an endpoint. Unfortunately, vLLM and TGI don't support this. Could anyone offer guidance on how to achieve this? I'd appreciate any suggestions you may have.

See "Chat Web Demo" at https://github.com/OpenGVLab/InternVL/blob/main/README.md

I want to deploy it as an inference not run it as a demo, Can you tell do InternVL-Chat-V1-5 requires flash attention?

Hi @whai362 @czczup what's the proper way to few-shot prompting (also called in-context learning? How do I give the previous context? I'm using lmdeploy to serve the inference can you help me, please?

Sign up or log in to comment