Where can I deploy this model for inference?
#7
by
catworld1212
- opened
Hi, I'm impressed with the work on InternVL and I'm interested in deploying its inference as an endpoint. Unfortunately, vLLM and TGI don't support this. Could anyone offer guidance on how to achieve this? I'd appreciate any suggestions you may have.
See "Chat Web Demo" at https://github.com/OpenGVLab/InternVL/blob/main/README.md
See "Chat Web Demo" at https://github.com/OpenGVLab/InternVL/blob/main/README.md
I want to deploy it as an inference not run it as a demo, Can you tell do InternVL-Chat-V1-5 requires flash attention?