Update README.md
Browse files
README.md
CHANGED
@@ -17,12 +17,12 @@ pipeline_tag: visual-question-answering
|
|
17 |
|
18 |
\[[Paper](https://arxiv.org/abs/2312.14238)\] \[[GitHub](https://github.com/OpenGVLab/InternVL)\] \[[Chat Demo](https://internvl.opengvlab.com/)\] \[[中文解读](https://zhuanlan.zhihu.com/p/675877376)]
|
19 |
|
20 |
-
| Model | Date | Download
|
21 |
-
| ----------------------- | ---------- |
|
22 |
-
| InternVL-Chat-V1.5 | 2024.04.18 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5)
|
23 |
-
| InternVL-Chat-V1.2-Plus | 2024.02.21 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-
|
24 |
-
| InternVL-Chat-V1.2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-
|
25 |
-
| InternVL-Chat-V1.1 | 2024.01.24 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-
|
26 |
|
27 |
## Model Details
|
28 |
- **Model Type:** multimodal large language model (MLLM)
|
@@ -53,7 +53,7 @@ from PIL import Image
|
|
53 |
from transformers import AutoModel, CLIPImageProcessor
|
54 |
from transformers import AutoTokenizer
|
55 |
|
56 |
-
path = "OpenGVLab/InternVL-Chat-
|
57 |
# If your GPU has more than 40G memory, you can put the entire model on a single GPU.
|
58 |
model = AutoModel.from_pretrained(
|
59 |
path,
|
|
|
17 |
|
18 |
\[[Paper](https://arxiv.org/abs/2312.14238)\] \[[GitHub](https://github.com/OpenGVLab/InternVL)\] \[[Chat Demo](https://internvl.opengvlab.com/)\] \[[中文解读](https://zhuanlan.zhihu.com/p/675877376)]
|
19 |
|
20 |
+
| Model | Date | Download | Note |
|
21 |
+
| ----------------------- | ---------- | --------------------------------------------------------------------------- | ---------------------------------- |
|
22 |
+
| InternVL-Chat-V1.5 | 2024.04.18 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5) | support 4K image; super strong OCR; Approaching the performance of GPT-4V and Gemini Pro on various benchmarks like MMMU, DocVQA, ChartQA, MathVista, etc. (🔥new)|
|
23 |
+
| InternVL-Chat-V1.2-Plus | 2024.02.21 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2-Plus) | more SFT data and stronger |
|
24 |
+
| InternVL-Chat-V1.2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2) | scaling up LLM to 34B |
|
25 |
+
| InternVL-Chat-V1.1 | 2024.01.24 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-1) | support Chinese and stronger OCR |
|
26 |
|
27 |
## Model Details
|
28 |
- **Model Type:** multimodal large language model (MLLM)
|
|
|
53 |
from transformers import AutoModel, CLIPImageProcessor
|
54 |
from transformers import AutoTokenizer
|
55 |
|
56 |
+
path = "OpenGVLab/InternVL-Chat-V1-1"
|
57 |
# If your GPU has more than 40G memory, you can put the entire model on a single GPU.
|
58 |
model = AutoModel.from_pretrained(
|
59 |
path,
|