Rodeszones
commited on
Commit
•
79b7662
1
Parent(s):
0943e9a
Update README.md
Browse files
README.md
CHANGED
@@ -6,6 +6,8 @@ pipeline_tag: visual-question-answering
|
|
6 |
|
7 |
# CogVLM
|
8 |
|
|
|
|
|
9 |
**CogVLM** is a powerful **open-source visual language model** (**VLM**). CogVLM-17B has 10 billion vision parameters and 7 billion language parameters. CogVLM-17B achieves state-of-the-art performance on 10 classic cross-modal benchmarks, including NoCaps, Flicker30k captioning, RefCOCO, RefCOCO+, RefCOCOg, Visual7W, GQA, ScienceQA, VizWiz VQA and TDIUC, and rank the 2nd on VQAv2, OKVQA, TextVQA, COCO captioning, etc., **surpassing or matching PaLI-X 55B**. CogVLM can also [chat with you](http://36.103.203.44:7861/) about images.
|
10 |
|
11 |
<div align="center">
|
|
|
6 |
|
7 |
# CogVLM
|
8 |
|
9 |
+
CogVLM Grounding generalist model quantized with bitsandbytes 4 bit precision
|
10 |
+
|
11 |
**CogVLM** is a powerful **open-source visual language model** (**VLM**). CogVLM-17B has 10 billion vision parameters and 7 billion language parameters. CogVLM-17B achieves state-of-the-art performance on 10 classic cross-modal benchmarks, including NoCaps, Flicker30k captioning, RefCOCO, RefCOCO+, RefCOCOg, Visual7W, GQA, ScienceQA, VizWiz VQA and TDIUC, and rank the 2nd on VQAv2, OKVQA, TextVQA, COCO captioning, etc., **surpassing or matching PaLI-X 55B**. CogVLM can also [chat with you](http://36.103.203.44:7861/) about images.
|
12 |
|
13 |
<div align="center">
|