Update README.md
Browse files
README.md
CHANGED
@@ -5,4 +5,10 @@ language:
|
|
5 |
metrics:
|
6 |
- accuracy
|
7 |
pipeline_tag: image-text-to-text
|
8 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
metrics:
|
6 |
- accuracy
|
7 |
pipeline_tag: image-text-to-text
|
8 |
+
---
|
9 |
+
|
10 |
+
We use the powerfull [TinyLLaVA Factory](https://github.com/TinyLLaVA/TinyLLaVA_Factory) to create a super small image-text-to-text model with only 296M params.
|
11 |
+
|
12 |
+
The goal is to make it possible to run LLaVA models on edge devices (with few gigabytes of memory).
|
13 |
+
|
14 |
+
For LLM and vision tower, we choose [OpenELM-270M-Instruct](apple/OpenELM-270M-Instruct) and [facebook/dinov2-small](facebook/dinov2-small), respectively.
|