Update README.md
Browse files
README.md
CHANGED
@@ -6,10 +6,11 @@ Update: PR is merged, llama.cpp now natively supports these models
|
|
6 |
Important: Verify that processing a simple question with any image at least uses 1200 tokens of prompt processing, that shows that the new PR is in use.
|
7 |
If your prompt is just 576 + a few tokens, you are using llava-1.5 code (or projector) and this is incompatible with llava-1.6
|
8 |
|
9 |
-
**note**
|
10 |
|
11 |
|
12 |
The mmproj files are the embedded ViT's that came with llava-1.6, I've not compared them but given the previous releases from the team I'd be surprised if the ViT has not been fine tuned this time.
|
13 |
-
If that's the case, using another ViT can cause issues.
|
|
|
14 |
|
15 |
Original models: https://github.com/haotian-liu/LLaVA
|
|
|
6 |
Important: Verify that processing a simple question with any image at least uses 1200 tokens of prompt processing, that shows that the new PR is in use.
|
7 |
If your prompt is just 576 + a few tokens, you are using llava-1.5 code (or projector) and this is incompatible with llava-1.6
|
8 |
|
9 |
+
**note** Keep in mind the different fine tunes as described in the llama.cpp llava readme, it's essential to use the non defaults for non vicuna models
|
10 |
|
11 |
|
12 |
The mmproj files are the embedded ViT's that came with llava-1.6, I've not compared them but given the previous releases from the team I'd be surprised if the ViT has not been fine tuned this time.
|
13 |
+
If that's the case, using another ViT can cause issues.
|
14 |
+
You need to use the mmproj of the right model but you can mix quantizations.
|
15 |
|
16 |
Original models: https://github.com/haotian-liu/LLaVA
|