cmp-nct commited on
Commit
f2c82be
1 Parent(s): 4d811ed

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -6,10 +6,11 @@ Update: PR is merged, llama.cpp now natively supports these models
6
  Important: Verify that processing a simple question with any image at least uses 1200 tokens of prompt processing, that shows that the new PR is in use.
7
  If your prompt is just 576 + a few tokens, you are using llava-1.5 code (or projector) and this is incompatible with llava-1.6
8
 
9
- **note** Currently llama.cpp has full lava-1.6 support but not the server example. The server example is still using llava-1.5 style hardcoded routines and needs a hotfix to adopt the llava.cpp functions instead.
10
 
11
 
12
  The mmproj files are the embedded ViT's that came with llava-1.6, I've not compared them but given the previous releases from the team I'd be surprised if the ViT has not been fine tuned this time.
13
- If that's the case, using another ViT can cause issues.
 
14
 
15
  Original models: https://github.com/haotian-liu/LLaVA
 
6
  Important: Verify that processing a simple question with any image at least uses 1200 tokens of prompt processing, that shows that the new PR is in use.
7
  If your prompt is just 576 + a few tokens, you are using llava-1.5 code (or projector) and this is incompatible with llava-1.6
8
 
9
+ **note** Keep in mind the different fine tunes as described in the llama.cpp llava readme, it's essential to use the non defaults for non vicuna models
10
 
11
 
12
  The mmproj files are the embedded ViT's that came with llava-1.6, I've not compared them but given the previous releases from the team I'd be surprised if the ViT has not been fine tuned this time.
13
+ If that's the case, using another ViT can cause issues.
14
+ You need to use the mmproj of the right model but you can mix quantizations.
15
 
16
  Original models: https://github.com/haotian-liu/LLaVA