Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,8 @@ Update: PR is merged, llama.cpp now natively supports these models
|
|
5 |
Important: Verify that processing a simple question with any image at least uses 1200 tokens of prompt processing, that shows that the new PR is in use.
|
6 |
If your prompt is just 576 + a few tokens, you are using llava-1.5 code (or projector) and this is incompatible with llava-1.6
|
7 |
|
|
|
|
|
8 |
|
9 |
The mmproj files are the embedded ViT's that came with llava-1.6, I've not compared them but given the previous releases from the team I'd be surprised if the ViT has not been fine tuned this time.
|
10 |
If that's the case, using another ViT can cause issues.
|
|
|
5 |
Important: Verify that processing a simple question with any image at least uses 1200 tokens of prompt processing, that shows that the new PR is in use.
|
6 |
If your prompt is just 576 + a few tokens, you are using llava-1.5 code (or projector) and this is incompatible with llava-1.6
|
7 |
|
8 |
+
**note** Currently llama.cpp has full lava-1.6 support but not the server example. The server example is still using llava-1.5 style hardcoded routines and needs a hotfix to adopt the llava.cpp functions instead.
|
9 |
+
|
10 |
|
11 |
The mmproj files are the embedded ViT's that came with llava-1.6, I've not compared them but given the previous releases from the team I'd be surprised if the ViT has not been fine tuned this time.
|
12 |
If that's the case, using another ViT can cause issues.
|