Update README.md
#1
by
alexsoft1
- opened
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
base_model: liuhaotian/llava-v1.5-7b
|
3 |
inference: false
|
4 |
library_name: transformers
|
5 |
-
license:
|
6 |
model_creator: liuhaotian
|
7 |
model_name: Llava v1.5 7B
|
8 |
quantized_by: Second State Inc.
|
@@ -68,4 +68,4 @@ quantized_by: Second State Inc.
|
|
68 |
| [llava-v1.5-7b-Q8_0.gguf](https://huggingface.co/second-state/Llava-v1.5-7B-GGUF/blob/main/llava-v1.5-7b-Q8_0.gguf) | Q8_0 | 8 | 7.16 GB| very large, extremely low quality loss - not recommended |
|
69 |
| [llava-v1.5-7b-mmproj-model-f16.gguf](https://huggingface.co/second-state/Llava-v1.5-7B-GGUF/blob/main/llava-v1.5-7b-mmproj-model-f16.gguf) | f16 | 8 | 624 MB| |
|
70 |
|
71 |
-
*Quantized with llama.cpp b2230*
|
|
|
2 |
base_model: liuhaotian/llava-v1.5-7b
|
3 |
inference: false
|
4 |
library_name: transformers
|
5 |
+
license: bigcode-openrail-m
|
6 |
model_creator: liuhaotian
|
7 |
model_name: Llava v1.5 7B
|
8 |
quantized_by: Second State Inc.
|
|
|
68 |
| [llava-v1.5-7b-Q8_0.gguf](https://huggingface.co/second-state/Llava-v1.5-7B-GGUF/blob/main/llava-v1.5-7b-Q8_0.gguf) | Q8_0 | 8 | 7.16 GB| very large, extremely low quality loss - not recommended |
|
69 |
| [llava-v1.5-7b-mmproj-model-f16.gguf](https://huggingface.co/second-state/Llava-v1.5-7B-GGUF/blob/main/llava-v1.5-7b-mmproj-model-f16.gguf) | f16 | 8 | 624 MB| |
|
70 |
|
71 |
+
*Quantized with llama.cpp b2230*
|