Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ Other quantized models are available from TheBloke: [GGML](https://huggingface.c
|
|
16 |
| - | 7 | 6.1056 | 2048 max context size for T4 GPU |
|
17 |
| - | 8 | 6.1027 | Just, why? |
|
18 |
|
19 |
-
I'll upload the 7 and 8 bits quant if someone request it. (Idk y the 5 bits quant preplexity is lower than higher bits quant,
|
20 |
|
21 |
## Prompt Format
|
22 |
|
|
|
16 |
| - | 7 | 6.1056 | 2048 max context size for T4 GPU |
|
17 |
| - | 8 | 6.1027 | Just, why? |
|
18 |
|
19 |
+
I'll upload the 7 and 8 bits quant if someone request it. (Idk y the 5 bits quant preplexity is lower than higher bits quant, I think I did something wrong?)
|
20 |
|
21 |
## Prompt Format
|
22 |
|