Update README.md
Browse files
README.md
CHANGED
@@ -32,7 +32,6 @@ It is the result of merging and/or converting the source repository to float16.
|
|
32 |
## Repositories available
|
33 |
|
34 |
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GPTQ)
|
35 |
-
* [GGML experimental 4, 5, 6 and 8-bit models for CPU only inference](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GGML)
|
36 |
* [Merged fp16 model in pytorch model format for GPU inference and further conversions](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-fp16)
|
37 |
* [Original QLoRA model](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora)
|
38 |
|
|
|
32 |
## Repositories available
|
33 |
|
34 |
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GPTQ)
|
|
|
35 |
* [Merged fp16 model in pytorch model format for GPU inference and further conversions](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-fp16)
|
36 |
* [Original QLoRA model](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora)
|
37 |
|