Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,10 @@ license: apache-2.0
|
|
10 |
|
11 |
<p><h1> speechless-mistral-moloras-7b </h1></p>
|
12 |
|
|
|
|
|
|
|
|
|
13 |
[4-bit GGUF models for CPU+GPU inference](https://huggingface.co/uukuguy/speechless-mistral-moloras-7b/tree/main/GGUF)
|
14 |
|
15 |
This model is the static version of moloras (Mixture-of-multi-LoRAs) based on the following 6 Mistral-based LoRa modules.
|
|
|
10 |
|
11 |
<p><h1> speechless-mistral-moloras-7b </h1></p>
|
12 |
|
13 |
+
* [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/speechless-mistral-moloras-7B-AWQ)
|
14 |
+
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/speechless-mistral-moloras-7B-GPTQ)
|
15 |
+
* [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/speechless-mistral-moloras-7B-GGUF)
|
16 |
+
|
17 |
[4-bit GGUF models for CPU+GPU inference](https://huggingface.co/uukuguy/speechless-mistral-moloras-7b/tree/main/GGUF)
|
18 |
|
19 |
This model is the static version of moloras (Mixture-of-multi-LoRAs) based on the following 6 Mistral-based LoRa modules.
|