hierholzer
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -1,13 +1,12 @@
|
|
1 |
-
---
|
2 |
-
license:
|
3 |
-
language:
|
4 |
-
- en
|
5 |
-
---
|
6 |
|
|
|
7 |
|
8 |
-
|
9 |
-
|
10 |
-
# Model<br>
|
11 |
|
12 |
|
13 |
Here is a Quantized version of Llama-3.1-70B-Instruct using GGUF<br>
|
@@ -19,39 +18,45 @@ Models initially developed in frameworks like PyTorch can be converted to GGUF f
|
|
19 |
|
20 |
## Uploaded Quantization Types<br>
|
21 |
|
22 |
-
Currently, I have uploaded 2 quantized versions
|
23 |
-
|
24 |
-
Q5_K_M : - large, very low quality loss<br>
|
25 |
-
|
26 |
-
and<br>
|
27 |
|
28 |
-
|
|
|
|
|
29 |
|
30 |
### All Quantization Types Possible
|
31 |
|
32 |
Here are all of the Quantization Types that are Possible. Let me know if you need any other versions
|
33 |
|
34 |
-
|
|
35 |
-
|
36 |
-
|
|
37 |
-
|
|
38 |
-
|
|
39 |
-
|
|
40 |
-
|
|
41 |
-
| 12
|
42 |
-
|
|
43 |
-
|
|
44 |
-
|
|
45 |
-
| 15
|
46 |
-
|
|
47 |
-
|
|
48 |
-
| 17
|
49 |
-
|
|
50 |
-
|
|
51 |
-
|
|
52 |
-
|
|
53 |
-
|
|
|
54 |
|
55 |
## Uses
|
56 |
|
57 |
-
By using the GGUF version of Llama-3.1-70B-Instruct, you will be able to run this LLM while having to use significantly less resources than you would using the non quantized version.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
---
|
6 |
|
7 |
+
[![Hierholzer Banner](https://tvtime.us/static/images/LLAMA3.1.jpg)](#)
|
8 |
|
9 |
+
# Model
|
|
|
|
|
10 |
|
11 |
|
12 |
Here is a Quantized version of Llama-3.1-70B-Instruct using GGUF<br>
|
|
|
18 |
|
19 |
## Uploaded Quantization Types<br>
|
20 |
|
21 |
+
Currently, I have uploaded 2 quantized versions:
|
|
|
|
|
|
|
|
|
22 |
|
23 |
+
- [x] Q5_K_M ~ *Recommended*
|
24 |
+
- [x] Q8_0
|
25 |
+
- [ ] Q4_K_M
|
26 |
|
27 |
### All Quantization Types Possible
|
28 |
|
29 |
Here are all of the Quantization Types that are Possible. Let me know if you need any other versions
|
30 |
|
31 |
+
| **#** | **or** | **Q#** | **:** | _Description Of Quantization Types_ |
|
32 |
+
|-------|:------:|:------:|:-----:|----------------------------------------------------------------|
|
33 |
+
| 2 | or | Q4_0 | : | small, very high quality loss - legacy, prefer using Q3_K_M |
|
34 |
+
| 3 | or | Q4_1 | : | small, substantial quality loss - legacy, prefer using Q3_K_L |
|
35 |
+
| 8 | or | Q5_0 | : | medium, balanced quality - legacy, prefer using Q4_K_M |
|
36 |
+
| 9 | or | Q5_1 | : | medium, low quality loss - legacy, prefer using Q5_K_M |
|
37 |
+
| 10 | or | Q2_K | : | smallest, extreme quality loss - *NOT Recommended* |
|
38 |
+
| 12 | or | Q3_K | : | alias for Q3_K_M |
|
39 |
+
| 11 | or | Q3_K_S | : | very small, very high quality loss |
|
40 |
+
| 12 | or | Q3_K_M | : | very small, high quality loss |
|
41 |
+
| 13 | or | Q3_K_L | : | small, high quality loss |
|
42 |
+
| 15 | or | Q4_K | : | alias for Q4_K_M |
|
43 |
+
| 14 | or | Q4_K_S | : | small, some quality loss |
|
44 |
+
| 15 | or | Q4_K_M | : | medium, balanced quality - *Recommended* |
|
45 |
+
| 17 | or | Q5_K | : | alias for Q5_K_M |
|
46 |
+
| 16 | or | Q5_K_S | : | large, low quality loss - *Recommended* |
|
47 |
+
| 17 | or | Q5_K_M | : | large, very low quality loss - *Recommended* |
|
48 |
+
| 18 | or | Q6_K | : | very large, very low quality loss |
|
49 |
+
| 7 | or | Q8_0 | : | very large, extremely low quality loss |
|
50 |
+
| 1 | or | F16 | : | extremely large, virtually no quality loss - *NOT Recommended* |
|
51 |
+
| 0 | or | F32 | : | absolutely huge, lossless - *NOT Recommended* |
|
52 |
|
53 |
## Uses
|
54 |
|
55 |
+
By using the GGUF version of Llama-3.1-70B-Instruct, you will be able to run this LLM while having to use significantly less resources than you would using the non quantized version.
|
56 |
+
|
57 |
+
[![Hugging Face](https://img.shields.io/badge/Hugging%20Face-FFD21E?logo=huggingface&logoColor=000)](#)
|
58 |
+
[![OS](https://img.shields.io/badge/OS-linux%2C%20windows%2C%20macOS-0078D4)](https://docs.abblix.com/docs/technical-requirements)
|
59 |
+
[![CPU](https://img.shields.io/badge/CPU-x86%2C%20x64%2C%20ARM%2C%20ARM64-FF8C00)](https://docs.abblix.com/docs/technical-requirements)
|
60 |
+
[![forthebadge](https://forthebadge.com/images/badges/license-mit.svg)](https://forthebadge.com)
|
61 |
+
[![forthebadge](https://forthebadge.com/images/badges/made-with-python.svg)](https://forthebadge.com)
|
62 |
+
[![forthebadge](https://forthebadge.com/images/badges/powered-by-electricity.svg)](https://forthebadge.com)
|