Update README.md
Browse files
README.md
CHANGED
@@ -7,4 +7,26 @@ tags:
|
|
7 |
- quantization
|
8 |
- shuvom/yuj-v1
|
9 |
license: apache-2.0
|
10 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
- quantization
|
8 |
- shuvom/yuj-v1
|
9 |
license: apache-2.0
|
10 |
+
---
|
11 |
+
# yuj-v1-GGUF
|
12 |
+
- Model creator: [shuvom_](https://huggingface.co/shuvom)
|
13 |
+
- Original model: [shuvom/yuj-v1](https://huggingface.co/shuvom/yuj-v1)
|
14 |
+
|
15 |
+
<!-- description start -->
|
16 |
+
## Description
|
17 |
+
|
18 |
+
This repo contains GGUF format model files for [shuvom/yuj-v1](https://huggingface.co/shuvom/yuj-v1).
|
19 |
+
|
20 |
+
|
21 |
+
<!-- description end -->
|
22 |
+
<!-- README_GGUF.md-about-gguf start -->
|
23 |
+
### About GGUF
|
24 |
+
|
25 |
+
GGUF and GGML are file formats used for storing models for inference, especially in the context of language models like GPT (Generative Pre-trained Transformer).
|
26 |
+
[more info](https://medium.com/@phillipgimmi/what-is-gguf-and-ggml-e364834d241c)
|
27 |
+
|
28 |
+
## Provided files
|
29 |
+
|
30 |
+
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
31 |
+
| ---- | ---- | ---- | ---- | ---- | ----- |
|
32 |
+
| [yuj-v1.Q4_K_M.gguf](https://huggingface.co/shuvom/yuj-v1-GGUF/blob/main/yuj-v1.Q4_K_M.gguf) | Q4_K_M | 4 | 4.37 GB| 6.87 GB | medium, balanced quality - recommended |
|