File size: 850 Bytes
18ab389
 
 
 
bb8a221
18ab389
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
---
license: other
license_name: glm-4-9b-license
license_link: https://huggingface.co/THUDM/glm-4-9b-chat/blob/main/LICENSE
base_model: THUDM/LongWriter-glm4-9b
datasets:
- THUDM/LongWriter-6k
language:
- en
pipeline_tag: text-generation
---

# LongWriter-glm4-9b

Original model link: https://huggingface.co/THUDM/LongWriter-glm4-9b

Model by: **THUDM**

Quants by: **QuantPanda**

GGUF quantization for [llama.cpp](https://github.com/ggerganov/llama.cpp) and similar applications.

**Example:**

``./llama-cli -m LongWriter-glm4-9B-Q5_K_M.gguf -p "You are a helpful AI assistant." --conversation``

If the model takes too long to load you can reduce the context size with ```--ctx-size```

**Example with smaller context size:**

``./llama-cli -m LongWriter-glm4-9B-Q5_K_M.gguf -p "You are a helpful AI assistant." --conversation --ctx-size 4096``