Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,34 @@
|
|
1 |
-
---
|
2 |
-
license: cc-by-nc-4.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-4.0
|
3 |
+
tags:
|
4 |
+
- not-for-all-audiences
|
5 |
+
- nsfw
|
6 |
+
---
|
7 |
+
<!-- description start -->
|
8 |
+
Exllamav2 4.65bpw quantization of CausalLM-RP-34B from [NeverSleep](https://huggingface.co/NeverSleep/CausalLM-RP-34B), quantized with default calibration dataset.
|
9 |
+
> [!IMPORTANT]
|
10 |
+
>This bpw is the perfect size for 24GB GPUs, and can fit 32k+ context. Make sure to enable 4-bit cache option or you'll run into OOM errors.
|
11 |
+
|
12 |
+
---
|
13 |
+
## Original Card
|
14 |
+
## Description
|
15 |
+
|
16 |
+
This repo contains fp16 files of CausalLM-RP-34B, a finetuned model of the CausalLM-34B Beta on multiple RP datasets.
|
17 |
+
|
18 |
+
<!-- description end -->
|
19 |
+
<!-- description start -->
|
20 |
+
## Model used
|
21 |
+
|
22 |
+
- [CausalLM/34b-beta](https://huggingface.co/CausalLM/34b-beta)
|
23 |
+
|
24 |
+
### Prompt template ChatML
|
25 |
+
|
26 |
+
|
27 |
+
```
|
28 |
+
<|im_start|>system
|
29 |
+
{system_prompt}<|im_end|>
|
30 |
+
<|im_start|>user
|
31 |
+
{prompt}<|im_end|>
|
32 |
+
<|im_start|>assistant
|
33 |
+
{output}<|im_end|>
|
34 |
+
```
|