benk04 commited on
Commit
962de4a
1 Parent(s): 45662f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -3
README.md CHANGED
@@ -1,3 +1,34 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - not-for-all-audiences
5
+ - nsfw
6
+ ---
7
+ <!-- description start -->
8
+ Exllamav2 4.65bpw quantization of CausalLM-RP-34B from [NeverSleep](https://huggingface.co/NeverSleep/CausalLM-RP-34B), quantized with default calibration dataset.
9
+ > [!IMPORTANT]
10
+ >This bpw is the perfect size for 24GB GPUs, and can fit 32k+ context. Make sure to enable 4-bit cache option or you'll run into OOM errors.
11
+
12
+ ---
13
+ ## Original Card
14
+ ## Description
15
+
16
+ This repo contains fp16 files of CausalLM-RP-34B, a finetuned model of the CausalLM-34B Beta on multiple RP datasets.
17
+
18
+ <!-- description end -->
19
+ <!-- description start -->
20
+ ## Model used
21
+
22
+ - [CausalLM/34b-beta](https://huggingface.co/CausalLM/34b-beta)
23
+
24
+ ### Prompt template ChatML
25
+
26
+
27
+ ```
28
+ <|im_start|>system
29
+ {system_prompt}<|im_end|>
30
+ <|im_start|>user
31
+ {prompt}<|im_end|>
32
+ <|im_start|>assistant
33
+ {output}<|im_end|>
34
+ ```