benk04 commited on
Commit
9435b22
1 Parent(s): 962de4a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  <!-- description start -->
8
  Exllamav2 4.65bpw quantization of CausalLM-RP-34B from [NeverSleep](https://huggingface.co/NeverSleep/CausalLM-RP-34B), quantized with default calibration dataset.
9
  > [!IMPORTANT]
10
- >This bpw is the perfect size for 24GB GPUs, and can fit 32k+ context. Make sure to enable 4-bit cache option or you'll run into OOM errors.
11
 
12
  ---
13
  ## Original Card
 
7
  <!-- description start -->
8
  Exllamav2 4.65bpw quantization of CausalLM-RP-34B from [NeverSleep](https://huggingface.co/NeverSleep/CausalLM-RP-34B), quantized with default calibration dataset.
9
  > [!IMPORTANT]
10
+ >Fits in 24GB VRAM with 32k+ context. Make sure to enable 4-bit cache option or you'll run into OOM errors.
11
 
12
  ---
13
  ## Original Card