DarshanDeshpande commited on
Commit
06b6a01
·
verified ·
1 Parent(s): da98e26

Upload Q8_0 GGUF model

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +23 -3
  3. glider_Q8_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ glider_Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,23 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ base_model:
4
+ - PatronusAI/glider
5
+ ---
6
+
7
+ Available GGUF versions for the [PatronusAI/glider](https://huggingface.co/PatronusAI/glider) model: [`BF16`, `Q8_0`, `Q5_K_M`, `Q4_K_M`]
8
+
9
+ How to load your desired quantized model:
10
+
11
+ 1. Select the appropraite GGUF quantization from the available list above
12
+ 2. Run the following code:
13
+ ```bash
14
+ from transformers import AutoModelForCausalLM
15
+ model = AutoModelForCausalLM.from_pretrained("PatronusAI/glider-gguf", gguf_file="glider_{version_from_list}.gguf")
16
+ ```
17
+ For loading the Q8_0 version, this script will change to:
18
+ ```bash
19
+ from transformers import AutoModelForCausalLM
20
+ model = AutoModelForCausalLM.from_pretrained("PatronusAI/glider-gguf", gguf_file="glider_Q8_0.gguf")
21
+ ```
22
+
23
+ For any issues or questions, reach out to [Darshan Deshpande](https://huggingface.co/darshandeshpande) or [Rebecca Qian](https://huggingface.co/RebeccaQian1)
glider_Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efddf1a35a4d54169f0022553577b83e6d0e5982b11c24cda71098431c7df610
3
+ size 4061222560