|
--- |
|
license: cc-by-nc-4.0 |
|
base_model: |
|
- PatronusAI/glider |
|
--- |
|
|
|
Available GGUF versions for the [PatronusAI/glider](https://huggingface.co/PatronusAI/glider) model: [`BF16`, `Q8_0`, `Q5_K_M`, `Q4_K_M`] |
|
|
|
How to load your desired quantized model: |
|
|
|
1. Select the appropraite GGUF quantization from the available list above |
|
2. Run the following code: |
|
```bash |
|
from transformers import AutoModelForCausalLM |
|
model = AutoModelForCausalLM.from_pretrained("PatronusAI/glider-gguf", gguf_file="glider_{version_from_list}.gguf") |
|
``` |
|
For loading the Q8_0 version, this script will change to: |
|
```bash |
|
from transformers import AutoModelForCausalLM |
|
model = AutoModelForCausalLM.from_pretrained("PatronusAI/glider-gguf", gguf_file="glider_Q8_0.gguf") |
|
``` |
|
|
|
For any issues or questions, reach out to [Darshan Deshpande](https://huggingface.co/darshandeshpande) or [Rebecca Qian](https://huggingface.co/RebeccaQian1) |