24GB VRAM Optimal Quants Collection When asked what I use locally on a 24GB card, this is what I point to. I favor exl2s for long context, GGUF for very short context. • 12 items • Updated Oct 31 • 3
openaccess-ai-collective/openhermes-2_5-dpo-no-robots Text Generation • Updated Dec 2, 2023 • 24 • 11