GGUF model

#1
by 3XTron - opened

Please upload the GGUF model.

llama.cpp doesn't support Llama 3.2 Vision yet. Last I heard, they're working on redoing their whole multimodal implementation, so that will take a while.

Hello! Could you show me way to quantization you tune for ollama format? App or github link?
Now, ollama have q4_k_m for base models 3.2 vision in self library.

This comment has been hidden

Sign up or log in to comment