GGUF model
#1
by
3XTron
- opened
Please upload the GGUF model.
llama.cpp doesn't support Llama 3.2 Vision yet. Last I heard, they're working on redoing their whole multimodal implementation, so that will take a while.
Hello! Could you show me way to quantization you tune for ollama format? App or github link?
Now, ollama have q4_k_m for base models 3.2 vision in self library.
This comment has been hidden