I ran a quantized gguf, Q5_K_M using llama.cpp on Mac M2 24 GB. Extremely coherent chat and quality! It recognizes Devanagri script as well as Roman transliterated text and correctly responds. Will dig in more. Some pictures, from my custom chat client.