Model produces garbage
#2
by
WaveCut
- opened
Unfortunately, there are ongoing issues with llama3 tokenization in GGUF and I'm regenerating quants as potential fixes come in. Until then, I would recommend the mlx quants if you're on max otherwise exl2.