iMatrix gguf quants of a newer finetune of Mixtral-8x22B

EdgeQuants still underway, IQ4XS version recommended. Make sure to combine/merge the parts back together before using

cat tessIQ4XS.gguf.part* > tessIQ4XS.gguf 

Then use with llama.cpp version from April 12 or older. April 13 release had massive changes and messed up inferene for MoE models

Downloads last month
11
GGUF
Model size
141B params
Architecture
llama
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for nisten/Tess-Mixtral-8x22B-imatrix-gguf

Quantized
(2)
this model