Edit model card

iMatrix gguf quants of a newer finetune of Mixtral-8x22B

EdgeQuants still underway, IQ4XS version recommended. Make sure to combine/merge the parts back together before using

cat tessIQ4XS.gguf.part* > tessIQ4XS.gguf 

Then use with llama.cpp version from April 12 or older. April 13 release had massive changes and messed up inferene for MoE models

Downloads last month
23
GGUF
Model size
141B params
Architecture
llama
Unable to determine this model's library. Check the docs .

Quantized from