InferenceIllusionist commited on
Commit
e1591f2
1 Parent(s): 36796e5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -26,10 +26,12 @@ PROUDLY PRESENTS
26
 
27
  ## Neophanis-8x7B-iMat-GGUF
28
 
 
 
29
  Quantized from fp16 with love.
30
  * Quantizations made possible using mixtral-8x7b-instruct-v0.1.imatrix file from [this](https://huggingface.co/datasets/ikawrakow/imatrix-from-wiki-train) repo (special thanks to [ikawrakow](https://huggingface.co/ikawrakow) again)
31
-
32
- For a brief rundown of iMatrix quant performance please see this [PR](https://github.com/ggerganov/llama.cpp/pull/5747)
33
 
34
  <i>All quants are verified working prior to uploading to repo for your safety and convenience. </i>
35
 
 
26
 
27
  ## Neophanis-8x7B-iMat-GGUF
28
 
29
+ <b>The Good, The Bad, And The Ugly iMats edition</b>
30
+
31
  Quantized from fp16 with love.
32
  * Quantizations made possible using mixtral-8x7b-instruct-v0.1.imatrix file from [this](https://huggingface.co/datasets/ikawrakow/imatrix-from-wiki-train) repo (special thanks to [ikawrakow](https://huggingface.co/ikawrakow) again)
33
+ * An analysis was run on mixtral-8x7b.imatrix that showed worse KL-Divergence than mixtral-8x7b-instruct-v0.1, hence the latter was used for the imatrixes instead
34
+ * For a brief rundown of iMatrix quant performance please see this [PR](https://github.com/ggerganov/llama.cpp/pull/5747)
35
 
36
  <i>All quants are verified working prior to uploading to repo for your safety and convenience. </i>
37