bartowski commited on
Commit
dc510ca
1 Parent(s): cdd27da

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -2
README.md CHANGED
@@ -188,9 +188,17 @@ extra_gated_button_content: Submit
188
  quantized_by: bartowski
189
  ---
190
 
191
- ## Llamacpp iMatrix Quantizations of Meta-Llama-3-8B-Instruct
192
 
193
- <b>Warning: This conversion is based on the merged Llama 3 support in llama.cpp (release b6710), and you will need to update your inference tool to be on at least version 6710 of llama.cpp, this will vary across tools. Until then, you can use the older style with <|eot_id|> special = false [here](https://huggingface.co/bartowski/Meta-Llama-3-8B-Instruct-GGUF-old)</b>
 
 
 
 
 
 
 
 
194
 
195
  Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b2710">b2710</a> for quantization.
196
 
 
188
  quantized_by: bartowski
189
  ---
190
 
191
+ ## Llamacpp imatrix Quantizations of Meta-Llama-3-8B-Instruct
192
 
193
+ <b>This conversion is based on the merged Llama 3 support in llama.cpp (release b6710)</b>
194
+
195
+ # Known working on:
196
+ - LM Studio 0.2.20
197
+
198
+ # Confirmed not working on (as of April 21):
199
+ - text-generation-webui master/dev
200
+
201
+ Any others unknown, feel free to comment
202
 
203
  Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b2710">b2710</a> for quantization.
204