Posting these Qwen-14B-Chat quantized models in GGUF format for use with llama.cpp due to a user request.

But, having used an importance matrix derived from English-only training data in the quantization, I have no idea how these models will perform in Chinese.

Downloads last month
15
GGUF
Model size
14.2B params
Architecture
qwen
Inference API
Unable to determine this model's library. Check the docs .