Model Details

This is Qwen/QwQ-32B quantized with bitsandbytes in 4-bit. The model has been created, tested, and evaluated by The Kaitchup. The model is compatible with vLLM and Transformers.

image/png

Details on the quantization process and how to use the model here: The Kaitchup

  • Developed by: The Kaitchup
  • Language(s) (NLP): English
  • License: Apache 2.0 license

How to Support My Work

Subscribe to The Kaitchup. This helps me a lot to continue quantizing and evaluating models for free.

Downloads last month
17
Safetensors
Model size
17.7B params
Tensor type
F32
·
BF16
·
U8
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for kaitchup/QwQ-32B-bnb-4bit

Base model

Qwen/Qwen2.5-32B
Finetuned
Qwen/QwQ-32B
Quantized
(93)
this model

Collection including kaitchup/QwQ-32B-bnb-4bit