OPEA
/

Safetensors
qwen2
4-bit precision
intel/auto-round
cicdatopea commited on
Commit
2d7521c
1 Parent(s): 20e4b60

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -5,7 +5,7 @@ datasets:
5
  ---
6
  ## Model Details
7
 
8
- This model is an int4 model with group_size 128 and symmetric quantization of [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview) generated by [intel/auto-round](https://github.com/intel/auto-round). We excluded 3 layers from quantization due to the overflow issue on some int4 backends.
9
 
10
  ## How To Use
11
 
 
5
  ---
6
  ## Model Details
7
 
8
+ This model is an int4 model with group_size 128 and symmetric quantization of [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview) generated by [intel/auto-round](https://github.com/intel/auto-round). We excluded 3 layers from quantization due to the overflow issue on some int4 backends. You could find AutoAWQ format [here](https://huggingface.co/OPEA/QwQ-32B-Preview-int4-sym-mixed-awq-inc),which is a little different from this one.
9
 
10
  ## How To Use
11