cicdatopea
commited on
Commit
•
2d7521c
1
Parent(s):
20e4b60
Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ datasets:
|
|
5 |
---
|
6 |
## Model Details
|
7 |
|
8 |
-
This model is an int4 model with group_size 128 and symmetric quantization of [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview) generated by [intel/auto-round](https://github.com/intel/auto-round). We excluded 3 layers from quantization due to the overflow issue on some int4 backends.
|
9 |
|
10 |
## How To Use
|
11 |
|
|
|
5 |
---
|
6 |
## Model Details
|
7 |
|
8 |
+
This model is an int4 model with group_size 128 and symmetric quantization of [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview) generated by [intel/auto-round](https://github.com/intel/auto-round). We excluded 3 layers from quantization due to the overflow issue on some int4 backends. You could find AutoAWQ format [here](https://huggingface.co/OPEA/QwQ-32B-Preview-int4-sym-mixed-awq-inc),which is a little different from this one.
|
9 |
|
10 |
## How To Use
|
11 |
|