Update README.md
Browse files
README.md
CHANGED
@@ -29,6 +29,7 @@ pipeline_tag: text-generation
|
|
29 |
- **Layers**: 36
|
30 |
- **Attention Heads (GQA)**: 24 for Q, 4 for KV
|
31 |
- **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
|
|
|
32 |
|
33 |
### Requirements
|
34 |
The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.
|
|
|
29 |
- **Layers**: 36
|
30 |
- **Attention Heads (GQA)**: 24 for Q, 4 for KV
|
31 |
- **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
|
32 |
+
- **Quantization**: AWQ 4 bit
|
33 |
|
34 |
### Requirements
|
35 |
The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.
|