Update README.md
Browse files
README.md
CHANGED
@@ -30,6 +30,7 @@ pipeline_tag: text-generation
|
|
30 |
- **Attention Heads (GQA)**: 24 for Q, 4 for KV
|
31 |
- **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
|
32 |
- **Quantization**: AWQ 4 bit
|
|
|
33 |
|
34 |
### Requirements
|
35 |
The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.
|
|
|
30 |
- **Attention Heads (GQA)**: 24 for Q, 4 for KV
|
31 |
- **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
|
32 |
- **Quantization**: AWQ 4 bit
|
33 |
+
- **Base model**: Qwen/Qwen2.5-3B-Instruct-AWQ
|
34 |
|
35 |
### Requirements
|
36 |
The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.
|