marattt commited on
Commit
fcc96b1
1 Parent(s): 6a32fcf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -29,6 +29,7 @@ pipeline_tag: text-generation
29
  - **Layers**: 36
30
  - **Attention Heads (GQA)**: 24 for Q, 4 for KV
31
  - **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
 
32
 
33
  ### Requirements
34
  The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.
 
29
  - **Layers**: 36
30
  - **Attention Heads (GQA)**: 24 for Q, 4 for KV
31
  - **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
32
+ - **Quantization**: AWQ 4 bit
33
 
34
  ### Requirements
35
  The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.