marattt commited on
Commit
f9d6809
1 Parent(s): fcc96b1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -30,6 +30,7 @@ pipeline_tag: text-generation
30
  - **Attention Heads (GQA)**: 24 for Q, 4 for KV
31
  - **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
32
  - **Quantization**: AWQ 4 bit
 
33
 
34
  ### Requirements
35
  The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.
 
30
  - **Attention Heads (GQA)**: 24 for Q, 4 for KV
31
  - **Context Length**: Supports a full context of 131,072 tokens and generation of up to 8,192 tokens
32
  - **Quantization**: AWQ 4 bit
33
+ - **Base model**: Qwen/Qwen2.5-3B-Instruct-AWQ
34
 
35
  ### Requirements
36
  The code of Qwen2.5 has been in the latest Hugging face transformers and we advise you to use the latest version of transformers.