bnjmnmarie
commited on
Commit
•
ef8b17f
1
Parent(s):
f414bdc
Update README.md
Browse files
README.md
CHANGED
@@ -2,4 +2,17 @@
|
|
2 |
license: mit
|
3 |
---
|
4 |
|
5 |
-
Llama 2 7B quantized in 2-bit with GPTQ.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
license: mit
|
3 |
---
|
4 |
|
5 |
+
Llama 2 7B quantized in 2-bit with GPTQ.
|
6 |
+
|
7 |
+
```
|
8 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
9 |
+
from optimum.gptq import GPTQQuantizer
|
10 |
+
import torch
|
11 |
+
w = 2
|
12 |
+
model_path = meta-llama/Llama-2-7b-chat-hf
|
13 |
+
|
14 |
+
tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True)
|
15 |
+
model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.float16)
|
16 |
+
quantizer = GPTQQuantizer(bits=w, dataset="c4", model_seqlen = 4096)
|
17 |
+
quantized_model = quantizer.quantize_model(model, tokenizer)
|
18 |
+
```
|