lrl-modelcloud
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -47,7 +47,7 @@ This model has been quantized using [GPTQModel](https://github.com/ModelCloud/GP
|
|
47 |
from transformers import AutoTokenizer
|
48 |
from gptqmodel import GPTQModel
|
49 |
|
50 |
-
model_name = "ModelCloud/Llama-3.2-3B-Instruct-gptqmodel-4bit-
|
51 |
|
52 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
53 |
model = GPTQModel.from_quantized(model_name)
|
|
|
47 |
from transformers import AutoTokenizer
|
48 |
from gptqmodel import GPTQModel
|
49 |
|
50 |
+
model_name = "ModelCloud/Llama-3.2-3B-Instruct-gptqmodel-4bit-vortex-v3"
|
51 |
|
52 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
53 |
model = GPTQModel.from_quantized(model_name)
|