Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -7,12 +7,18 @@ import torch
|
|
7 |
# Load the model and tokenizer
|
8 |
@st.cache_resource
|
9 |
def load_model():
|
10 |
-
model = AutoModelForCausalLM.from_pretrained(
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
tokenizer = AutoTokenizer.from_pretrained("TheBloke/Mistral-7B-Instruct-v0.2-GPTQ")
|
17 |
config = PeftConfig.from_pretrained("vignesh0007/Hubermangpt")
|
18 |
model = PeftModel.from_pretrained(model, "vignesh0007/Hubermangpt")
|
|
|
7 |
# Load the model and tokenizer
|
8 |
@st.cache_resource
|
9 |
def load_model():
|
10 |
+
# model = AutoModelForCausalLM.from_pretrained(
|
11 |
+
# "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ",
|
12 |
+
# device_map="auto",
|
13 |
+
# torch_dtype=torch.float16,
|
14 |
+
# load_in_8bit=True, # Enable 8-bit quantization
|
15 |
+
# )
|
16 |
+
model_name = "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ"
|
17 |
+
model = AutoModelForCausalLM.from_pretrained(model_name,
|
18 |
+
device_map="auto",
|
19 |
+
trust_remote_code=False,
|
20 |
+
revision="main")
|
21 |
+
|
22 |
tokenizer = AutoTokenizer.from_pretrained("TheBloke/Mistral-7B-Instruct-v0.2-GPTQ")
|
23 |
config = PeftConfig.from_pretrained("vignesh0007/Hubermangpt")
|
24 |
model = PeftModel.from_pretrained(model, "vignesh0007/Hubermangpt")
|