vignesh0007 commited on
Commit
6134863
·
verified ·
1 Parent(s): 4cea5b8

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +12 -6
app.py CHANGED
@@ -7,12 +7,18 @@ import torch
7
  # Load the model and tokenizer
8
  @st.cache_resource
9
  def load_model():
10
- model = AutoModelForCausalLM.from_pretrained(
11
- "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ",
12
- device_map="auto",
13
- torch_dtype=torch.float16,
14
- load_in_8bit=True, # Enable 8-bit quantization
15
- )
 
 
 
 
 
 
16
  tokenizer = AutoTokenizer.from_pretrained("TheBloke/Mistral-7B-Instruct-v0.2-GPTQ")
17
  config = PeftConfig.from_pretrained("vignesh0007/Hubermangpt")
18
  model = PeftModel.from_pretrained(model, "vignesh0007/Hubermangpt")
 
7
  # Load the model and tokenizer
8
  @st.cache_resource
9
  def load_model():
10
+ # model = AutoModelForCausalLM.from_pretrained(
11
+ # "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ",
12
+ # device_map="auto",
13
+ # torch_dtype=torch.float16,
14
+ # load_in_8bit=True, # Enable 8-bit quantization
15
+ # )
16
+ model_name = "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ"
17
+ model = AutoModelForCausalLM.from_pretrained(model_name,
18
+ device_map="auto",
19
+ trust_remote_code=False,
20
+ revision="main")
21
+
22
  tokenizer = AutoTokenizer.from_pretrained("TheBloke/Mistral-7B-Instruct-v0.2-GPTQ")
23
  config = PeftConfig.from_pretrained("vignesh0007/Hubermangpt")
24
  model = PeftModel.from_pretrained(model, "vignesh0007/Hubermangpt")