iman37 commited on
Commit
dd22465
·
verified ·
1 Parent(s): cc7c86a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -8
app.py CHANGED
@@ -6,17 +6,12 @@ import torch
6
  # Load the model and tokenizer with 4-bit quantization
7
  @st.cache_resource
8
  def load_model():
9
- tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2-7B")
10
  model = AutoModelForCausalLM.from_pretrained(
11
- "Qwen/Qwen2-7B",
12
  load_in_4bit=True,
13
  device_map="auto",
14
- quantization_config=bnb.QuantizationConfig(
15
- load_in_4bit=True,
16
- bnb_4bit_use_double_quant=True,
17
- bnb_4bit_quant_type="nf4",
18
- bnb_4bit_compute_dtype=torch.float16
19
- )
20
  )
21
  return tokenizer, model
22
 
 
6
  # Load the model and tokenizer with 4-bit quantization
7
  @st.cache_resource
8
  def load_model():
9
+ tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen-7B", trust_remote_code=True)
10
  model = AutoModelForCausalLM.from_pretrained(
11
+ "Qwen/Qwen-7B",
12
  load_in_4bit=True,
13
  device_map="auto",
14
+ trust_remote_code=True # Allows custom code execution
 
 
 
 
 
15
  )
16
  return tokenizer, model
17