dad1909 commited on
Commit
d4b1572
·
verified ·
1 Parent(s): 879fab0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -18,7 +18,7 @@ print(f"stage ${current_num}")
18
 
19
  api = HfApi(token=hf_token)
20
  # models = f"dad1909/cybersentinal-2.0-{current_num}"
21
- model_base = "unsloth/Meta-Llama-3.1-70B-Instruct-bnb-4bit"
22
 
23
  print("Starting model and tokenizer loading...")
24
 
@@ -28,8 +28,7 @@ model, tokenizer = FastLanguageModel.from_pretrained(
28
  max_seq_length=max_seq_length,
29
  dtype=dtype,
30
  load_in_4bit=load_in_4bit,
31
- token=hf_token,
32
- load_in_8bit_fp32_cpu_offload=True, # Enable CPU offloading
33
  )
34
 
35
  print("Model and tokenizer loaded successfully.")
 
18
 
19
  api = HfApi(token=hf_token)
20
  # models = f"dad1909/cybersentinal-2.0-{current_num}"
21
+ model_base = "unsloth/gemma-2-27b-bnb-4bit"
22
 
23
  print("Starting model and tokenizer loading...")
24
 
 
28
  max_seq_length=max_seq_length,
29
  dtype=dtype,
30
  load_in_4bit=load_in_4bit,
31
+ token=hf_token
 
32
  )
33
 
34
  print("Model and tokenizer loaded successfully.")