Ahren09 commited on
Commit
cd63fc2
·
verified ·
1 Parent(s): 975ce56

Upload builder.py

Browse files
Files changed (1) hide show
  1. llava/model/builder.py +2 -1
llava/model/builder.py CHANGED
@@ -25,6 +25,7 @@ from llava.constants import DEFAULT_IMAGE_PATCH_TOKEN, DEFAULT_IM_START_TOKEN, D
25
 
26
  def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda"):
27
  kwargs = {"device_map": device_map}
 
28
 
29
  if load_8bit:
30
  kwargs['load_in_8bit'] = True
@@ -100,7 +101,7 @@ def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, l
100
  model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)
101
  else:
102
  tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False)
103
- model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)
104
  else:
105
  # Load language model
106
  if model_base is not None:
 
25
 
26
  def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda"):
27
  kwargs = {"device_map": device_map}
28
+ os.makedirs("offload_folder", exist_ok=True)
29
 
30
  if load_8bit:
31
  kwargs['load_in_8bit'] = True
 
101
  model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)
102
  else:
103
  tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False)
104
+ model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, offload_folder="offload_folder",**kwargs)
105
  else:
106
  # Load language model
107
  if model_base is not None: