lym0302 commited on
Commit
c7895e7
·
1 Parent(s): 9ebfdc2
third_party/VideoLLaMA2/videollama2/model/__init__.py CHANGED
@@ -53,7 +53,7 @@ VLLMConfigs = {
53
 
54
 
55
  def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", use_flash_attn=False, **kwargs):
56
- print("00000000000000000000000000: ", device)
57
  if 'token' in kwargs:
58
  token = kwargs['token']
59
  else:
@@ -181,7 +181,7 @@ def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, l
181
  elif model_type in ['videollama2_mixtral']:
182
  model = Videollama2MixtralForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, config=config, **kwargs)
183
  elif model_type in ['videollama2_qwen2']:
184
- model = Videollama2Qwen2ForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, config=config, **kwargs)
185
  elif model_type in ['videollama2_gemma2']:
186
  model = Videollama2Gemma2ForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, config=config, **kwargs)
187
  elif model_type in ['videollama2_phi3']:
 
53
 
54
 
55
  def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", use_flash_attn=False, **kwargs):
56
+ print("00000000000000000000000000: ", device, use_flash_attn)
57
  if 'token' in kwargs:
58
  token = kwargs['token']
59
  else:
 
181
  elif model_type in ['videollama2_mixtral']:
182
  model = Videollama2MixtralForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, config=config, **kwargs)
183
  elif model_type in ['videollama2_qwen2']:
184
+ model = Videollama2Qwen2ForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, torch_dtype=torch.bfloat16, config=config, **kwargs)
185
  elif model_type in ['videollama2_gemma2']:
186
  model = Videollama2Gemma2ForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, config=config, **kwargs)
187
  elif model_type in ['videollama2_phi3']: