Wisdom Chen commited on
Commit
1d2ec06
·
unverified ·
1 Parent(s): c4a2b95

Update model.py

Browse files
Files changed (1) hide show
  1. model.py +1 -4
model.py CHANGED
@@ -54,7 +54,6 @@ def initialize_models() -> bool:
54
 
55
  # Initialize CLIP model with error handling and fallback
56
  try:
57
- # First try loading with device mapping
58
  clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
59
  'hf-hub:Marqo/marqo-fashionCLIP',
60
  device=device
@@ -66,7 +65,6 @@ def initialize_models() -> bool:
66
  print(f"CLIP initialization error: {str(e)}")
67
  print("Attempting to load CLIP model with CPU fallback...")
68
  try:
69
- # Fallback to CPU
70
  device = "cpu"
71
  clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
72
  'hf-hub:Marqo/marqo-fashionCLIP',
@@ -80,7 +78,6 @@ def initialize_models() -> bool:
80
 
81
  # Initialize LLM with optimized settings
82
  try:
83
- # Check for HF_TOKEN and authenticate
84
  hf_token = os.environ.get("HF_TOKEN")
85
  if not hf_token:
86
  raise RuntimeError("HF_TOKEN environment variable is not set")
@@ -109,7 +106,7 @@ def initialize_models() -> bool:
109
  device_map="auto",
110
  torch_dtype=torch.float16,
111
  token=hf_token,
112
- low_cpu_mem_usage=False # Disable low CPU memory usage to prevent meta tensor issues
113
  )
114
  llm_model.eval()
115
  print("LLM initialized successfully")
 
54
 
55
  # Initialize CLIP model with error handling and fallback
56
  try:
 
57
  clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
58
  'hf-hub:Marqo/marqo-fashionCLIP',
59
  device=device
 
65
  print(f"CLIP initialization error: {str(e)}")
66
  print("Attempting to load CLIP model with CPU fallback...")
67
  try:
 
68
  device = "cpu"
69
  clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
70
  'hf-hub:Marqo/marqo-fashionCLIP',
 
78
 
79
  # Initialize LLM with optimized settings
80
  try:
 
81
  hf_token = os.environ.get("HF_TOKEN")
82
  if not hf_token:
83
  raise RuntimeError("HF_TOKEN environment variable is not set")
 
106
  device_map="auto",
107
  torch_dtype=torch.float16,
108
  token=hf_token,
109
+ low_cpu_mem_usage=True # Set to True to allow device_map usage
110
  )
111
  llm_model.eval()
112
  print("LLM initialized successfully")