Spaces:
Build error
Build error
Wisdom Chen
commited on
Update model.py
Browse files
model.py
CHANGED
@@ -54,7 +54,6 @@ def initialize_models() -> bool:
|
|
54 |
|
55 |
# Initialize CLIP model with error handling and fallback
|
56 |
try:
|
57 |
-
# First try loading with device mapping
|
58 |
clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
|
59 |
'hf-hub:Marqo/marqo-fashionCLIP',
|
60 |
device=device
|
@@ -66,7 +65,6 @@ def initialize_models() -> bool:
|
|
66 |
print(f"CLIP initialization error: {str(e)}")
|
67 |
print("Attempting to load CLIP model with CPU fallback...")
|
68 |
try:
|
69 |
-
# Fallback to CPU
|
70 |
device = "cpu"
|
71 |
clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
|
72 |
'hf-hub:Marqo/marqo-fashionCLIP',
|
@@ -80,7 +78,6 @@ def initialize_models() -> bool:
|
|
80 |
|
81 |
# Initialize LLM with optimized settings
|
82 |
try:
|
83 |
-
# Check for HF_TOKEN and authenticate
|
84 |
hf_token = os.environ.get("HF_TOKEN")
|
85 |
if not hf_token:
|
86 |
raise RuntimeError("HF_TOKEN environment variable is not set")
|
@@ -109,7 +106,7 @@ def initialize_models() -> bool:
|
|
109 |
device_map="auto",
|
110 |
torch_dtype=torch.float16,
|
111 |
token=hf_token,
|
112 |
-
low_cpu_mem_usage=
|
113 |
)
|
114 |
llm_model.eval()
|
115 |
print("LLM initialized successfully")
|
|
|
54 |
|
55 |
# Initialize CLIP model with error handling and fallback
|
56 |
try:
|
|
|
57 |
clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
|
58 |
'hf-hub:Marqo/marqo-fashionCLIP',
|
59 |
device=device
|
|
|
65 |
print(f"CLIP initialization error: {str(e)}")
|
66 |
print("Attempting to load CLIP model with CPU fallback...")
|
67 |
try:
|
|
|
68 |
device = "cpu"
|
69 |
clip_model, _, clip_preprocess = open_clip.create_model_and_transforms(
|
70 |
'hf-hub:Marqo/marqo-fashionCLIP',
|
|
|
78 |
|
79 |
# Initialize LLM with optimized settings
|
80 |
try:
|
|
|
81 |
hf_token = os.environ.get("HF_TOKEN")
|
82 |
if not hf_token:
|
83 |
raise RuntimeError("HF_TOKEN environment variable is not set")
|
|
|
106 |
device_map="auto",
|
107 |
torch_dtype=torch.float16,
|
108 |
token=hf_token,
|
109 |
+
low_cpu_mem_usage=True # Set to True to allow device_map usage
|
110 |
)
|
111 |
llm_model.eval()
|
112 |
print("LLM initialized successfully")
|