Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -61,9 +61,10 @@ class ModelManager:
|
|
61 |
with open(cache_file, "rb") as f:
|
62 |
self.models[model_name] = pickle.load(f)
|
63 |
else:
|
64 |
-
|
|
|
65 |
with open(cache_file, "wb") as f:
|
66 |
-
pickle.dump(
|
67 |
except Exception as e:
|
68 |
print(f"Error loading {model_name}: {e}")
|
69 |
self.models[model_name] = None
|
@@ -132,7 +133,6 @@ async def api_generate_multimodel(request: Request):
|
|
132 |
except Exception as e:
|
133 |
return JSONResponse({"error": str(e)}, status_code=500)
|
134 |
|
135 |
-
|
136 |
iface = gr.Interface(
|
137 |
fn=process_message,
|
138 |
inputs=gr.Textbox(lines=2, placeholder="Enter your message here..."),
|
@@ -144,4 +144,4 @@ iface = gr.Interface(
|
|
144 |
|
145 |
if __name__ == "__main__":
|
146 |
port = int(os.environ.get("PORT", 7860))
|
147 |
-
iface.launch(server_port=port)
|
|
|
61 |
with open(cache_file, "rb") as f:
|
62 |
self.models[model_name] = pickle.load(f)
|
63 |
else:
|
64 |
+
llama_model = Llama.from_pretrained(repo_id=model_config['repo_id'], filename=model_config['filename'], use_auth_token=HUGGINGFACE_TOKEN)
|
65 |
+
self.models[model_name] = llama_model
|
66 |
with open(cache_file, "wb") as f:
|
67 |
+
pickle.dump(llama_model, f)
|
68 |
except Exception as e:
|
69 |
print(f"Error loading {model_name}: {e}")
|
70 |
self.models[model_name] = None
|
|
|
133 |
except Exception as e:
|
134 |
return JSONResponse({"error": str(e)}, status_code=500)
|
135 |
|
|
|
136 |
iface = gr.Interface(
|
137 |
fn=process_message,
|
138 |
inputs=gr.Textbox(lines=2, placeholder="Enter your message here..."),
|
|
|
144 |
|
145 |
if __name__ == "__main__":
|
146 |
port = int(os.environ.get("PORT", 7860))
|
147 |
+
iface.launch(server_port=port)
|