Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -250,8 +250,8 @@ def status():
|
|
250 |
return {"status": "active", "agents": ["planner", "executor", "critic"]}
|
251 |
|
252 |
from ctransformers import AutoModelForCausalLM
|
|
|
253 |
|
254 |
-
# Load Zephyr 7B GGUF model (quantized)
|
255 |
llm_model = AutoModelForCausalLM.from_pretrained(
|
256 |
"TheBloke/zephyr-7B-alpha-GGUF",
|
257 |
model_file="zephyr-7b-alpha.Q4_K_M.gguf",
|
@@ -259,7 +259,7 @@ llm_model = AutoModelForCausalLM.from_pretrained(
|
|
259 |
max_new_tokens=256,
|
260 |
temperature=0.7
|
261 |
)
|
262 |
-
|
263 |
cognitive_engine = CognitiveEngine(llm_model)
|
264 |
|
265 |
response = model.generate("Hello, how are you?")
|
|
|
250 |
return {"status": "active", "agents": ["planner", "executor", "critic"]}
|
251 |
|
252 |
from ctransformers import AutoModelForCausalLM
|
253 |
+
from src.core.cognitive_engine import CognitiveEngine
|
254 |
|
|
|
255 |
llm_model = AutoModelForCausalLM.from_pretrained(
|
256 |
"TheBloke/zephyr-7B-alpha-GGUF",
|
257 |
model_file="zephyr-7b-alpha.Q4_K_M.gguf",
|
|
|
259 |
max_new_tokens=256,
|
260 |
temperature=0.7
|
261 |
)
|
262 |
+
|
263 |
cognitive_engine = CognitiveEngine(llm_model)
|
264 |
|
265 |
response = model.generate("Hello, how are you?")
|