Spaces:
Running
on
L4
Running
on
L4
Commit
·
aee6c99
1
Parent(s):
1b1e5f4
Fix pydantic errors
Browse files- svc/router.py +8 -2
svc/router.py
CHANGED
@@ -9,7 +9,7 @@ from dotenv import load_dotenv
|
|
9 |
import os
|
10 |
import json
|
11 |
from src.deepeval.deepeval_task_manager import DeepEvalTaskManager
|
12 |
-
|
13 |
|
14 |
router = APIRouter()
|
15 |
|
@@ -34,6 +34,8 @@ async def protected_route(username: str = Depends(get_current_user)):
|
|
34 |
|
35 |
@router.post("/chat", response_model=TaskResponse)
|
36 |
def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_current_user)):
|
|
|
|
|
37 |
try:
|
38 |
logger.info("Inside")
|
39 |
results = evaluator.simple_evaluate(
|
@@ -49,6 +51,8 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
|
|
49 |
except Exception as e:
|
50 |
raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
|
51 |
|
|
|
|
|
52 |
results["config"]["model_dtype"] = request.precision
|
53 |
|
54 |
model_name = request.model_args.split(",")[0].split("=")[1]
|
@@ -67,6 +71,8 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
|
|
67 |
async def deep_eval_suite(request: DeepEvalSuiteRequest):
|
68 |
des = DeepEvalTaskManager(request.model_name, request.tasks)
|
69 |
results = des.run_tasks() #TODO: format should be different. Check metunlp/results repo for the correct format
|
70 |
-
|
|
|
|
|
71 |
|
72 |
|
|
|
9 |
import os
|
10 |
import json
|
11 |
from src.deepeval.deepeval_task_manager import DeepEvalTaskManager
|
12 |
+
import torch
|
13 |
|
14 |
router = APIRouter()
|
15 |
|
|
|
34 |
|
35 |
@router.post("/chat", response_model=TaskResponse)
|
36 |
def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_current_user)):
|
37 |
+
logger.info(request)
|
38 |
+
|
39 |
try:
|
40 |
logger.info("Inside")
|
41 |
results = evaluator.simple_evaluate(
|
|
|
51 |
except Exception as e:
|
52 |
raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
|
53 |
|
54 |
+
torch.cuda.empty_cache()
|
55 |
+
|
56 |
results["config"]["model_dtype"] = request.precision
|
57 |
|
58 |
model_name = request.model_args.split(",")[0].split("=")[1]
|
|
|
71 |
async def deep_eval_suite(request: DeepEvalSuiteRequest):
|
72 |
des = DeepEvalTaskManager(request.model_name, request.tasks)
|
73 |
results = des.run_tasks() #TODO: format should be different. Check metunlp/results repo for the correct format
|
74 |
+
json_results = json.dumps(results)
|
75 |
+
|
76 |
+
return TaskResponse(results=json_results)
|
77 |
|
78 |
|