ecemumutlu commited on
Commit
aee6c99
·
1 Parent(s): 1b1e5f4

Fix pydantic errors

Browse files
Files changed (1) hide show
  1. svc/router.py +8 -2
svc/router.py CHANGED
@@ -9,7 +9,7 @@ from dotenv import load_dotenv
9
  import os
10
  import json
11
  from src.deepeval.deepeval_task_manager import DeepEvalTaskManager
12
-
13
 
14
  router = APIRouter()
15
 
@@ -34,6 +34,8 @@ async def protected_route(username: str = Depends(get_current_user)):
34
 
35
  @router.post("/chat", response_model=TaskResponse)
36
  def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_current_user)):
 
 
37
  try:
38
  logger.info("Inside")
39
  results = evaluator.simple_evaluate(
@@ -49,6 +51,8 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
49
  except Exception as e:
50
  raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
51
 
 
 
52
  results["config"]["model_dtype"] = request.precision
53
 
54
  model_name = request.model_args.split(",")[0].split("=")[1]
@@ -67,6 +71,8 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
67
  async def deep_eval_suite(request: DeepEvalSuiteRequest):
68
  des = DeepEvalTaskManager(request.model_name, request.tasks)
69
  results = des.run_tasks() #TODO: format should be different. Check metunlp/results repo for the correct format
70
- return TaskResponse(results=results)
 
 
71
 
72
 
 
9
  import os
10
  import json
11
  from src.deepeval.deepeval_task_manager import DeepEvalTaskManager
12
+ import torch
13
 
14
  router = APIRouter()
15
 
 
34
 
35
  @router.post("/chat", response_model=TaskResponse)
36
  def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_current_user)):
37
+ logger.info(request)
38
+
39
  try:
40
  logger.info("Inside")
41
  results = evaluator.simple_evaluate(
 
51
  except Exception as e:
52
  raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
53
 
54
+ torch.cuda.empty_cache()
55
+
56
  results["config"]["model_dtype"] = request.precision
57
 
58
  model_name = request.model_args.split(",")[0].split("=")[1]
 
71
  async def deep_eval_suite(request: DeepEvalSuiteRequest):
72
  des = DeepEvalTaskManager(request.model_name, request.tasks)
73
  results = des.run_tasks() #TODO: format should be different. Check metunlp/results repo for the correct format
74
+ json_results = json.dumps(results)
75
+
76
+ return TaskResponse(results=json_results)
77
 
78