ecemumutlu commited on
Commit
38cf5ff
·
1 Parent(s): 6034d34

Fix errors on results

Browse files
Files changed (1) hide show
  1. svc/router.py +17 -2
svc/router.py CHANGED
@@ -2,12 +2,15 @@ from fastapi import APIRouter, HTTPException, Depends
2
  import logging
3
 
4
  from lm_eval import evaluator
5
- from svc.schemas import LMHarnessTaskRequest, LMHarnessTaskResponse
6
  from fastapi.security import OAuth2PasswordRequestForm
7
  from auth.authentication import get_current_user, create_access_token
8
  from dotenv import load_dotenv
9
  import os
10
  import json
 
 
 
11
  router = APIRouter()
12
 
13
  logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
@@ -47,7 +50,9 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
47
  raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
48
 
49
  results["config"]["model_dtype"] = request.precision
50
- results["config"]["model_name"] = request.model
 
 
51
  results["config"]["model_sha"] = request.model_sha
52
 
53
  dumped = json.dumps(results, indent=2)
@@ -56,3 +61,13 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
56
  logger.info("-------------------results end------------------\n")
57
  return LMHarnessTaskResponse(results=dumped)
58
 
 
 
 
 
 
 
 
 
 
 
 
2
  import logging
3
 
4
  from lm_eval import evaluator
5
+ from svc.schemas import LMHarnessTaskRequest, LMHarnessTaskResponse, LoadModelRequest
6
  from fastapi.security import OAuth2PasswordRequestForm
7
  from auth.authentication import get_current_user, create_access_token
8
  from dotenv import load_dotenv
9
  import os
10
  import json
11
+
12
+ from transformers import AutoModelForCausalLM, AutoTokenizer
13
+
14
  router = APIRouter()
15
 
16
  logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
 
50
  raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
51
 
52
  results["config"]["model_dtype"] = request.precision
53
+
54
+ model_name = request.model_args.split(",")[0].split("=")[1]
55
+ results["config"]["model_name"] = model_name
56
  results["config"]["model_sha"] = request.model_sha
57
 
58
  dumped = json.dumps(results, indent=2)
 
61
  logger.info("-------------------results end------------------\n")
62
  return LMHarnessTaskResponse(results=dumped)
63
 
64
+
65
+ # @router.post("/load_model")
66
+ # async def load_model(request: LoadModelRequest):
67
+ # model = AutoModelForCausalLM.from_pretrained(
68
+ # request.model_name,
69
+ # torch_dtype="auto",
70
+ # device_map="auto"
71
+ # )
72
+ # tokenizer = AutoTokenizer.from_pretrained(request.model_name)
73
+ # return model, tokenizer