Spaces:
Sleeping
Sleeping
Commit
·
38cf5ff
1
Parent(s):
6034d34
Fix errors on results
Browse files- svc/router.py +17 -2
svc/router.py
CHANGED
@@ -2,12 +2,15 @@ from fastapi import APIRouter, HTTPException, Depends
|
|
2 |
import logging
|
3 |
|
4 |
from lm_eval import evaluator
|
5 |
-
from svc.schemas import LMHarnessTaskRequest, LMHarnessTaskResponse
|
6 |
from fastapi.security import OAuth2PasswordRequestForm
|
7 |
from auth.authentication import get_current_user, create_access_token
|
8 |
from dotenv import load_dotenv
|
9 |
import os
|
10 |
import json
|
|
|
|
|
|
|
11 |
router = APIRouter()
|
12 |
|
13 |
logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
|
@@ -47,7 +50,9 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
|
|
47 |
raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
|
48 |
|
49 |
results["config"]["model_dtype"] = request.precision
|
50 |
-
|
|
|
|
|
51 |
results["config"]["model_sha"] = request.model_sha
|
52 |
|
53 |
dumped = json.dumps(results, indent=2)
|
@@ -56,3 +61,13 @@ def inference_model(request: LMHarnessTaskRequest, username: str = Depends(get_c
|
|
56 |
logger.info("-------------------results end------------------\n")
|
57 |
return LMHarnessTaskResponse(results=dumped)
|
58 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
import logging
|
3 |
|
4 |
from lm_eval import evaluator
|
5 |
+
from svc.schemas import LMHarnessTaskRequest, LMHarnessTaskResponse, LoadModelRequest
|
6 |
from fastapi.security import OAuth2PasswordRequestForm
|
7 |
from auth.authentication import get_current_user, create_access_token
|
8 |
from dotenv import load_dotenv
|
9 |
import os
|
10 |
import json
|
11 |
+
|
12 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
13 |
+
|
14 |
router = APIRouter()
|
15 |
|
16 |
logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
|
|
|
50 |
raise HTTPException(status_code=500, detail=f"lm-harness task execution failed for model: {request.model_args}")
|
51 |
|
52 |
results["config"]["model_dtype"] = request.precision
|
53 |
+
|
54 |
+
model_name = request.model_args.split(",")[0].split("=")[1]
|
55 |
+
results["config"]["model_name"] = model_name
|
56 |
results["config"]["model_sha"] = request.model_sha
|
57 |
|
58 |
dumped = json.dumps(results, indent=2)
|
|
|
61 |
logger.info("-------------------results end------------------\n")
|
62 |
return LMHarnessTaskResponse(results=dumped)
|
63 |
|
64 |
+
|
65 |
+
# @router.post("/load_model")
|
66 |
+
# async def load_model(request: LoadModelRequest):
|
67 |
+
# model = AutoModelForCausalLM.from_pretrained(
|
68 |
+
# request.model_name,
|
69 |
+
# torch_dtype="auto",
|
70 |
+
# device_map="auto"
|
71 |
+
# )
|
72 |
+
# tokenizer = AutoTokenizer.from_pretrained(request.model_name)
|
73 |
+
# return model, tokenizer
|