DSBench / data_modeling /evaluation /dont-overfit-ii_eval.py
liqiang888's picture
Upload 738 files
fe8d248 verified
raw
history blame
836 Bytes
import os.path
import numpy as np
import pandas as pd
import argparse
from sklearn.metrics import roc_auc_score
parser = argparse.ArgumentParser()
parser.add_argument('--path', type=str, required=True)
parser.add_argument('--name', type=str, required=True)
parser.add_argument('--answer_file', type=str, required=True)
parser.add_argument('--predict_file', type=str, required=True)
parser.add_argument('--value', type=str, default="target")
args = parser.parse_args()
actual = pd.read_csv(os.path.join(args.path, args.name, args.answer_file))
submission = pd.read_csv(os.path.join(args.path, args.name, args.predict_file))
# 计算平均绝对误差
performance = roc_auc_score(actual[args.value], submission[args.value])
with open(os.path.join(args.path, args.name, "result.txt"), "w") as f:
f.write(str(performance))