Commit
·
5812da8
1
Parent(s):
33ce85b
New version of dataset names
Browse files- app.py +2 -2
- src/display/about.py +8 -9
- src/populate.py +0 -1
app.py
CHANGED
@@ -42,14 +42,14 @@ def download_data():
|
|
42 |
global original_df
|
43 |
global leaderboard_df
|
44 |
try:
|
45 |
-
print(EVAL_REQUESTS_PATH)
|
46 |
snapshot_download(
|
47 |
repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30
|
48 |
)
|
49 |
except Exception:
|
50 |
restart_space()
|
51 |
try:
|
52 |
-
print(EVAL_RESULTS_PATH)
|
53 |
snapshot_download(
|
54 |
repo_id=RESULTS_REPO, local_dir=EVAL_RESULTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30
|
55 |
)
|
|
|
42 |
global original_df
|
43 |
global leaderboard_df
|
44 |
try:
|
45 |
+
print(EVAL_REQUESTS_PATH,QUEUE_REPO)
|
46 |
snapshot_download(
|
47 |
repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30
|
48 |
)
|
49 |
except Exception:
|
50 |
restart_space()
|
51 |
try:
|
52 |
+
print(EVAL_RESULTS_PATH, RESULTS_REPO)
|
53 |
snapshot_download(
|
54 |
repo_id=RESULTS_REPO, local_dir=EVAL_RESULTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30
|
55 |
)
|
src/display/about.py
CHANGED
@@ -19,15 +19,14 @@ class Tasks(Enum):
|
|
19 |
task4 = Task("belebele_cs", "accuracy", "belebele_cs")
|
20 |
task5 = Task("ctkfacts_cs", "accuracy", "ctkfacts_cs")
|
21 |
task6 = Task("czechnews_cs", "accuracy", "czechnews_cs")
|
22 |
-
task7 = Task("
|
23 |
-
task8 = Task("
|
24 |
-
task9 = Task("
|
25 |
-
task10 = Task("
|
26 |
-
task11 = Task("
|
27 |
-
task12 = Task("
|
28 |
-
task13 = Task("
|
29 |
-
task14 = Task("
|
30 |
-
task15 = Task("truthfulqa_cs", "accuracy", "truthfulqa_cs")
|
31 |
|
32 |
# Your leaderboard name
|
33 |
TITLE = """<h1 align="center" id="space-title">🇨🇿 CzechBench Leaderboard</h1>"""
|
|
|
19 |
task4 = Task("belebele_cs", "accuracy", "belebele_cs")
|
20 |
task5 = Task("ctkfacts_cs", "accuracy", "ctkfacts_cs")
|
21 |
task6 = Task("czechnews_cs", "accuracy", "czechnews_cs")
|
22 |
+
task7 = Task("fb_comments_cs", "accuracy", "fb_comments_cs")
|
23 |
+
task8 = Task("gsm8k_cs", "accuracy", "gsm8k_cs")
|
24 |
+
task9 = Task("klokanek_cs", "accuracy", "klokanek_cs")
|
25 |
+
task10 = Task("mall_reviews_cs", "accuracy", "mall_reviews_cs")
|
26 |
+
task11 = Task("mmlu_cs", "accuracy", "mmlu_cs")
|
27 |
+
task12 = Task("sqad_cs", "accuracy", "sqad_cs")
|
28 |
+
task13 = Task("subjectivity_cs", "accuracy", "subjectivity_cs")
|
29 |
+
task14 = Task("truthfulqa_cs", "accuracy", "truthfulqa_cs")
|
|
|
30 |
|
31 |
# Your leaderboard name
|
32 |
TITLE = """<h1 align="center" id="space-title">🇨🇿 CzechBench Leaderboard</h1>"""
|
src/populate.py
CHANGED
@@ -13,7 +13,6 @@ def get_leaderboard_df(results_path: str, requests_path: str, cols: list, benchm
|
|
13 |
#all_data_json = [v.to_dict() for v in raw_data]
|
14 |
df = pd.DataFrame.from_records(raw_data)
|
15 |
#df = df.sort_values(by=[AutoEvalColumn.average.name], ascending=False)
|
16 |
-
|
17 |
df = df[cols].round(decimals=2)
|
18 |
|
19 |
# filter out if any of the benchmarks have not been produced
|
|
|
13 |
#all_data_json = [v.to_dict() for v in raw_data]
|
14 |
df = pd.DataFrame.from_records(raw_data)
|
15 |
#df = df.sort_values(by=[AutoEvalColumn.average.name], ascending=False)
|
|
|
16 |
df = df[cols].round(decimals=2)
|
17 |
|
18 |
# filter out if any of the benchmarks have not been produced
|