sync from github
Browse files- src/leaderboard/read_evals.py +2 -2
- src/populate.py +1 -1
src/leaderboard/read_evals.py
CHANGED
@@ -116,9 +116,9 @@ class EvalResult:
|
|
116 |
multiplier = 1.0
|
117 |
if "time" in metric:
|
118 |
multiplier = 1.0
|
119 |
-
if "throughput" in metric:
|
120 |
multiplier = 1.0
|
121 |
-
if "batch_" in metric or "Mem" in metric or "Util" in metric
|
122 |
multiplier = 1
|
123 |
|
124 |
|
|
|
116 |
multiplier = 1.0
|
117 |
if "time" in metric:
|
118 |
multiplier = 1.0
|
119 |
+
if "throughput" in metric or "mfu" in metric or "mbu" in metric:
|
120 |
multiplier = 1.0
|
121 |
+
if "batch_" in metric or "Mem" in metric or "Util" in metric:
|
122 |
multiplier = 1
|
123 |
|
124 |
|
src/populate.py
CHANGED
@@ -75,7 +75,7 @@ def get_leaderboard_df(
|
|
75 |
df[col] = np.nan
|
76 |
|
77 |
if not df.empty:
|
78 |
-
df = df.round(decimals=
|
79 |
|
80 |
# filter out if any of the benchmarks have not been produced
|
81 |
# df = df[has_no_nan_values(df, benchmark_cols)]
|
|
|
75 |
df[col] = np.nan
|
76 |
|
77 |
if not df.empty:
|
78 |
+
df = df.round(decimals=4)
|
79 |
|
80 |
# filter out if any of the benchmarks have not been produced
|
81 |
# df = df[has_no_nan_values(df, benchmark_cols)]
|