|
import ast |
|
import inspect |
|
import json |
|
import multiprocessing |
|
import sys |
|
from concurrent.futures import ProcessPoolExecutor, as_completed |
|
|
|
from tqdm import tqdm |
|
|
|
from evalplus.data.humaneval import ( |
|
HUMANEVAL_PLUS_VERSION, |
|
get_human_eval_plus, |
|
get_human_eval_plus_hash, |
|
) |
|
from evalplus.eval import is_floats |
|
from evalplus.eval._special_oracle import _poly |
|
from evalplus.evaluate import get_groundtruth |
|
|
|
HUMANEVAL_TEST_TEMPLATE = """\ |
|
{imports} |
|
|
|
{aux_fn} |
|
|
|
def check(candidate): |
|
inputs = {inputs} |
|
results = {results} |
|
for i, (inp, exp) in enumerate(zip(inputs, results)): |
|
{assertion} |
|
""" |
|
|
|
HUMANEVAL_CROSSCHECK_TEMPLATE = """\ |
|
{aux_fn} |
|
|
|
{ref_func} |
|
|
|
def check(candidate): |
|
inputs = {inputs} |
|
for i, inp in enumerate(inputs): |
|
assertion(candidate(*inp), ref_func(*inp), {atol}) |
|
""" |
|
|
|
ASSERTION_FN = f"""\ |
|
import numpy as np |
|
|
|
{inspect.getsource(is_floats)} |
|
|
|
def assertion(out, exp, atol): |
|
exact_match = out == exp |
|
|
|
if atol == 0 and is_floats(exp): |
|
atol = 1e-6 |
|
if not exact_match and atol != 0: |
|
np.testing.assert_allclose(out, exp, atol=atol) |
|
else: |
|
assert exact_match |
|
""" |
|
|
|
|
|
def synthesize_test_code(task_id, entry_point, inputs, results, ref_func, atol): |
|
|
|
if entry_point in ( |
|
"tri", |
|
"string_sequence", |
|
"starts_one_ends", |
|
"make_a_pile", |
|
"special_factorial", |
|
"all_prefixes", |
|
): |
|
return task_id, HUMANEVAL_CROSSCHECK_TEMPLATE.format( |
|
aux_fn=ASSERTION_FN, |
|
inputs=inputs, |
|
ref_func=ref_func.replace(f" {entry_point}(", " ref_func("), |
|
atol=atol, |
|
) |
|
|
|
|
|
imports = set() |
|
aux_fn = ASSERTION_FN |
|
assertion = f"assertion(candidate(*inp), exp, {atol})" |
|
|
|
|
|
if entry_point == "find_zero": |
|
imports.add("import math") |
|
aux_fn = inspect.getsource(_poly) + "\n" |
|
assertion = f"assert _poly(*candidate(*inp), inp) <= {atol}" |
|
|
|
return task_id, HUMANEVAL_TEST_TEMPLATE.format( |
|
imports="\n".join(imports), |
|
aux_fn=aux_fn, |
|
inputs=inputs, |
|
results=results, |
|
assertion=assertion, |
|
) |
|
|
|
|
|
def deduplicate(inputs, results): |
|
assert len(inputs) == len(results) |
|
unique_input_strs = set([f"{x}" for x in inputs]) |
|
|
|
new_inputs, new_results = [], [] |
|
for inp, res in zip(inputs, results): |
|
inp_str = f"{inp}" |
|
if inp_str in unique_input_strs: |
|
new_inputs.append(inp) |
|
new_results.append(res) |
|
unique_input_strs.remove(inp_str) |
|
|
|
return new_inputs, new_results |
|
|
|
|
|
def main(): |
|
import argparse |
|
|
|
parser = argparse.ArgumentParser() |
|
parser.add_argument("--debug-tasks", nargs="+", default=[], type=int) |
|
|
|
args = parser.parse_args() |
|
|
|
if hasattr(sys, "set_int_max_str_digits"): |
|
sys.set_int_max_str_digits(int(10e8)) |
|
|
|
plus_problems = get_human_eval_plus(mini=False) |
|
dataset_hash = get_human_eval_plus_hash() |
|
|
|
compatible_problems = {} |
|
expected_outputs = get_groundtruth(plus_problems, dataset_hash, []) |
|
|
|
|
|
id2bytes = {} |
|
|
|
n_workers = max(1, multiprocessing.cpu_count() // 4) |
|
with ProcessPoolExecutor(max_workers=n_workers) as executor: |
|
futures = [] |
|
for task_id, plus_form in tqdm(plus_problems.items()): |
|
if args.debug_tasks and int(task_id.split("/")[-1]) not in args.debug_tasks: |
|
continue |
|
|
|
compatible_form = {} |
|
compatible_form["task_id"] = task_id |
|
compatible_form["prompt"] = plus_form["prompt"] |
|
compatible_form["canonical_solution"] = plus_form["canonical_solution"] |
|
compatible_form["entry_point"] = plus_form["entry_point"] |
|
compatible_problems[task_id] = compatible_form |
|
|
|
inputs = plus_form["base_input"] + plus_form["plus_input"] |
|
results = ( |
|
expected_outputs[task_id]["base"] + expected_outputs[task_id]["plus"] |
|
) |
|
|
|
inputs, results = deduplicate(inputs, results) |
|
|
|
assert len(inputs) == len(results) |
|
atol = plus_form["atol"] |
|
|
|
simplified_prompt = "" |
|
for line in compatible_form["prompt"].split("\n"): |
|
if not line: |
|
continue |
|
if '"""' in line or "'''" in line: |
|
break |
|
simplified_prompt += line + "\n" |
|
|
|
futures.append( |
|
executor.submit( |
|
synthesize_test_code, |
|
task_id, |
|
compatible_form["entry_point"], |
|
inputs, |
|
results, |
|
simplified_prompt + compatible_form["canonical_solution"], |
|
atol, |
|
) |
|
) |
|
|
|
for future in tqdm(as_completed(futures), total=len(plus_problems)): |
|
task_id, test_code = future.result() |
|
|
|
ast.parse(test_code) |
|
id2bytes[task_id] = len(test_code.encode("utf-8")) |
|
compatible_problems[task_id]["test"] = test_code |
|
|
|
|
|
print("Top-10 largest test code comes from problems (in megabytes):") |
|
for task_id, size in sorted(id2bytes.items(), key=lambda x: x[1], reverse=True)[ |
|
:10 |
|
]: |
|
print(f"{task_id}:\t{size / 1024 / 1024:.2f}mb") |
|
|
|
if args.debug_tasks: |
|
for problem in compatible_problems.values(): |
|
print("--- debugging:", problem["task_id"]) |
|
print(problem["prompt"] + problem["canonical_solution"]) |
|
test_code = problem["test"] |
|
if len(test_code) <= 2048 + 512: |
|
print(test_code) |
|
else: |
|
print(problem["test"][:1024], "...") |
|
print("...", problem["test"][-1024:]) |
|
else: |
|
with open(f"HumanEvalPlus-OriginFmt-{HUMANEVAL_PLUS_VERSION}.jsonl", "w") as f: |
|
for problem in compatible_problems.values(): |
|
f.write(json.dumps(problem) + "\n") |
|
|
|
|
|
if __name__ == "__main__": |
|
main() |
|
|