|
from dataclasses import dataclass |
|
from enum import Enum |
|
from src.envs import QUEUE_REPO |
|
|
|
|
|
@dataclass |
|
class Task: |
|
benchmark: str |
|
metric: str |
|
col_name: str |
|
|
|
|
|
|
|
TITLE = """<h1 align="center" id="space-title">LLVM APR Benchmark Leaderboard</h1>""" |
|
|
|
|
|
INTRODUCTION_TEXT = """ |
|
Leaderboard for the [LLVM APR Benchmark](https://huggingface.co/datasets/dtcxzyw/llvm-apr-benchmark). |
|
""" |
|
|
|
EVALUATION_QUEUE_TEXT = f""" |
|
With the provided evaluation environment, you can get a certificate by calling `env.dump()`. |
|
|
|
Please submit your evaluation results (generated by scripts/submit.py) to [{QUEUE_REPO}](https://huggingface.co/datasets/{QUEUE_REPO}). |
|
""" |
|
|
|
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results" |
|
CITATION_BUTTON_TEXT = r"""@misc{llvm-apr-benchmark, |
|
title = {LLVM APR Benchmark: A Large-Scale Automated Program Repair Benchmark of Real-World LLVM Middle-End Bugs}, |
|
url = {https://github.com/dtcxzyw/llvm-apr-benchmark}, |
|
author = {Yingwei Zheng}, |
|
year = {2025}, |
|
} |
|
""" |
|
|