Datasets:
ArXiv:
License:
Update README.md
#1
by
tjpxiaoming
- opened
README.md
CHANGED
@@ -6,6 +6,19 @@ configs:
|
|
6 |
- split: main
|
7 |
path: "instances/*.json"
|
8 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
# DEVAI dataset
|
10 |
<p align="center" width="100%">
|
11 |
<img src="dataset_stats.png" align="center" width="84%"/>
|
|
|
6 |
- split: main
|
7 |
path: "instances/*.json"
|
8 |
---
|
9 |
+
|
10 |
+
**GITHUB:** https://github.com/metauto-ai/agent-as-a-judge
|
11 |
+
|
12 |
+
> [!NOTE]
|
13 |
+
> Current evaluation techniques are often inadequate for advanced **agentic systems** due to their focus on final outcomes and labor-intensive manual reviews. To overcome this limitation, we introduce the **Agent-as-a-Judge** framework.
|
14 |
+
|
15 |
+
|
16 |
+
> [!IMPORTANT]
|
17 |
+
> As a **proof-of-concept**, we applied **Agent-as-a-Judge** to code generation tasks using **DevAI**, a benchmark consisting of 55 realistic AI development tasks with 365 hierarchical user requirements. The results demonstrate that **Agent-as-a-Judge** significantly outperforms traditional evaluation methods, delivering reliable reward signals for scalable self-improvement in agentic systems.
|
18 |
+
>
|
19 |
+
> Check out the dataset on [Hugging Face 🤗](https://huggingface.co/DEVAI-benchmark).
|
20 |
+
> See how to use this dataset in the [guidelines](benchmark/devai/README.md).
|
21 |
+
|
22 |
# DEVAI dataset
|
23 |
<p align="center" width="100%">
|
24 |
<img src="dataset_stats.png" align="center" width="84%"/>
|