Datasets:

Modalities:
Text
Formats:
csv
Languages:
Chinese
Size:
< 1K
Libraries:
Datasets
pandas
License:
File size: 3,001 Bytes
88fd12b
 
 
 
efdb8bd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6ed14c9
 
efdb8bd
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
---
license: mit
language:
- zh
---
# xbench-evals

🌐 [Website](https://xbench.org) | 📄 [Paper](https://xbench.org/files/xbench_profession_v2.4.pdf) | 🤗 [Dataset](https://huggingface.co/datasets/xbench)

Evergreen, contamination-free, real-world, domain-specific AI evaluation framework

xbench is more than just a scoreboard — it's a new evaluation framework with two complementary tracks, designed to measure both the intelligence frontier and real-world utility of AI systems:
- AGI Tracking: Measures core model capabilities like reasoning, tool-use, and memory
- Profession Aligned: A new class of evals grounded in workflows, environments, and business KPIs, co-designed with domain experts

We open source the dataset and evaluation code for two of our AGI Tracking benchmarks: ScienceQA and DeepSearch.

## xbench-ScienceQA
ScienceQA is part of xbench's AGI Tracking series, focused on evaluating fundamental knowledge capabilities across scientific domains. For detailed evaluation procedures and further information, please refer to the [website](https://xbench.org/#/agi/scienceqa) and Eval Card [xbench-ScienceQA.pdf](https://xbench.org/files/Eval%20Card%20xbench-ScienceQA.pdf) (Chinese version)

| Rank |                Model                |   Company    | Score | BoN (N=5) | Time cost (s) |
|------|:-----------------------------------:|:------------:|:-----:|:---------:|:-------------:|
| 1    |               o3-high               |    OpenAI    | 60.8  |   78.0    |     87.7      |
| 2    |           Gemini 2.5 Pro            |    Google    | 57.2  |   74.0    |     63.7      |
| 3    |       Doubao-1.5-thinking-pro       |  ByteDance   | 53.6	 |   69.0    |     116.9     |
| 4    |             DeepSeek-R1             |   DeepSeek   | 50.4  |   71.0    |     161.6     |
| 5    |            o4-mini-high             |    OpenAI    | 50.4	 |   67.0    |     48.2      |
| 6    |  Claude Opus 4 - Extended Thinking  |  Anthropic   | 46.6	 |   69.0    |     30.8      |
| 7    |          Gemini 2.5 flash           |    Google    | 46.2  |   70.0    |     24.1      |
| 8    |            Qwen3 - Think            |   Alibaba    | 45.4  |   66.0    |     105.9     |
| 9    |     Grok 3 Mini (with Thinking)     |     xAI      | 42.6	 |   53.0    |     193.1     |
| 10   | Claude Sonnet 4 - Extended Thinking |  Anthropic   | 39.4	 |   61.0    |     28.3      |

## Notes
Benchmark data is encrypted to prevent search engine crawling and contamination, please refer to the decrypt code in [xbench_evals github repo](https://github.com/xbench-ai/xbench-evals) to get the plain text data. Please don't upload the plain text online. 

## Submit your agent
If you are developing an AI agent and would like to evaluate it using the latest version of xbench, we welcome you to contact us. Please submit a public access link of your agent, and we will complete the evaluation within an agreed timeframe and share the results with you promptly.

Contact: [[email protected]]()