|
--- |
|
license: mit |
|
language: |
|
- zh |
|
--- |
|
# xbench-evals |
|
|
|
๐ [Website](https://xbench.org) | ๐ [Paper](https://xbench.org/files/xbench_profession_v2.4.pdf) | ๐ค [Dataset](https://huggingface.co/datasets/xbench) |
|
|
|
Evergreen, contamination-free, real-world, domain-specific AI evaluation framework |
|
|
|
xbench is more than just a scoreboard โ it's a new evaluation framework with two complementary tracks, designed to measure both the intelligence frontier and real-world utility of AI systems: |
|
- AGI Tracking: Measures core model capabilities like reasoning, tool-use, and memory |
|
- Profession Aligned: A new class of evals grounded in workflows, environments, and business KPIs, co-designed with domain experts |
|
|
|
We open source the dataset and evaluation code for two of our AGI Tracking benchmarks: ScienceQA and DeepSearch. |
|
|
|
## xbench-ScienceQA |
|
ScienceQA is part of xbench's AGI Tracking series, focused on evaluating fundamental knowledge capabilities across scientific domains. For detailed evaluation procedures and further information, please refer to the [website](https://xbench.org/#/agi/scienceqa) and Eval Card [xbench-ScienceQA.pdf](https://xbench.org/files/Eval%20Card%20xbench-ScienceQA.pdf) (Chinese version) |
|
|
|
| Rank | Model | Company | Score | BoN (N=5) | Time cost (s) | |
|
|------|:-----------------------------------:|:------------:|:-----:|:---------:|:-------------:| |
|
| 1 | o3-high | OpenAI | 60.8 | 78.0 | 87.7 | |
|
| 2 | Gemini 2.5 Pro | Google | 57.2 | 74.0 | 63.7 | |
|
| 3 | Doubao-1.5-thinking-pro | ByteDance | 53.6 | 69.0 | 116.9 | |
|
| 4 | DeepSeek-R1 | DeepSeek | 50.4 | 71.0 | 161.6 | |
|
| 5 | o4-mini-high | OpenAI | 50.4 | 67.0 | 48.2 | |
|
| 6 | Claude Opus 4 - Extended Thinking | Anthropic | 46.6 | 69.0 | 30.8 | |
|
| 7 | Gemini 2.5 flash | Google | 46.2 | 70.0 | 24.1 | |
|
| 8 | Qwen3 - Think | Alibaba | 45.4 | 66.0 | 105.9 | |
|
| 9 | Grok 3 Mini (with Thinking) | xAI | 42.6 | 53.0 | 193.1 | |
|
| 10 | Claude Sonnet 4 - Extended Thinking | Anthropic | 39.4 | 61.0 | 28.3 | |
|
|
|
## Notes |
|
Benchmark data is encrypted to prevent search engine crawling and contamination, please refer to the decrypt code in [xbench_evals github repo](https://github.com/xbench-ai/xbench-evals) to get the plain text data. Please don't upload the plain text online. |
|
|
|
## Submit your agent |
|
If you are developing an AI agent and would like to evaluate it using the latest version of xbench, we welcome you to contact us. Please submit a public access link of your agent, and we will complete the evaluation within an agreed timeframe and share the results with you promptly. |
|
|
|
Contact: [[email protected]]() |