Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
License:
File size: 2,934 Bytes
c74bbec
 
0ebc3d3
 
 
 
 
 
 
 
 
 
 
 
7c3e610
 
 
 
 
 
 
 
 
 
 
 
8448c6a
 
 
 
 
 
 
 
 
 
 
 
875d22f
 
 
 
 
 
 
 
 
 
 
 
af10278
 
 
 
 
 
 
 
b03a5bc
 
 
 
c74bbec
0ebc3d3
 
 
 
7c3e610
 
 
 
8448c6a
 
 
 
875d22f
 
 
 
af10278
 
 
 
5e6ffab
 
 
b03a5bc
d4d362e
0c9673b
67323f8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
---
dataset_info:
- config_name: imaginary-reference
  features:
  - name: role
    dtype: string
  - name: content
    dtype: string
  splits:
  - name: test
    num_bytes: 4485
    num_examples: 25
  download_size: 4391
  dataset_size: 4485
- config_name: indifferent
  features:
  - name: role
    dtype: string
  - name: content
    dtype: string
  splits:
  - name: test
    num_bytes: 11732
    num_examples: 25
  download_size: 10536
  dataset_size: 11732
- config_name: math
  features:
  - name: role
    dtype: string
  - name: content
    dtype: string
  splits:
  - name: test
    num_bytes: 5440
    num_examples: 25
  download_size: 4740
  dataset_size: 5440
- config_name: redundant
  features:
  - name: role
    dtype: string
  - name: content
    dtype: string
  splits:
  - name: test
    num_bytes: 5087
    num_examples: 25
  download_size: 4096
  dataset_size: 5087
- config_name: unanswerable
  features:
  - name: role
    dtype: string
  - name: content
    dtype: string
  splits:
  - name: test
    num_bytes: 12501
    num_examples: 50
  download_size: 8242
  dataset_size: 12501
configs:
- config_name: imaginary-reference
  data_files:
  - split: test
    path: imaginary-reference/test-*
- config_name: indifferent
  data_files:
  - split: test
    path: indifferent/test-*
- config_name: math
  data_files:
  - split: test
    path: math/test-*
- config_name: redundant
  data_files:
  - split: test
    path: redundant/test-*
- config_name: unanswerable
  data_files:
  - split: test
    path: unanswerable/test-*
license: cc-by-nc-4.0
language:
- en
---
# DNR Bench

Don’t Reason Bench (DNR Bench), a novel benchmark designed to expose a vulnerability in current RLMs: their tendency to over-reason by attempting to solve unsolvable 
problems, leading to excessively long responses.

# Data Summary
The DNR Bench dataset contains 150 adversarially crafted prompts divided into five distinct categories:
- Imaginary Reference
- Indifferent
- Math,
- Redundant,
- Unanswerable.

Each category targets a specific failure mode observed in reasoning-optimized LLMs, such as hallucinating nonexistent references, failing to remain neutral in ambiguous contexts, incorrectly solving flawed math problems, overanalyzing redundant information, or answering questions that lack sufficient data.

# Leaderboard
This dataset is used to test reasoning LLMs in [DNR Leaderboard on Huggingface](https://huggingface.co/spaces/ServiceNow-AI/Do-not-reason-bench)


# Citation
```bibtex
@misc{hashemi2025dnrbenchbenchmarkingoverreasoning,
      title={DNR Bench: Benchmarking Over-Reasoning in Reasoning LLMs}, 
      author={Masoud Hashemi and Oluwanifemi Bamgbose and Sathwik Tejaswi Madhusudhan and Jishnu Sethumadhavan Nair and Aman Tiwari and Vikas Yadav},
      year={2025},
      eprint={2503.15793},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2503.15793}, 
}
```