Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,20 @@
|
|
1 |
---
|
2 |
license: mit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
4 |
## Introduction
|
5 |
|
6 |
**M4LE** is a **M**ulti-ability, **M**ulti-range, **M**ulti-task, bilingual benchmark for long-context evaluation. We categorize long-context understanding into five distinct abilities by considering whether it is required to identify single or multiple spans in long contexts based on explicit or semantic hints. Specifically, these abilities are explicit single-span, semantic single-span, explicit multiple-span, semantic multiple-span, and global. Different from previous long-context benchmark that simply compile from a set of existing long NLP benchmarks, we introduce an automated method to transform short-sequence tasks into a comprehensive long-sequence scenario encompassing all these capabilities.
|
@@ -127,4 +141,4 @@ If you find our paper and resources useful, please consider citing our paper:
|
|
127 |
author = {Kwan, Wai-Chung and Zeng, Xingshan and Wang, Yufei and Sun, Yusen and Li, Liangyou and Shang, Lifeng and Liu, Qun and Wong, Kam-Fai},
|
128 |
year = {2023},
|
129 |
}
|
130 |
-
```
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
+
task_categories:
|
4 |
+
- question-answering
|
5 |
+
- translation
|
6 |
+
- summarization
|
7 |
+
- text-classification
|
8 |
+
- retrieval
|
9 |
+
language:
|
10 |
+
- en
|
11 |
+
- zh
|
12 |
+
tags:
|
13 |
+
- Long Context
|
14 |
+
size_categories:
|
15 |
+
- 1K<n<10K
|
16 |
---
|
17 |
+
|
18 |
## Introduction
|
19 |
|
20 |
**M4LE** is a **M**ulti-ability, **M**ulti-range, **M**ulti-task, bilingual benchmark for long-context evaluation. We categorize long-context understanding into five distinct abilities by considering whether it is required to identify single or multiple spans in long contexts based on explicit or semantic hints. Specifically, these abilities are explicit single-span, semantic single-span, explicit multiple-span, semantic multiple-span, and global. Different from previous long-context benchmark that simply compile from a set of existing long NLP benchmarks, we introduce an automated method to transform short-sequence tasks into a comprehensive long-sequence scenario encompassing all these capabilities.
|
|
|
141 |
author = {Kwan, Wai-Chung and Zeng, Xingshan and Wang, Yufei and Sun, Yusen and Li, Liangyou and Shang, Lifeng and Liu, Qun and Wong, Kam-Fai},
|
142 |
year = {2023},
|
143 |
}
|
144 |
+
```
|