Update README
Browse files
README.md
CHANGED
@@ -71,3 +71,47 @@ configs:
|
|
71 |
- split: test
|
72 |
path: synthetic_queries/test-*
|
73 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
71 |
- split: test
|
72 |
path: synthetic_queries/test-*
|
73 |
---
|
74 |
+
# ConTEB - MLDR (evaluation)
|
75 |
+
|
76 |
+
This dataset is part of *ConTEB* (Context-aware Text Embedding Benchmark), designed for evaluating contextual embedding model capabilities. It stems from the widely used [MLDR](https://huggingface.co/datasets/Shitao/MLDR) dataset.
|
77 |
+
|
78 |
+
## Dataset Summary
|
79 |
+
|
80 |
+
MLDR consists of long documents, associated to existing sets of question-answer pairs. To build the corpus, we start from the pre-existing collection documents, extract the text, and chunk them (using [LangChain](https://github.com/langchain-ai/langchain)'s RecursiveCharacterSplitter with a threshold of 1000 characters). Since chunking is done a posteriori without considering the questions, chunks are not always self-contained and eliciting document-wide context can help build meaningful representations. We use GPT-4o to annotate which chunk, among the gold document, best contains information needed to answer the query.
|
81 |
+
|
82 |
+
This dataset provides a focused benchmark for contextualized embeddings. It includes a set of original documents, chunks stemming from them, and queries.
|
83 |
+
|
84 |
+
|
85 |
+
* **Number of Documents:** 100
|
86 |
+
* **Number of Chunks:** 1536
|
87 |
+
* **Number of Queries:** 100
|
88 |
+
* **Average Number of Tokens per Chunk:** 164.2
|
89 |
+
|
90 |
+
## Dataset Structure (Hugging Face Datasets)
|
91 |
+
The dataset is structured into the following columns:
|
92 |
+
|
93 |
+
* **`documents`**: Contains chunk information:
|
94 |
+
* `"chunk_id"`: The ID of the chunk, of the form `doc-id_chunk-id`, where `doc-id` is the ID of the original document and `chunk-id` is the position of the chunk within that document.
|
95 |
+
* `"chunk"`: The text of the chunk
|
96 |
+
* **`queries`**: Contains query information:
|
97 |
+
* `"query"`: The text of the query.
|
98 |
+
* `"answer"`: The answer relevant to the query, from the original dataset.
|
99 |
+
* `"chunk_id"`: The ID of the chunk that the query is related to, of the form `doc-id_chunk-id`, where `doc-id` is the ID of the original document and `chunk-id` is the position of the chunk within that document.
|
100 |
+
|
101 |
+
## Usage
|
102 |
+
|
103 |
+
Use the `test` split for evaluation.
|
104 |
+
We will upload a Quickstart evaluation snippet soon.
|
105 |
+
|
106 |
+
## Citation
|
107 |
+
|
108 |
+
We will add the corresponding citation soon.
|
109 |
+
|
110 |
+
## Acknowledgments
|
111 |
+
|
112 |
+
This work is partially supported by [ILLUIN Technology](https://www.illuin.tech/), and by a grant from ANRT France.
|
113 |
+
|
114 |
+
## Copyright
|
115 |
+
|
116 |
+
All rights are reserved to the original authors of the documents.
|
117 |
+
|