|
--- |
|
license: mit |
|
--- |
|
|
|
# MS MARCO Distillation Scores for Translate-Distill |
|
|
|
This repository contains [MS MARCO](https://microsoft.github.io/msmarco/) training |
|
query-passage scores produced by MonoT5 reranker |
|
[`unicamp-dl/mt5-13b-mmarco-100k`](https://huggingface.co/unicamp-dl/mt5-13b-mmarco-100k) and |
|
[`castorini/monot5-3b-msmarco-10k`](https://huggingface.co/castorini/monot5-3b-msmarco-10k). |
|
|
|
Each training query is associated with the top-50 passages retrieved by the [ColBERTv2](https://arxiv.org/abs/2112.01488) model. |
|
|
|
Files are gzip compressed and with the naming scheme of `{teacher}-monot5-{msmarco, mmarco}-{qlang}{plang}.jsonl.gz`, |
|
which indicates the teacher reranker that inferenced using `qlang` queries and `plang` passages from MS MARCO. |
|
For languages other than English (eng), we use the translated text provided by mmarco and [neuMarco](https://ir-datasets.com/neumarco.html). |
|
|
|
We additionally provide the Persian translation of the MS MARCO training queries since they were not included in either neuMARCO or mMARCO. |
|
You can find the tsv files containing the translation in `msmarco.train.query.fas.tsv.gz`. |
|
|
|
|
|
## Usage |
|
|
|
We recommand downloading the files to incorporate with the training script in the [PLAID-X](https://github.com/hltcoe/ColBERT-X/tree/plaid-x) codebase. |
|
|
|
## Citation and Bibtex Info |
|
|
|
Please cite the following paper if you use the scores. |
|
|
|
```bibtext |
|
@inproceedings{translate-distill, |
|
author = {Eugene Yang and Dawn Lawrie and James Mayfield and Douglas W. Oard and Scott Miller}, |
|
title = {Translate-Distill: Learning Cross-Language \ Dense Retrieval by Translation and Distillation}, |
|
booktitle = {Proceedings of the 46th European Conference on Information Retrieval (ECIR)}, |
|
year = {2024}, |
|
url = {https://arxiv.org/abs/2401.04810} |
|
} |
|
``` |
|
|
|
|