YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Less is More: Pre-train a Strong Text Encoder for Dense Retrieval Using a Weak Decoder

Please check the official repository for more details and updates.

Fine-tuning on Marco passage/doc ranking tasks and NQ tasks

MSMARCO Dev Passage Retrieval MRR@10 Recall@1k
BM25 warmup checkpoint 0.329 0.953
ANCE Passage checkpoint 0.334 0.961
MSMARCO Document Retrieval MRR@10 (Dev) MRR@10 (Eval)
ANCE Document (FirstP) checkpoint 0.394 0.362
NQ Task Top-1 Top-5 Top-20 Top-100 MRR@20 P@20
DPR checkpoint 46.1 68.8 80.4 87.1 56.2 20.1
ANCE NQ checkpoint 52.5 73.1 83.1 88.7 61.5 22.5

Citation

If you find SEED-Encoder useful for your work, please cite the following paper:

 @article{lu2021less,
   title={Less is More: Pre-training a Strong Siamese Encoder Using a Weak Decoder},
   author={Lu, Shuqi and Xiong, Chenyan and He, Di and Ke, Guolin and Malik, Waleed and Dou, Zhicheng and Bennett, Paul and Liu, Tieyan and Overwijk, Arnold},
   journal={arXiv preprint arXiv:2102.09206},
   year={2021}
 }
Downloads last month
7
Inference API
Unable to determine this model’s pipeline type. Check the docs .