license: apache-2.0
datasets:
- dgslibisey/MuSiQue
- hotpotqa/hotpot_qa
- xanhho/2WikiMultihopQA
- hover-nlp/hover
language:
- en
base_model:
- GritLM/GritLM-7B
- mistralai/Mistral-7B-v0.1
pipeline_tag: sentence-similarity
GritHopper-7B is the first decoder-based multi-hop dense retrieval model and achieves state-of-the-art performance on both in-distribution and out-of-distribution benchmarks for decomposition-free multi-hop dense retrieval. Built on GRITLM, it is trained across diverse datasets spanning question-answering and fact-checking. Unlike traditional decomposition-based approaches, GritHopper iteratively retrieves passages without explicit sub-question decomposition, concatenating retrieved evidence with the query at each step.
Using the decoder model in an encoder-only approach (like MDR), it performs each retrieval step in a single forward pass. In contrast to previous SOTA BERT-based approaches (like BeamRetriever or MDR), GritHopper generalizes significantly better to out-of-distribution data.
Key Strengths of GritHopper
- Encoder-Only Efficiency: Each retrieval iteration requires only a single forward pass (rather than multiple autoregressive steps).
- Out-of-Distribution Robustness: Achieves state-of-the-art performance compared to other decomposition-free methods on multiple OOD benchmarks.
- Unified Training: Combines dense retrieval with generative objectives, exploring how post-retrieval information on the generation loss improves dense retrieval performance.
- Stopping: GritHopper utilizes its generative capabilities via ReAct to control its own state. This way, it can stop itself through causal next-token prediction.
Staring with GritHopper
GritHopper is trained on MuSiQue, 2WikiMultiHopQA, HotPotQA, EX-Fever and HoVer.
GritHopper Models
Model Name | Datasets | Description | Model Size |
---|---|---|---|
GritHopper-7B | All Datasets | GritHopper trained on Answers as Post-Retrieval information (SOTA) | 7B |
1. Installation
pip install grithopper
2. Initialization
from grithopper import GritHopper
# Initialize GritHopper with your GRITLM model checkpoint or huggingface path
hopper = GritHopper(
model_name_or_path="UKPLab/GritHopper-7B",
device="cuda" # or "cpu"
)
3. Load Document Candidates
You can either load from a list of (title, passage) pairs and optionally dump them to a file:
documents = [
("Title A", "Passage text for document A."),
("Title B", "Passage text for document B."),
# ...
]
hopper.load_document_candidates(
document_candidates=documents,
device="cuda",
output_directory_candidates_dump="my_candidates.pkl" # optional
)
Or load them from a pre-encoded dump:
hopper.load_candidates_from_file(
dump_filepath="my_candidates.pkl",
device="cuda"
)
4. Encode a Query
question = "Who wrote the novel that was adapted into the film Blade Runner?"
previous_evidences = [("Blade Runner (Movie)", " The Movie....")] # optional
query_vector = hopper.encode_query(
multi_hop_question=question,
previous_evidences=previous_evidences, # optional
instruction_type="multi-hop" # or "fact-check" alternatively you can provide a custom instruction with insruction="your_instruction"
)
5. Single-Step Retrieval
result = hopper.retrieve_(
query=query_vector,
top_k=1,
get_stopping_probability=True
)
# {
# "retrieved": [
# {
# "title": "Title B",
# "passage": "Passage text for document B.",
# "score": 0.873
# }
# ],
# "continue_probability": 0.65, # present if get_stopping_probability=True
# "stop_probability": 0.35
# }
If you prefer to pass the question string directly:
result = hopper.retrieve_(
query="Who is the mother of the writer who wrote the novel that was adapted into the film Blade Runner?",
# optional previous_evidences=[("Blade Runner (Movie)", " The Movie....")],
top_k=1,
get_stopping_probability=True,
)
# {
# "retrieved": [
# { "title": "Blade Runner (Movie)", "passage": "...", "score": 0.92 }
# ],
# "continue_probability": 0.75,
# "stop_probability": 0.25
# }
6. Iterative (Multi-Hop) Retrieval
chain_of_retrieval = hopper.iterative_retrieve(
multi_hop_question="Who wrote the novel that was adapted into the film Blade Runner?",
instruction_type="multi-hop",
automatic_stopping=True,
max_hops=4
)
# [
# {
# "retrieved": [
# { "title": "Blade Runner (Movie)", "passage": "...", "score": 0.92 }
# ],
# "continue_probability": 0.75,
# "stop_probability": 0.25
# },
# {
# "retrieved": [
# { "title": "Philip K.", "passage": "...", "score": 0.88 }
# ],
# "continue_probability": 0.65,
# "stop_probability": 0.35
# },
# ...
# ]
This process continues until either:
1. The model determines it should stop (if automatic_stopping=True and stop_probability > continue_probability).
2. It hits max_hops.
3. Or no documents can be retrieved at a given step.
Citation
If you use GritHopper in your research, please cite the following paper:
TBD
Contact
Contact person: Justus-Jonas Erker, [email protected]
https://www.ukp.tu-darmstadt.de/
Don't hesitate to send us an e-mail or report an issue, if something is broken (and it shouldn't be) or if you have further questions. This repository contains experimental software and is published for the sole purpose of giving additional background details on the respective publication.
License
GritHopper is licensed under the Apache License, Version 2.0. See LICENSE for the full license text.
Acknowledgement
this Model is based upon the GRITLM.