Hybrid and Collaborative Passage Reranking
- URL: http://arxiv.org/abs/2305.09313v1
- Date: Tue, 16 May 2023 09:38:52 GMT
- Title: Hybrid and Collaborative Passage Reranking
- Authors: Zongmeng Zhang, Wengang Zhou, Jiaxin Shi, Houqiang Li
- Abstract summary: We propose a Hybrid and Collaborative Passage Reranking (HybRank) method.
It incorporates the lexical and semantic properties of sparse and dense retrievers for reranking.
Built on off-the-shelf retriever features, HybRank is a plug-in reranker capable of enhancing arbitrary passage lists.
- Score: 144.83902343298112
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In passage retrieval system, the initial passage retrieval results may be
unsatisfactory, which can be refined by a reranking scheme. Existing solutions
to passage reranking focus on enriching the interaction between query and each
passage separately, neglecting the context among the top-ranked passages in the
initial retrieval list. To tackle this problem, we propose a Hybrid and
Collaborative Passage Reranking (HybRank) method, which leverages the
substantial similarity measurements of upstream retrievers for passage
collaboration and incorporates the lexical and semantic properties of sparse
and dense retrievers for reranking. Besides, built on off-the-shelf retriever
features, HybRank is a plug-in reranker capable of enhancing arbitrary passage
lists including previously reranked ones. Extensive experiments demonstrate the
stable improvements of performance over prevalent retrieval and reranking
methods, and verify the effectiveness of the core components of HybRank.
Related papers
- Gumbel Reranking: Differentiable End-to-End Reranker Optimization [61.16471123356738]
RAG systems rely on rerankers to identify relevant documents.
fine-tuning these models remains challenging due to the scarcity of annotated query-document pairs.
We propose Gumbel Reranking, an end-to-end training framework for rerankers aimed at minimizing the training-inference gap.
arXiv Detail & Related papers (2025-02-16T13:23:39Z) - Rankify: A Comprehensive Python Toolkit for Retrieval, Re-Ranking, and Retrieval-Augmented Generation [15.31883349259767]
Rankify is an open-source toolkit designed to unify retrieval, re-ranking, and RAG within a cohesive framework.
It supports a wide range of retrieval techniques, including dense and sparse retrievers, while incorporating state-of-the-art re-ranking models.
Rankify includes a collection of pre-retrieved datasets to facilitate benchmarking, available at Huggingface.
arXiv Detail & Related papers (2025-02-04T16:33:25Z) - Domain-specific Question Answering with Hybrid Search [39.85176264551715]
We show that a hybrid approach combining a fine-tuned dense retriever with keyword based sparse search methods significantly enhances performance.
Experimental results indicate that this hybrid method outperforms our single-retriever system.
arXiv Detail & Related papers (2024-12-04T22:04:13Z) - Improve Dense Passage Retrieval with Entailment Tuning [22.39221206192245]
Key to a retrieval system is to calculate relevance scores to query and passage pairs.
We observed that a major class of relevance aligns with the concept of entailment in NLI tasks.
We design a method called entailment tuning to improve the embedding of dense retrievers.
arXiv Detail & Related papers (2024-10-21T09:18:30Z) - Modeling Uncertainty and Using Post-fusion as Fallback Improves Retrieval Augmented Generation with LLMs [80.74263278847063]
The integration of retrieved passages and large language models (LLMs) has significantly contributed to improving open-domain question answering.
This paper investigates different methods of combining retrieved passages with LLMs to enhance answer generation.
arXiv Detail & Related papers (2023-08-24T05:26:54Z) - Retrieval Augmentation for Commonsense Reasoning: A Unified Approach [64.63071051375289]
We propose a unified framework of retrieval-augmented commonsense reasoning (called RACo)
Our proposed RACo can significantly outperform other knowledge-enhanced method counterparts.
arXiv Detail & Related papers (2022-10-23T23:49:08Z) - RocketQAv2: A Joint Training Method for Dense Passage Retrieval and
Passage Re-ranking [89.82301733609279]
We propose a novel joint training approach for dense passage retrieval and passage re-ranking.
A major contribution is that we introduce the dynamic listwise distillation, where we design a unified listwise training approach for both the retriever and the re-ranker.
During the dynamic distillation, the retriever and the re-ranker can be adaptively improved according to each other's relevance information.
arXiv Detail & Related papers (2021-10-14T13:52:55Z) - Joint Passage Ranking for Diverse Multi-Answer Retrieval [56.43443577137929]
We study multi-answer retrieval, an under-explored problem that requires retrieving passages to cover multiple distinct answers for a question.
This task requires joint modeling of retrieved passages, as models should not repeatedly retrieve passages containing the same answer at the cost of missing a different valid answer.
In this paper, we introduce JPR, a joint passage retrieval model focusing on reranking. To model the joint probability of the retrieved passages, JPR makes use of an autoregressive reranker that selects a sequence of passages, equipped with novel training and decoding algorithms.
arXiv Detail & Related papers (2021-04-17T04:48:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.