Constructing and Evaluating Declarative RAG Pipelines in PyTerrier
        - URL: http://arxiv.org/abs/2506.10802v1
 - Date: Thu, 12 Jun 2025 15:16:34 GMT
 - Title: Constructing and Evaluating Declarative RAG Pipelines in PyTerrier
 - Authors: Craig Macdonald, Jinyuan Fang, Andrew Parry, Zaiqiao Meng, 
 - Abstract summary: Retrieval augmented generation (RAG) is an exciting application of the pipeline architecture.<n>Our PyTerrier-RAG extension for PyTerrier provides easy access to standard RAG datasets.<n>We show how to build on the larger PyTerrier ecosystem with state-of-the-art sparse, learned-sparse, and dense retrievers.
 - Score: 27.90584159600631
 - License: http://creativecommons.org/licenses/by/4.0/
 - Abstract:   Search engines often follow a pipeline architecture, where complex but effective reranking components are used to refine the results of an initial retrieval. Retrieval augmented generation (RAG) is an exciting application of the pipeline architecture, where the final component generates a coherent answer for the users from the retrieved documents. In this demo paper, we describe how such RAG pipelines can be formulated in the declarative PyTerrier architecture, and the advantages of doing so. Our PyTerrier-RAG extension for PyTerrier provides easy access to standard RAG datasets and evaluation measures, state-of-the-art LLM readers, and using PyTerrier's unique operator notation, easy-to-build pipelines. We demonstrate the succinctness of indexing and RAG pipelines on standard datasets (including Natural Questions) and how to build on the larger PyTerrier ecosystem with state-of-the-art sparse, learned-sparse, and dense retrievers, and other neural rankers. 
 
       
      
        Related papers
        - LLM-Assisted Question-Answering on Technical Documents Using Structured   Data-Aware Retrieval Augmented Generation [0.432776344138537]
Large Language Models (LLMs) are capable of natural language understanding and generation.<n> Fine-tuning is one possible solution, but it is resource-intensive and must be repeated with every data update.<n>Retrieval-Augmented Generation (RAG) offers an efficient solution by allowing LLMs to access external knowledge sources.
arXiv  Detail & Related papers  (2025-06-29T08:22:03Z) - RAG Without the Lag: Interactive Debugging for Retrieval-Augmented   Generation Pipelines [1.5741300187949614]
Retrieval-augmented generation (RAG) pipelines have become the de-facto approach for building AI assistants with access to external, domain-specific knowledge.<n>RAGGY is a tool that combines a Python library of composable RAG primitives with an interactive interface for real-time debug.
arXiv  Detail & Related papers  (2025-04-18T09:38:49Z) - Rankify: A Comprehensive Python Toolkit for Retrieval, Re-Ranking, and   Retrieval-Augmented Generation [15.31883349259767]
Rankify is an open-source toolkit designed to unify retrieval, re-ranking, and RAG within a cohesive framework.<n>It supports a wide range of retrieval techniques, including dense and sparse retrievers, while incorporating state-of-the-art re-ranking models.<n>Rankify includes a collection of pre-retrieved datasets to facilitate benchmarking, available at Huggingface.
arXiv  Detail & Related papers  (2025-02-04T16:33:25Z) - Toward Optimal Search and Retrieval for RAG [39.69494982983534]
Retrieval-augmented generation (RAG) is a promising method for addressing some of the memory-related challenges associated with Large Language Models (LLMs)
Here, we work towards the goal of understanding how retrievers can be optimized for RAG pipelines for common tasks such as Question Answering (QA)
arXiv  Detail & Related papers  (2024-11-11T22:06:51Z) - Toward General Instruction-Following Alignment for Retrieval-Augmented   Generation [63.611024451010316]
Following natural instructions is crucial for the effective application of Retrieval-Augmented Generation (RAG) systems.
We propose VIF-RAG, the first automated, scalable, and verifiable synthetic pipeline for instruction-following alignment in RAG systems.
arXiv  Detail & Related papers  (2024-10-12T16:30:51Z) - Instrumentation and Analysis of Native ML Pipelines via Logical Query   Plans [3.2362171533623054]
We envision highly-automated software platforms to assist data scientists with developing, validating, monitoring, and analysing their Machine Learning pipelines.
We extract "logical query plans" from ML pipeline code relying on popular libraries.
Based on these plans, we automatically infer pipeline semantics and instrument and rewrite the ML pipelines to enable diverse use cases without requiring data scientists to manually annotate or rewrite their code.
arXiv  Detail & Related papers  (2024-07-10T11:35:02Z) - Blended RAG: Improving RAG (Retriever-Augmented Generation) Accuracy   with Semantic Search and Hybrid Query-Based Retrievers [0.0]
Retrieval-Augmented Generation (RAG) is a prevalent approach to infuse a private knowledge base of documents with Large Language Models (LLM) to build Generative Q&A (Question-Answering) systems.
We propose the 'Blended RAG' method of leveraging semantic search techniques, such as Vector indexes and Sparse indexes, blended with hybrid query strategies.
Our study achieves better retrieval results and sets new benchmarks for IR (Information Retrieval) datasets like NQ and TREC-COVID datasets.
arXiv  Detail & Related papers  (2024-03-22T17:13:46Z) - SPRINT: A Unified Toolkit for Evaluating and Demystifying Zero-shot
  Neural Sparse Retrieval [92.27387459751309]
We provide SPRINT, a unified Python toolkit for evaluating neural sparse retrieval.
We establish strong and reproducible zero-shot sparse retrieval baselines across the well-acknowledged benchmark, BEIR.
We show that SPLADEv2 produces sparse representations with a majority of tokens outside of the original query and document.
arXiv  Detail & Related papers  (2023-07-19T22:48:02Z) - torchgfn: A PyTorch GFlowNet library [56.071033896777784]
torchgfn is a PyTorch library that aims to address this need.
It provides users with a simple API for environments and useful abstractions for samplers and losses.
arXiv  Detail & Related papers  (2023-05-24T00:20:59Z) - Towards Personalized Preprocessing Pipeline Search [52.59156206880384]
ClusterP3S is a novel framework for Personalized Preprocessing Pipeline Search via Clustering.
We propose a hierarchical search strategy to jointly learn the clusters and search for the optimal pipelines.
 Experiments on benchmark classification datasets demonstrate the effectiveness of enabling feature-wise preprocessing pipeline search.
arXiv  Detail & Related papers  (2023-02-28T05:45:05Z) - Omnidata: A Scalable Pipeline for Making Multi-Task Mid-Level Vision
  Datasets from 3D Scans [103.92680099373567]
This paper introduces a pipeline to parametrically sample and render multi-task vision datasets from comprehensive 3D scans from the real world.
Changing the sampling parameters allows one to "steer" the generated datasets to emphasize specific information.
Common architectures trained on a generated starter dataset reached state-of-the-art performance on multiple common vision tasks and benchmarks.
arXiv  Detail & Related papers  (2021-10-11T04:21:46Z) - TODS: An Automated Time Series Outlier Detection System [70.88663649631857]
TODS is a highly modular system that supports easy pipeline construction.<n>Tods supports 70 primitives, including data processing, time series processing, feature analysis, detection algorithms, and a reinforcement module.
arXiv  Detail & Related papers  (2020-09-18T15:36:43Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.