FedRAG: A Framework for Fine-Tuning Retrieval-Augmented Generation Systems
- URL: http://arxiv.org/abs/2506.09200v2
- Date: Thu, 12 Jun 2025 13:48:57 GMT
- Title: FedRAG: A Framework for Fine-Tuning Retrieval-Augmented Generation Systems
- Authors: Val Andrei Fajardo, David B. Emerson, Amandeep Singh, Veronica Chatrath, Marcelo Lotif, Ravi Theja, Alex Cheung, Izuki Matsuba,
- Abstract summary: FedRAG is a framework for fine-tuning RAG systems across centralized and federated architectures.<n>FedRAG supports state-of-the-art fine-tuning methods, offering a simple and intuitive interface.
- Score: 3.2733670032760456
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Retrieval-augmented generation (RAG) systems have been shown to be effective in addressing many of the drawbacks of relying solely on the parametric memory of large language models. Recent work has demonstrated that RAG systems can be improved via fine-tuning of their retriever and generator models. In this work, we introduce FedRAG, a framework for fine-tuning RAG systems across centralized and federated architectures. FedRAG supports state-of-the-art fine-tuning methods, offering a simple and intuitive interface and a seamless conversion from centralized to federated training tasks. FedRAG is also deeply integrated with the modern RAG ecosystem, filling a critical gap in available tools.
Related papers
- Graph-R1: Towards Agentic GraphRAG Framework via End-to-end Reinforcement Learning [20.05893083101089]
Graph-R1 is an agentic GraphRAG framework via end-to-end reinforcement learning (RL)<n>It introduces lightweight knowledge hypergraph construction, models retrieval as a multi-turn agent-environment interaction.<n>Experiments on standard RAG datasets show that Graph-R1 outperforms traditional GraphRAG and RL-enhanced RAG methods in reasoning accuracy, retrieval efficiency, and generation quality.
arXiv Detail & Related papers (2025-07-29T15:01:26Z) - RGE-GS: Reward-Guided Expansive Driving Scene Reconstruction via Diffusion Priors [54.81109375939306]
RGE-GS is a novel expansive reconstruction framework that synergizes diffusion-based generation with reward-guided Gaussian integration.<n>We propose a reward network that learns to identify and prioritize consistently generated patterns prior to reconstruction phases.<n>During the reconstruction process, we devise a differentiated training strategy that automatically adjust Gaussian optimization progress according to scene converge metrics.
arXiv Detail & Related papers (2025-06-28T08:02:54Z) - FlexRAG: A Flexible and Comprehensive Framework for Retrieval-Augmented Generation [24.01783076521377]
RAG plays a pivotal role in modern large language model applications, with numerous existing frameworks offering a wide range of functionalities.<n>We have identified several persistent challenges in these frameworks, including difficulties in algorithm reproduction and sharing.<n>To address these limitations, we introduce textbfFlexRAG, an open-source framework specifically designed for research and prototyping.
arXiv Detail & Related papers (2025-06-14T13:16:31Z) - Privacy-Preserving Federated Embedding Learning for Localized Retrieval-Augmented Generation [60.81109086640437]
We propose a novel framework called Federated Retrieval-Augmented Generation (FedE4RAG)<n>FedE4RAG facilitates collaborative training of client-side RAG retrieval models.<n>We apply homomorphic encryption within federated learning to safeguard model parameters.
arXiv Detail & Related papers (2025-04-27T04:26:02Z) - A System for Comprehensive Assessment of RAG Frameworks [0.0]
Retrieval Augmented Generation (RAG) has emerged as a standard paradigm for enhancing the factual accuracy and contextual relevance of Large Language Models (LLMs)<n>Existing evaluation frameworks fail to provide a holistic black-box approach to assessing RAG systems.<n>We introduce SCARF, a modular and flexible evaluation framework designed to benchmark deployed RAG applications systematically.
arXiv Detail & Related papers (2025-04-10T14:41:34Z) - UltraRAG: A Modular and Automated Toolkit for Adaptive Retrieval-Augmented Generation [64.79921229760332]
Retrieval-Augmented Generation (RAG) significantly enhances the performance of large language models (LLMs) in downstream tasks.<n>Existing RAG toolkits lack support for knowledge adaptation tailored to specific application scenarios.<n>We propose UltraRAG, a RAG toolkit that automates knowledge adaptation throughout the entire workflow.
arXiv Detail & Related papers (2025-03-31T03:49:49Z) - RGL: A Graph-Centric, Modular Framework for Efficient Retrieval-Augmented Generation on Graphs [58.10503898336799]
We introduce the RAG-on-Graphs Library (RGL), a modular framework that seamlessly integrates the complete RAG pipeline.<n>RGL addresses key challenges by supporting a variety of graph formats and integrating optimized implementations for essential components.<n>Our evaluations demonstrate that RGL not only accelerates the prototyping process but also enhances the performance and applicability of graph-based RAG systems.
arXiv Detail & Related papers (2025-03-25T03:21:48Z) - Retrieval-Augmented Generation with Hierarchical Knowledge [38.500133410610495]
Graph-based Retrieval-Augmented Generation (RAG) methods have significantly enhanced the performance of large language models (LLMs) in domain-specific tasks.<n>Existing RAG methods do not adequately utilize the naturally inherent hierarchical knowledge in human cognition.<n>We introduce a new RAG approach, called HiRAG, which utilizes hierarchical knowledge to enhance the semantic understanding and structure capturing capabilities of RAG systems.
arXiv Detail & Related papers (2025-03-13T08:22:31Z) - Unanswerability Evaluation for Retrieval Augmented Generation [74.3022365715597]
UAEval4RAG is a framework designed to evaluate whether RAG systems can handle unanswerable queries effectively.<n>We define a taxonomy with six unanswerable categories, and UAEval4RAG automatically synthesizes diverse and challenging queries.
arXiv Detail & Related papers (2024-12-16T19:11:55Z) - Plan*RAG: Efficient Test-Time Planning for Retrieval Augmented Generation [20.5047654554575]
Plan*RAG is a framework that enables structured multi-hop reasoning in retrieval-augmented generation (RAG)<n>Plan*RAG consistently achieves improvements over recently proposed methods such as RQ-RAG and Self-RAG.
arXiv Detail & Related papers (2024-10-28T05:35:04Z) - Modular RAG: Transforming RAG Systems into LEGO-like Reconfigurable Frameworks [15.241520961365051]
Retrieval-augmented Generation (RAG) has markedly enhanced the capabilities of Large Language Models (LLMs)
This paper examines the limitations of the existing RAG paradigm and introduces the modular RAG framework.
arXiv Detail & Related papers (2024-07-26T03:45:30Z) - Pistis-RAG: Enhancing Retrieval-Augmented Generation with Human Feedback [41.88662700261036]
RAG systems face limitations when semantic relevance alone does not guarantee improved generation quality.
We propose Pistis-RAG, a new RAG framework designed with a content-centric approach to better align LLMs with human preferences.
arXiv Detail & Related papers (2024-06-21T08:52:11Z) - FlashRAG: A Modular Toolkit for Efficient Retrieval-Augmented Generation Research [70.6584488911715]
retrieval-augmented generation (RAG) has attracted considerable research attention.<n>Existing RAG toolkits are often heavy and inflexibly, failing to meet the customization needs of researchers.<n>Our toolkit has implemented 16 advanced RAG methods and gathered and organized 38 benchmark datasets.
arXiv Detail & Related papers (2024-05-22T12:12:40Z) - RAGGED: Towards Informed Design of Retrieval Augmented Generation Systems [51.171355532527365]
Retrieval-augmented generation (RAG) can significantly improve the performance of language models (LMs)
RAGGED is a framework for analyzing RAG configurations across various document-based question answering tasks.
arXiv Detail & Related papers (2024-03-14T02:26:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.