CIIR@LiveRAG 2025: Optimizing Multi-Agent Retrieval Augmented Generation   through Self-Training
        - URL: http://arxiv.org/abs/2506.10844v1
 - Date: Thu, 12 Jun 2025 16:02:29 GMT
 - Title: CIIR@LiveRAG 2025: Optimizing Multi-Agent Retrieval Augmented Generation   through Self-Training
 - Authors: Alireza Salemi, Mukta Maddipatla, Hamed Zamani, 
 - Abstract summary: mRAG is a multi-agent retrieval-augmented generation framework composed of specialized agents for subtasks such as planning, searching, reasoning, and coordination.<n> Evaluated on DataMorgana-derived datasets during the SIGIR 2025 LiveRAG competition, mRAG outperforms conventional RAG baselines.
 - Score: 18.787703082459046
 - License: http://creativecommons.org/licenses/by-nc-sa/4.0/
 - Abstract:   This paper presents mRAG, a multi-agent retrieval-augmented generation (RAG) framework composed of specialized agents for subtasks such as planning, searching, reasoning, and coordination. Our system uses a self-training paradigm with reward-guided trajectory sampling to optimize inter-agent collaboration and enhance response generation. Evaluated on DataMorgana-derived datasets during the SIGIR 2025 LiveRAG competition, mRAG outperforms conventional RAG baselines. We further analyze competition outcomes and showcase the framework's strengths with case studies, demonstrating its efficacy for complex, real-world RAG tasks. 
 
       
      
        Related papers
        - Towards Agentic RAG with Deep Reasoning: A Survey of RAG-Reasoning   Systems in LLMs [69.10441885629787]
Retrieval-Augmented Generation (RAG) lifts the factuality of Large Language Models (LLMs) by injecting external knowledge.<n>It falls short on problems that demand multi-step inference; conversely, purely reasoning-oriented approaches often hallucinate or mis-ground facts.<n>This survey synthesizes both strands under a unified reasoning-retrieval perspective.
arXiv  Detail & Related papers  (2025-07-13T03:29:41Z) - TopClustRAG at SIGIR 2025 LiveRAG Challenge [2.56711111236449]
TopClustRAG is a retrieval-augmented generation (RAG) system developed for the LiveRAG Challenge.<n>Our system employs a hybrid retrieval strategy combining sparse and dense indices, followed by K-Means clustering to group semantically similar passages.
arXiv  Detail & Related papers  (2025-06-18T08:24:27Z) - Reasoning RAG via System 1 or System 2: A Survey on Reasoning Agentic   Retrieval-Augmented Generation for Industry Challenges [6.615766570234612]
Retrieval-Augmented Generation (RAG) has emerged as a powerful framework to overcome the knowledge limitations of Large Language Models.<n>To address these challenges, the field has shifted toward Reasoning Agentic RAG, a paradigm that embeds decision-making and adaptive tool use directly into the retrieval process.
arXiv  Detail & Related papers  (2025-06-12T07:01:56Z) - ImpRAG: Retrieval-Augmented Generation with Implicit Queries [49.510101132093396]
ImpRAG is a query-free RAG system that integrates retrieval and generation into a unified model.<n>We show that ImpRAG achieves 3.6-11.5 improvements in exact match scores on unseen tasks with diverse formats.
arXiv  Detail & Related papers  (2025-06-02T21:38:21Z) - ReMA: Learning to Meta-think for LLMs with Multi-Agent Reinforcement   Learning [53.817538122688944]
We introduce Reinforced Meta-thinking Agents (ReMA) to elicit meta-thinking behaviors from Reasoning of Large Language Models (LLMs)<n>ReMA decouples the reasoning process into two hierarchical agents: a high-level meta-thinking agent responsible for generating strategic oversight and plans, and a low-level reasoning agent for detailed executions.<n> Empirical results from single-turn experiments demonstrate that ReMA outperforms single-agent RL baselines on complex reasoning tasks.
arXiv  Detail & Related papers  (2025-03-12T16:05:31Z) - RAG-Gym: Systematic Optimization of Language Agents for   Retrieval-Augmented Generation [43.50113345998687]
We introduce RAG-Gym, a comprehensive platform that explores three optimization dimensions: (1) prompt engineering, (2) actor tuning, and (3) critic training.<n>For prompt engineering, we propose Re$2$Search, a novel agent incorporating reflection reasoning that significantly outperforms standard prompts.<n>In actor tuning, we evaluate three popular post-training algorithms with fine-grained process supervision and identify direct preference optimization as the most effective.
arXiv  Detail & Related papers  (2025-02-19T18:56:03Z) - Improving Retrieval-Augmented Generation through Multi-Agent   Reinforcement Learning [51.54046200512198]
Retrieval-augmented generation (RAG) is extensively utilized to incorporate external, current knowledge into large language models.<n>A standard RAG pipeline may comprise several components, such as query rewriting, document retrieval, document filtering, and answer generation.<n>To overcome these challenges, we propose treating the RAG pipeline as a multi-agent cooperative task, with each component regarded as an RL agent.
arXiv  Detail & Related papers  (2025-01-25T14:24:50Z) - Chain-of-Retrieval Augmented Generation [72.06205327186069]
This paper introduces an approach for training o1-like RAG models that retrieve and reason over relevant information step by step before generating the final answer.<n>Our proposed method, CoRAG, allows the model to dynamically reformulate the query based on the evolving state.
arXiv  Detail & Related papers  (2025-01-24T09:12:52Z) - RAG-Reward: Optimizing RAG with Reward Modeling and RLHF [8.911260109659489]
Retrieval-augmented generation (RAG) enhances Large Language Models (LLMs) with relevant and up-to-date knowledge.<n>The role of reward models in reinforcement learning for optimizing RAG remains underexplored.<n>We introduce textbfRAG-Reward, a framework designed to develop reward models.
arXiv  Detail & Related papers  (2025-01-22T22:59:19Z) - Reward-RAG: Enhancing RAG with Reward Driven Supervision [43.66966457772646]
We introduce Reward-RAG, a novel approach designed to enhance the Retrieval-Augmented Generation (RAG) model through Reward-Driven Supervision.
Unlike previous RAG methodologies, our method adapts retrieval information to specific domains by employing CriticGPT to train a dedicated reward model.
This reward model generates synthesized datasets for fine-tuning the RAG, aligning its outputs more closely with human preferences.
arXiv  Detail & Related papers  (2024-10-03T15:26:50Z) - FlashRAG: A Modular Toolkit for Efficient Retrieval-Augmented Generation   Research [70.6584488911715]
retrieval-augmented generation (RAG) has attracted considerable research attention.<n>Existing RAG toolkits are often heavy and inflexibly, failing to meet the customization needs of researchers.<n>Our toolkit has implemented 16 advanced RAG methods and gathered and organized 38 benchmark datasets.
arXiv  Detail & Related papers  (2024-05-22T12:12:40Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.