CORD: Balancing COnsistency and Rank Distillation for Robust Retrieval-Augmented Generation
- URL: http://arxiv.org/abs/2412.14581v1
- Date: Thu, 19 Dec 2024 07:01:25 GMT
- Title: CORD: Balancing COnsistency and Rank Distillation for Robust Retrieval-Augmented Generation
- Authors: Youngwon Lee, Seung-won Hwang, Daniel Campos, Filip GraliĆski, Zhewei Yao, Yuxiong He,
- Abstract summary: We propose consistency regularization with augmentation and distillation.
We propose CORD, balancing COnsistency and Rank Distillation.
Empirical results show this balance enables CORD to outperform consistently in diverse RAG benchmarks.
- Score: 47.091086803980765
- License:
- Abstract: With the adoption of retrieval-augmented generation (RAG), large language models (LLMs) are expected to ground their generation to the retrieved contexts. Yet, this is hindered by position bias of LLMs, failing to evenly attend to all contexts. Previous work has addressed this by synthesizing contexts with perturbed positions of gold segment, creating a position-diversified train set. We extend this intuition to propose consistency regularization with augmentation and distillation. First, we augment each training instance with its position perturbation to encourage consistent predictions, regardless of ordering. We also distill behaviors of this pair, although it can be counterproductive in certain RAG scenarios where the given order from the retriever is crucial for generation quality. We thus propose CORD, balancing COnsistency and Rank Distillation. CORD adaptively samples noise-controlled perturbations from an interpolation space, ensuring both consistency and respect for the rank prior. Empirical results show this balance enables CORD to outperform consistently in diverse RAG benchmarks.
Related papers
- Chain-of-Retrieval Augmented Generation [72.06205327186069]
This paper introduces an approach for training o1-like RAG models that retrieve and reason over relevant information step by step before generating the final answer.
Our proposed method, CoRAG, allows the model to dynamically reformulate the query based on the evolving state.
arXiv Detail & Related papers (2025-01-24T09:12:52Z) - Adaptive Regularization of Representation Rank as an Implicit Constraint of Bellman Equation [40.88431931273978]
We find a guiding principle for adaptive control of the representation rank.
We propose a novel regularizer, namely BEllman Equation-based automatic rank Regularizer (BEER)
BEER adaptively regularizes the representation rank, thus improving the DRL agent's performance.
arXiv Detail & Related papers (2024-04-19T10:00:34Z) - Time-series Generation by Contrastive Imitation [87.51882102248395]
We study a generative framework that seeks to combine the strengths of both: Motivated by a moment-matching objective to mitigate compounding error, we optimize a local (but forward-looking) transition policy.
At inference, the learned policy serves as the generator for iterative sampling, and the learned energy serves as a trajectory-level measure for evaluating sample quality.
arXiv Detail & Related papers (2023-11-02T16:45:25Z) - Task-Balanced Distillation for Object Detection [18.939830805129787]
RetinaNet with ResNet-50 achieves 41.0 mAP under the benchmark, outperforming the recent FGD and COCO.
A novel Task-decoupled Feature Distillation (TFD) is proposed by flexibly balancing the contributions of classification and regression tasks.
arXiv Detail & Related papers (2022-08-05T06:43:40Z) - Inducing and Using Alignments for Transition-based AMR Parsing [51.35194383275297]
We propose a neural aligner for AMR that learns node-to-word alignments without relying on complex pipelines.
We attain a new state-of-the art for gold-only trained models, matching silver-trained performance without the need for beam search on AMR3.0.
arXiv Detail & Related papers (2022-05-03T12:58:36Z) - Reconcile Prediction Consistency for Balanced Object Detection [10.61438063305309]
We propose a Harmonic loss to harmonize the optimization of classification branch and localization branch.
The Harmonic loss enables these two branches to supervise and promote each other during training.
In order to prevent the localization loss from being dominated by outliers during training phase, a Harmonic IoU loss is proposed to harmonize the weight of the localization loss of different IoU-level samples.
arXiv Detail & Related papers (2021-08-24T15:52:11Z) - Parallelized Reverse Curriculum Generation [62.25453821794469]
For reinforcement learning, it is challenging for an agent to master a task that requires a specific series of actions due to sparse rewards.
reverse curriculum generation (RCG) provides a reverse expansion approach that automatically generates a curriculum for the agent to learn.
We propose a parallelized approach that simultaneously trains multiple AC pairs and periodically exchanges their critics.
arXiv Detail & Related papers (2021-08-04T15:58:35Z) - PCPL: Predicate-Correlation Perception Learning for Unbiased Scene Graph
Generation [58.98802062945709]
We propose a novel Predicate-Correlation Perception Learning scheme to adaptively seek out appropriate loss weights.
Our PCPL framework is further equipped with a graph encoder module to better extract context features.
arXiv Detail & Related papers (2020-09-02T08:30:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.