Chain-of-Context Learning: Dynamic Constraint Understanding for Multi-Task VRPs
- URL: http://arxiv.org/abs/2603.01667v1
- Date: Mon, 02 Mar 2026 09:57:15 GMT
- Title: Chain-of-Context Learning: Dynamic Constraint Understanding for Multi-Task VRPs
- Authors: Shuangchun Gui, Suyu Liu, Xuehe Wang, Zhiguang Cao,
- Abstract summary: Multi-task Vehicle Routing Problems (VRPs) aim to minimize routing costs while satisfying diverse constraints.<n>Existing solvers typically adopt a unified reinforcement learning (RL) framework to learn generalizable patterns across tasks.<n>We propose Chain-of-Context Learning (CCL), a novel framework that progressively captures the evolving context to guide fine-grained node adaptation.
- Score: 27.821803041439953
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-task Vehicle Routing Problems (VRPs) aim to minimize routing costs while satisfying diverse constraints. Existing solvers typically adopt a unified reinforcement learning (RL) framework to learn generalizable patterns across tasks. However, they often overlook the constraint and node dynamics during the decision process, making the model fail to accurately react to the current context. To address this limitation, we propose Chain-of-Context Learning (CCL), a novel framework that progressively captures the evolving context to guide fine-grained node adaptation. Specifically, CCL constructs step-wise contextual information via a Relevance-Guided Context Reformulation (RGCR) module, which adaptively prioritizes salient constraints. This context then guides node updates through a Trajectory-Shared Node Re-embedding (TSNR) module, which aggregates shared node features from all trajectories' contexts and uses them to update inputs for the next step. By modeling evolving preferences of the RL agent, CCL captures step-by-step dependencies in sequential decision-making. We evaluate CCL on 48 diverse VRP variants, including 16 in-distribution and 32 out-of-distribution (with unseen constraints) tasks. Experimental results show that CCL performs favorably against the state-of-the-art baselines, achieving the best performance on all in-distribution tasks and the majority of out-of-distribution tasks.
Related papers
- Structure Detection for Contextual Reinforcement Learning [6.56045575313744]
Contextual Reinforcement Learning tackles the problem of solving a set of related Contextual Markov Decision Processes (CMDPs) that vary across different context variables.<n>Traditional approaches--independent training and multi-task learning--struggle with excessive computational costs or negative transfer.<n>We introduce Structure Detection MBTL, a generic framework that dynamically identifies the underlying generalization structure of CMDP and selects an appropriate MBTL algorithm.
arXiv Detail & Related papers (2026-01-13T01:22:39Z) - KBQA-R1: Reinforcing Large Language Models for Knowledge Base Question Answering [64.62317305868264]
We present textbfKBQA-R1, a framework that shifts the paradigm from text imitation to interaction optimization via Reinforcement Learning.<n>Treating KBQA as a multi-turn decision process, our model learns to navigate the knowledge base using a list of actions.<n>Experiments on WebQSP, GrailQA, and GraphQuestions demonstrate that KBQA-R1 achieves state-of-the-art performance.
arXiv Detail & Related papers (2025-12-10T17:45:42Z) - CIR-CoT: Towards Interpretable Composed Image Retrieval via End-to-End Chain-of-Thought Reasoning [93.05917922306196]
Composed Image Retrieval (CIR) aims to find a target image from a reference image and a modification text.<n>CIR-CoT is the first end-to-end retrieval-oriented MLLM designed to integrate explicit Chain-of-Thought (CoT) reasoning.
arXiv Detail & Related papers (2025-10-09T09:41:45Z) - ContextNav: Towards Agentic Multimodal In-Context Learning [85.05420047017513]
ContextNav is an agentic framework that integrates the scalability of automated retrieval with the quality and adaptiveness of human-like curation.<n>It builds a resource-aware multimodal embedding pipeline, maintains a retrievable vector database, and applies agentic retrieval and structural alignment to construct noise-resilient contexts.<n> Experimental results demonstrate that ContextNav achieves state-of-the-art performance across various datasets.
arXiv Detail & Related papers (2025-10-06T07:49:52Z) - Submodular Context Partitioning and Compression for In-Context Learning [23.18825768870575]
Sub-CP is a block-aware context selection framework for large language models.<n>It supports a flexible spectrum of selection strategies, allowing each block to range from globally diverse to locally coherent.<n>Experiments show that Sub-CP consistently improves performance across model scales.
arXiv Detail & Related papers (2025-09-30T23:42:11Z) - Towards Generalizable Implicit In-Context Learning with Attention Routing [32.86419794893806]
In-Context Routing (ICR) is a novel implicit ICL method that internalizes generalizable ICL patterns at the attention logits level.<n>We evaluate ICR on 12 real-world datasets spanning diverse domains and multiple Large Language Models.
arXiv Detail & Related papers (2025-09-26T19:05:45Z) - RCR-Router: Efficient Role-Aware Context Routing for Multi-Agent LLM Systems with Structured Memory [57.449129198822476]
RCR is a role-aware context routing framework for multi-agent large language model (LLM) systems.<n>It dynamically selects semantically relevant memory subsets for each agent based on its role and task stage.<n>A lightweight scoring policy guides memory selection, and agent outputs are integrated into a shared memory store.
arXiv Detail & Related papers (2025-08-06T21:59:34Z) - DESign: Dynamic Context-Aware Convolution and Efficient Subnet Regularization for Continuous Sign Language Recognition [11.879737436401175]
We propose DESign, a novel framework that incorporates Dynamic Context-Aware Convolution (DCAC) and Subnet Regularization Connectionist Temporal Classification (SR-CTC)<n>DCAC dynamically captures the inter-frame motion cues that constitute signs and uniquely adapts convolutional weights based on contextual information.<n>SR-CTC regularizes training by applying supervision tovolutionworks, encouraging the model to explore diverse CTC alignment paths and effectively preventing overfitting.
arXiv Detail & Related papers (2025-07-04T06:56:28Z) - Scalable In-Context Q-Learning [68.9917436397079]
We propose textbfScalable textbfIn-textbfContext textbfQ-textbfLearning (textbfSICQL) to steer in-context reinforcement learning.<n>textbfSICQL harnesses dynamic programming and world modeling to steer ICRL toward efficient reward and task generalization.
arXiv Detail & Related papers (2025-06-02T04:21:56Z) - Self-regulating Prompts: Foundational Model Adaptation without
Forgetting [112.66832145320434]
We introduce a self-regularization framework for prompting called PromptSRC.
PromptSRC guides the prompts to optimize for both task-specific and task-agnostic general representations.
arXiv Detail & Related papers (2023-07-13T17:59:35Z) - Dynamics-Adaptive Continual Reinforcement Learning via Progressive
Contextualization [29.61829620717385]
Key challenge of continual reinforcement learning (CRL) in dynamic environments is to promptly adapt the RL agent's behavior as the environment changes over its lifetime.
DaCoRL learns a context-conditioned policy using progressive contextualization.
DaCoRL features consistent superiority over existing methods in terms of the stability, overall performance and generalization ability.
arXiv Detail & Related papers (2022-09-01T10:26:58Z) - Contextualize Me -- The Case for Context in Reinforcement Learning [49.794253971446416]
Contextual Reinforcement Learning (cRL) provides a framework to model such changes in a principled manner.
We show how cRL contributes to improving zero-shot generalization in RL through meaningful benchmarks and structured reasoning about generalization tasks.
arXiv Detail & Related papers (2022-02-09T15:01:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.