Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
- URL: http://arxiv.org/abs/2511.11992v1
- Date: Sat, 15 Nov 2025 02:11:31 GMT
- Title: Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
- Authors: Hung Du, Hy Nguyen, Srikanth Thudumu, Rajesh Vasa, Kon Mouzakis,
- Abstract summary: We propose a decentralized Multi-Agent Reinforcement Learning (MARL) framework that enables vehicles, acting as agents, to communicate selectively based on local goals and observations.<n>This goal-aware communication strategy allows agents to share only relevant information, enhancing collaboration while respecting visibility limitations.<n>Results show that our method significantly improves task success rates and reduces time-to-goal compared to non-cooperative baselines.
- Score: 0.6676697660506798
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Connected and autonomous vehicles across land, water, and air must often operate in dynamic, unpredictable environments with limited communication, no centralized control, and partial observability. These real-world constraints pose significant challenges for coordination, particularly when vehicles pursue individual objectives. To address this, we propose a decentralized Multi-Agent Reinforcement Learning (MARL) framework that enables vehicles, acting as agents, to communicate selectively based on local goals and observations. This goal-aware communication strategy allows agents to share only relevant information, enhancing collaboration while respecting visibility limitations. We validate our approach in complex multi-agent navigation tasks featuring obstacles and dynamic agent populations. Results show that our method significantly improves task success rates and reduces time-to-goal compared to non-cooperative baselines. Moreover, task performance remains stable as the number of agents increases, demonstrating scalability. These findings highlight the potential of decentralized, goal-driven MARL to support effective coordination in realistic multi-vehicle systems operating across diverse domains.
Related papers
- Advances and Innovations in the Multi-Agent Robotic System (MARS) Challenge [170.47383225329915]
Multi-agent system frameworks are becoming essential for achieving scalable, efficient, and collaborative solutions.<n>This shift is fueled by three primary factors: increasing agent capabilities, enhancing system efficiency through task delegation, and enabling advanced human-agent interactions.<n>We propose the Multi-Agent Robotic System (MARS) Challenge, held at the NeurIPS 2025 Workshop on SpaVLE.
arXiv Detail & Related papers (2026-01-26T17:56:19Z) - Scalable Multiagent Reinforcement Learning with Collective Influence Estimation [5.050035210247092]
This paper proposes a multiagent learning framework augmented with a Collective Influence Estimation Network.<n>By explicitly modeling the collective influence of other agents on the task object, each agent can infer critical interaction information.<n> Experimental results show that the proposed method achieves stable and efficient coordination under communication-limited environments.
arXiv Detail & Related papers (2026-01-13T04:24:11Z) - Multi-Agent Deep Reinforcement Learning for Collaborative UAV Relay Networks under Jamming Atatcks [36.380478794869234]
This paper formulates this challenge as a cooperative Multi-Agent Reinforcement Learning (MARL) problem, solved using the Training with Decentralized Execution (CTDE) framework.<n>Our proposed framework significantly outperforms baselines, increasing total system throughput by approximately 50% while simultaneously achieving a near-zero collision rate.<n>A key finding is that the agents develop an emergent anti-jamming strategy without explicit programming.
arXiv Detail & Related papers (2025-12-09T08:11:21Z) - Joint Optimization of Cooperation Efficiency and Communication Covertness for Target Detection with AUVs [105.81167650318054]
This paper investigates underwater cooperative target detection using autonomous underwater vehicles (AUVs)<n>We first formulate a joint trajectory and power control optimization problem, and then present an innovative hierarchical action management framework to solve it.<n>Under the centralized training and decentralized execution paradigm, our target detection framework enables adaptive covert cooperation while satisfying both energy and mobility constraints.
arXiv Detail & Related papers (2025-10-21T02:14:11Z) - Learning to Interact in World Latent for Team Coordination [53.51290193631586]
This work presents a novel representation learning framework, interactive world latent (IWoL), to facilitate team coordination in multi-agent reinforcement learning (MARL)<n>Our key insight is to construct a learnable representation space that jointly captures inter-agent relations and task-specific world information by directly modeling communication protocols.<n>Our representation can be used not only as an implicit latent for each agent, but also as an explicit message for communication.
arXiv Detail & Related papers (2025-09-29T22:13:39Z) - RALLY: Role-Adaptive LLM-Driven Yoked Navigation for Agentic UAV Swarms [17.931091064654904]
We develop a Role-Adaptive LLM-Driven Yoked navigation algorithm RALLY.<n>RALLY uses structured natural language for efficient semantic communication and collaborative reasoning.<n> Experiments show that RALLY outperforms conventional approaches in terms of task coverage, convergence speed, and generalization.
arXiv Detail & Related papers (2025-07-02T05:44:17Z) - Benchmarking LLMs' Swarm intelligence [51.648605206159125]
Large Language Models (LLMs) show potential for complex reasoning, yet their capacity for emergent coordination in Multi-Agent Systems (MAS) remains largely unexplored.<n>We introduce SwarmBench, a novel benchmark designed to systematically evaluate tasks of LLMs acting as decentralized agents.<n>We propose metrics for coordination effectiveness and analyze emergent group dynamics.
arXiv Detail & Related papers (2025-05-07T12:32:01Z) - Emergence of Roles in Robotic Teams with Model Sharing and Limited Communication [0.0]
We present a reinforcement learning strategy for use in multi-agent foraging systems in which the learning is centralised to a single agent.<n>This approach aims to significantly reduce the computational and energy demands compared to approaches such as MARL and centralised learning models.
arXiv Detail & Related papers (2025-05-01T14:05:46Z) - Contextual Knowledge Sharing in Multi-Agent Reinforcement Learning with Decentralized Communication and Coordination [0.9776703963093367]
Multi-Agent Reinforcement Learning (Dec-MARL) has emerged as a pivotal approach for addressing complex tasks in dynamic environments.<n>This paper presents a novel Dec-MARL framework that integrates peer-to-peer communication and coordination, incorporating goal-awareness and time-awareness into the agents' knowledge-sharing processes.
arXiv Detail & Related papers (2025-01-26T22:49:50Z) - Communication Learning in Multi-Agent Systems from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
We introduce a temporal gating mechanism for each agent, enabling dynamic decisions on whether to receive shared information at a given time.
arXiv Detail & Related papers (2024-11-01T05:56:51Z) - Decentralized and Lifelong-Adaptive Multi-Agent Collaborative Learning [57.652899266553035]
Decentralized and lifelong-adaptive multi-agent collaborative learning aims to enhance collaboration among multiple agents without a central server.
We propose DeLAMA, a decentralized multi-agent lifelong collaborative learning algorithm with dynamic collaboration graphs.
arXiv Detail & Related papers (2024-03-11T09:21:11Z) - AdverSAR: Adversarial Search and Rescue via Multi-Agent Reinforcement
Learning [4.843554492319537]
We propose an algorithm that allows robots to efficiently coordinate their strategies in the presence of adversarial inter-agent communications.
It is assumed that the robots have no prior knowledge of the target locations, and they can interact with only a subset of neighboring robots at any time.
The effectiveness of our approach is demonstrated on a collection of prototype grid-world environments.
arXiv Detail & Related papers (2022-12-20T08:13:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.