Coordinated Anti-Jamming Resilience in Swarm Networks via Multi-Agent Reinforcement Learning
- URL: http://arxiv.org/abs/2512.16813v1
- Date: Thu, 18 Dec 2025 17:54:20 GMT
- Title: Coordinated Anti-Jamming Resilience in Swarm Networks via Multi-Agent Reinforcement Learning
- Authors: Bahman Abolhassani, Tugba Erpek, Kemal Davaslioglu, Yalin E. Sagduyu, Sastry Kompella,
- Abstract summary: Reactive jammers pose a severe security threat to robotic-swarm networks by selectively disrupting inter-agent communications.<n>Conventional countermeasures such as fixed power control or static channel hopping are largely ineffective against such adaptive adversaries.<n>This paper presents a multi-agent reinforcement learning framework based on the QMIX algorithm to improve the resilience of swarm communications under reactive jamming.
- Score: 8.533838668681737
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reactive jammers pose a severe security threat to robotic-swarm networks by selectively disrupting inter-agent communications and undermining formation integrity and mission success. Conventional countermeasures such as fixed power control or static channel hopping are largely ineffective against such adaptive adversaries. This paper presents a multi-agent reinforcement learning (MARL) framework based on the QMIX algorithm to improve the resilience of swarm communications under reactive jamming. We consider a network of multiple transmitter-receiver pairs sharing channels while a reactive jammer with Markovian threshold dynamics senses aggregate power and reacts accordingly. Each agent jointly selects transmit frequency (channel) and power, and QMIX learns a centralized but factorizable action-value function that enables coordinated yet decentralized execution. We benchmark QMIX against a genie-aided optimal policy in a no-channel-reuse setting, and against local Upper Confidence Bound (UCB) and a stateless reactive policy in a more general fading regime with channel reuse enabled. Simulation results show that QMIX rapidly converges to cooperative policies that nearly match the genie-aided bound, while achieving higher throughput and lower jamming incidence than the baselines, thereby demonstrating MARL's effectiveness for securing autonomous swarms in contested environments.
Related papers
- Uncertainty-Aware Jamming Mitigation with Active RIS: A Robust Stackelberg Game Approach [65.06640919319413]
This paper investigates the jamming mitigation by leveraging an active reconfigurable intelligent surface (ARIS)<n>We adopt the Stackelberg game formulation to model the strategic interaction between the legitimate side and the adversary.<n>We first derive the optimal jamming policy as the follower's best response, which is then incorporated into the legitimate-side optimization for robust anti-jamming design.
arXiv Detail & Related papers (2026-02-20T12:02:01Z) - Multi-Agent-Driven Cognitive Secure Communications in Satellite-Terrestrial Networks [58.70163955407538]
Malicious eavesdroppers pose a serious threat to private information via satellite-terrestrial networks (STNs)<n>We propose a cognitive secure communication framework driven by multiple agents that coordinates spectrum scheduling and protection through real-time sensing.<n>We exploit generative adversarial networks to produce adversarial matrices, and employ learning-aided power control to set real and adversarial signal powers for protection layer.
arXiv Detail & Related papers (2026-01-06T10:30:41Z) - Deep Q-Network Based Resilient Drone Communication:Neutralizing First-Order Markov Jammers [0.0]
Deep Q Network based transmitter continuously selects the next frequency hopping channel while facing first order reactive jamming.<n>Through self training, the proposed agent learns a uniform random frequency hopping policy that effectively neutralizes the predictive advantage of the jamming.
arXiv Detail & Related papers (2026-01-01T14:16:40Z) - Multi-Agent Deep Reinforcement Learning for Collaborative UAV Relay Networks under Jamming Atatcks [36.380478794869234]
This paper formulates this challenge as a cooperative Multi-Agent Reinforcement Learning (MARL) problem, solved using the Training with Decentralized Execution (CTDE) framework.<n>Our proposed framework significantly outperforms baselines, increasing total system throughput by approximately 50% while simultaneously achieving a near-zero collision rate.<n>A key finding is that the agents develop an emergent anti-jamming strategy without explicit programming.
arXiv Detail & Related papers (2025-12-09T08:11:21Z) - SecDiff: Diffusion-Aided Secure Deep Joint Source-Channel Coding Against Adversarial Attacks [73.41290017870097]
SecDiff is a plug-and-play, diffusion-aided decoding framework.<n>It significantly enhances the security and robustness of deep J SCC under adversarial wireless environments.
arXiv Detail & Related papers (2025-11-03T11:24:06Z) - When UAV Swarm Meets IRS: Collaborative Secure Communications in Low-altitude Wireless Networks [68.45202147860537]
Low-altitude wireless networks (LAWNs) provide enhanced coverage, reliability, and throughput for diverse applications.<n>These networks face significant security vulnerabilities from both known and potential unknown eavesdroppers.<n>We propose a novel secure communication framework for LAWNs where the selected UAVs within a swarm function as a virtual antenna array.
arXiv Detail & Related papers (2025-10-25T02:02:14Z) - STAR-RIS-assisted Collaborative Beamforming for Low-altitude Wireless Networks [58.13757830013997]
Wireless networks based on uncrewed aerial vehicles (UAVs) offer high mobility, flexibility, and coverage for urban communications.<n>They face severe signal attenuation in dense environments due to obstructions.<n>To address this critical issue, we consider introducing collaborative beam of UAVs and omni-directional re-altitude beamforming.
arXiv Detail & Related papers (2025-10-25T01:28:37Z) - How to Combat Reactive and Dynamic Jamming Attacks with Reinforcement Learning [7.510555203834326]
This paper studies the problem of mitigating reactive jamming, where a jammer adopts a dynamic policy of selecting channels and sensing thresholds to detect and jam ongoing transmissions.<n>The transmitter-receiver pair learns to avoid jamming and optimize throughput over time by using reinforcement learning (RL) to adapt transmit power, modulation, and channel selection.
arXiv Detail & Related papers (2025-10-02T17:44:38Z) - Heterogeneous Multi-Agent Reinforcement Learning for Distributed Channel Access in WLANs [47.600901884970845]
This paper investigates the use of multi-agent reinforcement learning (MARL) to address distributed channel access in wireless local area networks.<n>In particular, we consider the challenging yet more practical case where the agents heterogeneously adopt value-based or policy-based reinforcement learning algorithms to train the model.<n>We propose a heterogeneous MARL training framework, named QPMIX, which adopts a centralized training with distributed execution paradigm to enable heterogeneous agents to collaborate.
arXiv Detail & Related papers (2024-12-18T13:50:31Z) - Distributed Autonomous Swarm Formation for Dynamic Network Bridging [40.27919181139919]
We formulate the problem of dynamic network bridging in a novel Decentralized Partially Observable Markov Decision Process (Dec-POMDP)
We propose a Multi-Agent Reinforcement Learning (MARL) approach for the problem based on Graph Convolutional Reinforcement Learning (DGN)
The proposed method is evaluated in a simulated environment and compared to a centralized baseline showing promising results.
arXiv Detail & Related papers (2024-04-02T01:45:03Z) - Optimal Power Allocation for Rate Splitting Communications with Deep
Reinforcement Learning [61.91604046990993]
This letter introduces a novel framework to optimize the power allocation for users in a Rate Splitting Multiple Access network.
In the network, messages intended for users are split into different parts that are a single common part and respective private parts.
arXiv Detail & Related papers (2021-07-01T06:32:49Z) - Monotonic Value Function Factorisation for Deep Multi-Agent
Reinforcement Learning [55.20040781688844]
QMIX is a novel value-based method that can train decentralised policies in a centralised end-to-end fashion.
We propose the StarCraft Multi-Agent Challenge (SMAC) as a new benchmark for deep multi-agent reinforcement learning.
arXiv Detail & Related papers (2020-03-19T16:51:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.