How to Combat Reactive and Dynamic Jamming Attacks with Reinforcement Learning
- URL: http://arxiv.org/abs/2510.02265v1
- Date: Thu, 02 Oct 2025 17:44:38 GMT
- Title: How to Combat Reactive and Dynamic Jamming Attacks with Reinforcement Learning
- Authors: Yalin E. Sagduyu, Tugba Erpek, Kemal Davaslioglu, Sastry Kompella,
- Abstract summary: This paper studies the problem of mitigating reactive jamming, where a jammer adopts a dynamic policy of selecting channels and sensing thresholds to detect and jam ongoing transmissions.<n>The transmitter-receiver pair learns to avoid jamming and optimize throughput over time by using reinforcement learning (RL) to adapt transmit power, modulation, and channel selection.
- Score: 7.510555203834326
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies the problem of mitigating reactive jamming, where a jammer adopts a dynamic policy of selecting channels and sensing thresholds to detect and jam ongoing transmissions. The transmitter-receiver pair learns to avoid jamming and optimize throughput over time (without prior knowledge of channel conditions or jamming strategies) by using reinforcement learning (RL) to adapt transmit power, modulation, and channel selection. Q-learning is employed for discrete jamming-event states, while Deep Q-Networks (DQN) are employed for continuous states based on received power. Through different reward functions and action sets, the results show that RL can adapt rapidly to spectrum dynamics and sustain high rates as channels and jamming policies change over time.
Related papers
- Deep Q-Network Based Resilient Drone Communication:Neutralizing First-Order Markov Jammers [0.0]
Deep Q Network based transmitter continuously selects the next frequency hopping channel while facing first order reactive jamming.<n>Through self training, the proposed agent learns a uniform random frequency hopping policy that effectively neutralizes the predictive advantage of the jamming.
arXiv Detail & Related papers (2026-01-01T14:16:40Z) - Coordinated Anti-Jamming Resilience in Swarm Networks via Multi-Agent Reinforcement Learning [8.533838668681737]
Reactive jammers pose a severe security threat to robotic-swarm networks by selectively disrupting inter-agent communications.<n>Conventional countermeasures such as fixed power control or static channel hopping are largely ineffective against such adaptive adversaries.<n>This paper presents a multi-agent reinforcement learning framework based on the QMIX algorithm to improve the resilience of swarm communications under reactive jamming.
arXiv Detail & Related papers (2025-12-18T17:54:20Z) - ImplicitRDP: An End-to-End Visual-Force Diffusion Policy with Structural Slow-Fast Learning [52.86018040861575]
We propose a unified end-to-end visual-force diffusion policy that integrates visual planning and reactive force control within a single network.<n>We introduce Structural Slow-Fast Learning, a mechanism utilizing causal attention to simultaneously process asynchronous visual and force tokens.<n>Experiments on contact-rich tasks demonstrate that ImplicitRDP significantly outperforms both vision-only and hierarchical baselines.
arXiv Detail & Related papers (2025-12-11T18:59:46Z) - Efficient Test-time Adaptive Object Detection via Sensitivity-Guided Pruning [73.40364018029673]
Continual test-time adaptive object detection (CTTA-OD) aims to online adapt a source pre-trained detector to ever-changing environments.<n>Our motivation stems from the observation that not all learned source features are beneficial.<n>Our method achieves superior adaptation performance while reducing computational overhead by 12% in FLOPs.
arXiv Detail & Related papers (2025-06-03T05:27:56Z) - ViaRL: Adaptive Temporal Grounding via Visual Iterated Amplification Reinforcement Learning [68.76048244253582]
We introduce ViaRL, the first framework to leverage rule-based reinforcement learning (RL) for optimizing frame selection in video understanding.<n>ViaRL utilizes the answer accuracy of a downstream model as a reward signal to train a frame selector through trial-and-error.<n>ViaRL consistently delivers superior temporal grounding performance and robust generalization across diverse video understanding tasks.
arXiv Detail & Related papers (2025-05-21T12:29:40Z) - Fast Adaptive Anti-Jamming Channel Access via Deep Q Learning and Coarse-Grained Spectrum Prediction [13.4498936010732]
This paper investigates the anti-jamming channel access problem in complex and unknown jamming environments.<n>Traditional channel hopping anti-jamming approaches using fixed patterns are ineffective against dynamic jamming attacks.<n>We propose a fast adaptive anti-jamming channel access approach guided by the intuition of learning faster than the jammer'
arXiv Detail & Related papers (2025-02-07T14:25:28Z) - Reinforcement Learning Based Goodput Maximization with Quantized Feedback in URLLC [25.794259368427277]
The study investigates a communication system, where the receiver provides quantized channel state information to the transmitter.<n>We introduce a novel Rician-$K$ factor estimation technique to enable the communication system to optimize the feedback scheme.
arXiv Detail & Related papers (2025-01-19T22:37:55Z) - Offline Reinforcement Learning and Sequence Modeling for Downlink Link Adaptation [3.687363450234871]
Link adaptation (LA) is an essential function in modern wireless communication systems.<n>LA dynamically adjusts the transmission rate of a communication link to match time- and frequency-varying radio link conditions.<n>Recent research has introduced online reinforcement learning approaches as an alternative to the more commonly used rule-based algorithms.
arXiv Detail & Related papers (2024-10-30T14:01:31Z) - Diffusion Policies creating a Trust Region for Offline Reinforcement Learning [66.17291150498276]
We introduce a dual policy approach, Diffusion Trusted Q-Learning (DTQL), which comprises a diffusion policy for pure behavior cloning and a practical one-step policy.
DTQL eliminates the need for iterative denoising sampling during both training and inference, making it remarkably computationally efficient.
We show that DTQL could not only outperform other methods on the majority of the D4RL benchmark tasks but also demonstrate efficiency in training and inference speeds.
arXiv Detail & Related papers (2024-05-30T05:04:33Z) - Actor-Critic Scheduling for Path-Aware Air-to-Ground Multipath
Multimedia Delivery [5.01187288554981]
We present a novel scheduler for real-time multimedia delivery in multipath systems based on an Actor-Critic (AC) RL algorithm.
The scheduler acting as an RL agent learns in real-time the optimal policy for path selection, path rate allocation and redundancy estimation for flow protection.
arXiv Detail & Related papers (2022-04-28T08:28:25Z) - Reinforcement Learning for Deceiving Reactive Jammers in Wireless
Networks [76.82565500647323]
A novel anti-jamming strategy is proposed based on the idea of deceiving the jammer into attacking a victim channel.
Since the jammer's channel information is not known to the users, an optimal channel selection scheme and a sub optimal power allocation are proposed.
Simulation results show that the proposed anti-jamming method outperforms the compared RL based anti-jamming methods and random search method.
arXiv Detail & Related papers (2021-03-25T18:12:41Z) - Accelerating Deep Reinforcement Learning With the Aid of Partial Model:
Energy-Efficient Predictive Video Streaming [97.75330397207742]
Predictive power allocation is conceived for energy-efficient video streaming over mobile networks using deep reinforcement learning.
To handle the continuous state and action spaces, we resort to deep deterministic policy gradient (DDPG) algorithm.
Our simulation results show that the proposed policies converge to the optimal policy that is derived based on perfect large-scale channel prediction.
arXiv Detail & Related papers (2020-03-21T17:36:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.