Learning Security Strategies through Game Play and Optimal Stopping
- URL: http://arxiv.org/abs/2205.14694v1
- Date: Sun, 29 May 2022 15:30:00 GMT
- Title: Learning Security Strategies through Game Play and Optimal Stopping
- Authors: Kim Hammar and Rolf Stadler
- Abstract summary: We study automated intrusion prevention using reinforcement learning.
We formulate the interaction between an attacker and a defender as an optimal stopping game.
To obtain the optimal defender strategies, we introduce T-FP, a fictitious self-play algorithm.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We study automated intrusion prevention using reinforcement learning.
Following a novel approach, we formulate the interaction between an attacker
and a defender as an optimal stopping game and let attack and defense
strategies evolve through reinforcement learning and self-play. The
game-theoretic perspective allows us to find defender strategies that are
effective against dynamic attackers. The optimal stopping formulation gives us
insight into the structure of optimal strategies, which we show to have
threshold properties. To obtain the optimal defender strategies, we introduce
T-FP, a fictitious self-play algorithm that learns Nash equilibria through
stochastic approximation. We show that T-FP outperforms a state-of-the-art
algorithm for our use case. Our overall method for learning and evaluating
strategies includes two systems: a simulation system where defender strategies
are incrementally learned and an emulation system where statistics are produced
that drive simulation runs and where learned strategies are evaluated. We
conclude that this approach can produce effective defender strategies for a
practical IT infrastructure.
Related papers
- Robust Image Classification: Defensive Strategies against FGSM and PGD Adversarial Attacks [0.0]
Adversarial attacks pose significant threats to the robustness of deep learning models in image classification.
This paper explores and refines defense mechanisms against these attacks to enhance the resilience of neural networks.
arXiv Detail & Related papers (2024-08-20T02:00:02Z) - Optimizing Cyber Defense in Dynamic Active Directories through Reinforcement Learning [10.601458163651582]
This paper addresses the absence of effective edge-blocking ACO strategies in dynamic, real-world networks.
It specifically targets the cybersecurity vulnerabilities of organizational Active Directory (AD) systems.
Unlike the existing literature on edge-blocking defenses which considers AD systems as static entities, our study counters this by recognizing their dynamic nature.
arXiv Detail & Related papers (2024-06-28T01:37:46Z) - Mutual-modality Adversarial Attack with Semantic Perturbation [81.66172089175346]
We propose a novel approach that generates adversarial attacks in a mutual-modality optimization scheme.
Our approach outperforms state-of-the-art attack methods and can be readily deployed as a plug-and-play solution.
arXiv Detail & Related papers (2023-12-20T05:06:01Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - Learning Near-Optimal Intrusion Responses Against Dynamic Attackers [0.0]
We study automated intrusion response and formulate the interaction between an attacker and a defender as an optimal stopping game.
To obtain near-optimal defender strategies, we develop a fictitious self-play algorithm that learns Nashlibria through approximation.
We argue that this approach can produce effective defender strategies for a practical IT infrastructure.
arXiv Detail & Related papers (2023-01-11T16:36:24Z) - LAS-AT: Adversarial Training with Learnable Attack Strategy [82.88724890186094]
"Learnable attack strategy", dubbed LAS-AT, learns to automatically produce attack strategies to improve the model robustness.
Our framework is composed of a target network that uses AEs for training to improve robustness and a strategy network that produces attack strategies to control the AE generation.
arXiv Detail & Related papers (2022-03-13T10:21:26Z) - Projective Ranking-based GNN Evasion Attacks [52.85890533994233]
Graph neural networks (GNNs) offer promising learning methods for graph-related tasks.
GNNs are at risk of adversarial attacks.
arXiv Detail & Related papers (2022-02-25T21:52:09Z) - Intrusion Prevention through Optimal Stopping [0.0]
We study automated intrusion prevention using reinforcement learning.
We show that our approach can produce effective defender policies for a practical IT infrastructure of limited size.
arXiv Detail & Related papers (2021-10-30T17:03:28Z) - Model-Agnostic Meta-Attack: Towards Reliable Evaluation of Adversarial
Robustness [53.094682754683255]
We propose a Model-Agnostic Meta-Attack (MAMA) approach to discover stronger attack algorithms automatically.
Our method learns the in adversarial attacks parameterized by a recurrent neural network.
We develop a model-agnostic training algorithm to improve the ability of the learned when attacking unseen defenses.
arXiv Detail & Related papers (2021-10-13T13:54:24Z) - Finding Effective Security Strategies through Reinforcement Learning and
Self-Play [0.0]
We show that effective security strategies can emerge from self-play.
We address known challenges of reinforcement learning in this domain.
Our method is superior to two baseline methods but that policy convergence in self-play remains a challenge.
arXiv Detail & Related papers (2020-09-17T07:41:27Z) - Boosting Adversarial Training with Hypersphere Embedding [53.75693100495097]
Adversarial training is one of the most effective defenses against adversarial attacks for deep learning models.
In this work, we advocate incorporating the hypersphere embedding mechanism into the AT procedure.
We validate our methods under a wide range of adversarial attacks on the CIFAR-10 and ImageNet datasets.
arXiv Detail & Related papers (2020-02-20T08:42:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.