Multi-agent Reinforcement Learning in Bayesian Stackelberg Markov Games
for Adaptive Moving Target Defense
- URL: http://arxiv.org/abs/2007.10457v1
- Date: Mon, 20 Jul 2020 20:34:53 GMT
- Title: Multi-agent Reinforcement Learning in Bayesian Stackelberg Markov Games
for Adaptive Moving Target Defense
- Authors: Sailik Sengupta, Subbarao Kambhampati
- Abstract summary: We argue that existing models are inadequate in sequential settings when there is incomplete information about a rational adversary.
We propose a unifying game-theoretic model, called the Bayesian Stackelberg Markov Games (BSMGs)
We show that our learning approach converges to an SSE of a BSMG and then highlight that the learned movement policy improves the state-of-the-art in MTD for web-application security.
- Score: 22.760124873882184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The field of cybersecurity has mostly been a cat-and-mouse game with the
discovery of new attacks leading the way. To take away an attacker's advantage
of reconnaissance, researchers have proposed proactive defense methods such as
Moving Target Defense (MTD). To find good movement strategies, researchers have
modeled MTD as leader-follower games between the defender and a
cyber-adversary. We argue that existing models are inadequate in sequential
settings when there is incomplete information about a rational adversary and
yield sub-optimal movement strategies. Further, while there exists an array of
work on learning defense policies in sequential settings for cyber-security,
they are either unpopular due to scalability issues arising out of incomplete
information or tend to ignore the strategic nature of the adversary simplifying
the scenario to use single-agent reinforcement learning techniques. To address
these concerns, we propose (1) a unifying game-theoretic model, called the
Bayesian Stackelberg Markov Games (BSMGs), that can model uncertainty over
attacker types and the nuances of an MTD system and (2) a Bayesian Strong
Stackelberg Q-learning (BSS-Q) approach that can, via interaction, learn the
optimal movement policy for BSMGs within a reasonable time. We situate BSMGs in
the landscape of incomplete-information Markov games and characterize the
notion of Strong Stackelberg Equilibrium (SSE) in them. We show that our
learning approach converges to an SSE of a BSMG and then highlight that the
learned movement policy (1) improves the state-of-the-art in MTD for
web-application security and (2) converges to an optimal policy in MTD domains
with incomplete information about adversaries even when prior information about
rewards and transitions is absent.
Related papers
- Meta Stackelberg Game: Robust Federated Learning against Adaptive and Mixed Poisoning Attacks [15.199885837603576]
Federated learning (FL) is susceptible to a range of security threats.
We develop an efficient meta-learning approach to solve the game, leading to a robust and adaptive FL defense.
arXiv Detail & Related papers (2024-10-22T21:08:28Z) - Toward Optimal LLM Alignments Using Two-Player Games [86.39338084862324]
In this paper, we investigate alignment through the lens of two-agent games, involving iterative interactions between an adversarial and a defensive agent.
We theoretically demonstrate that this iterative reinforcement learning optimization converges to a Nash Equilibrium for the game induced by the agents.
Experimental results in safety scenarios demonstrate that learning in such a competitive environment not only fully trains agents but also leads to policies with enhanced generalization capabilities for both adversarial and defensive agents.
arXiv Detail & Related papers (2024-06-16T15:24:50Z) - Mutual-modality Adversarial Attack with Semantic Perturbation [81.66172089175346]
We propose a novel approach that generates adversarial attacks in a mutual-modality optimization scheme.
Our approach outperforms state-of-the-art attack methods and can be readily deployed as a plug-and-play solution.
arXiv Detail & Related papers (2023-12-20T05:06:01Z) - Baseline Defenses for Adversarial Attacks Against Aligned Language
Models [109.75753454188705]
Recent work shows that text moderations can produce jailbreaking prompts that bypass defenses.
We look at three types of defenses: detection (perplexity based), input preprocessing (paraphrase and retokenization), and adversarial training.
We find that the weakness of existing discretes for text, combined with the relatively high costs of optimization, makes standard adaptive attacks more challenging for LLMs.
arXiv Detail & Related papers (2023-09-01T17:59:44Z) - A First Order Meta Stackelberg Method for Robust Federated Learning [19.130600532727062]
This work models adversarial federated learning as a Bayesian Stackelberg Markov game (BSMG)
We propose meta-Stackelberg learning (meta-SL), a provably efficient meta-learning algorithm, to solve the equilibrium strategy in BSMG.
We demonstrate that meta-SL converges to the first-order $varepsilon$-equilibrium point in $O(varepsilon-2)$ gradient, with $O(varepsilon-4)$ samples needed per iteration.
arXiv Detail & Related papers (2023-06-23T22:22:33Z) - Avoid Adversarial Adaption in Federated Learning by Multi-Metric
Investigations [55.2480439325792]
Federated Learning (FL) facilitates decentralized machine learning model training, preserving data privacy, lowering communication costs, and boosting model performance through diversified data sources.
FL faces vulnerabilities such as poisoning attacks, undermining model integrity with both untargeted performance degradation and targeted backdoor attacks.
We define a new notion of strong adaptive adversaries, capable of adapting to multiple objectives simultaneously.
MESAS is the first defense robust against strong adaptive adversaries, effective in real-world data scenarios, with an average overhead of just 24.37 seconds.
arXiv Detail & Related papers (2023-06-06T11:44:42Z) - Ares: A System-Oriented Wargame Framework for Adversarial ML [3.197282271064602]
Ares is an evaluation framework for adversarial ML that allows researchers to explore attacks and defenses in a realistic wargame-like environment.
Ares frames the conflict between the attacker and defender as two agents in a reinforcement learning environment with opposing objectives.
This allows the introduction of system-level evaluation metrics such as time to failure and evaluation of complex strategies.
arXiv Detail & Related papers (2022-10-24T04:55:18Z) - StratDef: Strategic Defense Against Adversarial Attacks in ML-based
Malware Detection [0.0]
StratDef is a strategic defense system based on a moving target defense approach.
We show that StratDef performs better than other defenses even when facing the peak adversarial threat.
arXiv Detail & Related papers (2022-02-15T16:51:53Z) - Fixed Points in Cyber Space: Rethinking Optimal Evasion Attacks in the
Age of AI-NIDS [70.60975663021952]
We study blackbox adversarial attacks on network classifiers.
We argue that attacker-defender fixed points are themselves general-sum games with complex phase transitions.
We show that a continual learning approach is required to study attacker-defender dynamics.
arXiv Detail & Related papers (2021-11-23T23:42:16Z) - Adversarial Attack and Defense in Deep Ranking [100.17641539999055]
We propose two attacks against deep ranking systems that can raise or lower the rank of chosen candidates by adversarial perturbations.
Conversely, an anti-collapse triplet defense is proposed to improve the ranking model robustness against all proposed attacks.
Our adversarial ranking attacks and defenses are evaluated on MNIST, Fashion-MNIST, CUB200-2011, CARS196 and Stanford Online Products datasets.
arXiv Detail & Related papers (2021-06-07T13:41:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.