Inverse-Inverse Reinforcement Learning. How to Hide Strategy from an
Adversarial Inverse Reinforcement Learner
- URL: http://arxiv.org/abs/2205.10802v1
- Date: Sun, 22 May 2022 11:54:44 GMT
- Title: Inverse-Inverse Reinforcement Learning. How to Hide Strategy from an
Adversarial Inverse Reinforcement Learner
- Authors: Kunal Pattanayak and Vikram Krishnamurthy and Christopher Berry
- Abstract summary: Inverse reinforcement learning deals with estimating an agent's utility function from its actions.
We consider how an agent can hide its strategy and mitigate an adversarial IRL attack.
- Score: 19.044614610714856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Inverse reinforcement learning (IRL) deals with estimating an agent's utility
function from its actions. In this paper, we consider how an agent can hide its
strategy and mitigate an adversarial IRL attack; we call this inverse IRL
(I-IRL). How should the decision maker choose its response to ensure a poor
reconstruction of its strategy by an adversary performing IRL to estimate the
agent's strategy? This paper comprises four results: First, we present an
adversarial IRL algorithm that estimates the agent's strategy while controlling
the agent's utility function. Our second result for I-IRL result spoofs the IRL
algorithm used by the adversary. Our I-IRL results are based on revealed
preference theory in micro-economics. The key idea is for the agent to
deliberately choose sub-optimal responses that sufficiently masks its true
strategy. Third, we give a sample complexity result for our main I-IRL result
when the agent has noisy estimates of the adversary specified utility function.
Finally, we illustrate our I-IRL scheme in a radar problem where a
meta-cognitive radar is trying to mitigate an adversarial target.
Related papers
- Non-Adversarial Inverse Reinforcement Learning via Successor Feature Matching [23.600285251963395]
In inverse reinforcement learning (IRL), an agent seeks to replicate expert demonstrations through interactions with the environment.
Traditionally, IRL is treated as an adversarial game, where an adversary searches over reward models, and a learner optimize the reward through repeated RL procedures.
We propose a novel approach to IRL by direct policy optimization, exploiting a linear factorization of the return as the inner product of successor features and a reward vector.
arXiv Detail & Related papers (2024-11-11T14:05:50Z) - Strategic Apple Tasting [35.25249063553063]
Algorithmic decision-making in high-stakes domains often involves assigning decisions to agents with incentives to strategically modify their input to the algorithm.
We formalize this setting as an online learning problem with apple-tasting feedback.
Our goal is to achieve sublinear strategic regret, which compares the performance of the principal to that of the best fixed policy in hindsight.
arXiv Detail & Related papers (2023-06-09T20:46:31Z) - Reinforcement Learning Agents in Colonel Blotto [0.0]
We focus on a specific instance of agent-based models, which uses reinforcement learning (RL) to train the agent how to act in its environment.
We find that the RL agent handily beats a single opponent, and still performs quite well when the number of opponents are increased.
We also analyze the RL agent and look at what strategies it has arrived by looking at the actions that it has given the highest and lowest Q-values.
arXiv Detail & Related papers (2022-04-04T16:18:01Z) - Inverse Reinforcement Learning for Strategy Identification [2.6572330982240935]
In adversarial environments, one side could gain an advantage by identifying the opponent's strategy.
This paper proposes to use inverse reinforcement learning (IRL) to identify strategies in adversarial environments.
arXiv Detail & Related papers (2021-07-31T17:22:52Z) - Explore and Control with Adversarial Surprise [78.41972292110967]
Reinforcement learning (RL) provides a framework for learning goal-directed policies given user-specified rewards.
We propose a new unsupervised RL technique based on an adversarial game which pits two policies against each other to compete over the amount of surprise an RL agent experiences.
We show that our method leads to the emergence of complex skills by exhibiting clear phase transitions.
arXiv Detail & Related papers (2021-07-12T17:58:40Z) - Disturbing Reinforcement Learning Agents with Corrupted Rewards [62.997667081978825]
We analyze the effects of different attack strategies based on reward perturbations on reinforcement learning algorithms.
We show that smoothly crafting adversarial rewards are able to mislead the learner, and that using low exploration probability values, the policy learned is more robust to corrupt rewards.
arXiv Detail & Related papers (2021-02-12T15:53:48Z) - Provably Efficient Algorithms for Multi-Objective Competitive RL [54.22598924633369]
We study multi-objective reinforcement learning (RL) where an agent's reward is represented as a vector.
In settings where an agent competes against opponents, its performance is measured by the distance of its average return vector to a target set.
We develop statistically and computationally efficient algorithms to approach the associated target set.
arXiv Detail & Related papers (2021-02-05T14:26:00Z) - Robust Reinforcement Learning on State Observations with Learned Optimal
Adversary [86.0846119254031]
We study the robustness of reinforcement learning with adversarially perturbed state observations.
With a fixed agent policy, we demonstrate that an optimal adversary to perturb state observations can be found.
For DRL settings, this leads to a novel empirical adversarial attack to RL agents via a learned adversary that is much stronger than previous ones.
arXiv Detail & Related papers (2021-01-21T05:38:52Z) - Robust Deep Reinforcement Learning through Adversarial Loss [74.20501663956604]
Recent studies have shown that deep reinforcement learning agents are vulnerable to small adversarial perturbations on the agent's inputs.
We propose RADIAL-RL, a principled framework to train reinforcement learning agents with improved robustness against adversarial attacks.
arXiv Detail & Related papers (2020-08-05T07:49:42Z) - Efficient exploration of zero-sum stochastic games [83.28949556413717]
We investigate the increasingly important and common game-solving setting where we do not have an explicit description of the game but only oracle access to it through gameplay.
During a limited-duration learning phase, the algorithm can control the actions of both players in order to try to learn the game and how to play it well.
Our motivation is to quickly learn strategies that have low exploitability in situations where evaluating the payoffs of a queried strategy profile is costly.
arXiv Detail & Related papers (2020-02-24T20:30:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.