Deep Reinforcement Learning Policies Learn Shared Adversarial Features
Across MDPs
- URL: http://arxiv.org/abs/2112.09025v1
- Date: Thu, 16 Dec 2021 17:10:41 GMT
- Title: Deep Reinforcement Learning Policies Learn Shared Adversarial Features
Across MDPs
- Authors: Ezgi Korkmaz
- Abstract summary: We propose a framework to investigate the decision boundary and loss landscape similarities across states and across MDPs.
We conduct experiments in various games from Arcade Learning Environment, and discover that high sensitivity directions for neural policies are correlated across MDPs.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The use of deep neural networks as function approximators has led to striking
progress for reinforcement learning algorithms and applications. Yet the
knowledge we have on decision boundary geometry and the loss landscape of
neural policies is still quite limited. In this paper we propose a framework to
investigate the decision boundary and loss landscape similarities across states
and across MDPs. We conduct experiments in various games from Arcade Learning
Environment, and discover that high sensitivity directions for neural policies
are correlated across MDPs. We argue that these high sensitivity directions
support the hypothesis that non-robust features are shared across training
environments of reinforcement learning agents. We believe our results reveal
fundamental properties of the environments used in deep reinforcement learning
training, and represent a tangible step towards building robust and reliable
deep reinforcement learning agents.
Related papers
- Understanding and Diagnosing Deep Reinforcement Learning [14.141453107129403]
Deep neural policies have recently been installed in a diverse range of settings, from biotechnology to automated financial systems.
We introduce a theoretically founded technique that provides a systematic analysis of the directions in the deep neural policy decision decision both time and space.
arXiv Detail & Related papers (2024-06-23T18:10:16Z) - Discovering Behavioral Modes in Deep Reinforcement Learning Policies
Using Trajectory Clustering in Latent Space [0.0]
We introduce a new approach for investigating the behavior modes of DRL policies.
Specifically, we use Pairwise Controlled Manifold Approximation Projection (PaCMAP) for dimensionality reduction and TRACLUS for trajectory clustering.
Our methodology helps identify diverse behavior patterns and suboptimal choices by the policy, thus allowing for targeted improvements.
arXiv Detail & Related papers (2024-02-20T11:50:50Z) - Detecting Adversarial Directions in Deep Reinforcement Learning to Make
Robust Decisions [8.173034693197351]
We propose a novel method to detect the presence of non-robust directions in MDPs.
Our method provides a theoretical basis for the fundamental cut-off between safe observations and adversarial observations.
Most significantly, we demonstrate the effectiveness of our approach even in the setting where non-robust directions are explicitly optimized to circumvent our proposed method.
arXiv Detail & Related papers (2023-06-09T13:11:05Z) - Adversarial Robust Deep Reinforcement Learning Requires Redefining
Robustness [7.6146285961466]
We show that high sensitivity directions are more abundant in the deep neural policy landscape and can be found via more natural means in a black-box setting.
We show that vanilla training techniques intriguingly result in learning more robust policies compared to the policies learnt via the state-of-the-art adversarial training techniques.
arXiv Detail & Related papers (2023-01-17T16:54:33Z) - Learning Dynamics and Generalization in Reinforcement Learning [59.530058000689884]
We show theoretically that temporal difference learning encourages agents to fit non-smooth components of the value function early in training.
We show that neural networks trained using temporal difference algorithms on dense reward tasks exhibit weaker generalization between states than randomly networks and gradient networks trained with policy methods.
arXiv Detail & Related papers (2022-06-05T08:49:16Z) - Rethinking Learning Dynamics in RL using Adversarial Networks [79.56118674435844]
We present a learning mechanism for reinforcement learning of closely related skills parameterized via a skill embedding space.
The main contribution of our work is to formulate an adversarial training regime for reinforcement learning with the help of entropy-regularized policy gradient formulation.
arXiv Detail & Related papers (2022-01-27T19:51:09Z) - Investigating Vulnerabilities of Deep Neural Policies [0.0]
Reinforcement learning policies based on deep neural networks are vulnerable to imperceptible adversarial perturbations to their inputs.
Recent work has proposed several methods to improve the robustness of deep reinforcement learning agents to adversarial perturbations.
We study the effects of adversarial training on the neural policy learned by the agent.
arXiv Detail & Related papers (2021-08-30T10:04:50Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Fusing the Old with the New: Learning Relative Camera Pose with
Geometry-Guided Uncertainty [91.0564497403256]
We present a novel framework that involves probabilistic fusion between the two families of predictions during network training.
Our network features a self-attention graph neural network, which drives the learning by enforcing strong interactions between different correspondences.
We propose motion parmeterizations suitable for learning and show that our method achieves state-of-the-art performance on the challenging DeMoN and ScanNet datasets.
arXiv Detail & Related papers (2021-04-16T17:59:06Z) - Optimism in the Face of Adversity: Understanding and Improving Deep
Learning through Adversarial Robustness [63.627760598441796]
We provide an in-depth review of the field of adversarial robustness in deep learning.
We highlight the intuitive connection between adversarial examples and the geometry of deep neural networks.
We provide an overview of the main emerging applications of adversarial robustness beyond security.
arXiv Detail & Related papers (2020-10-19T16:03:46Z) - Learn2Perturb: an End-to-end Feature Perturbation Learning to Improve
Adversarial Robustness [79.47619798416194]
Learn2Perturb is an end-to-end feature perturbation learning approach for improving the adversarial robustness of deep neural networks.
Inspired by the Expectation-Maximization, an alternating back-propagation training algorithm is introduced to train the network and noise parameters consecutively.
arXiv Detail & Related papers (2020-03-02T18:27:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.