Follow the Neurally-Perturbed Leader for Adversarial Training
- URL: http://arxiv.org/abs/2002.06476v2
- Date: Mon, 8 Jun 2020 04:54:53 GMT
- Title: Follow the Neurally-Perturbed Leader for Adversarial Training
- Authors: Ari Azarafrooz
- Abstract summary: We propose a novel leader algorithm for zeros-sum training to mixed equilibrium without behaviors without perturbations.
We validate our theoretical results by applying this training algorithm to games with convex and non-perturbed loss as well as generative adversarial architectures.
We customize the implementation of this algorithm for adversarial imitation learning applications.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Game-theoretic models of learning are a powerful set of models that optimize
multi-objective architectures. Among these models are zero-sum architectures
that have inspired adversarial learning frameworks. An important shortcoming of
these zeros-sum architectures is that gradient-based training leads to weak
convergence and cyclic dynamics.
We propose a novel follow the leader training algorithm for zeros-sum
architectures that guarantees convergence to mixed Nash equilibrium without
cyclic behaviors. It is a special type of follow the perturbed leader algorithm
where perturbations are the result of a neural mediating agent.
We validate our theoretical results by applying this training algorithm to
games with convex and non-convex loss as well as generative adversarial
architectures. Moreover, we customize the implementation of this algorithm for
adversarial imitation learning applications. At every step of the training, the
mediator agent perturbs the observations with generated codes. As a result of
these mediating codes, the proposed algorithm is also efficient for learning in
environments with various factors of variations. We validate our assertion by
using a procedurally generated game environment as well as synthetic data.
Github implementation is available.
Related papers
- Optimizing Neural Network Performance and Interpretability with Diophantine Equation Encoding [0.0]
We introduce a novel approach that enhances the precision and robustness of deep learning models.
Our method integrates a custom loss function that enforces Diophantine constraints during training, leading to better generalization, reduced error bounds, and enhanced resilience against adversarial attacks.
arXiv Detail & Related papers (2024-09-11T14:38:40Z) - The Edge-of-Reach Problem in Offline Model-Based Reinforcement Learning [37.387280102209274]
offline reinforcement learning aims to enable agents to be trained from pre-collected datasets, however, this comes with the added challenge of estimating the value of behavior not covered in the dataset.
Model-based methods offer a solution by allowing agents to collect additional synthetic data via rollouts in a learned dynamics model.
However, if the learned dynamics model is replaced by the true error-free dynamics, existing model-based methods completely fail.
We propose Reach-Aware Value Learning (RAVL), a simple and robust method that directly addresses the edge-of-reach problem.
arXiv Detail & Related papers (2024-02-19T20:38:00Z) - Robust Stochastically-Descending Unrolled Networks [85.6993263983062]
Deep unrolling is an emerging learning-to-optimize method that unrolls a truncated iterative algorithm in the layers of a trainable neural network.
We show that convergence guarantees and generalizability of the unrolled networks are still open theoretical problems.
We numerically assess unrolled architectures trained under the proposed constraints in two different applications.
arXiv Detail & Related papers (2023-12-25T18:51:23Z) - Training Neural Networks with Internal State, Unconstrained
Connectivity, and Discrete Activations [66.53734987585244]
True intelligence may require the ability of a machine learning model to manage internal state.
We show that we have not yet discovered the most effective algorithms for training such models.
We present one attempt to design such a training algorithm, applied to an architecture with binary activations and only a single matrix of weights.
arXiv Detail & Related papers (2023-12-22T01:19:08Z) - Robust Graph Representation Learning via Predictive Coding [46.22695915912123]
Predictive coding is a message-passing framework initially developed to model information processing in the brain.
In this work, we build models that rely on the message-passing rule of predictive coding.
We show that the proposed models are comparable to standard ones in terms of performance in both inductive and transductive tasks.
arXiv Detail & Related papers (2022-12-09T03:58:22Z) - Improved Model based Deep Learning using Monotone Operator Learning
(MOL) [25.077510176642807]
MoDL algorithms that rely on unrolling are emerging as powerful tools for image recovery.
We introduce a novel monotone operator learning framework to overcome some of the challenges associated with current unrolled frameworks.
We demonstrate the utility of the proposed scheme in the context of parallel MRI.
arXiv Detail & Related papers (2021-11-22T17:42:27Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork [73.94896986868146]
Phase retrieval is an important component in modern computational imaging systems.
Recent advances in deep learning have opened up a new possibility for robust and fast PR.
We develop a novel framework for deep unfolding to overcome the existing limitations.
arXiv Detail & Related papers (2021-01-12T08:36:23Z) - Predictive Coding Approximates Backprop along Arbitrary Computation
Graphs [68.8204255655161]
We develop a strategy to translate core machine learning architectures into their predictive coding equivalents.
Our models perform equivalently to backprop on challenging machine learning benchmarks.
Our method raises the potential that standard machine learning algorithms could in principle be directly implemented in neural circuitry.
arXiv Detail & Related papers (2020-06-07T15:35:47Z) - Learning to Rank Learning Curves [15.976034696758148]
We present a new method that saves computational budget by terminating poor configurations early on in the training.
We show that our model is able to effectively rank learning curves without having to observe many or very long learning curves.
arXiv Detail & Related papers (2020-06-05T10:49:52Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.