Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach
- URL: http://arxiv.org/abs/2208.03889v1
- Date: Mon, 8 Aug 2022 03:13:24 GMT
- Title: Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach
- Authors: Saber Jafarpour and Alexander Davydov and Matthew Abate and Francesco
Bullo and Samuel Coogan
- Abstract summary: This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
- Score: 64.23331120621118
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a theoretical and computational framework for training
and robustness verification of implicit neural networks based upon
non-Euclidean contraction theory. The basic idea is to cast the robustness
analysis of a neural network as a reachability problem and use (i) the
$\ell_{\infty}$-norm input-output Lipschitz constant and (ii) the tight
inclusion function of the network to over-approximate its reachable sets.
First, for a given implicit neural network, we use $\ell_{\infty}$-matrix
measures to propose sufficient conditions for its well-posedness, design an
iterative algorithm to compute its fixed points, and provide upper bounds for
its $\ell_\infty$-norm input-output Lipschitz constant. Second, we introduce a
related embedded network and show that the embedded network can be used to
provide an $\ell_\infty$-norm box over-approximation of the reachable sets of
the original network. Moreover, we use the embedded network to design an
iterative algorithm for computing the upper bounds of the original system's
tight inclusion function. Third, we use the upper bounds of the Lipschitz
constants and the upper bounds of the tight inclusion functions to design two
algorithms for the training and robustness verification of implicit neural
networks. Finally, we apply our algorithms to train implicit neural networks on
the MNIST dataset and compare the robustness of our models with the models
trained via existing approaches in the literature.
Related papers
- Computable Lipschitz Bounds for Deep Neural Networks [0.0]
We analyse three existing upper bounds written for the $l2$ norm.
We propose two novel bounds for both feed-forward fully-connected neural networks and convolutional neural networks.
arXiv Detail & Related papers (2024-10-28T14:09:46Z) - LinSATNet: The Positive Linear Satisfiability Neural Networks [116.65291739666303]
This paper studies how to introduce the popular positive linear satisfiability to neural networks.
We propose the first differentiable satisfiability layer based on an extension of the classic Sinkhorn algorithm for jointly encoding multiple sets of marginal distributions.
arXiv Detail & Related papers (2024-07-18T22:05:21Z) - Fixing the NTK: From Neural Network Linearizations to Exact Convex
Programs [63.768739279562105]
We show that for a particular choice of mask weights that do not depend on the learning targets, this kernel is equivalent to the NTK of the gated ReLU network on the training data.
A consequence of this lack of dependence on the targets is that the NTK cannot perform better than the optimal MKL kernel on the training set.
arXiv Detail & Related papers (2023-09-26T17:42:52Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Robustness Certificates for Implicit Neural Networks: A Mixed Monotone
Contractive Approach [60.67748036747221]
Implicit neural networks offer competitive performance and reduced memory consumption.
They can remain brittle with respect to input adversarial perturbations.
This paper proposes a theoretical and computational framework for robustness verification of implicit neural networks.
arXiv Detail & Related papers (2021-12-10T03:08:55Z) - Critical Initialization of Wide and Deep Neural Networks through Partial
Jacobians: General Theory and Applications [6.579523168465526]
We introduce emphpartial Jacobians of a network, defined as derivatives of preactivations in layer $l$ with respect to preactivations in layer $l_0leq l$.
We derive recurrence relations for the norms of partial Jacobians and utilize these relations to analyze criticality of deep fully connected neural networks with LayerNorm and/or residual connections.
arXiv Detail & Related papers (2021-11-23T20:31:42Z) - A quantum algorithm for training wide and deep classical neural networks [72.2614468437919]
We show that conditions amenable to classical trainability via gradient descent coincide with those necessary for efficiently solving quantum linear systems.
We numerically demonstrate that the MNIST image dataset satisfies such conditions.
We provide empirical evidence for $O(log n)$ training of a convolutional neural network with pooling.
arXiv Detail & Related papers (2021-07-19T23:41:03Z) - Certifying Incremental Quadratic Constraints for Neural Networks via
Convex Optimization [2.388501293246858]
We propose a convex program to certify incremental quadratic constraints on the map of neural networks over a region of interest.
certificates can capture several useful properties such as (local) Lipschitz continuity, one-sided Lipschitz continuity, invertibility, and contraction.
arXiv Detail & Related papers (2020-12-10T21:15:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.