LinSyn: Synthesizing Tight Linear Bounds for Arbitrary Neural Network
Activation Functions
- URL: http://arxiv.org/abs/2201.13351v1
- Date: Mon, 31 Jan 2022 17:00:50 GMT
- Title: LinSyn: Synthesizing Tight Linear Bounds for Arbitrary Neural Network
Activation Functions
- Authors: Brandon Paulsen and Chao Wang
- Abstract summary: LinSyn is the first approach that achieves tight bounds for any arbitrary activation function.
We show that our approach often achieves 2-5X tighter final output bounds and more than quadruple certified robustness.
- Score: 4.03308187036005
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The most scalable approaches to certifying neural network robustness depend
on computing sound linear lower and upper bounds for the network's activation
functions. Current approaches are limited in that the linear bounds must be
handcrafted by an expert, and can be sub-optimal, especially when the network's
architecture composes operations using, for example, multiplication such as in
LSTMs and the recently popular Swish activation. The dependence on an expert
prevents the application of robustness certification to developments in the
state-of-the-art of activation functions, and furthermore the lack of tightness
guarantees may give a false sense of insecurity about a particular model. To
the best of our knowledge, we are the first to consider the problem of
automatically computing tight linear bounds for arbitrary n-dimensional
activation functions. We propose LinSyn, the first approach that achieves tight
bounds for any arbitrary activation function, while only leveraging the
mathematical definition of the activation function itself. Our approach
leverages an efficient heuristic approach to synthesize bounds that are tight
and usually sound, and then verifies the soundness (and adjusts the bounds if
necessary) using the highly optimized branch-and-bound SMT solver, dReal. Even
though our approach depends on an SMT solver, we show that the runtime is
reasonable in practice, and, compared with state of the art, our approach often
achieves 2-5X tighter final output bounds and more than quadruple certified
robustness.
Related papers
- GloptiNets: Scalable Non-Convex Optimization with Certificates [61.50835040805378]
We present a novel approach to non-cube optimization with certificates, which handles smooth functions on the hypercube or on the torus.
By exploiting the regularity of the target function intrinsic in the decay of its spectrum, we allow at the same time to obtain precise certificates and leverage the advanced and powerful neural networks.
arXiv Detail & Related papers (2023-06-26T09:42:59Z) - Certifying Robustness of Convolutional Neural Networks with Tight Linear
Approximation [5.678314425261842]
Ti-Lin is a Tight Linear approximation approach for robustness verification of Conal Neural Networks.
We present a new linear constraints for S-shaped activation functions, which is better than both existing Neuron-wise Tightest and Network-wise Tightest tools.
We evaluate it with 48 different CNNs trained on MNIST, CIFAR-10, and Tiny ImageNet datasets.
arXiv Detail & Related papers (2022-11-13T08:37:13Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Reachability analysis of neural networks using mixed monotonicity [0.0]
We present a new reachability analysis tool to compute an interval over-approximation of the output set of a feedforward neural network under given input uncertainty.
The proposed approach adapts to neural networks an existing mixed-monotonicity method for the reachability analysis of dynamical systems.
arXiv Detail & Related papers (2021-11-15T11:35:18Z) - Training Certifiably Robust Neural Networks with Efficient Local
Lipschitz Bounds [99.23098204458336]
Certified robustness is a desirable property for deep neural networks in safety-critical applications.
We show that our method consistently outperforms state-of-the-art methods on MNIST and TinyNet datasets.
arXiv Detail & Related papers (2021-11-02T06:44:10Z) - Improved Branch and Bound for Neural Network Verification via Lagrangian
Decomposition [161.09660864941603]
We improve the scalability of Branch and Bound (BaB) algorithms for formally proving input-output properties of neural networks.
We present a novel activation-based branching strategy and a BaB framework, named Branch and Dual Network Bound (BaDNB)
BaDNB outperforms previous complete verification systems by a large margin, cutting average verification times by factors up to 50 on adversarial properties.
arXiv Detail & Related papers (2021-04-14T09:22:42Z) - Scaling the Convex Barrier with Sparse Dual Algorithms [141.4085318878354]
We present two novel dual algorithms for tight and efficient neural network bounding.
Both methods recover the strengths of the new relaxation: tightness and a linear separation oracle.
We can obtain better bounds than off-the-shelf solvers in only a fraction of their running time.
arXiv Detail & Related papers (2021-01-14T19:45:17Z) - Lagrangian Decomposition for Neural Network Verification [148.0448557991349]
A fundamental component of neural network verification is the computation of bounds on the values their outputs can take.
We propose a novel approach based on Lagrangian Decomposition.
We show that we obtain bounds comparable with off-the-shelf solvers in a fraction of their running time.
arXiv Detail & Related papers (2020-02-24T17:55:10Z) - Deep Neural Networks with Trainable Activations and Controlled Lipschitz
Constant [26.22495169129119]
We introduce a variational framework to learn the activation functions of deep neural networks.
Our aim is to increase the capacity of the network while controlling an upper-bound of the Lipschitz constant.
We numerically compare our scheme with standard ReLU network and its variations, PReLU and LeakyReLU.
arXiv Detail & Related papers (2020-01-17T12:32:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.