Tight Certified Robustness via Min-Max Representations of ReLU Neural
Networks
- URL: http://arxiv.org/abs/2310.04916v1
- Date: Sat, 7 Oct 2023 21:07:45 GMT
- Title: Tight Certified Robustness via Min-Max Representations of ReLU Neural
Networks
- Authors: Brendon G. Anderson, Samuel Pfrommer, Somayeh Sojoudi
- Abstract summary: The reliable deployment of neural networks in control systems requires rigorous robustness guarantees.
In this paper, we obtain tight robustness certificates over convex representations of ReLU neural networks.
- Score: 9.771011198361865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The reliable deployment of neural networks in control systems requires
rigorous robustness guarantees. In this paper, we obtain tight robustness
certificates over convex attack sets for min-max representations of ReLU neural
networks by developing a convex reformulation of the nonconvex certification
problem. This is done by "lifting" the problem to an infinite-dimensional
optimization over probability measures, leveraging recent results in
distributionally robust optimization to solve for an optimal discrete
distribution, and proving that solutions of the original nonconvex problem are
generated by the discrete distribution under mild boundedness, nonredundancy,
and Slater conditions. As a consequence, optimal (worst-case) attacks against
the model may be solved for exactly. This contrasts prior state-of-the-art that
either requires expensive branch-and-bound schemes or loose relaxation
techniques. Experiments on robust control and MNIST image classification
examples highlight the benefits of our approach.
Related papers
- Robust Stochastically-Descending Unrolled Networks [85.6993263983062]
Deep unrolling is an emerging learning-to-optimize method that unrolls a truncated iterative algorithm in the layers of a trainable neural network.
We show that convergence guarantees and generalizability of the unrolled networks are still open theoretical problems.
We numerically assess unrolled architectures trained under the proposed constraints in two different applications.
arXiv Detail & Related papers (2023-12-25T18:51:23Z) - The Convex Landscape of Neural Networks: Characterizing Global Optima
and Stationary Points via Lasso Models [75.33431791218302]
Deep Neural Network Network (DNN) models are used for programming purposes.
In this paper we examine the use of convex neural recovery models.
We show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
We also show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
arXiv Detail & Related papers (2023-12-19T23:04:56Z) - Optimization Guarantees of Unfolded ISTA and ADMM Networks With Smooth
Soft-Thresholding [57.71603937699949]
We study optimization guarantees, i.e., achieving near-zero training loss with the increase in the number of learning epochs.
We show that the threshold on the number of training samples increases with the increase in the network width.
arXiv Detail & Related papers (2023-09-12T13:03:47Z) - Robust Bayesian Satisficing [8.65552688277074]
We propose a novel robust satisficing algorithm called RoBOS for noisy black-box optimization.
Our algorithm guarantees sublinear lenient regret under certain assumptions on the amount of distribution shift.
In addition, we define a weaker notion of regret called robust satisficing regret, in which our algorithm achieves a sublinear upper bound independent of the amount of distribution shift.
arXiv Detail & Related papers (2023-08-16T11:31:18Z) - Conditional Denoising Diffusion for Sequential Recommendation [62.127862728308045]
Two prominent generative models, Generative Adversarial Networks (GANs) and Variational AutoEncoders (VAEs)
GANs suffer from unstable optimization, while VAEs are prone to posterior collapse and over-smoothed generations.
We present a conditional denoising diffusion model, which includes a sequence encoder, a cross-attentive denoising decoder, and a step-wise diffuser.
arXiv Detail & Related papers (2023-04-22T15:32:59Z) - Trust your neighbours: Penalty-based constraints for model calibration [19.437451462590108]
We present a constrained optimization perspective of SVLS and demonstrate that it enforces an implicit constraint on soft class proportions of surrounding pixels.
We propose a principled and simple solution based on equality constraints on the logit values, which enables to control explicitly both the enforced constraint and the weight of the penalty.
arXiv Detail & Related papers (2023-03-11T01:10:26Z) - Robust lEarned Shrinkage-Thresholding (REST): Robust unrolling for
sparse recover [87.28082715343896]
We consider deep neural networks for solving inverse problems that are robust to forward model mis-specifications.
We design a new robust deep neural network architecture by applying algorithm unfolding techniques to a robust version of the underlying recovery problem.
The proposed REST network is shown to outperform state-of-the-art model-based and data-driven algorithms in both compressive sensing and radar imaging problems.
arXiv Detail & Related papers (2021-10-20T06:15:45Z) - Proxy Convexity: A Unified Framework for the Analysis of Neural Networks
Trained by Gradient Descent [95.94432031144716]
We propose a unified non- optimization framework for the analysis of a learning network.
We show that existing guarantees can be trained unified through gradient descent.
arXiv Detail & Related papers (2021-06-25T17:45:00Z) - A Sequential Framework Towards an Exact SDP Verification of Neural
Networks [14.191310794366075]
A number of techniques based on convex optimization have been proposed in the literature to study the robustness of neural networks.
The challenge to a robust certification approach is that it is prone to a large relaxation gap.
In this work, we address the issue by developing a sequential programming framework to shrink this gap to zero.
arXiv Detail & Related papers (2020-10-16T19:45:11Z) - Online and Distribution-Free Robustness: Regression and Contextual
Bandits with Huber Contamination [29.85468294601847]
We revisit two classic high-dimensional online learning problems, namely linear regression and contextual bandits.
We show that our algorithms succeed where conventional methods fail.
arXiv Detail & Related papers (2020-10-08T17:59:05Z) - PEREGRiNN: Penalized-Relaxation Greedy Neural Network Verifier [1.1011268090482575]
We introduce a new approach to formally verify the most commonly considered safety specifications for ReLU NNs.
We use a convex solver not only as a linear feasibility checker, but also as a means of penalizing the amount of relaxation allowed in solutions.
arXiv Detail & Related papers (2020-06-18T21:33:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.