Controlling the Complexity and Lipschitz Constant improves polynomial
nets
- URL: http://arxiv.org/abs/2202.05068v1
- Date: Thu, 10 Feb 2022 14:54:29 GMT
- Title: Controlling the Complexity and Lipschitz Constant improves polynomial
nets
- Authors: Zhenyu Zhu, Fabian Latorre, Grigorios G Chrysos, Volkan Cevher
- Abstract summary: We derive new complexity bounds for the set of Coupled CP-Decomposition (CCP) and Nested Coupled CP-decomposition (NCP) models of Polynomial Nets.
We propose a principled regularization scheme that we evaluate experimentally in six datasets and show that it improves the accuracy as well as the robustness of the models to adversarial perturbations.
- Score: 55.121200972539114
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While the class of Polynomial Nets demonstrates comparable performance to
neural networks (NN), it currently has neither theoretical generalization
characterization nor robustness guarantees. To this end, we derive new
complexity bounds for the set of Coupled CP-Decomposition (CCP) and Nested
Coupled CP-decomposition (NCP) models of Polynomial Nets in terms of the
$\ell_\infty$-operator-norm and the $\ell_2$-operator norm. In addition, we
derive bounds on the Lipschitz constant for both models to establish a
theoretical certificate for their robustness. The theoretical results enable us
to propose a principled regularization scheme that we also evaluate
experimentally in six datasets and show that it improves the accuracy as well
as the robustness of the models to adversarial perturbations. We showcase how
this regularization can be combined with adversarial training, resulting in
further improvements.
Related papers
- Weakly Convex Regularisers for Inverse Problems: Convergence of Critical Points and Primal-Dual Optimisation [12.455342327482223]
We present a generalised formulation of convergent regularisation in terms of critical points.
We show that this is achieved by a class of weakly convex regularisers.
Applying this theory to learned regularisation, we prove universal approximation for input weakly convex neural networks.
arXiv Detail & Related papers (2024-02-01T22:54:45Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Efficient Bound of Lipschitz Constant for Convolutional Layers by Gram
Iteration [122.51142131506639]
We introduce a precise, fast, and differentiable upper bound for the spectral norm of convolutional layers using circulant matrix theory.
We show through a comprehensive set of experiments that our approach outperforms other state-of-the-art methods in terms of precision, computational cost, and scalability.
It proves highly effective for the Lipschitz regularization of convolutional neural networks, with competitive results against concurrent approaches.
arXiv Detail & Related papers (2023-05-25T15:32:21Z) - Certifying Ensembles: A General Certification Theory with
S-Lipschitzness [128.2881318211724]
Ensembling has shown to be beneficial for generalisation, uncertainty estimation, calibration, and mitigating the effects of concept drift.
In this work, we generalise Lipschitz continuity by introducing S-Lipschitz classifiers, which we use to analyse the theoretical robustness of ensembles.
arXiv Detail & Related papers (2023-04-25T17:50:45Z) - Lipschitz Continuity Retained Binary Neural Network [52.17734681659175]
We introduce the Lipschitz continuity as the rigorous criteria to define the model robustness for BNN.
We then propose to retain the Lipschitz continuity as a regularization term to improve the model robustness.
Our experiments prove that our BNN-specific regularization method can effectively strengthen the robustness of BNN.
arXiv Detail & Related papers (2022-07-13T22:55:04Z) - Structural Extensions of Basis Pursuit: Guarantees on Adversarial
Robustness [0.0]
We prove that the stability of BP holds upon the following generalizations.
We introduce classification based on the $ell$ norms of the groups and show numerically that it can be accurate and offers considerable speedups.
arXiv Detail & Related papers (2022-05-05T09:12:07Z) - Sparsest Univariate Learning Models Under Lipschitz Constraint [31.28451181040038]
We propose continuous-domain formulations for one-dimensional regression problems.
We control the Lipschitz constant explicitly using a user-defined upper-bound.
We show that both problems admit global minimizers that are continuous and piecewise-linear.
arXiv Detail & Related papers (2021-12-27T07:03:43Z) - Robust Implicit Networks via Non-Euclidean Contractions [63.91638306025768]
Implicit neural networks show improved accuracy and significant reduction in memory consumption.
They can suffer from ill-posedness and convergence instability.
This paper provides a new framework to design well-posed and robust implicit neural networks.
arXiv Detail & Related papers (2021-06-06T18:05:02Z) - Lipschitz Bounded Equilibrium Networks [3.2872586139884623]
This paper introduces new parameterizations of equilibrium neural networks, i.e. networks defined by implicit equations.
The new parameterization admits a Lipschitz bound during training via unconstrained optimization.
In image classification experiments we show that the Lipschitz bounds are very accurate and improve robustness to adversarial attacks.
arXiv Detail & Related papers (2020-10-05T01:00:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.