Monotone operator equilibrium networks
- URL: http://arxiv.org/abs/2006.08591v2
- Date: Mon, 3 May 2021 22:39:35 GMT
- Title: Monotone operator equilibrium networks
- Authors: Ezra Winston, J. Zico Kolter
- Abstract summary: We develop a new class of implicit-depth model based on the theory of monotone operators, the Monotone Operator Equilibrium Network (monDEQ)
We show the close connection between finding the equilibrium point of an implicit network and solving a form of monotone operator splitting problem.
We then develop a parameterization of the network which ensures that all operators remain monotone, which guarantees the existence of a unique equilibrium point.
- Score: 97.86610752856987
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Implicit-depth models such as Deep Equilibrium Networks have recently been
shown to match or exceed the performance of traditional deep networks while
being much more memory efficient. However, these models suffer from unstable
convergence to a solution and lack guarantees that a solution exists. On the
other hand, Neural ODEs, another class of implicit-depth models, do guarantee
existence of a unique solution but perform poorly compared with traditional
networks. In this paper, we develop a new class of implicit-depth model based
on the theory of monotone operators, the Monotone Operator Equilibrium Network
(monDEQ). We show the close connection between finding the equilibrium point of
an implicit network and solving a form of monotone operator splitting problem,
which admits efficient solvers with guaranteed, stable convergence. We then
develop a parameterization of the network which ensures that all operators
remain monotone, which guarantees the existence of a unique equilibrium point.
Finally, we show how to instantiate several versions of these models, and
implement the resulting iterative solvers, for structured linear operators such
as multi-scale convolutions. The resulting models vastly outperform the Neural
ODE-based models while also being more computationally efficient. Code is
available at http://github.com/locuslab/monotone_op_net.
Related papers
- Vertical Layering of Quantized Neural Networks for Heterogeneous
Inference [57.42762335081385]
We study a new vertical-layered representation of neural network weights for encapsulating all quantized models into a single one.
We can theoretically achieve any precision network for on-demand service while only needing to train and maintain one model.
arXiv Detail & Related papers (2022-12-10T15:57:38Z) - Robustness Certificates for Implicit Neural Networks: A Mixed Monotone
Contractive Approach [60.67748036747221]
Implicit neural networks offer competitive performance and reduced memory consumption.
They can remain brittle with respect to input adversarial perturbations.
This paper proposes a theoretical and computational framework for robustness verification of implicit neural networks.
arXiv Detail & Related papers (2021-12-10T03:08:55Z) - Stabilizing Equilibrium Models by Jacobian Regularization [151.78151873928027]
Deep equilibrium networks (DEQs) are a new class of models that eschews traditional depth in favor of finding the fixed point of a single nonlinear layer.
We propose a regularization scheme for DEQ models that explicitly regularizes the Jacobian of the fixed-point update equations to stabilize the learning of equilibrium models.
We show that this regularization adds only minimal computational cost, significantly stabilizes the fixed-point convergence in both forward and backward passes, and scales well to high-dimensional, realistic domains.
arXiv Detail & Related papers (2021-06-28T00:14:11Z) - Robust Implicit Networks via Non-Euclidean Contractions [63.91638306025768]
Implicit neural networks show improved accuracy and significant reduction in memory consumption.
They can suffer from ill-posedness and convergence instability.
This paper provides a new framework to design well-posed and robust implicit neural networks.
arXiv Detail & Related papers (2021-06-06T18:05:02Z) - Certified Monotonic Neural Networks [15.537695725617576]
We propose to certify the monotonicity of the general piece-wise linear neural networks by solving a mixed integer linear programming problem.
Our approach does not require human-designed constraints on the weight space and also yields more accurate approximation.
arXiv Detail & Related papers (2020-11-20T04:58:13Z) - Consistency of Spectral Clustering on Hierarchical Stochastic Block
Models [5.983753938303726]
We study the hierarchy of communities in real-world networks under a generic block model.
We prove the strong consistency of this method under a wide range of model parameters.
Unlike most of existing work, our theory covers multiscale networks where the connection probabilities may differ by orders of magnitude.
arXiv Detail & Related papers (2020-04-30T01:08:59Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.