Fixed Points of Cone Mapping with the Application to Neural Networks
- URL: http://arxiv.org/abs/2207.09947v1
- Date: Wed, 20 Jul 2022 14:43:45 GMT
- Title: Fixed Points of Cone Mapping with the Application to Neural Networks
- Authors: Grzegorz Gabor and Krzysztof Rykaczewski
- Abstract summary: We derive conditions for the existence of fixed points of cone mappings without assuming scalability of functions.
In the case of specific non-negative data, it cannot be said that if the mapping is non-negative, it has only non-negative weights.
- Score: 1.0660480034605242
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We derive conditions for the existence of fixed points of cone mappings
without assuming scalability of functions. Monotonicity and scalability are
often inseparable in the literature in the context of searching for fixed
points of interference mappings. In applications, such mappings are
approximated by non-negative neural networks. It turns out, however, that the
process of training non-negative networks requires imposing an artificial
constraint on the weights of the model. However, in the case of specific
non-negative data, it cannot be said that if the mapping is non-negative, it
has only non-negative weights. Therefore, we considered the problem of the
existence of fixed points for general neural networks, assuming the conditions
of tangency conditions with respect to specific cones. This does not relax the
physical assumptions, because even assuming that the input and output are to be
non-negative, the weights can have (small, but) less than zero values. Such
properties (often found in papers on the interpretability of weights of neural
networks) lead to the weakening of the assumptions about the monotonicity or
scalability of the mapping associated with the neural network. To the best of
our knowledge, this paper is the first to study this phenomenon.
Related papers
- Approximation with Random Shallow ReLU Networks with Applications to Model Reference Adaptive Control [0.0]
We show that ReLU networks with randomly generated weights and biases achieve $L_infty$ error of $O(m-1/2)$ with high probability.
We show how the result can be used to get approximations of required accuracy in a model reference adaptive control application.
arXiv Detail & Related papers (2024-03-25T19:39:17Z) - Semantic Strengthening of Neuro-Symbolic Learning [85.6195120593625]
Neuro-symbolic approaches typically resort to fuzzy approximations of a probabilistic objective.
We show how to compute this efficiently for tractable circuits.
We test our approach on three tasks: predicting a minimum-cost path in Warcraft, predicting a minimum-cost perfect matching, and solving Sudoku puzzles.
arXiv Detail & Related papers (2023-02-28T00:04:22Z) - Consistency of Neural Networks with Regularization [0.0]
This paper proposes the general framework of neural networks with regularization and prove its consistency.
Two types of activation functions: hyperbolic function(Tanh) and rectified linear unit(ReLU) have been taken into consideration.
arXiv Detail & Related papers (2022-06-22T23:33:39Z) - Benign Overfitting in Two-layer Convolutional Neural Networks [90.75603889605043]
We study the benign overfitting phenomenon in training a two-layer convolutional neural network (CNN)
We show that when the signal-to-noise ratio satisfies a certain condition, a two-layer CNN trained by gradient descent can achieve arbitrarily small training and test loss.
On the other hand, when this condition does not hold, overfitting becomes harmful and the obtained CNN can only achieve constant level test loss.
arXiv Detail & Related papers (2022-02-14T07:45:51Z) - Non-Vacuous Generalisation Bounds for Shallow Neural Networks [5.799808780731661]
We focus on a specific class of shallow neural networks with a single hidden layer.
We derive new generalisation bounds through the PAC-Bayesian theory.
Our bounds are empirically non-vacuous when the network is trained with vanilla gradient descent on MNIST and Fashion-MNIST.
arXiv Detail & Related papers (2022-02-03T14:59:51Z) - On Convergence of Training Loss Without Reaching Stationary Points [62.41370821014218]
We show that Neural Network weight variables do not converge to stationary points where the gradient the loss function vanishes.
We propose a new perspective based on ergodic theory dynamical systems.
arXiv Detail & Related papers (2021-10-12T18:12:23Z) - Nonperturbative renormalization for the neural network-QFT
correspondence [0.0]
We study the concepts of locality and power-counting in this context.
We provide an analysis in terms of the nonperturbative renormalization group using the Wetterich-Morris equation.
Our aim is to provide a useful formalism to investigate neural networks behavior beyond the large-width limit.
arXiv Detail & Related papers (2021-08-03T10:36:04Z) - Fixed points of nonnegative neural networks [6.7113569772720565]
We first show that nonnegative neural networks with nonnegative weights and biases can be recognized as monotonic and (weakly) scalable mappings.
We prove that the shape of the fixed point set of nonnegative neural networks with nonnegative weights and biases is an interval, which under mild conditions degenerates to a point.
arXiv Detail & Related papers (2021-06-30T17:49:55Z) - Robust Implicit Networks via Non-Euclidean Contractions [63.91638306025768]
Implicit neural networks show improved accuracy and significant reduction in memory consumption.
They can suffer from ill-posedness and convergence instability.
This paper provides a new framework to design well-posed and robust implicit neural networks.
arXiv Detail & Related papers (2021-06-06T18:05:02Z) - Non-Singular Adversarial Robustness of Neural Networks [58.731070632586594]
Adrial robustness has become an emerging challenge for neural network owing to its over-sensitivity to small input perturbations.
We formalize the notion of non-singular adversarial robustness for neural networks through the lens of joint perturbations to data inputs as well as model weights.
arXiv Detail & Related papers (2021-02-23T20:59:30Z) - Implicit Regularization in ReLU Networks with the Square Loss [56.70360094597169]
We show that it is impossible to characterize the implicit regularization with the square loss by any explicit function of the model parameters.
Our results suggest that a more general framework may be needed to understand implicit regularization for nonlinear predictors.
arXiv Detail & Related papers (2020-12-09T16:48:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.