Stability of the scattering transform for deformations with minimal
regularity
- URL: http://arxiv.org/abs/2205.11142v1
- Date: Mon, 23 May 2022 09:08:21 GMT
- Title: Stability of the scattering transform for deformations with minimal
regularity
- Authors: Fabio Nicola and S. Ivan Trapasso
- Abstract summary: We investigate the relationship between the scattering structure and the regularity of the deformation in the H"older regularity scale $Calpha$, $alpha >0$.
We are able to precisely identify the stability threshold, proving that stability is still achievable for deformations of class $Calpha$, $alpha>1$.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Within the mathematical analysis of deep convolutional neural networks, the
wavelet scattering transform introduced by St\'ephane Mallat is a unique
example of how the ideas of multiscale analysis can be combined with a cascade
of modulus nonlinearities to build a nonexpansive, translation invariant signal
representation with provable geometric stability properties, namely Lipschitz
continuity to the action of small $C^2$ diffeomorphisms - a remarkable result
for both theoretical and practical purposes, inherently depending on the choice
of the filters and their arrangement into a hierarchical architecture. In this
note, we further investigate the intimate relationship between the scattering
structure and the regularity of the deformation in the H\"older regularity
scale $C^\alpha$, $\alpha >0$. We are able to precisely identify the stability
threshold, proving that stability is still achievable for deformations of class
$C^{\alpha}$, $\alpha>1$, whereas instability phenomena can occur at lower
regularity levels modelled by $C^\alpha$, $0\le \alpha <1$. While the behaviour
at the threshold given by Lipschitz (or even $C^1$) regularity remains beyond
reach, we are able to prove a stability bound in that case, up to $\varepsilon$
losses.
Related papers
- Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks [54.177130905659155]
Recent studies show that a reproducing kernel Hilbert space (RKHS) is not a suitable space to model functions by neural networks.
In this paper, we study a suitable function space for over- parameterized two-layer neural networks with bounded norms.
arXiv Detail & Related papers (2024-04-29T15:04:07Z) - Breaking the Heavy-Tailed Noise Barrier in Stochastic Optimization Problems [56.86067111855056]
We consider clipped optimization problems with heavy-tailed noise with structured density.
We show that it is possible to get faster rates of convergence than $mathcalO(K-(alpha - 1)/alpha)$, when the gradients have finite moments of order.
We prove that the resulting estimates have negligible bias and controllable variance.
arXiv Detail & Related papers (2023-11-07T17:39:17Z) - A Robustness Analysis of Blind Source Separation [91.3755431537592]
Blind source separation (BSS) aims to recover an unobserved signal from its mixture $X=f(S)$ under the condition that the transformation $f$ is invertible but unknown.
We present a general framework for analysing such violations and quantifying their impact on the blind recovery of $S$ from $X$.
We show that a generic BSS-solution in response to general deviations from its defining structural assumptions can be profitably analysed in the form of explicit continuity guarantees.
arXiv Detail & Related papers (2023-03-17T16:30:51Z) - A PDE-based Explanation of Extreme Numerical Sensitivities and Edge of Stability in Training Neural Networks [12.355137704908042]
We show restrained numerical instabilities in current training practices of deep networks with gradient descent (SGD)
We do this by presenting a theoretical framework using numerical analysis of partial differential equations (PDE), and analyzing the gradient descent PDE of convolutional neural networks (CNNs)
We show this is a consequence of the non-linear PDE associated with the descent of the CNN, whose local linearization changes when over-driving the step size of the discretization resulting in a stabilizing effect.
arXiv Detail & Related papers (2022-06-04T14:54:05Z) - Single Trajectory Nonparametric Learning of Nonlinear Dynamics [8.438421942654292]
Given a single trajectory of a dynamical system, we analyze the performance of the nonparametric least squares estimator (LSE)
We leverage recently developed information-theoretic methods to establish the optimality of the LSE for non hypotheses classes.
We specialize our results to a number of scenarios of practical interest, such as Lipschitz dynamics, generalized linear models, and dynamics described by functions in certain classes of Reproducing Kernel Hilbert Spaces (RKHS)
arXiv Detail & Related papers (2022-02-16T19:38:54Z) - Stability of Neural Networks on Manifolds to Relative Perturbations [118.84154142918214]
Graph Neural Networks (GNNs) show impressive performance in many practical scenarios.
GNNs can scale well on large size graphs, but this is contradicted by the fact that existing stability bounds grow with the number of nodes.
arXiv Detail & Related papers (2021-10-10T04:37:19Z) - On the stability of deep convolutional neural networks under irregular
or random deformations [0.0]
robustness under location deformations for deep convolutional neural networks (DCNNs) is of great theoretical and practical interest.
Here we address this issue for any field $tauin Linfty(mathbbRd;mathbbRd)$, without any additional regularity assumption.
We prove that for signals in multiresolution approximation spaces $U_s$ at scale $s$, stability in $|tau|_Linfty/sll 1$ holds in the regime $|tau|_
arXiv Detail & Related papers (2021-04-24T16:16:30Z) - Quantum Phase Transition of Many Interacting Spins Coupled to a Bosonic
Bath: static and dynamical properties [0.0]
We show that in the Ohmic regime, a Beretzinski-Thouless-Kosterlitz quantum phase transition occurs.
For the observed quantum phase transition we also establish a criterion analogous to that of the metal-insulator transition in solids.
arXiv Detail & Related papers (2021-03-30T10:07:11Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z) - Learning Dynamics Models with Stable Invariant Sets [17.63040340961143]
We propose a method to ensure that a dynamics model has a stable invariant set of general classes.
We compute the projection easily, and at the same time, we can maintain the model's flexibility using various invertible neural networks.
arXiv Detail & Related papers (2020-06-16T05:32:38Z) - Fine-Grained Analysis of Stability and Generalization for Stochastic
Gradient Descent [55.85456985750134]
We introduce a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iterates.
This yields generalization bounds depending on the behavior of the best model, and leads to the first-ever-known fast bounds in the low-noise setting.
To our best knowledge, this gives the firstever-known stability and generalization for SGD with even non-differentiable loss functions.
arXiv Detail & Related papers (2020-06-15T06:30:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.