Doubly-Stochastic Normalization of the Gaussian Kernel is Robust to
Heteroskedastic Noise
- URL: http://arxiv.org/abs/2006.00402v2
- Date: Mon, 25 Jan 2021 20:58:51 GMT
- Title: Doubly-Stochastic Normalization of the Gaussian Kernel is Robust to
Heteroskedastic Noise
- Authors: Boris Landa, Ronald R.Coifman, Yuval Kluger
- Abstract summary: We show that the doubly-stochastic normalization of the Gaussian kernel with zero main diagonal is robust to heteroskedastic noise.
We provide examples of simulated and experimental single-cell RNA sequence data with intrinsic heteroskedasticity.
- Score: 3.5429774642987915
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A fundamental step in many data-analysis techniques is the construction of an
affinity matrix describing similarities between data points. When the data
points reside in Euclidean space, a widespread approach is to from an affinity
matrix by the Gaussian kernel with pairwise distances, and to follow with a
certain normalization (e.g. the row-stochastic normalization or its symmetric
variant). We demonstrate that the doubly-stochastic normalization of the
Gaussian kernel with zero main diagonal (i.e., no self loops) is robust to
heteroskedastic noise. That is, the doubly-stochastic normalization is
advantageous in that it automatically accounts for observations with different
noise variances. Specifically, we prove that in a suitable high-dimensional
setting where heteroskedastic noise does not concentrate too much in any
particular direction in space, the resulting (doubly-stochastic) noisy affinity
matrix converges to its clean counterpart with rate $m^{-1/2}$, where $m$ is
the ambient dimension. We demonstrate this result numerically, and show that in
contrast, the popular row-stochastic and symmetric normalizations behave
unfavorably under heteroskedastic noise. Furthermore, we provide examples of
simulated and experimental single-cell RNA sequence data with intrinsic
heteroskedasticity, where the advantage of the doubly-stochastic normalization
for exploratory analysis is evident.
Related papers
- A Bayesian Approach Toward Robust Multidimensional Ellipsoid-Specific Fitting [0.0]
This work presents a novel and effective method for fitting multidimensional ellipsoids to scattered data in the contamination of noise and outliers.
We incorporate a uniform prior distribution to constrain the search for primitive parameters within an ellipsoidal domain.
We apply it to a wide range of practical applications such as microscopy cell counting, 3D reconstruction, geometric shape approximation, and magnetometer calibration tasks.
arXiv Detail & Related papers (2024-07-27T14:31:51Z) - High-Dimensional Kernel Methods under Covariate Shift: Data-Dependent Implicit Regularization [83.06112052443233]
This paper studies kernel ridge regression in high dimensions under covariate shifts.
By a bias-variance decomposition, we theoretically demonstrate that the re-weighting strategy allows for decreasing the variance.
For bias, we analyze the regularization of the arbitrary or well-chosen scale, showing that the bias can behave very differently under different regularization scales.
arXiv Detail & Related papers (2024-06-05T12:03:27Z) - Matrix Denoising with Doubly Heteroscedastic Noise: Fundamental Limits and Optimal Spectral Methods [24.06775799553418]
We study the matrix denoising problem of estimating the singular vectors of a rank-$1$ signal corrupted by noise with both column and row correlations.
Our work establishes the information-theoretic and algorithmic limits of matrix denoising with doubly heteroscedastic noise.
arXiv Detail & Related papers (2024-05-22T18:38:10Z) - SNEkhorn: Dimension Reduction with Symmetric Entropic Affinities [14.919246099820548]
Entropic affinities (EAs) are used in the popular Dimensionality Reduction (DR) algorithm t-SNE.
EAs are inherently asymmetric and row-wise, but they are used in DR approaches after undergoing symmetrization methods.
In this work, we uncover a novel characterization of EA as an optimal transport problem, allowing a natural symmetrization that can be computed efficiently.
arXiv Detail & Related papers (2023-05-23T08:08:10Z) - Sharp analysis of EM for learning mixtures of pairwise differences [14.01151780845689]
We consider a symmetric mixture of linear regressions with random samples from the pairwise comparison design.
We establish that the sequence converges linearly, providing an $ell_infty$-norm guarantee on the estimation error of the iterates.
We show that the limit of the EM sequence achieves the sharp rate of estimation in the $ell$-norm, matching the information-theoretically optimal constant.
arXiv Detail & Related papers (2023-02-20T16:13:19Z) - Robust Inference of Manifold Density and Geometry by Doubly Stochastic
Scaling [8.271859911016719]
We develop tools for robust inference under high-dimensional noise.
We show that our approach is robust to variability in technical noise levels across cell types.
arXiv Detail & Related papers (2022-09-16T15:39:11Z) - Nonconvex Stochastic Scaled-Gradient Descent and Generalized Eigenvector
Problems [98.34292831923335]
Motivated by the problem of online correlation analysis, we propose the emphStochastic Scaled-Gradient Descent (SSD) algorithm.
We bring these ideas together in an application to online correlation analysis, deriving for the first time an optimal one-time-scale algorithm with an explicit rate of local convergence to normality.
arXiv Detail & Related papers (2021-12-29T18:46:52Z) - On the Double Descent of Random Features Models Trained with SGD [78.0918823643911]
We study properties of random features (RF) regression in high dimensions optimized by gradient descent (SGD)
We derive precise non-asymptotic error bounds of RF regression under both constant and adaptive step-size SGD setting.
We observe the double descent phenomenon both theoretically and empirically.
arXiv Detail & Related papers (2021-10-13T17:47:39Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - Benign Overfitting of Constant-Stepsize SGD for Linear Regression [122.70478935214128]
inductive biases are central in preventing overfitting empirically.
This work considers this issue in arguably the most basic setting: constant-stepsize SGD for linear regression.
We reflect on a number of notable differences between the algorithmic regularization afforded by (unregularized) SGD in comparison to ordinary least squares.
arXiv Detail & Related papers (2021-03-23T17:15:53Z) - Random extrapolation for primal-dual coordinate descent [61.55967255151027]
We introduce a randomly extrapolated primal-dual coordinate descent method that adapts to sparsity of the data matrix and the favorable structures of the objective function.
We show almost sure convergence of the sequence and optimal sublinear convergence rates for the primal-dual gap and objective values, in the general convex-concave case.
arXiv Detail & Related papers (2020-07-13T17:39:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.