Deep Nonparametric Estimation of Intrinsic Data Structures by Chart
Autoencoders: Generalization Error and Robustness
- URL: http://arxiv.org/abs/2303.09863v3
- Date: Wed, 25 Oct 2023 09:06:44 GMT
- Title: Deep Nonparametric Estimation of Intrinsic Data Structures by Chart
Autoencoders: Generalization Error and Robustness
- Authors: Hao Liu, Alex Havrilla, Rongjie Lai and Wenjing Liao
- Abstract summary: We employ chart autoencoders to encode data into low-dimensional latent features on a collection of charts.
By training autoencoders, we show that chart autoencoders can effectively denoise the input data with normal noise.
As a special case, our theory also applies to classical autoencoders, as long as the data manifold has a global parametrization.
- Score: 11.441464617936173
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autoencoders have demonstrated remarkable success in learning low-dimensional
latent features of high-dimensional data across various applications. Assuming
that data are sampled near a low-dimensional manifold, we employ chart
autoencoders, which encode data into low-dimensional latent features on a
collection of charts, preserving the topology and geometry of the data
manifold. Our paper establishes statistical guarantees on the generalization
error of chart autoencoders, and we demonstrate their denoising capabilities by
considering $n$ noisy training samples, along with their noise-free
counterparts, on a $d$-dimensional manifold. By training autoencoders, we show
that chart autoencoders can effectively denoise the input data with normal
noise. We prove that, under proper network architectures, chart autoencoders
achieve a squared generalization error in the order of $\displaystyle
n^{-\frac{2}{d+2}}\log^4 n$, which depends on the intrinsic dimension of the
manifold and only weakly depends on the ambient dimension and noise level. We
further extend our theory on data with noise containing both normal and
tangential components, where chart autoencoders still exhibit a denoising
effect for the normal component. As a special case, our theory also applies to
classical autoencoders, as long as the data manifold has a global
parametrization. Our results provide a solid theoretical foundation for the
effectiveness of autoencoders, which is further validated through several
numerical experiments.
Related papers
- Compression of Structured Data with Autoencoders: Provable Benefit of
Nonlinearities and Depth [83.15263499262824]
We prove that gradient descent converges to a solution that completely disregards the sparse structure of the input.
We show how to improve upon Gaussian performance for the compression of sparse data by adding a denoising function to a shallow architecture.
We validate our findings on image datasets, such as CIFAR-10 and MNIST.
arXiv Detail & Related papers (2024-02-07T16:32:29Z) - ADA-GAD: Anomaly-Denoised Autoencoders for Graph Anomaly Detection [84.0718034981805]
We introduce a novel framework called Anomaly-Denoised Autoencoders for Graph Anomaly Detection (ADA-GAD)
In the first stage, we design a learning-free anomaly-denoised augmentation method to generate graphs with reduced anomaly levels.
In the next stage, the decoders are retrained for detection on the original graph.
arXiv Detail & Related papers (2023-12-22T09:02:01Z) - Are We Using Autoencoders in a Wrong Way? [3.110260251019273]
Autoencoders are used for dimensionality reduction, anomaly detection and feature extraction.
We revisited the standard training for the undercomplete Autoencoder modifying the shape of the latent space.
We also explored the behaviour of the latent space in the case of reconstruction of a random sample from the whole dataset.
arXiv Detail & Related papers (2023-09-04T11:22:43Z) - Fundamental Limits of Two-layer Autoencoders, and Achieving Them with
Gradient Methods [91.54785981649228]
This paper focuses on non-linear two-layer autoencoders trained in the challenging proportional regime.
Our results characterize the minimizers of the population risk, and show that such minimizers are achieved by gradient methods.
For the special case of a sign activation function, our analysis establishes the fundamental limits for the lossy compression of Gaussian sources via (shallow) autoencoders.
arXiv Detail & Related papers (2022-12-27T12:37:34Z) - Dimension Reduction for time series with Variational AutoEncoders [2.905751301655124]
We conduct a comparison between wavelet decomposition and convolutional variational autoencoders for dimension reduction.
We show that variational autoencoders are a good option for reducing the dimension of high dimensional data like ECG.
arXiv Detail & Related papers (2022-04-23T12:26:01Z) - COIN++: Data Agnostic Neural Compression [55.27113889737545]
COIN++ is a neural compression framework that seamlessly handles a wide range of data modalities.
We demonstrate the effectiveness of our method by compressing various data modalities.
arXiv Detail & Related papers (2022-01-30T20:12:04Z) - Neural Distributed Source Coding [59.630059301226474]
We present a framework for lossy DSC that is agnostic to the correlation structure and can scale to high dimensions.
We evaluate our method on multiple datasets and show that our method can handle complex correlations and state-of-the-art PSNR.
arXiv Detail & Related papers (2021-06-05T04:50:43Z) - An Introduction to Robust Graph Convolutional Networks [71.68610791161355]
We propose a novel Robust Graph Convolutional Neural Networks for possible erroneous single-view or multi-view data.
By incorporating an extra layers via Autoencoders into traditional graph convolutional networks, we characterize and handle typical error models explicitly.
arXiv Detail & Related papers (2021-03-27T04:47:59Z) - Isometric Autoencoders [36.947436313489746]
We advocate an isometry (i.e., local distance preserving) regularizer.
Our regularizer encourages: (i.e., the decoder to be an isometry; and (ii) the encoder to be the decoder's pseudo-inverse, that is, the encoder extends the inverse of the decoder to the ambient space by projection.
arXiv Detail & Related papers (2020-06-16T16:31:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.