Convergent autoencoder approximation of low bending and low distortion
manifold embeddings
- URL: http://arxiv.org/abs/2208.10193v2
- Date: Wed, 10 Jan 2024 12:15:26 GMT
- Title: Convergent autoencoder approximation of low bending and low distortion
manifold embeddings
- Authors: Juliane Braunsmann, Marko Rajkovi\'c, Martin Rumpf, Benedikt Wirth
- Abstract summary: We propose and analyze a novel regularization for learning the encoder component of an autoencoder.
The loss functional is computed via Monte Carlo integration with different sampling strategies for pairs of points on the input manifold.
Our main theorem identifies a loss functional of the embedding map as the $Gamma$-limit of the sampling-dependent loss functionals.
- Score: 5.5711773076846365
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autoencoders, which consist of an encoder and a decoder, are widely used in
machine learning for dimension reduction of high-dimensional data. The encoder
embeds the input data manifold into a lower-dimensional latent space, while the
decoder represents the inverse map, providing a parametrization of the data
manifold by the manifold in latent space. A good regularity and structure of
the embedded manifold may substantially simplify further data processing tasks
such as cluster analysis or data interpolation. We propose and analyze a novel
regularization for learning the encoder component of an autoencoder: a loss
functional that prefers isometric, extrinsically flat embeddings and allows to
train the encoder on its own. To perform the training it is assumed that for
pairs of nearby points on the input manifold their local Riemannian distance
and their local Riemannian average can be evaluated. The loss functional is
computed via Monte Carlo integration with different sampling strategies for
pairs of points on the input manifold. Our main theorem identifies a geometric
loss functional of the embedding map as the $\Gamma$-limit of the
sampling-dependent loss functionals. Numerical tests, using image data that
encodes different explicitly given data manifolds, show that smooth manifold
embeddings into latent space are obtained. Due to the promotion of extrinsic
flatness, these embeddings are regular enough such that interpolation between
not too distant points on the manifold is well approximated by linear
interpolation in latent space as one possible postprocessing.
Related papers
- Compression of Structured Data with Autoencoders: Provable Benefit of
Nonlinearities and Depth [83.15263499262824]
We prove that gradient descent converges to a solution that completely disregards the sparse structure of the input.
We show how to improve upon Gaussian performance for the compression of sparse data by adding a denoising function to a shallow architecture.
We validate our findings on image datasets, such as CIFAR-10 and MNIST.
arXiv Detail & Related papers (2024-02-07T16:32:29Z) - Improving embedding of graphs with missing data by soft manifolds [51.425411400683565]
The reliability of graph embeddings depends on how much the geometry of the continuous space matches the graph structure.
We introduce a new class of manifold, named soft manifold, that can solve this situation.
Using soft manifold for graph embedding, we can provide continuous spaces to pursue any task in data analysis over complex datasets.
arXiv Detail & Related papers (2023-11-29T12:48:33Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - Fundamental Limits of Two-layer Autoencoders, and Achieving Them with
Gradient Methods [91.54785981649228]
This paper focuses on non-linear two-layer autoencoders trained in the challenging proportional regime.
Our results characterize the minimizers of the population risk, and show that such minimizers are achieved by gradient methods.
For the special case of a sign activation function, our analysis establishes the fundamental limits for the lossy compression of Gaussian sources via (shallow) autoencoders.
arXiv Detail & Related papers (2022-12-27T12:37:34Z) - Convolutional Filtering on Sampled Manifolds [122.06927400759021]
We show that convolutional filtering on a sampled manifold converges to continuous manifold filtering.
Our findings are further demonstrated empirically on a problem of navigation control.
arXiv Detail & Related papers (2022-11-20T19:09:50Z) - Semi-Supervised Manifold Learning with Complexity Decoupled Chart Autoencoders [45.29194877564103]
This work introduces a chart autoencoder with an asymmetric encoding-decoding process that can incorporate additional semi-supervised information such as class labels.
We discuss the approximation power of such networks and derive a bound that essentially depends on the intrinsic dimension of the data manifold rather than the dimension of ambient space.
arXiv Detail & Related papers (2022-08-22T19:58:03Z) - Learning low bending and low distortion manifold embeddings [1.8046244926068666]
The encoder provides an embedding from the input data manifold into a latent space which may then be used for further processing.
In this article, the embedding into latent space is regularized via a loss function that promotes an as isometric and as flat embedding.
The loss functional is computed via a Monte Carlo integration which is shown to be consistent with a geometric loss functional defined directly on the embedding map.
arXiv Detail & Related papers (2021-04-27T13:51:12Z) - Encoded Prior Sliced Wasserstein AutoEncoder for learning latent
manifold representations [0.7614628596146599]
We introduce an Encoded Prior Sliced Wasserstein AutoEncoder.
An additional prior-encoder network learns an embedding of the data manifold.
We show that the prior encodes the geometry underlying the data unlike conventional autoencoders.
arXiv Detail & Related papers (2020-10-02T14:58:54Z) - Autoencoder Image Interpolation by Shaping the Latent Space [12.482988592988868]
Autoencoders represent an effective approach for computing the underlying factors characterizing datasets of different types.
We propose a regularization technique that shapes the latent representation to follow a manifold consistent with the training images.
arXiv Detail & Related papers (2020-08-04T12:32:54Z) - Manifold Learning via Manifold Deflation [105.7418091051558]
dimensionality reduction methods provide a valuable means to visualize and interpret high-dimensional data.
Many popular methods can fail dramatically, even on simple two-dimensional Manifolds.
This paper presents an embedding method for a novel, incremental tangent space estimator that incorporates global structure as coordinates.
Empirically, we show our algorithm recovers novel and interesting embeddings on real-world and synthetic datasets.
arXiv Detail & Related papers (2020-07-07T10:04:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.