Information-Ordered Bottlenecks for Adaptive Semantic Compression
- URL: http://arxiv.org/abs/2305.11213v1
- Date: Thu, 18 May 2023 18:00:00 GMT
- Title: Information-Ordered Bottlenecks for Adaptive Semantic Compression
- Authors: Matthew Ho, Xiaosheng Zhao, Benjamin Wandelt
- Abstract summary: We present a neural layer designed to adaptively compress data into variables ordered by likelihood.
We show that IOBs achieve near-optimal compression for a given architecture and can assign encoding signals in a manner that is semantically meaningful.
We introduce a novel theory for estimating global dimensionality with IOBs and show that they recover SOTA dimensionality estimates for complex synthetic data.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present the information-ordered bottleneck (IOB), a neural layer designed
to adaptively compress data into latent variables ordered by likelihood
maximization. Without retraining, IOB nodes can be truncated at any bottleneck
width, capturing the most crucial information in the first latent variables.
Unifying several previous approaches, we show that IOBs achieve near-optimal
compression for a given encoding architecture and can assign ordering to latent
signals in a manner that is semantically meaningful. IOBs demonstrate a
remarkable ability to compress embeddings of image and text data, leveraging
the performance of SOTA architectures such as CNNs, transformers, and diffusion
models. Moreover, we introduce a novel theory for estimating global intrinsic
dimensionality with IOBs and show that they recover SOTA dimensionality
estimates for complex synthetic data. Furthermore, we showcase the utility of
these models for exploratory analysis through applications on heterogeneous
datasets, enabling computer-aided discovery of dataset complexity.
Related papers
- Hierarchical Features Matter: A Deep Exploration of GAN Priors for Improved Dataset Distillation [51.44054828384487]
We propose a novel parameterization method dubbed Hierarchical Generative Latent Distillation (H-GLaD)
This method systematically explores hierarchical layers within the generative adversarial networks (GANs)
In addition, we introduce a novel class-relevant feature distance metric to alleviate the computational burden associated with synthetic dataset evaluation.
arXiv Detail & Related papers (2024-06-09T09:15:54Z) - Distribution-Aware Data Expansion with Diffusion Models [55.979857976023695]
We propose DistDiff, a training-free data expansion framework based on the distribution-aware diffusion model.
DistDiff consistently enhances accuracy across a diverse range of datasets compared to models trained solely on original data.
arXiv Detail & Related papers (2024-03-11T14:07:53Z) - On Characterizing the Evolution of Embedding Space of Neural Networks
using Algebraic Topology [9.537910170141467]
We study how the topology of feature embedding space changes as it passes through the layers of a well-trained deep neural network (DNN) through Betti numbers.
We demonstrate that as depth increases, a topologically complicated dataset is transformed into a simple one, resulting in Betti numbers attaining their lowest possible value.
arXiv Detail & Related papers (2023-11-08T10:45:12Z) - Interpretable Spectral Variational AutoEncoder (ISVAE) for time series
clustering [48.0650332513417]
We introduce a novel model that incorporates an interpretable bottleneck-termed the Filter Bank (FB)-at the outset of a Variational Autoencoder (VAE)
This arrangement compels the VAE to attend on the most informative segments of the input signal.
By deliberately constraining the VAE with this FB, we promote the development of an encoding that is discernible, separable, and of reduced dimensionality.
arXiv Detail & Related papers (2023-10-18T13:06:05Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - RENs: Relevance Encoding Networks [0.0]
This paper proposes relevance encoding networks (RENs): a novel probabilistic VAE-based framework that uses the automatic relevance determination (ARD) prior in the latent space to learn the data-specific bottleneck dimensionality.
We show that the proposed model learns the relevant latent bottleneck dimensionality without compromising the representation and generation quality of the samples.
arXiv Detail & Related papers (2022-05-25T21:53:48Z) - Visualizing the Finer Cluster Structure of Large-Scale and
High-Dimensional Data [7.400745342582259]
We propose using a generalized sigmoid function to model the distance similarity in both high- and low-dimensional spaces.
Using both simulated and real-world data sets, we show that our proposed method can generate visualization results comparable to those of uniform manifold approximation and projection.
arXiv Detail & Related papers (2020-07-17T01:36:45Z) - Deep Autoencoding Topic Model with Scalable Hybrid Bayesian Inference [55.35176938713946]
We develop deep autoencoding topic model (DATM) that uses a hierarchy of gamma distributions to construct its multi-stochastic-layer generative network.
We propose a Weibull upward-downward variational encoder that deterministically propagates information upward via a deep neural network, followed by a downward generative model.
The efficacy and scalability of our models are demonstrated on both unsupervised and supervised learning tasks on big corpora.
arXiv Detail & Related papers (2020-06-15T22:22:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.