Category-Learning with Context-Augmented Autoencoder
- URL: http://arxiv.org/abs/2010.05007v1
- Date: Sat, 10 Oct 2020 14:04:44 GMT
- Title: Category-Learning with Context-Augmented Autoencoder
- Authors: Denis Kuzminykh, Laida Kushnareva, Timofey Grigoryev, Alexander
Zatolokin
- Abstract summary: Finding an interpretable non-redundant representation of real-world data is one of the key problems in Machine Learning.
We propose a novel method of using data augmentations when training autoencoders.
We train a Variational Autoencoder in such a way, that it makes transformation outcome predictable by auxiliary network.
- Score: 63.05016513788047
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Finding an interpretable non-redundant representation of real-world data is
one of the key problems in Machine Learning. Biological neural networks are
known to solve this problem quite well in unsupervised manner, yet unsupervised
artificial neural networks either struggle to do it or require fine tuning for
each task individually. We associate this with the fact that a biological brain
learns in the context of the relationships between observations, while an
artificial network does not. We also notice that, though a naive data
augmentation technique can be very useful for supervised learning problems,
autoencoders typically fail to generalize transformations from data
augmentations. Thus, we believe that providing additional knowledge about
relationships between data samples will improve model's capability of finding
useful inner data representation. More formally, we consider a dataset not as a
manifold, but as a category, where the examples are objects. Two these objects
are connected by a morphism, if they actually represent different
transformations of the same entity. Following this formalism, we propose a
novel method of using data augmentations when training autoencoders. We train a
Variational Autoencoder in such a way, that it makes transformation outcome
predictable by auxiliary network in terms of the hidden representation. We
believe that the classification accuracy of a linear classifier on the learned
representation is a good metric to measure its interpretability. In our
experiments, present approach outperforms $\beta$-VAE and is comparable with
Gaussian-mixture VAE.
Related papers
- Unsupervised Learning of Invariance Transformations [105.54048699217668]
We develop an algorithmic framework for finding approximate graph automorphisms.
We discuss how this framework can be used to find approximate automorphisms in weighted graphs in general.
arXiv Detail & Related papers (2023-07-24T17:03:28Z) - Mutual Information Learned Classifiers: an Information-theoretic
Viewpoint of Training Deep Learning Classification Systems [9.660129425150926]
Cross entropy loss can easily lead us to find models which demonstrate severe overfitting behavior.
In this paper, we prove that the existing cross entropy loss minimization for training DNN classifiers essentially learns the conditional entropy of the underlying data distribution.
We propose a mutual information learning framework where we train DNN classifiers via learning the mutual information between the label and input.
arXiv Detail & Related papers (2022-10-03T15:09:19Z) - Automatic Data Augmentation via Invariance-Constrained Learning [94.27081585149836]
Underlying data structures are often exploited to improve the solution of learning tasks.
Data augmentation induces these symmetries during training by applying multiple transformations to the input data.
This work tackles these issues by automatically adapting the data augmentation while solving the learning task.
arXiv Detail & Related papers (2022-09-29T18:11:01Z) - Learning sparse features can lead to overfitting in neural networks [9.2104922520782]
We show that feature learning can perform worse than lazy training.
Although sparsity is known to be essential for learning anisotropic data, it is detrimental when the target function is constant or smooth.
arXiv Detail & Related papers (2022-06-24T14:26:33Z) - How do Variational Autoencoders Learn? Insights from Representational
Similarity [2.969705152497174]
We study the internal behaviour of Variational Autoencoders (VAEs) using representational similarity techniques.
Using the CKA and Procrustes similarities, we found that the encoders' representations are learned long before the decoders'.
arXiv Detail & Related papers (2022-05-17T14:31:57Z) - Deep invariant networks with differentiable augmentation layers [87.22033101185201]
Methods for learning data augmentation policies require held-out data and are based on bilevel optimization problems.
We show that our approach is easier and faster to train than modern automatic data augmentation techniques.
arXiv Detail & Related papers (2022-02-04T14:12:31Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Malicious Network Traffic Detection via Deep Learning: An Information
Theoretic View [0.0]
We study how homeomorphism affects learned representation of a malware traffic dataset.
Our results suggest that although the details of learned representations and the specific coordinate system defined over the manifold of all parameters differ slightly, the functional approximations are the same.
arXiv Detail & Related papers (2020-09-16T15:37:44Z) - Laplacian Denoising Autoencoder [114.21219514831343]
We propose to learn data representations with a novel type of denoising autoencoder.
The noisy input data is generated by corrupting latent clean data in the gradient domain.
Experiments on several visual benchmarks demonstrate that better representations can be learned with the proposed approach.
arXiv Detail & Related papers (2020-03-30T16:52:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.