An analysis on the use of autoencoders for representation learning:
fundamentals, learning task case studies, explainability and challenges
- URL: http://arxiv.org/abs/2005.10516v1
- Date: Thu, 21 May 2020 08:41:57 GMT
- Title: An analysis on the use of autoencoders for representation learning:
fundamentals, learning task case studies, explainability and challenges
- Authors: David Charte, Francisco Charte, Mar\'ia J. del Jesus, Francisco
Herrera
- Abstract summary: In many machine learning tasks, learning a good representation of the data can be the key to building a well-performant solution.
We present a series of learning tasks: data embedding for visualization, image denoising, semantic hashing, detection of abnormal behaviors and instance generation.
A solution is proposed for each task employing autoencoders as the only learning method.
- Score: 11.329636084818778
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many machine learning tasks, learning a good representation of the data
can be the key to building a well-performant solution. This is because most
learning algorithms operate with the features in order to find models for the
data. For instance, classification performance can improve if the data is
mapped to a space where classes are easily separated, and regression can be
facilitated by finding a manifold of data in the feature space. As a general
rule, features are transformed by means of statistical methods such as
principal component analysis, or manifold learning techniques such as Isomap or
locally linear embedding. From a plethora of representation learning methods,
one of the most versatile tools is the autoencoder. In this paper we aim to
demonstrate how to influence its learned representations to achieve the desired
learning behavior. To this end, we present a series of learning tasks: data
embedding for visualization, image denoising, semantic hashing, detection of
abnormal behaviors and instance generation. We model them from the
representation learning perspective, following the state of the art
methodologies in each field. A solution is proposed for each task employing
autoencoders as the only learning method. The theoretical developments are put
into practice using a selection of datasets for the different problems and
implementing each solution, followed by a discussion of the results in each
case study and a brief explanation of other six learning applications. We also
explore the current challenges and approaches to explainability in the context
of autoencoders. All of this helps conclude that, thanks to alterations in
their structure as well as their objective function, autoencoders may be the
core of a possible solution to many problems which can be modeled as a
transformation of the feature space.
Related papers
- Speech representation learning: Learning bidirectional encoders with
single-view, multi-view, and multi-task methods [7.1345443932276424]
This thesis focuses on representation learning for sequence data over time or space.
It aims to improve downstream sequence prediction tasks by using the learned representations.
arXiv Detail & Related papers (2023-07-25T20:38:55Z) - A Survey of Learning on Small Data: Generalization, Optimization, and
Challenge [101.27154181792567]
Learning on small data that approximates the generalization ability of big data is one of the ultimate purposes of AI.
This survey follows the active sampling theory under a PAC framework to analyze the generalization error and label complexity of learning on small data.
Multiple data applications that may benefit from efficient small data representation are surveyed.
arXiv Detail & Related papers (2022-07-29T02:34:19Z) - X-Learner: Learning Cross Sources and Tasks for Universal Visual
Representation [71.51719469058666]
We propose a representation learning framework called X-Learner.
X-Learner learns the universal feature of multiple vision tasks supervised by various sources.
X-Learner achieves strong performance on different tasks without extra annotations, modalities and computational costs.
arXiv Detail & Related papers (2022-03-16T17:23:26Z) - Object Pursuit: Building a Space of Objects via Discriminative Weight
Generation [23.85039747700698]
We propose a framework to continuously learn object-centric representations for visual learning and understanding.
We leverage interactions to sample diverse variations of an object and the corresponding training signals while learning the object-centric representations.
We perform an extensive study of the key features of the proposed framework and analyze the characteristics of the learned representations.
arXiv Detail & Related papers (2021-12-15T08:25:30Z) - Few-shot Sequence Learning with Transformers [79.87875859408955]
Few-shot algorithms aim at learning new tasks provided only a handful of training examples.
In this work we investigate few-shot learning in the setting where the data points are sequences of tokens.
We propose an efficient learning algorithm based on Transformers.
arXiv Detail & Related papers (2020-12-17T12:30:38Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z) - On the Robustness of Active Learning [0.7340017786387767]
Active Learning is concerned with how to identify the most useful samples for a Machine Learning algorithm to be trained with.
We find that it is often applied with not enough care and domain knowledge.
We propose the new "Sum of Squared Logits" method based on the Simpson diversity index and investigate the effect of using the confusion matrix for balancing in sample selection.
arXiv Detail & Related papers (2020-06-18T09:07:23Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z) - Inverse Feature Learning: Feature learning based on Representation
Learning of Error [12.777440204911022]
This paper proposes inverse feature learning as a novel supervised feature learning technique that learns a set of high-level features for classification based on an error representation approach.
The proposed method results in significantly better performance compared to the state-of-the-art classification techniques for several popular data sets.
arXiv Detail & Related papers (2020-03-08T00:22:26Z) - Provable Meta-Learning of Linear Representations [114.656572506859]
We provide fast, sample-efficient algorithms to address the dual challenges of learning a common set of features from multiple, related tasks, and transferring this knowledge to new, unseen tasks.
We also provide information-theoretic lower bounds on the sample complexity of learning these linear features.
arXiv Detail & Related papers (2020-02-26T18:21:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.