Unsupervised Learning of slow features for Data Efficient Regression
- URL: http://arxiv.org/abs/2012.06279v1
- Date: Fri, 11 Dec 2020 12:19:45 GMT
- Title: Unsupervised Learning of slow features for Data Efficient Regression
- Authors: Oliver Struckmeier, Kshitij Tiwari, Ville Kyrki
- Abstract summary: We propose the slow variational autoencoder (S-VAE), an extension to the $beta$-VAE which applies a temporal similarity constraint to the latent representations.
We evaluate the three methods against their data-efficiency on down-stream tasks using a synthetic 2D ball tracking dataset, a dataset from a reinforcent learning environment and a dataset generated using the DeepMind Lab environment.
- Score: 15.73372211126635
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Research in computational neuroscience suggests that the human brain's
unparalleled data efficiency is a result of highly efficient mechanisms to
extract and organize slowly changing high level features from continuous
sensory inputs. In this paper, we apply this slowness principle to a state of
the art representation learning method with the goal of performing data
efficient learning of down-stream regression tasks. To this end, we propose the
slow variational autoencoder (S-VAE), an extension to the $\beta$-VAE which
applies a temporal similarity constraint to the latent representations. We
empirically compare our method to the $\beta$-VAE and the Temporal Difference
VAE (TD-VAE), a state-of-the-art method for next frame prediction in latent
space with temporal abstraction. We evaluate the three methods against their
data-efficiency on down-stream tasks using a synthetic 2D ball tracking
dataset, a dataset from a reinforcent learning environment and a dataset
generated using the DeepMind Lab environment. In all tasks, the proposed method
outperformed the baselines both with dense and especially sparse labeled data.
The S-VAE achieved similar or better performance compared to the baselines with
$20\%$ to $93\%$ less data.
Related papers
- fastHDMI: Fast Mutual Information Estimation for High-Dimensional Data [2.9901605297536027]
We introduce fastHDMI, a Python package designed for efficient variable screening in high-dimensional datasets.
This work pioneers the application of three mutual information estimation methods for neuroimaging variable selection.
arXiv Detail & Related papers (2024-10-14T01:49:53Z) - Data Augmentation for Sparse Multidimensional Learning Performance Data Using Generative AI [17.242331892899543]
Learning performance data describe correct and incorrect answers or problem-solving attempts in adaptive learning.
Learning performance data tend to be highly sparse (80%(sim)90% missing observations) in most real-world applications due to adaptive item selection.
This article proposes a systematic framework for augmenting learner data to address data sparsity in learning performance data.
arXiv Detail & Related papers (2024-09-24T00:25:07Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - ContrasInver: Ultra-Sparse Label Semi-supervised Regression for
Multi-dimensional Seismic Inversion [7.356328937024184]
ContrasInver is a method that achieves seismic inversion using as few as two or three well logs.
In experiments, ContrasInver achieved state-of-the-art performance in the synthetic data SEAM I.
It's the first data-driven approach yielding reliable results on the Netherlands F3 and Delft, using only three and two well logs respectively.
arXiv Detail & Related papers (2023-02-13T15:19:51Z) - Towards Robust Dataset Learning [90.2590325441068]
We propose a principled, tri-level optimization to formulate the robust dataset learning problem.
Under an abstraction model that characterizes robust vs. non-robust features, the proposed method provably learns a robust dataset.
arXiv Detail & Related papers (2022-11-19T17:06:10Z) - Open-Set Semi-Supervised Learning for 3D Point Cloud Understanding [62.17020485045456]
It is commonly assumed in semi-supervised learning (SSL) that the unlabeled data are drawn from the same distribution as that of the labeled ones.
We propose to selectively utilize unlabeled data through sample weighting, so that only conducive unlabeled data would be prioritized.
arXiv Detail & Related papers (2022-05-02T16:09:17Z) - DEALIO: Data-Efficient Adversarial Learning for Imitation from
Observation [57.358212277226315]
In imitation learning from observation IfO, a learning agent seeks to imitate a demonstrating agent using only observations of the demonstrated behavior without access to the control signals generated by the demonstrator.
Recent methods based on adversarial imitation learning have led to state-of-the-art performance on IfO problems, but they typically suffer from high sample complexity due to a reliance on data-inefficient, model-free reinforcement learning algorithms.
This issue makes them impractical to deploy in real-world settings, where gathering samples can incur high costs in terms of time, energy, and risk.
We propose a more data-efficient IfO algorithm
arXiv Detail & Related papers (2021-03-31T23:46:32Z) - AutoSimulate: (Quickly) Learning Synthetic Data Generation [70.82315853981838]
We propose an efficient alternative for optimal synthetic data generation based on a novel differentiable approximation of the objective.
We demonstrate that the proposed method finds the optimal data distribution faster (up to $50times$), with significantly reduced training data generation (up to $30times$) and better accuracy ($+8.7%$) on real-world test datasets than previous methods.
arXiv Detail & Related papers (2020-08-16T11:36:11Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.