Analytic Manifold Learning: Unifying and Evaluating Representations for
Continuous Control
- URL: http://arxiv.org/abs/2006.08718v2
- Date: Tue, 6 Oct 2020 19:43:50 GMT
- Title: Analytic Manifold Learning: Unifying and Evaluating Representations for
Continuous Control
- Authors: Rika Antonova, Maksim Maydanskiy, Danica Kragic, Sam Devlin, Katja
Hofmann
- Abstract summary: We address the problem of learning reusable state representations from streaming high-dimensional observations.
This is important for areas like Reinforcement Learning (RL), which yields non-stationary data distributions during training.
- Score: 32.773203015440075
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address the problem of learning reusable state representations from
streaming high-dimensional observations. This is important for areas like
Reinforcement Learning (RL), which yields non-stationary data distributions
during training. We make two key contributions. First, we propose an evaluation
suite that measures alignment between latent and true low-dimensional states.
We benchmark several widely used unsupervised learning approaches. This
uncovers the strengths and limitations of existing approaches that impose
additional constraints/objectives on the latent space. Our second contribution
is a unifying mathematical formulation for learning latent relations. We learn
analytic relations on source domains, then use these relations to help
structure the latent space when learning on target domains. This formulation
enables a more general, flexible and principled way of shaping the latent
space. It formalizes the notion of learning independent relations, without
imposing restrictive simplifying assumptions or requiring domain-specific
information. We present mathematical properties, concrete algorithms for
implementation and experimental validation of successful learning and transfer
of latent relations.
Related papers
- Zero-shot Model-based Reinforcement Learning using Large Language Models [12.930241182192988]
We investigate how pre-trained Large Language Models can be leveraged to predict in context the dynamics of continuous Markov decision processes.
We present proof-of-concept applications in two reinforcement learning settings: model-based policy evaluation and data-augmented off-policy reinforcement learning.
arXiv Detail & Related papers (2024-10-15T15:46:53Z) - Learning Physics-Consistent Material Behavior Without Prior Knowledge [6.691537914484337]
We introduce a machine learning approach called uLED, which overcomes the limitations by using the convex input neural network (ICNN) as a surrogate model.
We demonstrate that it is robust to a significant level of noise and that it converges to the ground truth with increasing data resolution.
arXiv Detail & Related papers (2024-07-25T08:24:04Z) - Towards Human-Like Machine Comprehension: Few-Shot Relational Learning in Visually-Rich Documents [16.78371134590167]
Key-value relations are prevalent in Visually-Rich Documents (VRDs)
These non-textual cues serve as important indicators that greatly enhance human comprehension and acquisition of such relation triplets.
Our research focuses on few-shot relational learning, specifically targeting the extraction of key-value relation triplets in VRDs.
arXiv Detail & Related papers (2024-03-23T08:40:35Z) - Sample-Efficient Reinforcement Learning in the Presence of Exogenous
Information [77.19830787312743]
In real-world reinforcement learning applications the learner's observation space is ubiquitously high-dimensional with both relevant and irrelevant information about the task at hand.
We introduce a new problem setting for reinforcement learning, the Exogenous Decision Process (ExoMDP), in which the state space admits an (unknown) factorization into a small controllable component and a large irrelevant component.
We provide a new algorithm, ExoRL, which learns a near-optimal policy with sample complexity in the size of the endogenous component.
arXiv Detail & Related papers (2022-06-09T05:19:32Z) - On Generalizing Beyond Domains in Cross-Domain Continual Learning [91.56748415975683]
Deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.
Our proposed approach learns new tasks under domain shift with accuracy boosts up to 10% on challenging datasets such as DomainNet and OfficeHome.
arXiv Detail & Related papers (2022-03-08T09:57:48Z) - Towards Robust Bisimulation Metric Learning [3.42658286826597]
Bisimulation metrics offer one solution to representation learning problem.
We generalize value function approximation bounds for on-policy bisimulation metrics to non-optimal policies.
We find that these issues stem from an underconstrained dynamics model and an unstable dependence of the embedding norm on the reward signal.
arXiv Detail & Related papers (2021-10-27T00:32:07Z) - Which Mutual-Information Representation Learning Objectives are
Sufficient for Control? [80.2534918595143]
Mutual information provides an appealing formalism for learning representations of data.
This paper formalizes the sufficiency of a state representation for learning and representing the optimal policy.
Surprisingly, we find that two of these objectives can yield insufficient representations given mild and common assumptions on the structure of the MDP.
arXiv Detail & Related papers (2021-06-14T10:12:34Z) - Bounding Information Leakage in Machine Learning [26.64770573405079]
This paper investigates fundamental bounds on information leakage.
We identify and bound the success rate of the worst-case membership inference attack.
We derive bounds on the mutual information between the sensitive attributes and model parameters.
arXiv Detail & Related papers (2021-05-09T08:49:14Z) - Fundamental Limits and Tradeoffs in Invariant Representation Learning [99.2368462915979]
Many machine learning applications involve learning representations that achieve two competing goals.
Minimax game-theoretic formulation represents a fundamental tradeoff between accuracy and invariance.
We provide an information-theoretic analysis of this general and important problem under both classification and regression settings.
arXiv Detail & Related papers (2020-12-19T15:24:04Z) - Learning Relation Prototype from Unlabeled Texts for Long-tail Relation
Extraction [84.64435075778988]
We propose a general approach to learn relation prototypes from unlabeled texts.
We learn relation prototypes as an implicit factor between entities.
We conduct experiments on two publicly available datasets: New York Times and Google Distant Supervision.
arXiv Detail & Related papers (2020-11-27T06:21:12Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.