An Improved Semi-Supervised VAE for Learning Disentangled
Representations
- URL: http://arxiv.org/abs/2006.07460v2
- Date: Mon, 22 Jun 2020 18:20:10 GMT
- Title: An Improved Semi-Supervised VAE for Learning Disentangled
Representations
- Authors: Weili Nie, Zichao Wang, Ankit B. Patel, Richard G. Baraniuk
- Abstract summary: We introduce another source of supervision that we denote as label replacement.
During training, we replace the inferred representation associated with a data point with its ground-truth representation whenever it is available.
Our extension is theoretically inspired by our proposed general framework of semi-supervised disentanglement learning.
- Score: 29.38345769998613
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning interpretable and disentangled representations is a crucial yet
challenging task in representation learning. In this work, we focus on
semi-supervised disentanglement learning and extend work by Locatello et al.
(2019) by introducing another source of supervision that we denote as label
replacement. Specifically, during training, we replace the inferred
representation associated with a data point with its ground-truth
representation whenever it is available. Our extension is theoretically
inspired by our proposed general framework of semi-supervised disentanglement
learning in the context of VAEs which naturally motivates the supervised terms
commonly used in existing semi-supervised VAEs (but not for disentanglement
learning). Extensive experiments on synthetic and real datasets demonstrate
both quantitatively and qualitatively the ability of our extension to
significantly and consistently improve disentanglement with very limited
supervision.
Related papers
- Zero-Shot Generalization during Instruction Tuning: Insights from Similarity and Granularity [84.12126298229866]
We show that zero-shot generalization during instruction tuning happens very early.
We also show that encountering highly similar and fine-grained training data earlier during instruction tuning, without the constraints of defined "tasks", enables better generalization.
For the first time, we show that zero-shot generalization during instruction tuning is a form of similarity-based generalization between training and test data at the instance level.
arXiv Detail & Related papers (2024-06-17T16:40:21Z) - Learning Interpretable Policies in Hindsight-Observable POMDPs through
Partially Supervised Reinforcement Learning [57.67629402360924]
We introduce the Partially Supervised Reinforcement Learning (PSRL) framework.
At the heart of PSRL is the fusion of both supervised and unsupervised learning.
We show that PSRL offers a potent balance, enhancing model interpretability while preserving, and often significantly outperforming, the performance benchmarks set by traditional methods.
arXiv Detail & Related papers (2024-02-14T16:23:23Z) - Harnessing Discrete Representations For Continual Reinforcement Learning [8.61539229796467]
We investigate the advantages of representing observations as vectors of categorical values within the context of reinforcement learning.
We find that, when compared to traditional continuous representations, world models learned over discrete representations accurately model more of the world with less capacity.
arXiv Detail & Related papers (2023-12-02T18:55:26Z) - Hierarchical Decomposition of Prompt-Based Continual Learning:
Rethinking Obscured Sub-optimality [55.88910947643436]
Self-supervised pre-training is essential for handling vast quantities of unlabeled data in practice.
HiDe-Prompt is an innovative approach that explicitly optimize the hierarchical components with an ensemble of task-specific prompts and statistics.
Our experiments demonstrate the superior performance of HiDe-Prompt and its robustness to pre-training paradigms in continual learning.
arXiv Detail & Related papers (2023-10-11T06:51:46Z) - Generalization Across Observation Shifts in Reinforcement Learning [13.136140831757189]
We extend the bisimulation framework to account for context dependent observation shifts.
Specifically, we focus on the simulator based learning setting and use alternate observations to learn a representation space.
This allows us to deploy the agent to varying observation settings during test time and generalize to unseen scenarios.
arXiv Detail & Related papers (2023-06-07T16:49:03Z) - Fair Representation Learning using Interpolation Enabled Disentanglement [9.043741281011304]
We propose a novel method to address two key issues: (a) Can we simultaneously learn fair disentangled representations while ensuring the utility of the learned representation for downstream tasks, and (b)Can we provide theoretical insights into when the proposed approach will be both fair and accurate.
To address the former, we propose the method FRIED, Fair Representation learning using Interpolation Enabled Disentanglement.
arXiv Detail & Related papers (2021-07-31T17:32:12Z) - Weakly Supervised Temporal Adjacent Network for Language Grounding [96.09453060585497]
We introduce a novel weakly supervised temporal adjacent network (WSTAN) for temporal language grounding.
WSTAN learns cross-modal semantic alignment by exploiting temporal adjacent network in a multiple instance learning (MIL) paradigm.
An additional self-discriminating loss is devised on both the MIL branch and the complementary branch, aiming to enhance semantic discrimination by self-supervising.
arXiv Detail & Related papers (2021-06-30T15:42:08Z) - Co$^2$L: Contrastive Continual Learning [69.46643497220586]
Recent breakthroughs in self-supervised learning show that such algorithms learn visual representations that can be transferred better to unseen tasks.
We propose a rehearsal-based continual learning algorithm that focuses on continually learning and maintaining transferable representations.
arXiv Detail & Related papers (2021-06-28T06:14:38Z) - Reinforcement Learning with Prototypical Representations [114.35801511501639]
Proto-RL is a self-supervised framework that ties representation learning with exploration through prototypical representations.
These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations.
This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks.
arXiv Detail & Related papers (2021-02-22T18:56:34Z) - Odd-One-Out Representation Learning [1.6822770693792826]
We show that a weakly-supervised downstream task based on odd-one-out observations is suitable for model selection.
We also show that a bespoke metric-learning VAE model which performs highly on this task also out-performs other standard unsupervised and a weakly-supervised disentanglement model.
arXiv Detail & Related papers (2020-12-14T22:01:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.