Unsupervised Learning of Structured Representations via Closed-Loop
Transcription
- URL: http://arxiv.org/abs/2210.16782v1
- Date: Sun, 30 Oct 2022 09:09:05 GMT
- Title: Unsupervised Learning of Structured Representations via Closed-Loop
Transcription
- Authors: Shengbang Tong, Xili Dai, Yubei Chen, Mingyang Li, Zengyi Li, Brent
Yi, Yann LeCun, Yi Ma
- Abstract summary: This paper proposes an unsupervised method for learning a unified representation that serves both discriminative and generative purposes.
We show that a unified representation can enjoy the mutual benefits of having both.
These structured representations enable classification close to state-of-the-art unsupervised discriminative representations.
- Score: 21.78655495464155
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes an unsupervised method for learning a unified
representation that serves both discriminative and generative purposes. While
most existing unsupervised learning approaches focus on a representation for
only one of these two goals, we show that a unified representation can enjoy
the mutual benefits of having both. Such a representation is attainable by
generalizing the recently proposed \textit{closed-loop transcription}
framework, known as CTRL, to the unsupervised setting. This entails solving a
constrained maximin game over a rate reduction objective that expands features
of all samples while compressing features of augmentations of each sample.
Through this process, we see discriminative low-dimensional structures emerge
in the resulting representations. Under comparable experimental conditions and
network complexities, we demonstrate that these structured representations
enable classification performance close to state-of-the-art unsupervised
discriminative representations, and conditionally generated image quality
significantly higher than that of state-of-the-art unsupervised generative
models. Source code can be found at https://github.com/Delay-Xili/uCTRL.
Related papers
- Disentangling Multi-view Representations Beyond Inductive Bias [32.15900989696017]
We propose a novel multi-view representation disentangling method that ensures both interpretability and generalizability of the resulting representations.
Our experiments on four multi-view datasets demonstrate that our proposed method outperforms 12 comparison methods in terms of clustering and classification performance.
arXiv Detail & Related papers (2023-08-03T09:09:28Z) - A Clustering-guided Contrastive Fusion for Multi-view Representation
Learning [7.630965478083513]
We propose a deep fusion network to fuse view-specific representations into the view-common representation.
We also design an asymmetrical contrastive strategy that aligns the view-common representation and each view-specific representation.
In the incomplete view scenario, our proposed method resists noise interference better than those of our competitors.
arXiv Detail & Related papers (2022-12-28T07:21:05Z) - Synergies between Disentanglement and Sparsity: Generalization and
Identifiability in Multi-Task Learning [79.83792914684985]
We prove a new identifiability result that provides conditions under which maximally sparse base-predictors yield disentangled representations.
Motivated by this theoretical result, we propose a practical approach to learn disentangled representations based on a sparsity-promoting bi-level optimization problem.
arXiv Detail & Related papers (2022-11-26T21:02:09Z) - Consistency Regularization for Deep Face Anti-Spoofing [69.70647782777051]
Face anti-spoofing (FAS) plays a crucial role in securing face recognition systems.
Motivated by this exciting observation, we conjecture that encouraging feature consistency of different views may be a promising way to boost FAS models.
We enhance both Embedding-level and Prediction-level Consistency Regularization (EPCR) in FAS.
arXiv Detail & Related papers (2021-11-24T08:03:48Z) - Self-Supervised Learning Disentangled Group Representation as Feature [82.07737719232972]
We show that existing Self-Supervised Learning (SSL) only disentangles simple augmentation features such as rotation and colorization.
We propose an iterative SSL algorithm: Iterative Partition-based Invariant Risk Minimization (IP-IRM)
We prove that IP-IRM converges to a fully disentangled representation and show its effectiveness on various benchmarks.
arXiv Detail & Related papers (2021-10-28T16:12:33Z) - Weakly Supervised Contrastive Learning [68.47096022526927]
We introduce a weakly supervised contrastive learning framework (WCL) to tackle this issue.
WCL achieves 65% and 72% ImageNet Top-1 Accuracy using ResNet50, which is even higher than SimCLRv2 with ResNet101.
arXiv Detail & Related papers (2021-10-10T12:03:52Z) - Neighborhood Contrastive Learning for Novel Class Discovery [79.14767688903028]
We build a new framework, named Neighborhood Contrastive Learning, to learn discriminative representations that are important to clustering performance.
We experimentally demonstrate that these two ingredients significantly contribute to clustering performance and lead our model to outperform state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2021-06-20T17:34:55Z) - Disentangling Semantic-to-visual Confusion for Zero-shot Learning [13.610995960100869]
We develop a novel model called Disentangling Class Representation Generative Adrial Network (DCR-GAN)
Benefiting from the disentangled representations, DCR-GAN could fit a more realistic distribution over both seen and unseen features.
Our proposed model can lead to superior performance to the state-of-the-arts on four benchmark datasets.
arXiv Detail & Related papers (2021-06-16T08:04:11Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Classify and Generate: Using Classification Latent Space Representations
for Image Generations [17.184760662429834]
We propose a discriminative modeling framework that employs manipulated supervised latent representations to reconstruct and generate new samples belonging to a given class.
ReGene has higher classification accuracy than existing conditional generative models while being competitive in terms of FID.
arXiv Detail & Related papers (2020-04-16T09:13:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.