Separating common from salient patterns with Contrastive Representation
Learning
- URL: http://arxiv.org/abs/2402.11928v1
- Date: Mon, 19 Feb 2024 08:17:13 GMT
- Title: Separating common from salient patterns with Contrastive Representation
Learning
- Authors: Robin Louiset, Edouard Duchesnay, Antoine Grigis, Pietro Gori
- Abstract summary: Contrastive Analysis aims at separating common factors of variation between two datasets.
Current models based on Variational Auto-Encoders have shown poor performance in learning semantically-expressive representations.
We propose to leverage the ability of Contrastive Learning to learn semantically expressive representations well adapted for Contrastive Analysis.
- Score: 2.250968907999846
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Contrastive Analysis is a sub-field of Representation Learning that aims at
separating common factors of variation between two datasets, a background
(i.e., healthy subjects) and a target (i.e., diseased subjects), from the
salient factors of variation, only present in the target dataset. Despite their
relevance, current models based on Variational Auto-Encoders have shown poor
performance in learning semantically-expressive representations. On the other
hand, Contrastive Representation Learning has shown tremendous performance
leaps in various applications (classification, clustering, etc.). In this work,
we propose to leverage the ability of Contrastive Learning to learn
semantically expressive representations well adapted for Contrastive Analysis.
We reformulate it under the lens of the InfoMax Principle and identify two
Mutual Information terms to maximize and one to minimize. We decompose the
first two terms into an Alignment and a Uniformity term, as commonly done in
Contrastive Learning. Then, we motivate a novel Mutual Information minimization
strategy to prevent information leakage between common and salient
distributions. We validate our method, called SepCLR, on three visual datasets
and three medical datasets, specifically conceived to assess the pattern
separation capability in Contrastive Analysis. Code available at
https://github.com/neurospin-projects/2024_rlouiset_sep_clr.
Related papers
- A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - Visual Commonsense based Heterogeneous Graph Contrastive Learning [79.22206720896664]
We propose a heterogeneous graph contrastive learning method to better finish the visual reasoning task.
Our method is designed as a plug-and-play way, so that it can be quickly and easily combined with a wide range of representative methods.
arXiv Detail & Related papers (2023-11-11T12:01:18Z) - SepVAE: a contrastive VAE to separate pathological patterns from healthy ones [2.619659560375341]
Contrastive Analysis VAE (CA-VAEs) is a family of Variational auto-encoders (VAEs) that aims at separating the common factors of variation between a background dataset (BG) and a target dataset (TG)
We show a better performance than previous CA-VAEs methods on three medical applications and a natural images dataset (CelebA)
arXiv Detail & Related papers (2023-07-12T14:52:21Z) - Mine yOur owN Anatomy: Revisiting Medical Image Segmentation with Extremely Limited Labels [54.58539616385138]
We introduce a novel semi-supervised 2D medical image segmentation framework termed Mine yOur owN Anatomy (MONA)
First, prior work argues that every pixel equally matters to the model training; we observe empirically that this alone is unlikely to define meaningful anatomical features.
Second, we construct a set of objectives that encourage the model to be capable of decomposing medical images into a collection of anatomical features.
arXiv Detail & Related papers (2022-09-27T15:50:31Z) - DUET: Cross-modal Semantic Grounding for Contrastive Zero-shot Learning [37.48292304239107]
We present a transformer-based end-to-end ZSL method named DUET.
We develop a cross-modal semantic grounding network to investigate the model's capability of disentangling semantic attributes from the images.
We find that DUET can often achieve state-of-the-art performance, its components are effective and its predictions are interpretable.
arXiv Detail & Related papers (2022-07-04T11:12:12Z) - Self-Supervised Visual Representation Learning with Semantic Grouping [50.14703605659837]
We tackle the problem of learning visual representations from unlabeled scene-centric data.
We propose contrastive learning from data-driven semantic slots, namely SlotCon, for joint semantic grouping and representation learning.
arXiv Detail & Related papers (2022-05-30T17:50:59Z) - Dense Contrastive Visual-Linguistic Pretraining [53.61233531733243]
Several multimodal representation learning approaches have been proposed that jointly represent image and text.
These approaches achieve superior performance by capturing high-level semantic information from large-scale multimodal pretraining.
We propose unbiased Dense Contrastive Visual-Linguistic Pretraining to replace the region regression and classification with cross-modality region contrastive learning.
arXiv Detail & Related papers (2021-09-24T07:20:13Z) - Lookahead Adversarial Learning for Near Real-Time Semantic Segmentation [2.538209532048867]
We build a conditional adversarial network with a state-of-the-art segmentation model (DeepLabv3+) at its core.
We focus on semantic segmentation models that run fast at inference for near real-time field applications.
arXiv Detail & Related papers (2020-06-19T17:04:38Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z) - Learning to Compare Relation: Semantic Alignment for Few-Shot Learning [48.463122399494175]
We present a novel semantic alignment model to compare relations, which is robust to content misalignment.
We conduct extensive experiments on several few-shot learning datasets.
arXiv Detail & Related papers (2020-02-29T08:37:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.