High Fidelity Visualization of What Your Self-Supervised Representation
Knows About
- URL: http://arxiv.org/abs/2112.09164v1
- Date: Thu, 16 Dec 2021 19:23:33 GMT
- Title: High Fidelity Visualization of What Your Self-Supervised Representation
Knows About
- Authors: Florian Bordes, Randall Balestriero, Pascal Vincent
- Abstract summary: In this work, we showcase the use of a conditional diffusion based generative model (RCDM) to visualize representations learned with self-supervised models.
We demonstrate how this model's generation quality is on par with state-of-the-art generative models while being faithful to the representation used as conditioning.
- Score: 22.982471878833362
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Discovering what is learned by neural networks remains a challenge. In
self-supervised learning, classification is the most common task used to
evaluate how good a representation is. However, relying only on such downstream
task can limit our understanding of how much information is retained in the
representation of a given input. In this work, we showcase the use of a
conditional diffusion based generative model (RCDM) to visualize
representations learned with self-supervised models. We further demonstrate how
this model's generation quality is on par with state-of-the-art generative
models while being faithful to the representation used as conditioning. By
using this new tool to analyze self-supervised models, we can show visually
that i) SSL (backbone) representation are not really invariant to many data
augmentation they were trained on. ii) SSL projector embedding appear too
invariant for tasks like classifications. iii) SSL representations are more
robust to small adversarial perturbation of their inputs iv) there is an
inherent structure learned with SSL model that can be used for image
manipulation.
Related papers
- A Closer Look at Benchmarking Self-Supervised Pre-training with Image Classification [51.35500308126506]
Self-supervised learning (SSL) is a machine learning approach where the data itself provides supervision, eliminating the need for external labels.
We study how classification-based evaluation protocols for SSL correlate and how well they predict downstream performance on different dataset types.
arXiv Detail & Related papers (2024-07-16T23:17:36Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Neural Clustering based Visual Representation Learning [61.72646814537163]
Clustering is one of the most classic approaches in machine learning and data analysis.
We propose feature extraction with clustering (FEC), which views feature extraction as a process of selecting representatives from data.
FEC alternates between grouping pixels into individual clusters to abstract representatives and updating the deep features of pixels with current representatives.
arXiv Detail & Related papers (2024-03-26T06:04:50Z) - Can Generative Models Improve Self-Supervised Representation Learning? [0.7999703756441756]
We introduce a novel framework that enriches the self-supervised learning paradigm by utilizing generative models to produce semantically consistent image augmentations.
Our results show that our framework significantly enhances the quality of learned visual representations by up to 10% Top-1 accuracy in downstream tasks.
arXiv Detail & Related papers (2024-03-09T17:17:07Z) - A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - Explaining, Analyzing, and Probing Representations of Self-Supervised
Learning Models for Sensor-based Human Activity Recognition [2.2082422928825136]
Self-supervised learning (SSL) frameworks have been extensively applied to sensor-based Human Activity Recognition (HAR)
In this paper, we aim to analyze deep representations of two recent SSL frameworks, namely SimCLR and VICReg.
arXiv Detail & Related papers (2023-04-14T07:53:59Z) - CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - Improving VAE-based Representation Learning [26.47244578124654]
We study what properties are required for good representations and how different VAE structure choices could affect the learned properties.
We show that by using a decoder that prefers to learn local features, the remaining global features can be well captured by the latent.
arXiv Detail & Related papers (2022-05-28T23:00:18Z) - A Study of the Generalizability of Self-Supervised Representations [0.0]
Recent advancements in self-supervised learning (SSL) made it possible to learn generalizable visual representations from unlabeled data.
We study generalizability of the SSL and SL-based models via their prediction accuracy as well as prediction confidence.
We show that the SSL representations are more generalizable as compared to the SL representations.
arXiv Detail & Related papers (2021-09-19T15:57:37Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z) - Guided Generative Adversarial Neural Network for Representation Learning
and High Fidelity Audio Generation using Fewer Labelled Audio Data [31.00018800567942]
Recent improvements in Generative Adversarial Neural Networks (GANs) have shown their ability to generate higher quality samples.
Most of the representation learning methods based on GANs learn representations ignoring their post-use scenario.
We propose a novel GAN framework: Guided Generative Neural Network (GGAN), which guides a GAN to focus on learning desired representations.
arXiv Detail & Related papers (2020-03-05T11:01:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.