Evaluating The Robustness of Self-Supervised Representations to
Background/Foreground Removal
- URL: http://arxiv.org/abs/2306.01398v1
- Date: Fri, 2 Jun 2023 09:46:22 GMT
- Title: Evaluating The Robustness of Self-Supervised Representations to
Background/Foreground Removal
- Authors: Xavier F. Cadet, Ranya Aloufi, Alain Miranville, Sara Ahmadi-Abhari,
Hamed Haddadi
- Abstract summary: We consider state-of-the-art SSL pretrained models, such as DINOv2, MAE, and SwaV, and analyzed changes at the representation levels across 4 Image Classification datasets.
Empirically, we show that not all models lead to representations that separate foreground, background, and complete images.
- Score: 4.007351600492541
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite impressive empirical advances of SSL in solving various tasks, the
problem of understanding and characterizing SSL representations learned from
input data remains relatively under-explored. We provide a comparative analysis
of how the representations produced by SSL models differ when masking parts of
the input. Specifically, we considered state-of-the-art SSL pretrained models,
such as DINOv2, MAE, and SwaV, and analyzed changes at the representation
levels across 4 Image Classification datasets. First, we generate variations of
the datasets by applying foreground and background segmentation. Then, we
conduct statistical analysis using Canonical Correlation Analysis (CCA) and
Centered Kernel Alignment (CKA) to evaluate the robustness of the
representations learned in SSL models. Empirically, we show that not all models
lead to representations that separate foreground, background, and complete
images. Furthermore, we test different masking strategies by occluding the
center regions of the images to address cases where foreground and background
are difficult. For example, the DTD dataset that focuses on texture rather
specific objects.
Related papers
- A Closer Look at Benchmarking Self-Supervised Pre-training with Image Classification [51.35500308126506]
Self-supervised learning (SSL) is a machine learning approach where the data itself provides supervision, eliminating the need for external labels.
We study how classification-based evaluation protocols for SSL correlate and how well they predict downstream performance on different dataset types.
arXiv Detail & Related papers (2024-07-16T23:17:36Z) - Visual Self-supervised Learning Scheme for Dense Prediction Tasks on X-ray Images [3.782392436834913]
Self-supervised learning (SSL) has led to considerable progress in natural language processing (NLP)
However, the incorporation of contrastive learning into existing visual SSL models has led to considerable progress, often surpassing supervised counterparts.
Here, we focus on dense prediction tasks using security inspection x-ray images to evaluate our proposed model, Segment localization (SegLoc)
Based upon the Instance localization (InsLoc) model, SegLoc addresses one of the key challenges of contrastive learning, i.e., false negative pairs of query embeddings.
arXiv Detail & Related papers (2023-10-12T15:42:17Z) - De-coupling and De-positioning Dense Self-supervised Learning [65.56679416475943]
Dense Self-Supervised Learning (SSL) methods address the limitations of using image-level feature representations when handling images with multiple objects.
We show that they suffer from coupling and positional bias, which arise from the receptive field increasing with layer depth and zero-padding.
We demonstrate the benefits of our method on COCO and on a new challenging benchmark, OpenImage-MINI, for object classification, semantic segmentation, and object detection.
arXiv Detail & Related papers (2023-03-29T18:07:25Z) - Evaluating Self-Supervised Learning via Risk Decomposition [100.73914689472507]
Self-supervised learning (SSL) pipelines differ in many design choices such as the architecture, augmentations, or pretraining data.
This does not provide much insight into why or when a model is better, now how to improve it.
We propose an SSL risk decomposition, which generalizes the classical supervised approximation-estimation decomposition.
arXiv Detail & Related papers (2023-02-06T19:09:00Z) - Consistency Regularisation in Varying Contexts and Feature Perturbations
for Semi-Supervised Semantic Segmentation of Histology Images [14.005379068469361]
We present a consistency based semi-supervised learning (SSL) approach that can help mitigate this challenge.
SSL models might also be susceptible to changing context and features perturbations exhibiting poor generalisation due to the limited training data.
We show that cross-consistency training makes the encoder features invariant to different perturbations and improves the prediction confidence.
arXiv Detail & Related papers (2023-01-30T18:21:57Z) - Deciphering the Projection Head: Representation Evaluation
Self-supervised Learning [6.375931203397043]
Self-supervised learning (SSL) aims to learn intrinsic features without labels.
Projection head always plays an important role in improving the performance of the downstream task.
We propose a Representation Evaluation Design (RED) in SSL models in which a shortcut connection between the representation and the projection vectors is built.
arXiv Detail & Related papers (2023-01-28T13:13:53Z) - A Study of the Generalizability of Self-Supervised Representations [0.0]
Recent advancements in self-supervised learning (SSL) made it possible to learn generalizable visual representations from unlabeled data.
We study generalizability of the SSL and SL-based models via their prediction accuracy as well as prediction confidence.
We show that the SSL representations are more generalizable as compared to the SL representations.
arXiv Detail & Related papers (2021-09-19T15:57:37Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - PGL: Prior-Guided Local Self-supervised Learning for 3D Medical Image
Segmentation [87.50205728818601]
We propose a PriorGuided Local (PGL) self-supervised model that learns the region-wise local consistency in the latent feature space.
Our PGL model learns the distinctive representations of local regions, and hence is able to retain structural information.
arXiv Detail & Related papers (2020-11-25T11:03:11Z) - Remote Sensing Image Scene Classification with Self-Supervised Paradigm
under Limited Labeled Samples [11.025191332244919]
We introduce new self-supervised learning (SSL) mechanism to obtain the high-performance pre-training model for RSIs scene classification from large unlabeled data.
Experiments on three commonly used RSIs scene classification datasets demonstrated that this new learning paradigm outperforms the traditional dominant ImageNet pre-trained model.
The insights distilled from our studies can help to foster the development of SSL in the remote sensing community.
arXiv Detail & Related papers (2020-10-02T09:27:19Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.