Improving Contrastive Learning on Visually Homogeneous Mars Rover Images
- URL: http://arxiv.org/abs/2210.09234v1
- Date: Mon, 17 Oct 2022 16:26:56 GMT
- Title: Improving Contrastive Learning on Visually Homogeneous Mars Rover Images
- Authors: Isaac Ronald Ward and Charles Moore and Kai Pak and Jingdao Chen and
Edwin Goh
- Abstract summary: We show how contrastive learning can be applied to hundreds of thousands of unlabeled Mars terrain images.
Contrastive learning assumes that any given pair of distinct images contain distinct semantic content.
We propose two approaches to resolve this: 1) an unsupervised deep clustering step on the Mars datasets, which identifies clusters of images containing similar semantic content and corrects false negative errors during training, and 2) a simple approach which mixes data from different domains to increase visual diversity of the total training dataset.
- Score: 3.206547922373737
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive learning has recently demonstrated superior performance to
supervised learning, despite requiring no training labels. We explore how
contrastive learning can be applied to hundreds of thousands of unlabeled Mars
terrain images, collected from the Mars rovers Curiosity and Perseverance, and
from the Mars Reconnaissance Orbiter. Such methods are appealing since the vast
majority of Mars images are unlabeled as manual annotation is labor intensive
and requires extensive domain knowledge. Contrastive learning, however, assumes
that any given pair of distinct images contain distinct semantic content. This
is an issue for Mars image datasets, as any two pairs of Mars images are far
more likely to be semantically similar due to the lack of visual diversity on
the planet's surface. Making the assumption that pairs of images will be in
visual contrast - when they are in fact not - results in pairs that are falsely
considered as negatives, impacting training performance. In this study, we
propose two approaches to resolve this: 1) an unsupervised deep clustering step
on the Mars datasets, which identifies clusters of images containing similar
semantic content and corrects false negative errors during training, and 2) a
simple approach which mixes data from different domains to increase visual
diversity of the total training dataset. Both cases reduce the rate of false
negative pairs, thus minimizing the rate in which the model is incorrectly
penalized during contrastive training. These modified approaches remain fully
unsupervised end-to-end. To evaluate their performance, we add a single linear
layer trained to generate class predictions based on these
contrastively-learned features and demonstrate increased performance compared
to supervised models; observing an improvement in classification accuracy of
3.06% using only 10% of the labeled data.
Related papers
- Learning to Rank Patches for Unbiased Image Redundancy Reduction [80.93989115541966]
Images suffer from heavy spatial redundancy because pixels in neighboring regions are spatially correlated.
Existing approaches strive to overcome this limitation by reducing less meaningful image regions.
We propose a self-supervised framework for image redundancy reduction called Learning to Rank Patches.
arXiv Detail & Related papers (2024-03-31T13:12:41Z) - SatSynth: Augmenting Image-Mask Pairs through Diffusion Models for Aerial Semantic Segmentation [69.42764583465508]
We explore the potential of generative image diffusion to address the scarcity of annotated data in earth observation tasks.
To the best of our knowledge, we are the first to generate both images and corresponding masks for satellite segmentation.
arXiv Detail & Related papers (2024-03-25T10:30:22Z) - Semi-Supervised Learning for Mars Imagery Classification and
Segmentation [35.103989798891476]
We introduce a semi-supervised framework for machine vision on Mars.
We try to resolve two specific tasks: classification and segmentation.
Our learning strategies can improve the classification and segmentation models by a large margin and outperform state-of-the-art approaches.
arXiv Detail & Related papers (2022-06-05T13:55:10Z) - Mix-up Self-Supervised Learning for Contrast-agnostic Applications [33.807005669824136]
We present the first mix-up self-supervised learning framework for contrast-agnostic applications.
We address the low variance across images based on cross-domain mix-up and build the pretext task based on image reconstruction and transparency prediction.
arXiv Detail & Related papers (2022-04-02T16:58:36Z) - Mars Terrain Segmentation with Less Labels [1.1745324895296465]
This research proposes a semi-supervised learning framework for Mars terrain segmentation.
It incorporates a backbone module which is trained using a contrastive loss function and an output atrous convolution module.
The proposed model is able to achieve a segmentation accuracy of 91.1% using only 161 training images.
arXiv Detail & Related papers (2022-02-01T22:25:15Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - An Empirical Study of the Collapsing Problem in Semi-Supervised 2D Human
Pose Estimation [80.02124918255059]
Semi-supervised learning aims to boost the accuracy of a model by exploring unlabeled images.
We learn two networks to mutually teach each other.
The more reliable predictions on easy images in each network are used to teach the other network to learn about the corresponding hard images.
arXiv Detail & Related papers (2020-11-25T03:29:52Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z) - Un-Mix: Rethinking Image Mixtures for Unsupervised Visual Representation
Learning [108.999497144296]
Recently advanced unsupervised learning approaches use the siamese-like framework to compare two "views" from the same image for learning representations.
This work aims to involve the distance concept on label space in the unsupervised learning and let the model be aware of the soft degree of similarity between positive or negative pairs.
Despite its conceptual simplicity, we show empirically that with the solution -- Unsupervised image mixtures (Un-Mix), we can learn subtler, more robust and generalized representations from the transformed input and corresponding new label space.
arXiv Detail & Related papers (2020-03-11T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.