GeoCLR: Georeference Contrastive Learning for Efficient Seafloor Image
Interpretation
- URL: http://arxiv.org/abs/2108.06421v1
- Date: Fri, 13 Aug 2021 22:42:34 GMT
- Title: GeoCLR: Georeference Contrastive Learning for Efficient Seafloor Image
Interpretation
- Authors: Takaki Yamada, Adam Pr\"ugel-Bennett, Stefan B. Williams, Oscar
Pizarro, Blair Thornton
- Abstract summary: This paper describes Georeference Contrastive Learning of visual Representation (GeoCLR) for efficient training of CNNs.
GeoCLR generates a similar image pair using images taken of nearby locations, and contrasting these with an image pair that is far apart.
A key advantage of this method is that it is self-supervised and does not require any human input for CNN training.
We demonstrate how the latent representations generated by GeoCLR can be used to efficiently guide human annotation efforts.
- Score: 8.837172743444249
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper describes Georeference Contrastive Learning of visual
Representation (GeoCLR) for efficient training of deep-learning Convolutional
Neural Networks (CNNs). The method leverages georeference information by
generating a similar image pair using images taken of nearby locations, and
contrasting these with an image pair that is far apart. The underlying
assumption is that images gathered within a close distance are more likely to
have similar visual appearance, where this can be reasonably satisfied in
seafloor robotic imaging applications where image footprints are limited to
edge lengths of a few metres and are taken so that they overlap along a
vehicle's trajectory, whereas seafloor substrates and habitats have patch sizes
that are far larger. A key advantage of this method is that it is
self-supervised and does not require any human input for CNN training. The
method is computationally efficient, where results can be generated between
dives during multi-day AUV missions using computational resources that would be
accessible during most oceanic field trials. We apply GeoCLR to habitat
classification on a dataset that consists of ~86k images gathered using an
Autonomous Underwater Vehicle (AUV). We demonstrate how the latent
representations generated by GeoCLR can be used to efficiently guide human
annotation efforts, where the semi-supervised framework improves classification
accuracy by an average of 11.8 % compared to state-of-the-art transfer learning
using the same CNN and equivalent number of human annotations for training.
Related papers
- Evaluation of Deep Learning Semantic Segmentation for Land Cover Mapping on Multispectral, Hyperspectral and High Spatial Aerial Imagery [0.0]
In the rise of climate change, land cover mapping has become such an urgent need in environmental monitoring.
This research implemented a semantic segmentation method such as Unet, Linknet, FPN, and PSPnet for categorizing vegetation, water, and others.
The LinkNet model obtained high accuracy in IoU at 0.92 in all datasets, which is comparable with other mentioned techniques.
arXiv Detail & Related papers (2024-06-20T11:40:12Z) - Annotation Cost-Efficient Active Learning for Deep Metric Learning Driven Remote Sensing Image Retrieval [3.2109665109975696]
ANNEAL aims to create a small but informative training set made up of similar and dissimilar image pairs.
The informativeness of image pairs is evaluated by combining uncertainty and diversity criteria.
This way of annotating images significantly reduces the annotation cost compared to annotating images with land-use land-cover class labels.
arXiv Detail & Related papers (2024-06-14T15:08:04Z) - CSP: Self-Supervised Contrastive Spatial Pre-Training for
Geospatial-Visual Representations [90.50864830038202]
We present Contrastive Spatial Pre-Training (CSP), a self-supervised learning framework for geo-tagged images.
We use a dual-encoder to separately encode the images and their corresponding geo-locations, and use contrastive objectives to learn effective location representations from images.
CSP significantly boosts the model performance with 10-34% relative improvement with various labeled training data sampling ratios.
arXiv Detail & Related papers (2023-05-01T23:11:18Z) - Enhancing Self-Supervised Learning for Remote Sensing with Elevation
Data: A Case Study with Scarce And High Level Semantic Labels [1.534667887016089]
This work proposes a hybrid unsupervised and supervised learning method to pre-train models applied in Earth observation downstream tasks.
We combine a contrastive approach to pre-train models with a pixel-wise regression pre-text task to predict coarse elevation maps.
arXiv Detail & Related papers (2023-04-13T23:01:11Z) - Semantic Segmentation of Vegetation in Remote Sensing Imagery Using Deep
Learning [77.34726150561087]
We propose an approach for creating a multi-modal and large-temporal dataset comprised of publicly available Remote Sensing data.
We use Convolutional Neural Networks (CNN) models that are capable of separating different classes of vegetation.
arXiv Detail & Related papers (2022-09-28T18:51:59Z) - Learning Cross-Scale Visual Representations for Real-Time Image
Geo-Localization [21.375640354558044]
State estimation approaches based on local sensors are drifting-prone for long-range missions as error accumulates.
We introduce the cross-scale dataset and a methodology to produce additional data from cross-modality sources.
We propose a framework that learns cross-scale visual representations without supervision.
arXiv Detail & Related papers (2021-09-09T08:08:54Z) - Learning Hybrid Representations for Automatic 3D Vessel Centerline
Extraction [57.74609918453932]
Automatic blood vessel extraction from 3D medical images is crucial for vascular disease diagnoses.
Existing methods may suffer from discontinuities of extracted vessels when segmenting such thin tubular structures from 3D images.
We argue that preserving the continuity of extracted vessels requires to take into account the global geometry.
We propose a hybrid representation learning approach to address this challenge.
arXiv Detail & Related papers (2020-12-14T05:22:49Z) - PGL: Prior-Guided Local Self-supervised Learning for 3D Medical Image
Segmentation [87.50205728818601]
We propose a PriorGuided Local (PGL) self-supervised model that learns the region-wise local consistency in the latent feature space.
Our PGL model learns the distinctive representations of local regions, and hence is able to retain structural information.
arXiv Detail & Related papers (2020-11-25T11:03:11Z) - Geography-Aware Self-Supervised Learning [79.4009241781968]
We show that due to their different characteristics, a non-trivial gap persists between contrastive and supervised learning on standard benchmarks.
We propose novel training methods that exploit the spatially aligned structure of remote sensing data.
Our experiments show that our proposed method closes the gap between contrastive and supervised learning on image classification, object detection and semantic segmentation for remote sensing.
arXiv Detail & Related papers (2020-11-19T17:29:13Z) - DeepEMD: Differentiable Earth Mover's Distance for Few-Shot Learning [122.51237307910878]
We develop methods for few-shot image classification from a new perspective of optimal matching between image regions.
We employ the Earth Mover's Distance (EMD) as a metric to compute a structural distance between dense image representations.
To generate the important weights of elements in the formulation, we design a cross-reference mechanism.
arXiv Detail & Related papers (2020-03-15T08:13:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.