Representation Learning for Remote Sensing: An Unsupervised Sensor
Fusion Approach
- URL: http://arxiv.org/abs/2108.05094v1
- Date: Wed, 11 Aug 2021 08:32:58 GMT
- Title: Representation Learning for Remote Sensing: An Unsupervised Sensor
Fusion Approach
- Authors: Aidan M. Swope, Xander H. Rudelis, Kyle T. Story
- Abstract summary: We propose Contrastive Sensor Fusion, which exploits coterminous data from multiple sources to learn useful representations of every possible combination of those sources.
Using a dataset of 47 million unlabeled coterminous image triplets, we train an encoder to produce meaningful representations from any possible combination of channels from the input sensors.
These representations outperform fully supervised ImageNet weights on a remote sensing classification task and improve as more sensors are fused.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In the application of machine learning to remote sensing, labeled data is
often scarce or expensive, which impedes the training of powerful models like
deep convolutional neural networks. Although unlabeled data is abundant, recent
self-supervised learning approaches are ill-suited to the remote sensing
domain. In addition, most remote sensing applications currently use only a
small subset of the multi-sensor, multi-channel information available,
motivating the need for fused multi-sensor representations. We propose a new
self-supervised training objective, Contrastive Sensor Fusion, which exploits
coterminous data from multiple sources to learn useful representations of every
possible combination of those sources. This method uses information common
across multiple sensors and bands by training a single model to produce a
representation that remains similar when any subset of its input channels is
used. Using a dataset of 47 million unlabeled coterminous image triplets, we
train an encoder to produce semantically meaningful representations from any
possible combination of channels from the input sensors. These representations
outperform fully supervised ImageNet weights on a remote sensing classification
task and improve as more sensors are fused. Our code is available at
https://storage.cloud.google.com/public-published-datasets/csf_code.zip.
Related papers
- Physical-Layer Semantic-Aware Network for Zero-Shot Wireless Sensing [74.12670841657038]
Device-free wireless sensing has recently attracted significant interest due to its potential to support a wide range of immersive human-machine interactive applications.
Data heterogeneity in wireless signals and data privacy regulation of distributed sensing have been considered as the major challenges that hinder the wide applications of wireless sensing in large area networking systems.
We propose a novel zero-shot wireless sensing solution that allows models constructed in one or a limited number of locations to be directly transferred to other locations without any labeled data.
arXiv Detail & Related papers (2023-12-08T13:50:30Z) - USat: A Unified Self-Supervised Encoder for Multi-Sensor Satellite
Imagery [5.671254904219855]
We develop a new encoder architecture called USat that can input multi-spectral data from multiple sensors for self-supervised pre-training.
We integrate USat into a Masked Autoencoder (MAE) self-supervised pre-training procedure and find that a pre-trained USat outperforms state-of-the-art MAE models trained on remote sensing data.
arXiv Detail & Related papers (2023-12-02T19:17:04Z) - UnLoc: A Universal Localization Method for Autonomous Vehicles using
LiDAR, Radar and/or Camera Input [51.150605800173366]
UnLoc is a novel unified neural modeling approach for localization with multi-sensor input in all weather conditions.
Our method is extensively evaluated on Oxford Radar RobotCar, ApolloSouthBay and Perth-WA datasets.
arXiv Detail & Related papers (2023-07-03T04:10:55Z) - Joint multi-modal Self-Supervised pre-training in Remote Sensing:
Application to Methane Source Classification [10.952006057356714]
In earth observation, there are opportunities to exploit domain-specific remote sensing image data.
We briefly review the core principles behind so-called joint-embeddings methods and investigate the usage of multiple remote sensing modalities in self-supervised pre-training.
arXiv Detail & Related papers (2023-06-16T14:01:57Z) - Learning Online Multi-Sensor Depth Fusion [100.84519175539378]
SenFuNet is a depth fusion approach that learns sensor-specific noise and outlier statistics.
We conduct experiments with various sensor combinations on the real-world CoRBS and Scene3D datasets.
arXiv Detail & Related papers (2022-04-07T10:45:32Z) - Self-supervised Audiovisual Representation Learning for Remote Sensing Data [96.23611272637943]
We propose a self-supervised approach for pre-training deep neural networks in remote sensing.
By exploiting the correspondence between geo-tagged audio recordings and remote sensing, this is done in a completely label-free manner.
We show that our approach outperforms existing pre-training strategies for remote sensing imagery.
arXiv Detail & Related papers (2021-08-02T07:50:50Z) - Seasonal Contrast: Unsupervised Pre-Training from Uncurated Remote
Sensing Data [64.40187171234838]
Seasonal Contrast (SeCo) is an effective pipeline to leverage unlabeled data for in-domain pre-training of re-mote sensing representations.
SeCo will be made public to facilitate transfer learning and enable rapid progress in re-mote sensing applications.
arXiv Detail & Related papers (2021-03-30T18:26:39Z) - Handling Variable-Dimensional Time Series with Graph Neural Networks [20.788813485815698]
Internet of Things (IoT) technology involves capturing data from multiple sensors resulting in multi-sensor time series.
Existing neural networks based approaches for such multi-sensor time series modeling assume fixed input dimension or number of sensors.
We consider training neural network models from such multi-sensor time series, where the time series have varying input dimensionality owing to availability or installation of a different subset of sensors at each source of time series.
arXiv Detail & Related papers (2020-07-01T12:11:16Z) - Laplacian Denoising Autoencoder [114.21219514831343]
We propose to learn data representations with a novel type of denoising autoencoder.
The noisy input data is generated by corrupting latent clean data in the gradient domain.
Experiments on several visual benchmarks demonstrate that better representations can be learned with the proposed approach.
arXiv Detail & Related papers (2020-03-30T16:52:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.