Continual Barlow Twins: continual self-supervised learning for remote
sensing semantic segmentation
- URL: http://arxiv.org/abs/2205.11319v1
- Date: Mon, 23 May 2022 14:02:12 GMT
- Title: Continual Barlow Twins: continual self-supervised learning for remote
sensing semantic segmentation
- Authors: Valerio Marsocci, Simone Scardapane
- Abstract summary: We propose a new algorithm for merging Self-Supervised Learning (SSL) and Continual Learning (CL) for remote sensing applications, that we call Continual Barlow Twins (CBT)
CBT combines the advantages of one of the simplest self-supervision techniques, i.e., Barlow Twins, with the Elastic Weight Consolidation method to avoid catastrophic forgetting.
For the first time we evaluate SSL methods on a highly heterogeneous Earth Observation dataset, showing the effectiveness of these strategies.
- Score: 8.775728170359024
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the field of Earth Observation (EO), Continual Learning (CL) algorithms
have been proposed to deal with large datasets by decomposing them into several
subsets and processing them incrementally. The majority of these algorithms
assume that data is (a) coming from a single source, and (b) fully labeled.
Real-world EO datasets are instead characterized by a large heterogeneity
(e.g., coming from aerial, satellite, or drone scenarios), and for the most
part they are unlabeled, meaning they can be fully exploited only through the
emerging Self-Supervised Learning (SSL) paradigm. For these reasons, in this
paper we propose a new algorithm for merging SSL and CL for remote sensing
applications, that we call Continual Barlow Twins (CBT). It combines the
advantages of one of the simplest self-supervision techniques, i.e., Barlow
Twins, with the Elastic Weight Consolidation method to avoid catastrophic
forgetting. In addition, for the first time we evaluate SSL methods on a highly
heterogeneous EO dataset, showing the effectiveness of these strategies on a
novel combination of three almost non-overlapping domains datasets (airborne
Potsdam dataset, satellite US3D dataset, and drone UAVid dataset), on a crucial
downstream task in EO, i.e., semantic segmentation. Encouraging results show
the superiority of SSL in this setting, and the effectiveness of creating an
incremental effective pretrained feature extractor, based on ResNet50, without
the need of relying on the complete availability of all the data, with a
valuable saving of time and resources.
Related papers
- Towards Satellite Non-IID Imagery: A Spectral Clustering-Assisted Federated Learning Approach [29.593406320684448]
Low Earth orbit (LEO) satellites are capable of gathering abundant Earth observation data (EOD) to enable different Internet of Things (IoT) applications.
To accomplish an effective EOD processing mechanism, it is imperative to investigate: 1) the challenge of processing the observed data without transmitting those large-size data to the ground because the connection between the satellites and the ground stations is intermittent, and 2) the challenge of processing the non-independent and identically distributed (non-IID) satellite data.
We propose an orbit-based spectral clustering-assisted clustered federated self-knowledge distillation (OSC-FSKD)
arXiv Detail & Related papers (2024-10-17T14:36:58Z) - Semi-Supervised One-Shot Imitation Learning [83.94646047695412]
One-shot Imitation Learning aims to imbue AI agents with the ability to learn a new task from a single demonstration.
We introduce the semi-supervised OSIL problem setting, where the learning agent is presented with a large dataset of trajectories.
We develop an algorithm specifically applicable to this semi-supervised OSIL setting.
arXiv Detail & Related papers (2024-08-09T18:11:26Z) - Does Decentralized Learning with Non-IID Unlabeled Data Benefit from
Self Supervision? [51.00034621304361]
We study decentralized learning with unlabeled data through the lens of self-supervised learning (SSL)
We study the effectiveness of contrastive learning algorithms under decentralized learning settings.
arXiv Detail & Related papers (2022-10-20T01:32:41Z) - Semantic Segmentation with Active Semi-Supervised Representation
Learning [23.79742108127707]
We train an effective semantic segmentation algorithm with significantly lesser labeled data.
We extend the prior state-of-the-art S4AL algorithm by replacing its mean teacher approach for semi-supervised learning with a self-training approach.
We evaluate our method on CamVid and CityScapes datasets, the de-facto standards for active learning for semantic segmentation.
arXiv Detail & Related papers (2022-10-16T00:21:43Z) - Bi-level Alignment for Cross-Domain Crowd Counting [113.78303285148041]
Current methods rely on external data for training an auxiliary task or apply an expensive coarse-to-fine estimation.
We develop a new adversarial learning based method, which is simple and efficient to apply.
We evaluate our approach on five real-world crowd counting benchmarks, where we outperform existing approaches by a large margin.
arXiv Detail & Related papers (2022-05-12T02:23:25Z) - Open-Set Semi-Supervised Learning for 3D Point Cloud Understanding [62.17020485045456]
It is commonly assumed in semi-supervised learning (SSL) that the unlabeled data are drawn from the same distribution as that of the labeled ones.
We propose to selectively utilize unlabeled data through sample weighting, so that only conducive unlabeled data would be prioritized.
arXiv Detail & Related papers (2022-05-02T16:09:17Z) - Meta Clustering Learning for Large-scale Unsupervised Person
Re-identification [124.54749810371986]
We propose a "small data for big task" paradigm dubbed Meta Clustering Learning (MCL)
MCL only pseudo-labels a subset of the entire unlabeled data via clustering to save computing for the first-phase training.
Our method significantly saves computational cost while achieving a comparable or even better performance compared to prior works.
arXiv Detail & Related papers (2021-11-19T04:10:18Z) - Relieving the Plateau: Active Semi-Supervised Learning for a Better
Landscape [2.3046646540823916]
Semi-supervised learning (SSL) leverages unlabeled data that are more accessible than their labeled counterparts.
Active learning (AL) selects unlabeled instances to be annotated by a human-in-the-loop in hopes of better performance with less labeled data.
We propose convergence rate control (CRC), an AL algorithm that selects unlabeled data to improve the problem conditioning upon inclusion to the labeled set.
arXiv Detail & Related papers (2021-04-08T06:03:59Z) - Learning to Count in the Crowd from Limited Labeled Data [109.2954525909007]
We focus on reducing the annotation efforts by learning to count in the crowd from limited number of labeled samples.
Specifically, we propose a Gaussian Process-based iterative learning mechanism that involves estimation of pseudo-ground truth for the unlabeled data.
arXiv Detail & Related papers (2020-07-07T04:17:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.