Inherit with Distillation and Evolve with Contrast: Exploring Class
Incremental Semantic Segmentation Without Exemplar Memory
- URL: http://arxiv.org/abs/2309.15413v1
- Date: Wed, 27 Sep 2023 05:38:31 GMT
- Title: Inherit with Distillation and Evolve with Contrast: Exploring Class
Incremental Semantic Segmentation Without Exemplar Memory
- Authors: Danpei Zhao, Bo Yuan, Zhenwei Shi
- Abstract summary: We present IDEC, which consists of a Dense Knowledge Distillation on all Aspects (DADA) and an Asymmetric Region-wise Contrastive Learning (ARCL) module.
We demonstrate the effectiveness of our method on multiple CISS tasks by state-of-the-art performance, including Pascal VOC 2012, ADE20K and ISPRS datasets.
- Score: 23.730424035141155
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a front-burner problem in incremental learning, class incremental semantic
segmentation (CISS) is plagued by catastrophic forgetting and semantic drift.
Although recent methods have utilized knowledge distillation to transfer
knowledge from the old model, they are still unable to avoid pixel confusion,
which results in severe misclassification after incremental steps due to the
lack of annotations for past and future classes. Meanwhile data-replay-based
approaches suffer from storage burdens and privacy concerns. In this paper, we
propose to address CISS without exemplar memory and resolve catastrophic
forgetting as well as semantic drift synchronously. We present Inherit with
Distillation and Evolve with Contrast (IDEC), which consists of a Dense
Knowledge Distillation on all Aspects (DADA) manner and an Asymmetric
Region-wise Contrastive Learning (ARCL) module. Driven by the devised dynamic
class-specific pseudo-labelling strategy, DADA distils intermediate-layer
features and output-logits collaboratively with more emphasis on
semantic-invariant knowledge inheritance. ARCL implements region-wise
contrastive learning in the latent space to resolve semantic drift among known
classes, current classes, and unknown classes. We demonstrate the effectiveness
of our method on multiple CISS tasks by state-of-the-art performance, including
Pascal VOC 2012, ADE20K and ISPRS datasets. Our method also shows superior
anti-forgetting ability, particularly in multi-step CISS tasks.
Related papers
- Learning at a Glance: Towards Interpretable Data-limited Continual Semantic Segmentation via Semantic-Invariance Modelling [21.114359437665364]
Continual semantic segmentation (CSS) based on incremental learning (IL) is a great endeavour in developing human-like segmentation models.
Current CSS approaches encounter challenges in the trade-off between preserving old knowledge and learning new ones.
We present Learning at a Glance (LAG), an efficient, robust, human-like and interpretable approach for CSS.
arXiv Detail & Related papers (2024-07-22T07:17:52Z) - ECLIPSE: Efficient Continual Learning in Panoptic Segmentation with Visual Prompt Tuning [54.68180752416519]
Panoptic segmentation is a cutting-edge computer vision task.
We introduce a novel and efficient method for continual panoptic segmentation based on Visual Prompt Tuning, dubbed ECLIPSE.
Our approach involves freezing the base model parameters and fine-tuning only a small set of prompt embeddings, addressing both catastrophic forgetting and plasticity.
arXiv Detail & Related papers (2024-03-29T11:31:12Z) - OrCo: Towards Better Generalization via Orthogonality and Contrast for Few-Shot Class-Incremental Learning [57.43911113915546]
Few-Shot Class-Incremental Learning (FSCIL) introduces a paradigm in which the problem space expands with limited data.
FSCIL methods inherently face the challenge of catastrophic forgetting as data arrives incrementally.
We propose the OrCo framework built on two core principles: features' orthogonality in the representation space, and contrastive learning.
arXiv Detail & Related papers (2024-03-27T13:30:48Z) - Class Gradient Projection For Continual Learning [99.105266615448]
Catastrophic forgetting is one of the most critical challenges in Continual Learning (CL)
We propose Class Gradient Projection (CGP), which calculates the gradient subspace from individual classes rather than tasks.
arXiv Detail & Related papers (2023-11-25T02:45:56Z) - Enhancing Few-shot CLIP with Semantic-Aware Fine-Tuning [61.902254546858465]
Methods based on Contrastive Language-Image Pre-training have exhibited promising performance in few-shot adaptation tasks.
We propose fine-tuning the parameters of the attention pooling layer during the training process to encourage the model to focus on task-specific semantics.
arXiv Detail & Related papers (2023-11-08T05:18:57Z) - Gradient-Semantic Compensation for Incremental Semantic Segmentation [43.00965727428193]
Incremental semantic segmentation aims to continually learn the segmentation of new coming classes without accessing the training data of previously learned classes.
We propose a Gradient-Semantic Compensation model, which surmounts incremental semantic segmentation from both gradient and semantic perspectives.
arXiv Detail & Related papers (2023-07-20T12:32:25Z) - SATS: Self-Attention Transfer for Continual Semantic Segmentation [50.51525791240729]
continual semantic segmentation suffers from the same catastrophic forgetting issue as in continual classification learning.
This study proposes to transfer a new type of information relevant to knowledge, i.e. the relationships between elements within each image.
The relationship information can be effectively obtained from the self-attention maps in a Transformer-style segmentation model.
arXiv Detail & Related papers (2022-03-15T06:09:28Z) - A Contrastive Distillation Approach for Incremental Semantic
Segmentation in Aerial Images [15.75291664088815]
A major issue concerning current deep neural architectures is known as catastrophic forgetting.
We propose a contrastive regularization, where any given input is compared with its augmented version.
We show the effectiveness of our solution on the Potsdam dataset, outperforming the incremental baseline in every test.
arXiv Detail & Related papers (2021-12-07T16:44:45Z) - SSUL: Semantic Segmentation with Unknown Label for Exemplar-based
Class-Incremental Learning [19.152041362805985]
We consider a class-incremental semantic segmentation (CISS) problem.
We propose a new method, dubbed as SSUL-M (Semantic with Unknown Label with Memory), by carefully combining several techniques tailored for semantic segmentation.
We show our method achieves significantly better performance than the recent state-of-the-art baselines on the standard benchmark datasets.
arXiv Detail & Related papers (2021-06-22T06:40:26Z) - Anti-aliasing Semantic Reconstruction for Few-Shot Semantic Segmentation [66.85202434812942]
We reformulate few-shot segmentation as a semantic reconstruction problem.
We convert base class features into a series of basis vectors which span a class-level semantic space for novel class reconstruction.
Our proposed approach, referred to as anti-aliasing semantic reconstruction (ASR), provides a systematic yet interpretable solution for few-shot learning problems.
arXiv Detail & Related papers (2021-06-01T02:17:36Z) - Continual Semantic Segmentation via Repulsion-Attraction of Sparse and
Disentangled Latent Representations [18.655840060559168]
This paper focuses on class incremental continual learning in semantic segmentation.
New categories are made available over time while previous training data is not retained.
The proposed continual learning scheme shapes the latent space to reduce forgetting whilst improving the recognition of novel classes.
arXiv Detail & Related papers (2021-03-10T21:02:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.