FAKD: Feature Augmented Knowledge Distillation for Semantic Segmentation
- URL: http://arxiv.org/abs/2208.14143v1
- Date: Tue, 30 Aug 2022 10:55:31 GMT
- Title: FAKD: Feature Augmented Knowledge Distillation for Semantic Segmentation
- Authors: Jianlong Yuan, Qian Qi, Fei Du, Zhibin Wang, Fan Wang, Yifan Liu
- Abstract summary: We explore data augmentations for knowledge distillation on semantic segmentation.
Inspired by the recent progress on semantic directions on feature-space, we propose to include augmentations in feature space for efficient distillation.
- Score: 17.294737459735675
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we explore data augmentations for knowledge distillation on
semantic segmentation. To avoid over-fitting to the noise in the teacher
network, a large number of training examples is essential for knowledge
distillation. Imagelevel argumentation techniques like flipping, translation or
rotation are widely used in previous knowledge distillation framework. Inspired
by the recent progress on semantic directions on feature-space, we propose to
include augmentations in feature space for efficient distillation.
Specifically, given a semantic direction, an infinite number of augmentations
can be obtained for the student in the feature space. Furthermore, the analysis
shows that those augmentations can be optimized simultaneously by minimizing an
upper bound for the losses defined by augmentations. Based on the observation,
a new algorithm is developed for knowledge distillation in semantic
segmentation. Extensive experiments on four semantic segmentation benchmarks
demonstrate that the proposed method can boost the performance of current
knowledge distillation methods without any significant overhead. Code is
available at: https://github.com/jianlong-yuan/FAKD.
Related papers
- Knowledge Diffusion for Distillation [53.908314960324915]
The representation gap between teacher and student is an emerging topic in knowledge distillation (KD)
We state that the essence of these methods is to discard the noisy information and distill the valuable information in the feature.
We propose a novel KD method dubbed DiffKD, to explicitly denoise and match features using diffusion models.
arXiv Detail & Related papers (2023-05-25T04:49:34Z) - Multi-to-Single Knowledge Distillation for Point Cloud Semantic
Segmentation [41.02741249858771]
We propose a novel multi-to-single knowledge distillation framework for the 3D point cloud semantic segmentation task.
Instead of fusing all the points of multi-scans directly, only the instances that belong to the previously defined hard classes are fused.
arXiv Detail & Related papers (2023-04-28T12:17:08Z) - Exploring Inconsistent Knowledge Distillation for Object Detection with
Data Augmentation [66.25738680429463]
Knowledge Distillation (KD) for object detection aims to train a compact detector by transferring knowledge from a teacher model.
We propose inconsistent knowledge distillation (IKD) which aims to distill knowledge inherent in the teacher model's counter-intuitive perceptions.
Our method outperforms state-of-the-art KD baselines on one-stage, two-stage and anchor-free object detectors.
arXiv Detail & Related papers (2022-09-20T16:36:28Z) - Normalized Feature Distillation for Semantic Segmentation [6.882655287146012]
We propose a simple yet effective feature distillation method called normalized feature distillation (NFD)
Our method achieves state-of-the-art distillation results for semantic segmentation on Cityscapes, VOC 2012, and ADE20K datasets.
arXiv Detail & Related papers (2022-07-12T01:54:25Z) - Knowledge Distillation Meets Open-Set Semi-Supervised Learning [69.21139647218456]
We propose a novel em modelname (bfem shortname) method dedicated for distilling representational knowledge semantically from a pretrained teacher to a target student.
At the problem level, this establishes an interesting connection between knowledge distillation with open-set semi-supervised learning (SSL)
Our shortname outperforms significantly previous state-of-the-art knowledge distillation methods on both coarse object classification and fine face recognition tasks.
arXiv Detail & Related papers (2022-05-13T15:15:27Z) - Energy-based Latent Aligner for Incremental Learning [83.0135278697976]
Deep learning models tend to forget their earlier knowledge while incrementally learning new tasks.
This behavior emerges because the parameter updates optimized for the new tasks may not align well with the updates suitable for older tasks.
We propose ELI: Energy-based Latent Aligner for Incremental Learning.
arXiv Detail & Related papers (2022-03-28T17:57:25Z) - Uncertainty-aware Contrastive Distillation for Incremental Semantic
Segmentation [46.14545656625703]
catastrophic forgetting is the tendency of neural networks to fail to preserve the knowledge acquired from old tasks when learning new tasks.
We propose a novel distillation framework, Uncertainty-aware Contrastive Distillation (method)
Our results demonstrate the advantage of the proposed distillation technique, which can be used in synergy with previous IL approaches.
arXiv Detail & Related papers (2022-03-26T15:32:12Z) - A Simple Baseline for Semi-supervised Semantic Segmentation with Strong
Data Augmentation [74.8791451327354]
We propose a simple yet effective semi-supervised learning framework for semantic segmentation.
A set of simple design and training techniques can collectively improve the performance of semi-supervised semantic segmentation significantly.
Our method achieves state-of-the-art results in the semi-supervised settings on the Cityscapes and Pascal VOC datasets.
arXiv Detail & Related papers (2021-04-15T06:01:39Z) - Refine Myself by Teaching Myself: Feature Refinement via Self-Knowledge
Distillation [12.097302014936655]
This paper proposes a novel self-knowledge distillation method, Feature Refinement via Self-Knowledge Distillation (FRSKD)
Our proposed method, FRSKD, can utilize both soft label and feature-map distillations for the self-knowledge distillation.
We demonstrate the effectiveness of FRSKD by enumerating its performance improvements in diverse tasks and benchmark datasets.
arXiv Detail & Related papers (2021-03-15T10:59:43Z) - Deep Semi-supervised Knowledge Distillation for Overlapping Cervical
Cell Instance Segmentation [54.49894381464853]
We propose to leverage both labeled and unlabeled data for instance segmentation with improved accuracy by knowledge distillation.
We propose a novel Mask-guided Mean Teacher framework with Perturbation-sensitive Sample Mining.
Experiments show that the proposed method improves the performance significantly compared with the supervised method learned from labeled data only.
arXiv Detail & Related papers (2020-07-21T13:27:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.