DMKD: Improving Feature-based Knowledge Distillation for Object
Detection Via Dual Masking Augmentation
- URL: http://arxiv.org/abs/2309.02719v2
- Date: Thu, 7 Sep 2023 03:36:37 GMT
- Title: DMKD: Improving Feature-based Knowledge Distillation for Object
Detection Via Dual Masking Augmentation
- Authors: Guang Yang, Yin Tang, Zhijian Wu, Jun Li, Jianhua Xu, Xili Wan
- Abstract summary: We devise a Dual Masked Knowledge Distillation (DMKD) framework which can capture both spatially important and channel-wise informative clues.
Our experiments on object detection task demonstrate that the student networks achieve performance gains of 4.1% and 4.3% with the help of our method.
- Score: 10.437237606721222
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent mainstream masked distillation methods function by reconstructing
selectively masked areas of a student network from the feature map of its
teacher counterpart. In these methods, the masked regions need to be properly
selected, such that reconstructed features encode sufficient discrimination and
representation capability like the teacher feature. However, previous masked
distillation methods only focus on spatial masking, making the resulting masked
areas biased towards spatial importance without encoding informative channel
clues. In this study, we devise a Dual Masked Knowledge Distillation (DMKD)
framework which can capture both spatially important and channel-wise
informative clues for comprehensive masked feature reconstruction. More
specifically, we employ dual attention mechanism for guiding the respective
masking branches, leading to reconstructed feature encoding dual significance.
Furthermore, fusing the reconstructed features is achieved by self-adjustable
weighting strategy for effective feature distillation. Our experiments on
object detection task demonstrate that the student networks achieve performance
gains of 4.1% and 4.3% with the help of our method when RetinaNet and Cascade
Mask R-CNN are respectively used as the teacher networks, while outperforming
the other state-of-the-art distillation methods.
Related papers
- Semantic Refocused Tuning for Open-Vocabulary Panoptic Segmentation [42.020470627552136]
Open-vocabulary panoptic segmentation is an emerging task aiming to accurately segment the image into semantically meaningful masks.
mask classification is the main performance bottleneck for open-vocab panoptic segmentation.
We propose Semantic Refocused Tuning, a novel framework that greatly enhances open-vocab panoptic segmentation.
arXiv Detail & Related papers (2024-09-24T17:50:28Z) - DFMSD: Dual Feature Masking Stage-wise Knowledge Distillation for Object Detection [6.371066478190595]
A novel dual feature-masking heterogeneous distillation framework termed DFMSD is proposed for object detection.
A masking enhancement strategy is combined with stage-wise learning to improve feature-masking reconstruction.
Experiments for the object detection task demonstrate the promise of our approach.
arXiv Detail & Related papers (2024-07-18T04:19:14Z) - Understanding Masked Autoencoders From a Local Contrastive Perspective [80.57196495601826]
Masked AutoEncoder (MAE) has revolutionized the field of self-supervised learning with its simple yet effective masking and reconstruction strategies.
We introduce a new empirical framework, called Local Contrastive MAE, to analyze both reconstructive and contrastive aspects of MAE.
arXiv Detail & Related papers (2023-10-03T12:08:15Z) - Masking Improves Contrastive Self-Supervised Learning for ConvNets, and Saliency Tells You Where [63.61248884015162]
We aim to alleviate the burden of including masking operation into the contrastive-learning framework for convolutional neural networks.
We propose to explicitly take the saliency constraint into consideration in which the masked regions are more evenly distributed among the foreground and background.
arXiv Detail & Related papers (2023-09-22T09:58:38Z) - Cross Architecture Distillation for Face Recognition [49.55061794917994]
We develop an Adaptable Prompting Teacher network (APT) that integrates prompts into the teacher, enabling it to manage distillation-specific knowledge.
Experiments on popular face benchmarks and two large-scale verification sets demonstrate the superiority of our method.
arXiv Detail & Related papers (2023-06-26T12:54:28Z) - AMD: Adaptive Masked Distillation for Object [8.668808292258706]
We propose a spatial-channel adaptive masked distillation (AMD) network for object detection.
We employ a simple and efficient module to allow the student network channel to be adaptive.
With the help of our proposed distillation method, the student networks report 41.3%, 42.4%, and 42.7% mAP scores.
arXiv Detail & Related papers (2023-01-31T10:32:13Z) - ADPS: Asymmetric Distillation Post-Segmentation for Image Anomaly
Detection [75.68023968735523]
Knowledge Distillation-based Anomaly Detection (KDAD) methods rely on the teacher-student paradigm to detect and segment anomalous regions.
We propose an innovative approach called Asymmetric Distillation Post-Segmentation (ADPS)
Our ADPS employs an asymmetric distillation paradigm that takes distinct forms of the same image as the input of the teacher-student networks.
We show that ADPS significantly improves Average Precision (AP) metric by 9% and 20% on the MVTec AD and KolektorSDD2 datasets.
arXiv Detail & Related papers (2022-10-19T12:04:47Z) - What You See is What You Classify: Black Box Attributions [61.998683569022006]
We train a deep network, the Explainer, to predict attributions for a pre-trained black-box classifier, the Explanandum.
Unlike most existing approaches, ours is capable of directly generating very distinct class-specific masks.
We show that our attributions are superior to established methods both visually and quantitatively.
arXiv Detail & Related papers (2022-05-23T12:30:04Z) - CoupleFace: Relation Matters for Face Recognition Distillation [26.2626768462705]
We propose an effective face recognition distillation method called CoupleFace.
We first propose to mine the informative mutual relations, and then introduce the Relation-Aware Distillation (RAD) loss to transfer the mutual relation knowledge of the teacher model to the student model.
Based on our proposed CoupleFace, we have won the first place in the ICCV21 Masked Face Recognition Challenge (MS1M track)
arXiv Detail & Related papers (2022-04-12T03:25:42Z) - HMFS: Hybrid Masking for Few-Shot Segmentation [27.49000348046462]
We develop a simple, effective, and efficient approach to enhance feature masking (FM)
We compensate for the loss of fine-grained spatial details in FM technique by investigating and leveraging a complementary basic input masking method.
Experimental results on three publicly available benchmarks reveal that HMFS outperforms the current state-of-the-art methods by visible margins.
arXiv Detail & Related papers (2022-03-24T03:07:20Z) - Image Inpainting by End-to-End Cascaded Refinement with Mask Awareness [66.55719330810547]
Inpainting arbitrary missing regions is challenging because learning valid features for various masked regions is nontrivial.
We propose a novel mask-aware inpainting solution that learns multi-scale features for missing regions in the encoding phase.
Our framework is validated both quantitatively and qualitatively via extensive experiments on three public datasets.
arXiv Detail & Related papers (2021-04-28T13:17:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.