SEA: Bridging the Gap Between One- and Two-stage Detector Distillation
via SEmantic-aware Alignment
- URL: http://arxiv.org/abs/2203.00862v1
- Date: Wed, 2 Mar 2022 04:24:05 GMT
- Title: SEA: Bridging the Gap Between One- and Two-stage Detector Distillation
via SEmantic-aware Alignment
- Authors: Yixin Chen, Zhuotao Tian, Pengguang Chen, Shu Liu, Jiaya Jia
- Abstract summary: We name our method SEA (SEmantic-aware Alignment) distillation given the nature of abstracting dense fine-grained information.
It achieves new state-of-the-art results on the challenging object detection task on both one- and two-stage detectors.
- Score: 76.80165589520385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We revisit the one- and two-stage detector distillation tasks and present a
simple and efficient semantic-aware framework to fill the gap between them. We
address the pixel-level imbalance problem by designing the category anchor to
produce a representative pattern for each category and regularize the
topological distance between pixels and category anchors to further tighten
their semantic bonds. We name our method SEA (SEmantic-aware Alignment)
distillation given the nature of abstracting dense fine-grained information by
semantic reliance to well facilitate distillation efficacy. SEA is well adapted
to either detection pipeline and achieves new state-of-the-art results on the
challenging COCO object detection task on both one- and two-stage detectors.
Its superior performance on instance segmentation further manifests the
generalization ability. Both 2x-distilled RetinaNet and FCOS with ResNet50-FPN
outperform their corresponding 3x ResNet101-FPN teacher, arriving 40.64 and
43.06 AP, respectively. Code will be made publicly available.
Related papers
- Auxiliary Tasks Enhanced Dual-affinity Learning for Weakly Supervised
Semantic Segmentation [79.05949524349005]
We propose AuxSegNet+, a weakly supervised auxiliary learning framework to explore the rich information from saliency maps.
We also propose a cross-task affinity learning mechanism to learn pixel-level affinities from the saliency and segmentation feature maps.
arXiv Detail & Related papers (2024-03-02T10:03:21Z) - PETDet: Proposal Enhancement for Two-Stage Fine-Grained Object Detection [26.843891792018447]
We present PETDet (Proposal Enhancement for Two-stage fine-grained object detection) to better handle the sub-tasks in two-stage FGOD methods.
An anchor-free Quality Oriented Proposal Network (QOPN) is proposed with dynamic label assignment and attention-based decomposition.
A novel Adaptive Recognition Loss (ARL) offers guidance for the R-CNN head to focus on high-quality proposals.
arXiv Detail & Related papers (2023-12-16T18:04:56Z) - Adversarial Score Distillation: When score distillation meets GAN [3.2794321281011394]
We decipher existing score distillation with the Wasserstein Generative Adversarial Network (WGAN) paradigm.
With the WGAN paradigm, we find that existing score distillation either employs a fixed sub-optimal discriminator or conducts incomplete discriminator optimization.
We propose the Adversarial Score Distillation (ASD), which maintains an optimizable discriminator and updates it using the complete optimization objective.
arXiv Detail & Related papers (2023-12-01T17:20:47Z) - G-DetKD: Towards General Distillation Framework for Object Detectors via
Contrastive and Semantic-guided Feature Imitation [49.421099172544196]
We propose a novel semantic-guided feature imitation technique, which automatically performs soft matching between feature pairs across all pyramid levels.
We also introduce contrastive distillation to effectively capture the information encoded in the relationship between different feature regions.
Our method consistently outperforms the existing detection KD techniques, and works when (1) components in the framework are used separately and in conjunction.
arXiv Detail & Related papers (2021-08-17T07:44:27Z) - Disentangle Your Dense Object Detector [82.22771433419727]
Deep learning-based dense object detectors have achieved great success in the past few years and have been applied to numerous multimedia applications such as video understanding.
However, the current training pipeline for dense detectors is compromised to lots of conjunctions that may not hold.
We propose Disentangled Dense Object Detector (DDOD), in which simple and effective disentanglement mechanisms are designed and integrated into the current state-of-the-art detectors.
arXiv Detail & Related papers (2021-07-07T00:52:16Z) - Modulating Localization and Classification for Harmonized Object
Detection [40.82723262074911]
We propose a mutual learning framework to modulate the two tasks.
In particular, the two tasks are forced to learn from each other with a novel mutual labeling strategy.
We achieve a significant performance gain over the baseline detectors on the COCO dataset.
arXiv Detail & Related papers (2021-03-16T10:36:02Z) - AFD-Net: Adaptive Fully-Dual Network for Few-Shot Object Detection [8.39479809973967]
Few-shot object detection (FSOD) aims at learning a detector that can fast adapt to previously unseen objects with scarce examples.
Existing methods solve this problem by performing subtasks of classification and localization utilizing a shared component.
We present that a general few-shot detector should consider the explicit decomposition of two subtasks, as well as leveraging information from both of them to enhance feature representations.
arXiv Detail & Related papers (2020-11-30T10:21:32Z) - Scope Head for Accurate Localization in Object Detection [135.9979405835606]
We propose a novel detector coined as ScopeNet, which models anchors of each location as a mutually dependent relationship.
With our concise and effective design, the proposed ScopeNet achieves state-of-the-art results on COCO.
arXiv Detail & Related papers (2020-05-11T04:00:09Z) - EHSOD: CAM-Guided End-to-end Hybrid-Supervised Object Detection with
Cascade Refinement [53.69674636044927]
We present EHSOD, an end-to-end hybrid-supervised object detection system.
It can be trained in one shot on both fully and weakly-annotated data.
It achieves comparable results on multiple object detection benchmarks with only 30% fully-annotated data.
arXiv Detail & Related papers (2020-02-18T08:04:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.