Instance-Conditional Knowledge Distillation for Object Detection
- URL: http://arxiv.org/abs/2110.12724v1
- Date: Mon, 25 Oct 2021 08:23:29 GMT
- Title: Instance-Conditional Knowledge Distillation for Object Detection
- Authors: Zijian Kang, Peizhen Zhang, Xiangyu Zhang, Jian Sun, Nanning Zheng
- Abstract summary: We propose an instance-conditional distillation framework to find desired knowledge.
We use observed instances as condition information and formulate the retrieval process as an instance-conditional decoding process.
- Score: 59.56780046291835
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the success of Knowledge Distillation (KD) on image classification,
it is still challenging to apply KD on object detection due to the difficulty
in locating knowledge. In this paper, we propose an instance-conditional
distillation framework to find desired knowledge. To locate knowledge of each
instance, we use observed instances as condition information and formulate the
retrieval process as an instance-conditional decoding process. Specifically,
information of each instance that specifies a condition is encoded as query,
and teacher's information is presented as key, we use the attention between
query and key to measure the correlation, formulated by the transformer
decoder. To guide this module, we further introduce an auxiliary task that
directs to instance localization and identification, which are fundamental for
detection. Extensive experiments demonstrate the efficacy of our method: we
observe impressive improvements under various settings. Notably, we boost
RetinaNet with ResNet-50 backbone from 37.4 to 40.7 mAP (+3.3) under 1x
schedule, that even surpasses the teacher (40.4 mAP) with ResNet-101 backbone
under 3x schedule. Code will be released soon.
Related papers
- Learning Lightweight Object Detectors via Multi-Teacher Progressive
Distillation [56.053397775016755]
We propose a sequential approach to knowledge distillation that progressively transfers the knowledge of a set of teacher detectors to a given lightweight student.
To the best of our knowledge, we are the first to successfully distill knowledge from Transformer-based teacher detectors to convolution-based students.
arXiv Detail & Related papers (2023-08-17T17:17:08Z) - Learning to Detect Instance-level Salient Objects Using Complementary
Image Labels [55.049347205603304]
We present the first weakly-supervised approach to the salient instance detection problem.
We propose a novel weakly-supervised network with three branches: a Saliency Detection Branch leveraging class consistency information to locate candidate objects; a Boundary Detection Branch exploiting class discrepancy information to delineate object boundaries; and a Centroid Detection Branch using subitizing information to detect salient instance centroids.
arXiv Detail & Related papers (2021-11-19T10:15:22Z) - Beyond Classification: Knowledge Distillation using Multi-Object
Impressions [17.214664783818687]
Knowledge Distillation (KD) utilizes training data as a transfer set to transfer knowledge from a complex network (Teacher) to a smaller network (Student)
Several works have recently identified many scenarios where the training data may not be available due to data privacy or sensitivity concerns.
We, for the first time, solve a much more challenging problem, i.e., "KD for object detection with zero knowledge about the training data and its statistics"
arXiv Detail & Related papers (2021-10-27T06:59:27Z) - Deep Structured Instance Graph for Distilling Object Detectors [82.16270736573176]
We present a simple knowledge structure to exploit and encode information inside the detection system to facilitate detector knowledge distillation.
We achieve new state-of-the-art results on the challenging COCO object detection task with diverse student-teacher pairs on both one- and two-stage detectors.
arXiv Detail & Related papers (2021-09-27T08:26:00Z) - Distilling Object Detectors via Decoupled Features [69.62967325617632]
We present a novel distillation algorithm via decoupled features (DeFeat) for learning a better student detector.
Experiments on various detectors with different backbones show that the proposed DeFeat is able to surpass the state-of-the-art distillation methods for object detection.
arXiv Detail & Related papers (2021-03-26T13:58:49Z) - General Instance Distillation for Object Detection [12.720908566642812]
RetinaNet with ResNet-50 achieves 39.1% in mAP with GID on dataset, which surpasses the baseline 36.2% by 2.9%, and even better than the ResNet-101 based teacher model with 38.1% AP.
arXiv Detail & Related papers (2021-03-03T11:41:26Z) - Weakly-supervised Salient Instance Detection [65.0408760733005]
We present the first weakly-supervised approach to the salient instance detection problem.
We propose a novel weakly-supervised network with three branches: a Saliency Detection Branch leveraging class consistency information to locate candidate objects; a Boundary Detection Branch exploiting class discrepancy information to delineate object boundaries; and a Centroid Detection Branch using subitizing information to detect salient instance centroids.
arXiv Detail & Related papers (2020-09-29T09:47:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.