PointDistiller: Structured Knowledge Distillation Towards Efficient and
Compact 3D Detection
- URL: http://arxiv.org/abs/2205.11098v1
- Date: Mon, 23 May 2022 07:40:07 GMT
- Title: PointDistiller: Structured Knowledge Distillation Towards Efficient and
Compact 3D Detection
- Authors: Linfeng Zhang, Runpei Dong, Hung-Shuo Tai, Kaisheng Ma
- Abstract summary: This paper proposes PointDistiller, a structured knowledge distillation framework for point clouds-based 3D detection.
PointDistiller includes local distillation which extracts and distills the local geometric structure of point clouds with dynamic graph convolution and reweighted learning strategy.
Our 4X compressed PointPillars student achieves 2.8 and 3.4 mAP improvements on BEV and 3D object detection, outperforming its teacher by 0.9 and 1.8 mAP, respectively.
- Score: 15.79799516495951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The remarkable breakthroughs in point cloud representation learning have
boosted their usage in real-world applications such as self-driving cars and
virtual reality. However, these applications usually have an urgent requirement
for not only accurate but also efficient 3D object detection. Recently,
knowledge distillation has been proposed as an effective model compression
technique, which transfers the knowledge from an over-parameterized teacher to
a lightweight student and achieves consistent effectiveness in 2D vision.
However, due to point clouds' sparsity and irregularity, directly applying
previous image-based knowledge distillation methods to point cloud detectors
usually leads to unsatisfactory performance. To fill the gap, this paper
proposes PointDistiller, a structured knowledge distillation framework for
point clouds-based 3D detection. Concretely, PointDistiller includes local
distillation which extracts and distills the local geometric structure of point
clouds with dynamic graph convolution and reweighted learning strategy, which
highlights student learning on the crucial points or voxels to improve
knowledge distillation efficiency. Extensive experiments on both voxels-based
and raw points-based detectors have demonstrated the effectiveness of our
method over seven previous knowledge distillation methods. For instance, our 4X
compressed PointPillars student achieves 2.8 and 3.4 mAP improvements on BEV
and 3D object detection, outperforming its teacher by 0.9 and 1.8 mAP,
respectively. Codes have been released at
https://github.com/RunpeiDong/PointDistiller.
Related papers
- Topology-Guided Knowledge Distillation for Efficient Point Cloud Processing [3.3903891679981593]
This work introduces a novel distillation framework to transfer knowledge from a high-capacity teacher to a lightweight student model.<n>Our approach captures the underlying geometric structures of point clouds while selectively guiding the student model's learning process.<n>Our method achieves state-of-the-art performance among knowledge distillation techniques trained solely on LiDAR data.
arXiv Detail & Related papers (2025-05-12T22:15:54Z) - CLoCKDistill: Consistent Location-and-Context-aware Knowledge Distillation for DETRs [2.7624021966289605]
This paper proposes Consistent Location-and-Context-aware Knowledge Distillation (CLoCKDistill) for DETR detectors.
We distill the transformer encoder output (i.e., memory) that contains valuable global context and long-range dependencies.
Our method boosts student detector performance by 2.2% to 6.4%.
arXiv Detail & Related papers (2025-02-15T06:02:51Z) - Gradient-Guided Knowledge Distillation for Object Detectors [3.236217153362305]
We propose a novel approach for knowledge distillation in object detection, named Gradient-guided Knowledge Distillation (GKD)
Our GKD uses gradient information to identify and assign more weights to features that significantly impact the detection loss, allowing the student to learn the most relevant features from the teacher.
Experiments on the KITTI and COCO-Traffic datasets demonstrate our method's efficacy in knowledge distillation for object detection.
arXiv Detail & Related papers (2023-03-07T21:09:09Z) - 3D Point Cloud Pre-training with Knowledge Distillation from 2D Images [128.40422211090078]
We propose a knowledge distillation method for 3D point cloud pre-trained models to acquire knowledge directly from the 2D representation learning model.
Specifically, we introduce a cross-attention mechanism to extract concept features from 3D point cloud and compare them with the semantic information from 2D images.
In this scheme, the point cloud pre-trained models learn directly from rich information contained in 2D teacher models.
arXiv Detail & Related papers (2022-12-17T23:21:04Z) - Let Images Give You More:Point Cloud Cross-Modal Training for Shape
Analysis [43.13887916301742]
This paper introduces a simple but effective point cloud cross-modality training (PointCMT) strategy to boost point cloud analysis.
To effectively acquire auxiliary knowledge from view images, we develop a teacher-student framework and formulate the cross modal learning as a knowledge distillation problem.
We verify significant gains on various datasets using appealing backbones, i.e., equipped with PointCMT, PointNet++ and PointMLP.
arXiv Detail & Related papers (2022-10-09T09:35:22Z) - Exploring Inconsistent Knowledge Distillation for Object Detection with
Data Augmentation [66.25738680429463]
Knowledge Distillation (KD) for object detection aims to train a compact detector by transferring knowledge from a teacher model.
We propose inconsistent knowledge distillation (IKD) which aims to distill knowledge inherent in the teacher model's counter-intuitive perceptions.
Our method outperforms state-of-the-art KD baselines on one-stage, two-stage and anchor-free object detectors.
arXiv Detail & Related papers (2022-09-20T16:36:28Z) - itKD: Interchange Transfer-based Knowledge Distillation for 3D Object
Detection [3.735965959270874]
We propose an autoencoder-style framework comprising channel-wise compression and decompression.
To learn the map-view feature of a teacher network, the features from teacher and student networks are independently passed through the shared autoencoder.
We present an head attention loss to match the 3D object detection information drawn by the multi-head self-attention mechanism.
arXiv Detail & Related papers (2022-05-31T04:25:37Z) - Localization Distillation for Object Detection [134.12664548771534]
Previous knowledge distillation (KD) methods for object detection mostly focus on feature imitation instead of mimicking the classification logits.
We present a novel localization distillation (LD) method which can efficiently transfer the localization knowledge from the teacher to the student.
We show that logit mimicking can outperform feature imitation and the absence of localization distillation is a critical reason for why logit mimicking underperforms for years.
arXiv Detail & Related papers (2022-04-12T17:14:34Z) - Learning-based Point Cloud Registration for 6D Object Pose Estimation in
the Real World [55.7340077183072]
We tackle the task of estimating the 6D pose of an object from point cloud data.
Recent learning-based approaches to addressing this task have shown great success on synthetic datasets.
We analyze the causes of these failures, which we trace back to the difference between the feature distributions of the source and target point clouds.
arXiv Detail & Related papers (2022-03-29T07:55:04Z) - What Stops Learning-based 3D Registration from Working in the Real
World? [53.68326201131434]
This work identifies the sources of 3D point cloud registration failures, analyze the reasons behind them, and propose solutions.
Ultimately, this translates to a best-practice 3D registration network (BPNet), constituting the first learning-based method able to handle previously-unseen objects in real-world data.
Our model generalizes to real data without any fine-tuning, reaching an accuracy of up to 67% on point clouds of unseen objects obtained with a commercial sensor.
arXiv Detail & Related papers (2021-11-19T19:24:27Z) - Empowering Knowledge Distillation via Open Set Recognition for Robust 3D
Point Cloud Classification [20.591508284285368]
We propose a joint Knowledge Distillation and Open Set recognition training methodology for three-dimensional object recognition.
We demonstrate the effectiveness of the proposed method via various experiments on how it allows us to obtain a much smaller model.
arXiv Detail & Related papers (2020-10-25T13:26:48Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.