itKD: Interchange Transfer-based Knowledge Distillation for 3D Object
Detection
- URL: http://arxiv.org/abs/2205.15531v2
- Date: Mon, 27 Mar 2023 04:30:25 GMT
- Title: itKD: Interchange Transfer-based Knowledge Distillation for 3D Object
Detection
- Authors: Hyeon Cho, Junyong Choi, Geonwoo Baek, Wonjun Hwang
- Abstract summary: We propose an autoencoder-style framework comprising channel-wise compression and decompression.
To learn the map-view feature of a teacher network, the features from teacher and student networks are independently passed through the shared autoencoder.
We present an head attention loss to match the 3D object detection information drawn by the multi-head self-attention mechanism.
- Score: 3.735965959270874
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Point-cloud based 3D object detectors recently have achieved remarkable
progress. However, most studies are limited to the development of network
architectures for improving only their accuracy without consideration of the
computational efficiency. In this paper, we first propose an autoencoder-style
framework comprising channel-wise compression and decompression via interchange
transfer-based knowledge distillation. To learn the map-view feature of a
teacher network, the features from teacher and student networks are
independently passed through the shared autoencoder; here, we use a compressed
representation loss that binds the channel-wised compression knowledge from
both student and teacher networks as a kind of regularization. The decompressed
features are transferred in opposite directions to reduce the gap in the
interchange reconstructions. Lastly, we present an head attention loss to match
the 3D object detection information drawn by the multi-head self-attention
mechanism. Through extensive experiments, we verify that our method can train
the lightweight model that is well-aligned with the 3D point cloud detection
task and we demonstrate its superiority using the well-known public datasets;
e.g., Waymo and nuScenes.
Related papers
- Cross-Cluster Shifting for Efficient and Effective 3D Object Detection
in Autonomous Driving [69.20604395205248]
We present a new 3D point-based detector model, named Shift-SSD, for precise 3D object detection in autonomous driving.
We introduce an intriguing Cross-Cluster Shifting operation to unleash the representation capacity of the point-based detector.
We conduct extensive experiments on the KITTI, runtime, and nuScenes datasets, and the results demonstrate the state-of-the-art performance of Shift-SSD.
arXiv Detail & Related papers (2024-03-10T10:36:32Z) - Unleash the Potential of Image Branch for Cross-modal 3D Object
Detection [67.94357336206136]
We present a new cross-modal 3D object detector, namely UPIDet, which aims to unleash the potential of the image branch from two aspects.
First, UPIDet introduces a new 2D auxiliary task called normalized local coordinate map estimation.
Second, we discover that the representational capability of the point cloud backbone can be enhanced through the gradients backpropagated from the training objectives of the image branch.
arXiv Detail & Related papers (2023-01-22T08:26:58Z) - ALSO: Automotive Lidar Self-supervision by Occupancy estimation [70.70557577874155]
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds.
The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled.
The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information.
arXiv Detail & Related papers (2022-12-12T13:10:19Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - Paint and Distill: Boosting 3D Object Detection with Semantic Passing
Network [70.53093934205057]
3D object detection task from lidar or camera sensors is essential for autonomous driving.
We propose a novel semantic passing framework, named SPNet, to boost the performance of existing lidar-based 3D detection models.
arXiv Detail & Related papers (2022-07-12T12:35:34Z) - Self-Ensemling for 3D Point Cloud Domain Adaption [29.330315360307374]
We propose an end-to-end self-ensembling network (SEN) for 3D point cloud domain adaption tasks.
Our SEN resorts to the advantages of Mean Teacher and semi-supervised learning, and introduces a soft classification loss and a consistency loss.
Our SEN outperforms the state-of-the-art methods on both classification and segmentation tasks.
arXiv Detail & Related papers (2021-12-10T02:18:09Z) - AttDLNet: Attention-based DL Network for 3D LiDAR Place Recognition [0.6352264764099531]
This paper proposes a novel 3D LiDAR-based deep learning network named AttDLNet.
It exploits an attention mechanism to selectively focus on long-range context and interfeature relationships.
Results show that the encoder network features are already very descriptive, but adding attention to the network further improves performance.
arXiv Detail & Related papers (2021-06-17T16:34:37Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.