FastPillars: A Deployment-friendly Pillar-based 3D Detector
- URL: http://arxiv.org/abs/2302.02367v6
- Date: Wed, 13 Dec 2023 08:55:27 GMT
- Title: FastPillars: A Deployment-friendly Pillar-based 3D Detector
- Authors: Sifan Zhou, Zhi Tian, Xiangxiang Chu, Xinyu Zhang, Bo Zhang, Xiaobo
Lu, Chengjian Feng, Zequn Jie, Patrick Yin Chiang, Lin Ma
- Abstract summary: Existing BEV-based (i.e., Bird Eye View) detectors favor sparse convolutions (known as SPConv) to speed up training and inference.
FastPillars delivers state-of-the-art accuracy on Open dataset with 1.8X speed up and 3.8 mAPH/L2 improvement over CenterPoint (SPConv-based)
- Score: 63.0697065653061
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The deployment of 3D detectors strikes one of the major challenges in
real-world self-driving scenarios. Existing BEV-based (i.e., Bird Eye View)
detectors favor sparse convolutions (known as SPConv) to speed up training and
inference, which puts a hard barrier for deployment, especially for on-device
applications. In this paper, to tackle the challenge of efficient 3D object
detection from an industry perspective, we devise a deployment-friendly
pillar-based 3D detector, termed FastPillars. First, we introduce a novel
lightweight Max-and-Attention Pillar Encoding (MAPE) module specially for
enhancing small 3D objects. Second, we propose a simple yet effective principle
for designing a backbone in pillar-based 3D detection. We construct FastPillars
based on these designs, achieving high performance and low latency without
SPConv. Extensive experiments on two large-scale datasets demonstrate the
effectiveness and efficiency of FastPillars for on-device 3D detection
regarding both performance and speed. Specifically, FastPillars delivers
state-of-the-art accuracy on Waymo Open Dataset with 1.8X speed up and 3.8
mAPH/L2 improvement over CenterPoint (SPConv-based). Our code is publicly
available at: https://github.com/StiphyJay/FastPillars.
Related papers
- PillarTrack: Redesigning Pillar-based Transformer Network for Single Object Tracking on Point Clouds [5.524413892353708]
LiDAR-based 3D single object tracking (3D SOT) is a critical issue in robotics and autonomous driving.
We propose PillarTrack, a pillar-based 3D single object tracking framework.
PillarTrack achieves state-of-the-art performance on the KITTI and nuScenes dataset and enables real-time tracking speed.
arXiv Detail & Related papers (2024-04-11T06:06:56Z) - FocalFormer3D : Focusing on Hard Instance for 3D Object Detection [97.56185033488168]
False negatives (FN) in 3D object detection can lead to potentially dangerous situations in autonomous driving.
In this work, we propose Hard Instance Probing (HIP), a general pipeline that identifies textitFN in a multi-stage manner.
We instantiate this method as FocalFormer3D, a simple yet effective detector that excels at excavating difficult objects.
arXiv Detail & Related papers (2023-08-08T20:06:12Z) - PVNAS: 3D Neural Architecture Search with Point-Voxel Convolution [26.059213743430192]
We study 3D deep learning from the efficiency perspective.
We propose a novel hardware-efficient 3D primitive, Point-Voxel Convolution (PVConv)
arXiv Detail & Related papers (2022-04-25T17:13:55Z) - A Lightweight and Detector-free 3D Single Object Tracker on Point Clouds [50.54083964183614]
It is non-trivial to perform accurate target-specific detection since the point cloud of objects in raw LiDAR scans is usually sparse and incomplete.
We propose DMT, a Detector-free Motion prediction based 3D Tracking network that totally removes the usage of complicated 3D detectors.
arXiv Detail & Related papers (2022-03-08T17:49:07Z) - PiFeNet: Pillar-Feature Network for Real-Time 3D Pedestrian Detection
from Point Cloud [64.12626752721766]
We present PiFeNet, an efficient real-time 3D detector for pedestrian detection from point clouds.
We address two challenges that 3D object detection frameworks encounter when detecting pedestrians: low of pillar features and small occupation areas of pedestrians in point clouds.
Our approach is ranked 1st in KITTI pedestrian BEV and 3D leaderboards while running at 26 frames per second (FPS), and achieves state-of-the-art performance on Nuscenes detection benchmark.
arXiv Detail & Related papers (2021-12-31T13:41:37Z) - Embracing Single Stride 3D Object Detector with Sparse Transformer [63.179720817019096]
In LiDAR-based 3D object detection for autonomous driving, the ratio of the object size to input scene size is significantly smaller compared to 2D detection cases.
Many 3D detectors directly follow the common practice of 2D detectors, which downsample the feature maps even after quantizing the point clouds.
We propose Single-stride Sparse Transformer (SST) to maintain the original resolution from the beginning to the end of the network.
arXiv Detail & Related papers (2021-12-13T02:12:02Z) - Fast and Furious: Real Time End-to-End 3D Detection, Tracking and Motion
Forecasting with a Single Convolutional Net [93.51773847125014]
We propose a novel deep neural network that is able to jointly reason about 3D detection, tracking and motion forecasting given data captured by a 3D sensor.
Our approach performs 3D convolutions across space and time over a bird's eye view representation of the 3D world.
arXiv Detail & Related papers (2020-12-22T22:43:35Z) - Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution [34.713667358316286]
Self-driving cars need to understand 3D scenes efficiently and accurately in order to drive safely.
Existing 3D perception models are not able to recognize small instances very well due to the low-resolution voxelization and aggressive downsampling.
We propose Sparse Point-Voxel Convolution (SPVConv), a lightweight 3D module that equips the vanilla Sparse Convolution with the high-resolution point-based branch.
arXiv Detail & Related papers (2020-07-31T14:27:27Z) - Boundary-Aware Dense Feature Indicator for Single-Stage 3D Object
Detection from Point Clouds [32.916690488130506]
We propose a universal module that helps 3D detectors focus on the densest region of the point clouds in a boundary-aware manner.
Experiments on KITTI dataset show that DENFI improves the performance of the baseline single-stage detector remarkably.
arXiv Detail & Related papers (2020-04-01T01:21:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.