VINet: Lightweight, Scalable, and Heterogeneous Cooperative Perception
for 3D Object Detection
- URL: http://arxiv.org/abs/2212.07060v2
- Date: Wed, 22 Mar 2023 02:44:57 GMT
- Title: VINet: Lightweight, Scalable, and Heterogeneous Cooperative Perception
for 3D Object Detection
- Authors: Zhengwei Bai, Guoyuan Wu, Matthew J. Barth, Yongkang Liu, Emrah Akin
Sisbot, Kentaro Oguchi
- Abstract summary: Cooperative Perception (CP) has emerged to significantly advance the perception of automated driving.
We introduce VINet, a unified deep learning-based CP network for scalable, lightweight, and heterogeneous cooperative 3D object detection.
VINet can reduce 84% system-level computational cost and 94% system-level communication cost while improving the 3D detection accuracy.
- Score: 15.195933965761645
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Utilizing the latest advances in Artificial Intelligence (AI), the computer
vision community is now witnessing an unprecedented evolution in all kinds of
perception tasks, particularly in object detection. Based on multiple spatially
separated perception nodes, Cooperative Perception (CP) has emerged to
significantly advance the perception of automated driving. However, current
cooperative object detection methods mainly focus on ego-vehicle efficiency
without considering the practical issues of system-wide costs. In this paper,
we introduce VINet, a unified deep learning-based CP network for scalable,
lightweight, and heterogeneous cooperative 3D object detection. VINet is the
first CP method designed from the standpoint of large-scale system-level
implementation and can be divided into three main phases: 1) Global
Pre-Processing and Lightweight Feature Extraction which prepare the data into
global style and extract features for cooperation in a lightweight manner; 2)
Two-Stream Fusion which fuses the features from scalable and heterogeneous
perception nodes; and 3) Central Feature Backbone and 3D Detection Head which
further process the fused features and generate cooperative detection results.
An open-source data experimental platform is designed and developed for CP
dataset acquisition and model evaluation. The experimental analysis shows that
VINet can reduce 84% system-level computational cost and 94% system-level
communication cost while improving the 3D detection accuracy.
Related papers
- PVAFN: Point-Voxel Attention Fusion Network with Multi-Pooling Enhancing for 3D Object Detection [59.355022416218624]
integration of point and voxel representations is becoming more common in LiDAR-based 3D object detection.
We propose a novel two-stage 3D object detector, called Point-Voxel Attention Fusion Network (PVAFN)
PVAFN uses a multi-pooling strategy to integrate both multi-scale and region-specific information effectively.
arXiv Detail & Related papers (2024-08-26T19:43:01Z) - Cross-Cluster Shifting for Efficient and Effective 3D Object Detection
in Autonomous Driving [69.20604395205248]
We present a new 3D point-based detector model, named Shift-SSD, for precise 3D object detection in autonomous driving.
We introduce an intriguing Cross-Cluster Shifting operation to unleash the representation capacity of the point-based detector.
We conduct extensive experiments on the KITTI, runtime, and nuScenes datasets, and the results demonstrate the state-of-the-art performance of Shift-SSD.
arXiv Detail & Related papers (2024-03-10T10:36:32Z) - Towards Unified 3D Object Detection via Algorithm and Data Unification [70.27631528933482]
We build the first unified multi-modal 3D object detection benchmark MM- Omni3D and extend the aforementioned monocular detector to its multi-modal version.
We name the designed monocular and multi-modal detectors as UniMODE and MM-UniMODE, respectively.
arXiv Detail & Related papers (2024-02-28T18:59:31Z) - SiCP: Simultaneous Individual and Cooperative Perception for 3D Object Detection in Connected and Automated Vehicles [18.23919432049492]
Cooperative perception for connected and automated vehicles is traditionally achieved through the fusion of feature maps from two or more vehicles.
This drawback impedes the adoption of cooperative perception as vehicle resources are often insufficient to concurrently employ two perception models.
We present Simultaneous Individual and Cooperative Perception (SiCP), a generic framework that supports a wide range of the state-of-the-art standalone perception backbones.
arXiv Detail & Related papers (2023-12-08T04:12:26Z) - V2X-AHD:Vehicle-to-Everything Cooperation Perception via Asymmetric
Heterogenous Distillation Network [13.248981195106069]
We propose a multi-view vehicle-road cooperation perception system, vehicle-to-everything cooperative perception (V2X-AHD)
The V2X-AHD can effectively improve the accuracy of 3D object detection and reduce the number of network parameters, according to this study.
arXiv Detail & Related papers (2023-10-10T13:12:03Z) - DeepSTEP -- Deep Learning-Based Spatio-Temporal End-To-End Perception
for Autonomous Vehicles [0.0]
We present our concept for an end-to-end perception architecture, named DeepSTEP.
DeepSTEP processes raw sensor data from the camera, LiDAR, and RaDAR, and combines the extracted data in a deep fusion network.
The architecture's end-to-end design, time-aware attention mechanism, and integration of multiple perception tasks make it a promising solution for real-world deployment.
arXiv Detail & Related papers (2023-05-11T14:13:37Z) - Multi-level and multi-modal feature fusion for accurate 3D object
detection in Connected and Automated Vehicles [0.8701566919381223]
This paper presents a Deep Neural Network based 3D object detection model that leverages a three-stage feature extractor.
The proposed feature extractor extracts high-level features from two input sensory modalities and recovers the important features discarded during the convolutional process.
The novel fusion scheme effectively fuses features across sensory modalities and convolutional layers to find the best representative global features.
arXiv Detail & Related papers (2022-12-15T00:25:05Z) - The Devil is in the Task: Exploiting Reciprocal Appearance-Localization
Features for Monocular 3D Object Detection [62.1185839286255]
Low-cost monocular 3D object detection plays a fundamental role in autonomous driving.
We introduce a Dynamic Feature Reflecting Network, named DFR-Net.
We rank 1st among all the monocular 3D object detectors in the KITTI test set.
arXiv Detail & Related papers (2021-12-28T07:31:18Z) - PC-RGNN: Point Cloud Completion and Graph Neural Network for 3D Object
Detection [57.49788100647103]
LiDAR-based 3D object detection is an important task for autonomous driving.
Current approaches suffer from sparse and partial point clouds of distant and occluded objects.
In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions.
arXiv Detail & Related papers (2020-12-18T18:06:43Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z) - Selective Convolutional Network: An Efficient Object Detector with
Ignoring Background [28.591619763438054]
We introduce an efficient object detector called Selective Convolutional Network (SCN), which selectively calculates only on the locations that contain meaningful and conducive information.
To solve it, we design an elaborate structure with negligible overheads to guide the network where to look next.
arXiv Detail & Related papers (2020-02-04T10:07:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.