Radar Voxel Fusion for 3D Object Detection
- URL: http://arxiv.org/abs/2106.14087v1
- Date: Sat, 26 Jun 2021 20:34:12 GMT
- Title: Radar Voxel Fusion for 3D Object Detection
- Authors: Felix Nobis, Ehsan Shafiei, Phillip Karle, Johannes Betz and Markus
Lienkamp
- Abstract summary: This paper develops a low-level sensor fusion network for 3D object detection.
The radar sensor fusion proves especially beneficial in inclement conditions such as rain and night scenes.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automotive traffic scenes are complex due to the variety of possible
scenarios, objects, and weather conditions that need to be handled. In contrast
to more constrained environments, such as automated underground trains,
automotive perception systems cannot be tailored to a narrow field of specific
tasks but must handle an ever-changing environment with unforeseen events. As
currently no single sensor is able to reliably perceive all relevant activity
in the surroundings, sensor data fusion is applied to perceive as much
information as possible. Data fusion of different sensors and sensor modalities
on a low abstraction level enables the compensation of sensor weaknesses and
misdetections among the sensors before the information-rich sensor data are
compressed and thereby information is lost after a sensor-individual object
detection. This paper develops a low-level sensor fusion network for 3D object
detection, which fuses lidar, camera, and radar data. The fusion network is
trained and evaluated on the nuScenes data set. On the test set, fusion of
radar data increases the resulting AP (Average Precision) detection score by
about 5.1% in comparison to the baseline lidar network. The radar sensor fusion
proves especially beneficial in inclement conditions such as rain and night
scenes. Fusing additional camera data contributes positively only in
conjunction with the radar fusion, which shows that interdependencies of the
sensors are important for the detection result. Additionally, the paper
proposes a novel loss to handle the discontinuity of a simple yaw
representation for object detection. Our updated loss increases the detection
and orientation estimation performance for all sensor input configurations. The
code for this research has been made available on GitHub.
Related papers
- Learning 3D Perception from Others' Predictions [64.09115694891679]
We investigate a new scenario to construct 3D object detectors: learning from the predictions of a nearby unit that is equipped with an accurate detector.
For example, when a self-driving car enters a new area, it may learn from other traffic participants whose detectors have been optimized for that area.
arXiv Detail & Related papers (2024-10-03T16:31:28Z) - Multi-Modal 3D Object Detection by Box Matching [109.43430123791684]
We propose a novel Fusion network by Box Matching (FBMNet) for multi-modal 3D detection.
With the learned assignments between 3D and 2D object proposals, the fusion for detection can be effectively performed by combing their ROI features.
arXiv Detail & Related papers (2023-05-12T18:08:51Z) - Multimodal Dataset from Harsh Sub-Terranean Environment with Aerosol
Particles for Frontier Exploration [55.41644538483948]
This paper introduces a multimodal dataset from the harsh and unstructured underground environment with aerosol particles.
It contains synchronized raw data measurements from all onboard sensors in Robot Operating System (ROS) format.
The focus of this paper is not only to capture both temporal and spatial data diversities but also to present the impact of harsh conditions on captured data.
arXiv Detail & Related papers (2023-04-27T20:21:18Z) - On the Importance of Accurate Geometry Data for Dense 3D Vision Tasks [61.74608497496841]
Training on inaccurate or corrupt data induces model bias and hampers generalisation capabilities.
This paper investigates the effect of sensor errors for the dense 3D vision tasks of depth estimation and reconstruction.
arXiv Detail & Related papers (2023-03-26T22:32:44Z) - Edge-Aided Sensor Data Sharing in Vehicular Communication Networks [8.67588704947974]
We consider sensor data sharing and fusion in a vehicular network with both, vehicle-to-infrastructure and vehicle-to-vehicle communication.
We propose a method, named Bidirectional Feedback Noise Estimation (BiFNoE), in which an edge server collects and caches sensor measurement data from vehicles.
We show that the perception accuracy is on average improved by around 80 % with only 12 kbps uplink and 28 kbps downlink bandwidth.
arXiv Detail & Related papers (2022-06-17T16:30:56Z) - Learning Online Multi-Sensor Depth Fusion [100.84519175539378]
SenFuNet is a depth fusion approach that learns sensor-specific noise and outlier statistics.
We conduct experiments with various sensor combinations on the real-world CoRBS and Scene3D datasets.
arXiv Detail & Related papers (2022-04-07T10:45:32Z) - Multi-Modal 3D Object Detection in Autonomous Driving: a Survey [10.913958563906931]
Self-driving cars are equipped with a suite of sensors to conduct robust and accurate environment perception.
As the number and type of sensors keep increasing, combining them for better perception is becoming a natural trend.
This survey devotes to review recent fusion-based 3D detection deep learning models that leverage multiple sensor data sources.
arXiv Detail & Related papers (2021-06-24T02:52:12Z) - Complex-valued Convolutional Neural Networks for Enhanced Radar Signal
Denoising and Interference Mitigation [73.0103413636673]
We propose the use of Complex-Valued Convolutional Neural Networks (CVCNNs) to address the issue of mutual interference between radar sensors.
CVCNNs increase data efficiency, speeds up network training and substantially improves the conservation of phase information during interference removal.
arXiv Detail & Related papers (2021-04-29T10:06:29Z) - On the Role of Sensor Fusion for Object Detection in Future Vehicular
Networks [25.838878314196375]
We evaluate how using a combination of different sensors affects the detection of the environment in which the vehicles move and operate.
The final objective is to identify the optimal setup that would minimize the amount of data to be distributed over the channel.
arXiv Detail & Related papers (2021-04-23T18:58:37Z) - YOdar: Uncertainty-based Sensor Fusion for Vehicle Detection with Camera
and Radar Sensors [4.396860522241306]
We present an uncertainty-based method for sensor fusion with camera and radar data.
In our experiments we combine the YOLOv3 object detection network with a customized $1D$ radar segmentation network.
Our experiments show, that this approach of uncertainty aware fusion significantly gains performance compared to single sensor baselines.
arXiv Detail & Related papers (2020-10-07T10:40:02Z) - A Deep Learning-based Radar and Camera Sensor Fusion Architecture for
Object Detection [0.0]
This research aims to enhance current 2D object detection networks by fusing camera data and projected sparse radar data in the network layers.
The proposed CameraRadarFusionNet (CRF-Net) automatically learns at which level the fusion of the sensor data is most beneficial for the detection result.
BlackIn, a training strategy inspired by Dropout, focuses the learning on a specific sensor type.
arXiv Detail & Related papers (2020-05-15T09:28:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.