Towards Autonomous Driving: a Multi-Modal 360$^{\circ}$ Perception
Proposal
- URL: http://arxiv.org/abs/2008.09672v1
- Date: Fri, 21 Aug 2020 20:36:21 GMT
- Title: Towards Autonomous Driving: a Multi-Modal 360$^{\circ}$ Perception
Proposal
- Authors: Jorge Beltr\'an, Carlos Guindel, Irene Cort\'es, Alejandro Barrera,
Armando Astudillo, Jes\'us Urdiales, Mario \'Alvarez, Farid Bekka, Vicente
Milan\'es, and Fernando Garc\'ia
- Abstract summary: This paper presents a framework for 3D object detection and tracking for autonomous vehicles.
The solution, based on a novel sensor fusion configuration, provides accurate and reliable road environment detection.
A variety of tests of the system, deployed in an autonomous vehicle, have successfully assessed the suitability of the proposed perception stack.
- Score: 87.11988786121447
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, a multi-modal 360$^{\circ}$ framework for 3D object detection
and tracking for autonomous vehicles is presented. The process is divided into
four main stages. First, images are fed into a CNN network to obtain instance
segmentation of the surrounding road participants. Second, LiDAR-to-image
association is performed for the estimated mask proposals. Then, the isolated
points of every object are processed by a PointNet ensemble to compute their
corresponding 3D bounding boxes and poses. Lastly, a tracking stage based on
Unscented Kalman Filter is used to track the agents along time. The solution,
based on a novel sensor fusion configuration, provides accurate and reliable
road environment detection. A wide variety of tests of the system, deployed in
an autonomous vehicle, have successfully assessed the suitability of the
proposed perception stack in a real autonomous driving application.
Related papers
- Multi-Object Tracking with Camera-LiDAR Fusion for Autonomous Driving [0.764971671709743]
The proposed MOT algorithm comprises a three-step association process, an Extended Kalman filter for estimating the motion of each detected dynamic obstacle, and a track management phase.
Unlike most state-of-the-art multi-modal MOT approaches, the proposed algorithm does not rely on maps or knowledge of the ego global pose.
The algorithm is validated both in simulation and with real-world data, with satisfactory results.
arXiv Detail & Related papers (2024-03-06T23:49:16Z) - End-to-End 3D Object Detection using LiDAR Point Cloud [0.0]
We present an approach wherein, using a novel encoding of the LiDAR point cloud we infer the location of different classes near the autonomous vehicles.
The output is predictions about the location and orientation of objects in the scene in form of 3D bounding boxes and labels of scene objects.
arXiv Detail & Related papers (2023-12-24T00:52:14Z) - Multi-Modal 3D Object Detection by Box Matching [109.43430123791684]
We propose a novel Fusion network by Box Matching (FBMNet) for multi-modal 3D detection.
With the learned assignments between 3D and 2D object proposals, the fusion for detection can be effectively performed by combing their ROI features.
arXiv Detail & Related papers (2023-05-12T18:08:51Z) - 3DMODT: Attention-Guided Affinities for Joint Detection & Tracking in 3D
Point Clouds [95.54285993019843]
We propose a method for joint detection and tracking of multiple objects in 3D point clouds.
Our model exploits temporal information employing multiple frames to detect objects and track them in a single network.
arXiv Detail & Related papers (2022-11-01T20:59:38Z) - A Simple Baseline for Multi-Camera 3D Object Detection [94.63944826540491]
3D object detection with surrounding cameras has been a promising direction for autonomous driving.
We present SimMOD, a Simple baseline for Multi-camera Object Detection.
We conduct extensive experiments on the 3D object detection benchmark of nuScenes to demonstrate the effectiveness of SimMOD.
arXiv Detail & Related papers (2022-08-22T03:38:01Z) - Exploring Simple 3D Multi-Object Tracking for Autonomous Driving [10.921208239968827]
3D multi-object tracking in LiDAR point clouds is a key ingredient for self-driving vehicles.
Existing methods are predominantly based on the tracking-by-detection pipeline and inevitably require a matching step for the detection association.
We present SimTrack to simplify the hand-crafted tracking paradigm by proposing an end-to-end trainable model for joint detection and tracking from raw point clouds.
arXiv Detail & Related papers (2021-08-23T17:59:22Z) - CFTrack: Center-based Radar and Camera Fusion for 3D Multi-Object
Tracking [9.62721286522053]
We propose an end-to-end network for joint object detection and tracking based on radar and camera sensor fusion.
Our proposed method uses a center-based radar-camera fusion algorithm for object detection and utilizes a greedy algorithm for object association.
We evaluate our method on the challenging nuScenes dataset, where it achieves 20.0 AMOTA and outperforms all vision-based 3D tracking methods in the benchmark.
arXiv Detail & Related papers (2021-07-11T23:56:53Z) - PerMO: Perceiving More at Once from a Single Image for Autonomous
Driving [76.35684439949094]
We present a novel approach to detect, segment, and reconstruct complete textured 3D models of vehicles from a single image.
Our approach combines the strengths of deep learning and the elegance of traditional techniques.
We have integrated these algorithms with an autonomous driving system.
arXiv Detail & Related papers (2020-07-16T05:02:45Z) - MVLidarNet: Real-Time Multi-Class Scene Understanding for Autonomous
Driving Using Multiple Views [60.538802124885414]
We present Multi-View LidarNet (MVLidarNet), a two-stage deep neural network for multi-class object detection and drivable space segmentation.
MVLidarNet is able to detect and classify objects while simultaneously determining the drivable space using a single LiDAR scan as input.
We show results on both KITTI and a much larger internal dataset, thus demonstrating the method's ability to scale by an order of magnitude.
arXiv Detail & Related papers (2020-06-09T21:28:17Z) - Ego-motion and Surrounding Vehicle State Estimation Using a Monocular
Camera [11.29865843123467]
We propose a novel machine learning method to estimate ego-motion and surrounding vehicle state using a single monocular camera.
Our approach is based on a combination of three deep neural networks to estimate the 3D vehicle bounding box, depth, and optical flow from a sequence of images.
arXiv Detail & Related papers (2020-05-04T16:41:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.