Scalable Primitives for Generalized Sensor Fusion in Autonomous Vehicles
- URL: http://arxiv.org/abs/2112.00219v1
- Date: Wed, 1 Dec 2021 01:43:15 GMT
- Title: Scalable Primitives for Generalized Sensor Fusion in Autonomous Vehicles
- Authors: Sammy Sidhu, Linda Wang, Tayyab Naseer, Ashish Malhotra, Jay Chia,
Aayush Ahuja, Ella Rasmussen, Qiangui Huang, and Ray Gao
- Abstract summary: Generalized Sensor Fusion (GSF) is designed in such a way that both sensor inputs and target tasks are modular and modifiable.
This enables AV system designers to easily experiment with different sensor configurations and methods and opens up the ability to deploy on heterogeneous fleets.
- Score: 3.7543422202019427
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In autonomous driving, there has been an explosion in the use of deep neural
networks for perception, prediction and planning tasks. As autonomous vehicles
(AVs) move closer to production, multi-modal sensor inputs and heterogeneous
vehicle fleets with different sets of sensor platforms are becoming
increasingly common in the industry. However, neural network architectures
typically target specific sensor platforms and are not robust to changes in
input, making the problem of scaling and model deployment particularly
difficult. Furthermore, most players still treat the problem of optimizing
software and hardware as entirely independent problems. We propose a new end to
end architecture, Generalized Sensor Fusion (GSF), which is designed in such a
way that both sensor inputs and target tasks are modular and modifiable. This
enables AV system designers to easily experiment with different sensor
configurations and methods and opens up the ability to deploy on heterogeneous
fleets using the same models that are shared across a large engineering
organization. Using this system, we report experimental results where we
demonstrate near-parity of an expensive high-density (HD) LiDAR sensor with a
cheap low-density (LD) LiDAR plus camera setup in the 3D object detection task.
This paves the way for the industry to jointly design hardware and software
architectures as well as large fleets with heterogeneous configurations.
Related papers
- Graph-Based Multi-Modal Sensor Fusion for Autonomous Driving [3.770103075126785]
We introduce a novel approach to multi-modal sensor fusion, focusing on developing a graph-based state representation.
We present a Sensor-Agnostic Graph-Aware Kalman Filter, the first online state estimation technique designed to fuse multi-modal graphs.
We validate the effectiveness of our proposed framework through extensive experiments conducted on both synthetic and real-world driving datasets.
arXiv Detail & Related papers (2024-11-06T06:58:17Z) - Condition-Aware Multimodal Fusion for Robust Semantic Perception of Driving Scenes [56.52618054240197]
We propose a novel, condition-aware multimodal fusion approach for robust semantic perception of driving scenes.
Our method, CAFuser, uses an RGB camera input to classify environmental conditions and generate a Condition Token that guides the fusion of multiple sensor modalities.
We set the new state of the art with CAFuser on the MUSES dataset with 59.7 PQ for multimodal panoptic segmentation and 78.2 mIoU for semantic segmentation, ranking first on the public benchmarks.
arXiv Detail & Related papers (2024-10-14T17:56:20Z) - Orbital AI-based Autonomous Refuelling Solution [6.776059370975249]
This paper documents the development of a proposed AI-based navigation algorithm intending to mature the use of on-board visible wavelength cameras as a main sensor for docking and on-orbit servicing (OOS)
Multiple convolutional neural network backbone architectures are benchmarked on synthetically generated data of docking manoeuvres with the International Space Station (ISS)
The integration of the solution with a physical prototype of the refuelling mechanism is validated in laboratory using a robotic arm to simulate a berthing procedure.
arXiv Detail & Related papers (2023-09-20T21:25:52Z) - RMMDet: Road-Side Multitype and Multigroup Sensor Detection System for
Autonomous Driving [3.8917150802484994]
RMMDet is a road-side multitype and multigroup sensor detection system for autonomous driving.
We use a ROS-based virtual environment to simulate real-world conditions.
We produce local datasets and real sand table field, and conduct various experiments.
arXiv Detail & Related papers (2023-03-09T12:13:39Z) - HUM3DIL: Semi-supervised Multi-modal 3D Human Pose Estimation for
Autonomous Driving [95.42203932627102]
3D human pose estimation is an emerging technology, which can enable the autonomous vehicle to perceive and understand the subtle and complex behaviors of pedestrians.
Our method efficiently makes use of these complementary signals, in a semi-supervised fashion and outperforms existing methods with a large margin.
Specifically, we embed LiDAR points into pixel-aligned multi-modal features, which we pass through a sequence of Transformer refinement stages.
arXiv Detail & Related papers (2022-12-15T11:15:14Z) - FUTR3D: A Unified Sensor Fusion Framework for 3D Detection [18.70932813595532]
We propose the first unified end-to-end sensor fusion framework for 3D detection, namedR3D, which can be used in almost any sensor configuration.
R3D employs a query-based Modality-Agnostic Feature Sampler (MAFS), together with a transformer decoder with a set-to-set loss for 3D detection.
On NuScenes dataset,R3D achieves better performance over specifically designed methods across different sensor combinations.
arXiv Detail & Related papers (2022-03-20T20:41:55Z) - Bayesian Imitation Learning for End-to-End Mobile Manipulation [80.47771322489422]
Augmenting policies with additional sensor inputs, such as RGB + depth cameras, is a straightforward approach to improving robot perception capabilities.
We show that using the Variational Information Bottleneck to regularize convolutional neural networks improves generalization to held-out domains.
We demonstrate that our method is able to help close the sim-to-real gap and successfully fuse RGB and depth modalities.
arXiv Detail & Related papers (2022-02-15T17:38:30Z) - Lite-HDSeg: LiDAR Semantic Segmentation Using Lite Harmonic Dense
Convolutions [2.099922236065961]
We present Lite-HDSeg, a novel real-time convolutional neural network for semantic segmentation of full $3$D LiDAR point clouds.
Our experimental results show that the proposed method outperforms state-of-the-art semantic segmentation approaches which can run real-time.
arXiv Detail & Related papers (2021-03-16T04:54:57Z) - SensiX: A Platform for Collaborative Machine Learning on the Edge [69.1412199244903]
We present SensiX, a personal edge platform that stays between sensor data and sensing models.
We demonstrate its efficacy in developing motion and audio-based multi-device sensing systems.
Our evaluation shows that SensiX offers a 7-13% increase in overall accuracy and up to 30% increase across different environment dynamics at the expense of 3mW power overhead.
arXiv Detail & Related papers (2020-12-04T23:06:56Z) - Deep Soft Procrustes for Markerless Volumetric Sensor Alignment [81.13055566952221]
In this work, we improve markerless data-driven correspondence estimation to achieve more robust multi-sensor spatial alignment.
We incorporate geometric constraints in an end-to-end manner into a typical segmentation based model and bridge the intermediate dense classification task with the targeted pose estimation one.
Our model is experimentally shown to achieve similar results with marker-based methods and outperform the markerless ones, while also being robust to the pose variations of the calibration structure.
arXiv Detail & Related papers (2020-03-23T10:51:32Z) - siaNMS: Non-Maximum Suppression with Siamese Networks for Multi-Camera
3D Object Detection [65.03384167873564]
A siamese network is integrated into the pipeline of a well-known 3D object detector approach.
associations are exploited to enhance the 3D box regression of the object.
The experimental evaluation on the nuScenes dataset shows that the proposed method outperforms traditional NMS approaches.
arXiv Detail & Related papers (2020-02-19T15:32:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.