Uncertainty-aware Perception Models for Off-road Autonomous Unmanned
Ground Vehicles
- URL: http://arxiv.org/abs/2209.11115v1
- Date: Thu, 22 Sep 2022 15:59:33 GMT
- Title: Uncertainty-aware Perception Models for Off-road Autonomous Unmanned
Ground Vehicles
- Authors: Zhaoyuan Yang, Yewteck Tan, Shiraj Sen, Johan Reimann, John
Karigiannis, Mohammed Yousefhussien, Nurali Virani
- Abstract summary: Off-road autonomous unmanned ground vehicles (UGVs) are being developed for military and commercial use to deliver crucial supplies in remote locations.
Current datasets used to train perception models for off-road autonomous navigation lack of diversity in seasons, locations, semantic classes, as well as time of day.
We investigate how to combine multiple datasets to train a semantic segmentation-based environment perception model.
We show that training the model to capture uncertainty could improve the model performance by a significant margin.
- Score: 6.2574402913714575
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Off-road autonomous unmanned ground vehicles (UGVs) are being developed for
military and commercial use to deliver crucial supplies in remote locations,
help with mapping and surveillance, and to assist war-fighters in contested
environments. Due to complexity of the off-road environments and variability in
terrain, lighting conditions, diurnal and seasonal changes, the models used to
perceive the environment must handle a lot of input variability. Current
datasets used to train perception models for off-road autonomous navigation
lack of diversity in seasons, locations, semantic classes, as well as time of
day. We test the hypothesis that model trained on a single dataset may not
generalize to other off-road navigation datasets and new locations due to the
input distribution drift. Additionally, we investigate how to combine multiple
datasets to train a semantic segmentation-based environment perception model
and we show that training the model to capture uncertainty could improve the
model performance by a significant margin. We extend the Masksembles approach
for uncertainty quantification to the semantic segmentation task and compare it
with Monte Carlo Dropout and standard baselines. Finally, we test the approach
against data collected from a UGV platform in a new testing environment. We
show that the developed perception model with uncertainty quantification can be
feasibly deployed on an UGV to support online perception and navigation tasks.
Related papers
- Hard Cases Detection in Motion Prediction by Vision-Language Foundation Models [16.452638202694246]
This work explores the potential of Vision-Language Foundation Models (VLMs) in detecting hard cases in autonomous driving.
We introduce a feasible pipeline where VLMs, fed with sequential image frames with designed prompts, effectively identify challenging agents or scenarios.
We show the effectiveness and feasibility of incorporating our pipeline with state-of-the-art methods on NuScenes datasets.
arXiv Detail & Related papers (2024-05-31T16:35:41Z) - JRDB-Traj: A Dataset and Benchmark for Trajectory Forecasting in Crowds [79.00975648564483]
Trajectory forecasting models, employed in fields such as robotics, autonomous vehicles, and navigation, face challenges in real-world scenarios.
This dataset provides comprehensive data, including the locations of all agents, scene images, and point clouds, all from the robot's perspective.
The objective is to predict the future positions of agents relative to the robot using raw sensory input data.
arXiv Detail & Related papers (2023-11-05T18:59:31Z) - Drive Anywhere: Generalizable End-to-end Autonomous Driving with
Multi-modal Foundation Models [114.69732301904419]
We present an approach to apply end-to-end open-set (any environment/scene) autonomous driving that is capable of providing driving decisions from representations queryable by image and text.
Our approach demonstrates unparalleled results in diverse tests while achieving significantly greater robustness in out-of-distribution situations.
arXiv Detail & Related papers (2023-10-26T17:56:35Z) - On Transferability of Driver Observation Models from Simulated to Real
Environments in Autonomous Cars [23.514129229090987]
This paper investigates the viability of transferring video-based driver observation models from simulation to real-world scenarios in autonomous vehicles.
We record a dataset featuring actual autonomous driving conditions and involving seven participants engaged in highly distracting secondary activities.
Our dataset was designed in accordance with an existing large-scale simulator dataset used as the training source.
arXiv Detail & Related papers (2023-07-31T10:18:49Z) - METAVerse: Meta-Learning Traversability Cost Map for Off-Road Navigation [5.036362492608702]
This paper presents METAVerse, a meta-learning framework for learning a global model that accurately predicts terrain traversability.
We train the traversability prediction network to generate a dense and continuous-terrain cost map from a sparse LiDAR point cloud.
Online adaptation is performed to rapidly adapt the network to the local environment by exploiting recent interaction experiences.
arXiv Detail & Related papers (2023-07-26T06:58:19Z) - Safe Navigation in Unstructured Environments by Minimizing Uncertainty
in Control and Perception [5.46262127926284]
Uncertainty in control and perception poses challenges for autonomous vehicle navigation in unstructured environments.
This paper introduces a framework that minimizes control and perception uncertainty to ensure safe and reliable navigation.
arXiv Detail & Related papers (2023-06-26T11:24:03Z) - TrafficBots: Towards World Models for Autonomous Driving Simulation and
Motion Prediction [149.5716746789134]
We show data-driven traffic simulation can be formulated as a world model.
We present TrafficBots, a multi-agent policy built upon motion prediction and end-to-end driving.
Experiments on the open motion dataset show TrafficBots can simulate realistic multi-agent behaviors.
arXiv Detail & Related papers (2023-03-07T18:28:41Z) - LOPR: Latent Occupancy PRediction using Generative Models [49.15687400958916]
LiDAR generated occupancy grid maps (L-OGMs) offer a robust bird's eye-view scene representation.
We propose a framework that decouples occupancy prediction into: representation learning and prediction within the learned latent space.
arXiv Detail & Related papers (2022-10-03T22:04:00Z) - Federated Deep Learning Meets Autonomous Vehicle Perception: Design and
Verification [168.67190934250868]
Federated learning empowered connected autonomous vehicle (FLCAV) has been proposed.
FLCAV preserves privacy while reducing communication and annotation costs.
It is challenging to determine the network resources and road sensor poses for multi-stage training.
arXiv Detail & Related papers (2022-06-03T23:55:45Z) - PSE-Match: A Viewpoint-free Place Recognition Method with Parallel
Semantic Embedding [9.265785042748158]
PSE-Match is a viewpoint-free place recognition method based on parallel semantic analysis of isolated semantic attributes from 3D point-cloud models.
PSE-Match incorporates a divergence place learning network to capture different semantic attributes parallelly through the spherical harmonics domain.
arXiv Detail & Related papers (2021-08-01T22:16:40Z) - SMART: Simultaneous Multi-Agent Recurrent Trajectory Prediction [72.37440317774556]
We propose advances that address two key challenges in future trajectory prediction.
multimodality in both training data and predictions and constant time inference regardless of number of agents.
arXiv Detail & Related papers (2020-07-26T08:17:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.