Generating Evidential BEV Maps in Continuous Driving Space
- URL: http://arxiv.org/abs/2302.02928v2
- Date: Mon, 4 Sep 2023 13:41:50 GMT
- Title: Generating Evidential BEV Maps in Continuous Driving Space
- Authors: Yunshuang Yuan, Hao Cheng, Michael Ying Yang and Monika Sester
- Abstract summary: We propose a complete probabilistic model named GevBEV.
It interprets the 2D driving space as a probabilistic Bird's Eye View (BEV) map with point-based spatial Gaussian distributions.
GevBEV helps reduce communication overhead by selecting only the most important information to share from the learned uncertainty.
- Score: 13.073542165482566
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Safety is critical for autonomous driving, and one aspect of improving safety
is to accurately capture the uncertainties of the perception system, especially
knowing the unknown. Different from only providing deterministic or
probabilistic results, e.g., probabilistic object detection, that only provide
partial information for the perception scenario, we propose a complete
probabilistic model named GevBEV. It interprets the 2D driving space as a
probabilistic Bird's Eye View (BEV) map with point-based spatial Gaussian
distributions, from which one can draw evidence as the parameters for the
categorical Dirichlet distribution of any new sample point in the continuous
driving space. The experimental results show that GevBEV not only provides more
reliable uncertainty quantification but also outperforms the previous works on
the benchmarks OPV2V and V2V4Real of BEV map interpretation for cooperative
perception in simulated and real-world driving scenarios, respectively. A
critical factor in cooperative perception is the data transmission size through
the communication channels. GevBEV helps reduce communication overhead by
selecting only the most important information to share from the learned
uncertainty, reducing the average information communicated by 87% with only a
slight performance drop. Our code is published at
https://github.com/YuanYunshuang/GevBEV.
Related papers
- BEVPose: Unveiling Scene Semantics through Pose-Guided Multi-Modal BEV Alignment [8.098296280937518]
We present BEVPose, a framework that integrates BEV representations from camera and lidar data, using sensor pose as a guiding supervisory signal.
By leveraging pose information, we align and fuse multi-modal sensory inputs, facilitating the learning of latent BEV embeddings that capture both geometric and semantic aspects of the environment.
arXiv Detail & Related papers (2024-10-28T12:40:27Z) - Unlocking Past Information: Temporal Embeddings in Cooperative Bird's
Eye View Prediction [34.68695222573004]
This paper introduces TempCoBEV, a temporal module designed to incorporate historical cues into current observations.
We show the efficacy of TempCoBEV and its capability to integrate historical cues into the current BEV map, improving predictions under optimal communication conditions by up to 2% and under communication failures by up to 19%.
arXiv Detail & Related papers (2024-01-25T17:21:35Z) - Diffusion-Based Particle-DETR for BEV Perception [94.88305708174796]
Bird-Eye-View (BEV) is one of the most widely-used scene representations for visual perception in Autonomous Vehicles (AVs)
Recent diffusion-based methods offer a promising approach to uncertainty modeling for visual perception but fail to effectively detect small objects in the large coverage of the BEV.
Here, we address this problem by combining the diffusion paradigm with current state-of-the-art 3D object detectors in BEV.
arXiv Detail & Related papers (2023-12-18T09:52:14Z) - U-BEV: Height-aware Bird's-Eye-View Segmentation and Neural Map-based Relocalization [81.76044207714637]
Relocalization is essential for intelligent vehicles when GPS reception is insufficient or sensor-based localization fails.
Recent advances in Bird's-Eye-View (BEV) segmentation allow for accurate estimation of local scene appearance.
This paper presents U-BEV, a U-Net inspired architecture that extends the current state-of-the-art by allowing the BEV to reason about the scene on multiple height layers before flattening the BEV features.
arXiv Detail & Related papers (2023-10-20T18:57:38Z) - ProbVLM: Probabilistic Adapter for Frozen Vision-Language Models [69.50316788263433]
We propose ProbVLM, a probabilistic adapter that estimates probability distributions for the embeddings of pre-trained vision-language models.
We quantify the calibration of embedding uncertainties in retrieval tasks and show that ProbVLM outperforms other methods.
We present a novel technique for visualizing the embedding distributions using a large-scale pre-trained latent diffusion model.
arXiv Detail & Related papers (2023-07-01T18:16:06Z) - Integrated Sensing, Computation, and Communication for UAV-assisted
Federated Edge Learning [52.7230652428711]
Federated edge learning (FEEL) enables privacy-preserving model training through periodic communication between edge devices and the server.
Unmanned Aerial Vehicle (UAV)mounted edge devices are particularly advantageous for FEEL due to their flexibility and mobility in efficient data collection.
arXiv Detail & Related papers (2023-06-05T16:01:33Z) - DeepAccident: A Motion and Accident Prediction Benchmark for V2X
Autonomous Driving [76.29141888408265]
We propose a large-scale dataset containing diverse accident scenarios that frequently occur in real-world driving.
The proposed DeepAccident dataset includes 57K annotated frames and 285K annotated samples, approximately 7 times more than the large-scale nuScenes dataset.
arXiv Detail & Related papers (2023-04-03T17:37:00Z) - DiffBEV: Conditional Diffusion Model for Bird's Eye View Perception [14.968177102647783]
We propose an end-to-end framework, named DiffBEV, to exploit the potential of diffusion model to generate a more comprehensive BEV representation.
In practice, we design three types of conditions to guide the training of the diffusion model which denoises the coarse samples and refines the semantic feature.
We show that DiffBEV achieves a 25.9% mIoU on the nuScenes dataset, which is 6.2% higher than the best-performing existing approach.
arXiv Detail & Related papers (2023-03-15T02:42:48Z) - Flexible Amortized Variational Inference in qBOLD MRI [56.4324135502282]
Oxygen extraction fraction (OEF) and deoxygenated blood volume (DBV) are more ambiguously determined from the data.
Existing inference methods tend to yield very noisy and underestimated OEF maps, while overestimating DBV.
This work describes a novel probabilistic machine learning approach that can infer plausible distributions of OEF and DBV.
arXiv Detail & Related papers (2022-03-11T10:47:16Z) - Estimation of Driver's Gaze Region from Head Position and Orientation
using Probabilistic Confidence Regions [43.9008720663172]
A smart vehicle should be able to understand human behavior and predict their actions to avoid hazardous situations.
One of the most important aspects pertaining to the driving task is the driver's visual attention.
This paper proposes a formulation based on probabilistic models to create salient regions describing the visual attention of the driver.
arXiv Detail & Related papers (2020-12-23T15:48:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.