Towards Vehicle-to-everything Autonomous Driving: A Survey on
Collaborative Perception
- URL: http://arxiv.org/abs/2308.16714v1
- Date: Thu, 31 Aug 2023 13:28:32 GMT
- Title: Towards Vehicle-to-everything Autonomous Driving: A Survey on
Collaborative Perception
- Authors: Si Liu, Chen Gao, Yuan Chen, Xingyu Peng, Xianghao Kong, Kun Wang,
Runsheng Xu, Wentao Jiang, Hao Xiang, Jiaqi Ma, Miao Wang
- Abstract summary: Vehicle-to-everything (V2X) autonomous driving opens up a promising direction for developing a new generation of intelligent transportation systems.
Collaborative perception (CP) as an essential component to achieve V2X can overcome the inherent limitations of individual perception.
We provide a comprehensive review of CP methods for V2X scenarios, bringing a profound and in-depth understanding to the community.
- Score: 40.90789787242417
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vehicle-to-everything (V2X) autonomous driving opens up a promising direction
for developing a new generation of intelligent transportation systems.
Collaborative perception (CP) as an essential component to achieve V2X can
overcome the inherent limitations of individual perception, including occlusion
and long-range perception. In this survey, we provide a comprehensive review of
CP methods for V2X scenarios, bringing a profound and in-depth understanding to
the community. Specifically, we first introduce the architecture and workflow
of typical V2X systems, which affords a broader perspective to understand the
entire V2X system and the role of CP within it. Then, we thoroughly summarize
and analyze existing V2X perception datasets and CP methods. Particularly, we
introduce numerous CP methods from various crucial perspectives, including
collaboration stages, roadside sensors placement, latency compensation,
performance-bandwidth trade-off, attack/defense, pose alignment, etc. Moreover,
we conduct extensive experimental analyses to compare and examine current CP
methods, revealing some essential and unexplored insights. Specifically, we
analyze the performance changes of different methods under different
bandwidths, providing a deep insight into the performance-bandwidth trade-off
issue. Also, we examine methods under different LiDAR ranges. To study the
model robustness, we further investigate the effects of various simulated
real-world noises on the performance of different CP methods, covering
communication latency, lossy communication, localization errors, and mixed
noises. In addition, we look into the sim-to-real generalization ability of
existing CP methods. At last, we thoroughly discuss issues and challenges,
highlighting promising directions for future efforts. Our codes for
experimental analysis will be public at
https://github.com/memberRE/Collaborative-Perception.
Related papers
- Deep Learning for Video Anomaly Detection: A Review [52.74513211976795]
Video anomaly detection (VAD) aims to discover behaviors or events deviating from the normality in videos.
In the era of deep learning, a great variety of deep learning based methods are constantly emerging for the VAD task.
This review covers the spectrum of five different categories, namely, semi-supervised, weakly supervised, fully supervised, unsupervised and open-set supervised VAD.
arXiv Detail & Related papers (2024-09-09T07:31:16Z) - CooTest: An Automated Testing Approach for V2X Communication Systems [15.570594621752834]
We design and implement CooTest, the first automated testing tool of the V2X-oriented cooperative perception module.
CooTest can effectively detect erroneous behaviors under various V2X-oriented driving conditions.
arXiv Detail & Related papers (2024-08-29T12:01:21Z) - 360 Layout Estimation via Orthogonal Planes Disentanglement and Multi-view Geometric Consistency Perception [56.84921040837699]
Existing panoramic layout estimation solutions tend to recover room boundaries from a vertically compressed sequence, yielding imprecise results.
We propose an orthogonal plane disentanglement network (termed DOPNet) to distinguish ambiguous semantics.
We also present an unsupervised adaptation technique tailored for horizon-depth and ratio representations.
Our solution outperforms other SoTA models on both monocular layout estimation and multi-view layout estimation tasks.
arXiv Detail & Related papers (2023-12-26T12:16:03Z) - V2X Cooperative Perception for Autonomous Driving: Recent Advances and Challenges [32.11627955649814]
Cooperative Perception with Vehicle-to-Everything (V2X) has emerged as a solution to overcome obstacles and enhance driving automation systems.
This paper provides a comprehensive overview of the evolution of CP technologies, spanning from early explorations to recent developments.
An extensive literature review is conducted within this taxonomy, evaluating existing datasets and simulators.
arXiv Detail & Related papers (2023-10-05T13:19:48Z) - CoPEM: Cooperative Perception Error Models for Autonomous Driving [20.60246432605745]
We focus on the (onboard) perception of Autonomous Vehicles (AV), which can manifest as misdetection errors on the occluded objects.
We introduce the notion of Cooperative Perception Error Models (coPEMs) towards achieving an effective integration of V2X solutions within a virtual test environment.
arXiv Detail & Related papers (2022-11-21T04:40:27Z) - A Survey and Framework of Cooperative Perception: From Heterogeneous
Singleton to Hierarchical Cooperation [14.525705886707089]
This paper reviews the research progress on Cooperative Perception (CP) and proposes a unified CP framework.
CP is born to unlock the bottleneck of perception for driving automation.
A Hierarchical CP framework is proposed, followed by a review of existing datasets and Simulators to sketch an overall landscape of CP.
arXiv Detail & Related papers (2022-08-22T20:47:35Z) - V2X-ViT: Vehicle-to-Everything Cooperative Perception with Vision
Transformer [58.71845618090022]
We build a holistic attention model, namely V2X-ViT, to fuse information across on-road agents.
V2X-ViT consists of alternating layers of heterogeneous multi-agent self-attention and multi-scale window self-attention.
To validate our approach, we create a large-scale V2X perception dataset.
arXiv Detail & Related papers (2022-03-20T20:18:25Z) - V2X-Sim: A Virtual Collaborative Perception Dataset for Autonomous
Driving [26.961213523096948]
Vehicle-to-everything (V2X) denotes the collaboration between a vehicle and any entity in its surrounding.
We present the V2X-Sim dataset, the first public large-scale collaborative perception dataset in autonomous driving.
arXiv Detail & Related papers (2022-02-17T05:14:02Z) - Consistency Regularization for Deep Face Anti-Spoofing [69.70647782777051]
Face anti-spoofing (FAS) plays a crucial role in securing face recognition systems.
Motivated by this exciting observation, we conjecture that encouraging feature consistency of different views may be a promising way to boost FAS models.
We enhance both Embedding-level and Prediction-level Consistency Regularization (EPCR) in FAS.
arXiv Detail & Related papers (2021-11-24T08:03:48Z) - Divide-and-Conquer for Lane-Aware Diverse Trajectory Prediction [71.97877759413272]
Trajectory prediction is a safety-critical tool for autonomous vehicles to plan and execute actions.
Recent methods have achieved strong performances using Multi-Choice Learning objectives like winner-takes-all (WTA) or best-of-many.
Our work addresses two key challenges in trajectory prediction, learning outputs, and better predictions by imposing constraints using driving knowledge.
arXiv Detail & Related papers (2021-04-16T17:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.