Learning Cooperative Trajectory Representations for Motion Forecasting
- URL: http://arxiv.org/abs/2311.00371v1
- Date: Wed, 1 Nov 2023 08:53:05 GMT
- Title: Learning Cooperative Trajectory Representations for Motion Forecasting
- Authors: Hongzhi Ruan, Haibao Yu, Wenxian Yang, Siqi Fan, Yingjuan Tang,
Zaiqing Nie
- Abstract summary: We present V2X-Graph, the first interpretable and end-to-end learning framework for cooperative motion forecasting.
V2X-Graph employs an interpretable graph to fully leverage the cooperative motion and interaction contexts.
We construct the first real-world vehicle-to-everything (V2X) motion forecasting dataset.
- Score: 4.380073528690906
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motion forecasting is an essential task for autonomous driving, and the
effective information utilization from infrastructure and other vehicles can
enhance motion forecasting capabilities. Existing research have primarily
focused on leveraging single-frame cooperative information to enhance the
limited perception capability of the ego vehicle, while underutilizing the
motion and interaction information of traffic participants observed from
cooperative devices. In this paper, we first propose the cooperative trajectory
representations learning paradigm. Specifically, we present V2X-Graph, the
first interpretable and end-to-end learning framework for cooperative motion
forecasting. V2X-Graph employs an interpretable graph to fully leverage the
cooperative motion and interaction contexts. Experimental results on the
vehicle-to-infrastructure (V2I) motion forecasting dataset, V2X-Seq,
demonstrate the effectiveness of V2X-Graph. To further evaluate on V2X
scenario, we construct the first real-world vehicle-to-everything (V2X) motion
forecasting dataset V2X-Traj, and the performance shows the advantage of our
method. We hope both V2X-Graph and V2X-Traj can facilitate the further
development of cooperative motion forecasting. Find project at
https://github.com/AIR-THU/V2X-Graph, find data at
https://github.com/AIR-THU/DAIR-V2X-Seq.
Related papers
- CRUISE: Cooperative Reconstruction and Editing in V2X Scenarios using Gaussian Splatting [48.883658011726915]
Vehicle-to-everything (V2X) communication plays a crucial role in autonomous driving, enabling cooperation between vehicles and infrastructure.<n>In this paper, we introduce CRUISE, a comprehensive reconstruction-and-synthesis framework designed for V2X driving environments.
arXiv Detail & Related papers (2025-07-24T14:48:44Z) - Co-MTP: A Cooperative Trajectory Prediction Framework with Multi-Temporal Fusion for Autonomous Driving [16.479343520119073]
Co-MTP is a general cooperative trajectory prediction framework with multi-temporal fusion for autonomous driving.
In the future domain, V2X can provide the prediction results of surrounding objects.
We evaluate the Co-MTP framework on the real-world dataset V2X-Seq.
arXiv Detail & Related papers (2025-02-23T14:38:13Z) - CooPre: Cooperative Pretraining for V2X Cooperative Perception [47.00472259100765]
We present a self-supervised learning method for V2X cooperative perception.
We utilize the vast amount of unlabeled 3D V2X data to enhance the perception performance.
arXiv Detail & Related papers (2024-08-20T23:39:26Z) - V2X-VLM: End-to-End V2X Cooperative Autonomous Driving Through Large Vision-Language Models [13.716889927164383]
Vehicle-to-everything (V2X) cooperation has emerged as a promising paradigm to overcome the perception limitations of classical autonomous driving.<n>This paper introduces V2X-VLM, a novel end-to-end (E2E) cooperative autonomous driving framework based on vision-language models (VLMs)<n>V2X-VLM integrates multiperspective camera views from vehicles and infrastructure with text-based scene descriptions to enable a more comprehensive understanding of driving environments.
arXiv Detail & Related papers (2024-08-17T16:42:13Z) - Conformal Trajectory Prediction with Multi-View Data Integration in Cooperative Driving [4.628774934971078]
Current research on trajectory prediction primarily relies on data collected by onboard sensors of an ego vehicle.
We introduce V2INet, a novel trajectory prediction framework designed to model multi-view data by extending existing single-view models.
Our results demonstrate superior performance in terms of Final Displacement Error (FDE) and Miss Rate (MR) using a single GPU.
arXiv Detail & Related papers (2024-08-01T08:32:03Z) - End-to-End Autonomous Driving through V2X Cooperation [23.44597411612664]
We introduce UniV2X, a pioneering cooperative autonomous driving framework.
UniV2X seamlessly integrates all key driving modules across diverse views into a unified network.
arXiv Detail & Related papers (2024-03-31T15:22:11Z) - Vanishing-Point-Guided Video Semantic Segmentation of Driving Scenes [70.08318779492944]
We are the first to harness vanishing point (VP) priors for more effective segmentation.
Our novel, efficient network for VSS, named VPSeg, incorporates two modules that utilize exactly this pair of static and dynamic VP priors.
arXiv Detail & Related papers (2024-01-27T01:01:58Z) - DeepAccident: A Motion and Accident Prediction Benchmark for V2X
Autonomous Driving [76.29141888408265]
We propose a large-scale dataset containing diverse accident scenarios that frequently occur in real-world driving.
The proposed DeepAccident dataset includes 57K annotated frames and 285K annotated samples, approximately 7 times more than the large-scale nuScenes dataset.
arXiv Detail & Related papers (2023-04-03T17:37:00Z) - V2X-ViT: Vehicle-to-Everything Cooperative Perception with Vision
Transformer [58.71845618090022]
We build a holistic attention model, namely V2X-ViT, to fuse information across on-road agents.
V2X-ViT consists of alternating layers of heterogeneous multi-agent self-attention and multi-scale window self-attention.
To validate our approach, we create a large-scale V2X perception dataset.
arXiv Detail & Related papers (2022-03-20T20:18:25Z) - V2X-Sim: A Virtual Collaborative Perception Dataset for Autonomous
Driving [26.961213523096948]
Vehicle-to-everything (V2X) denotes the collaboration between a vehicle and any entity in its surrounding.
We present the V2X-Sim dataset, the first public large-scale collaborative perception dataset in autonomous driving.
arXiv Detail & Related papers (2022-02-17T05:14:02Z) - Visual Relationship Forecasting in Videos [56.122037294234865]
We present a new task named Visual Relationship Forecasting (VRF) in videos to explore the prediction of visual relationships in a manner of reasoning.
Given a subject-object pair with H existing frames, VRF aims to predict their future interactions for the next T frames without visual evidence.
To evaluate the VRF task, we introduce two video datasets named VRF-AG and VRF-VidOR, with a series oftemporally localized visual relation annotations in a video.
arXiv Detail & Related papers (2021-07-02T16:43:19Z) - Implicit Latent Variable Model for Scene-Consistent Motion Forecasting [78.74510891099395]
In this paper, we aim to learn scene-consistent motion forecasts of complex urban traffic directly from sensor data.
We model the scene as an interaction graph and employ powerful graph neural networks to learn a distributed latent representation of the scene.
arXiv Detail & Related papers (2020-07-23T14:31:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.