CRUISE: Cooperative Reconstruction and Editing in V2X Scenarios using   Gaussian Splatting
        - URL: http://arxiv.org/abs/2507.18473v1
 - Date: Thu, 24 Jul 2025 14:48:44 GMT
 - Title: CRUISE: Cooperative Reconstruction and Editing in V2X Scenarios using   Gaussian Splatting
 - Authors: Haoran Xu, Saining Zhang, Peishuo Li, Baijun Ye, Xiaoxue Chen, Huan-ang Gao, Jv Zheng, Xiaowei Song, Ziqiao Peng, Run Miao, Jinrang Jia, Yifeng Shi, Guangqi Yi, Hang Zhao, Hao Tang, Hongyang Li, Kaicheng Yu, Hao Zhao, 
 - Abstract summary: Vehicle-to-everything (V2X) communication plays a crucial role in autonomous driving, enabling cooperation between vehicles and infrastructure.<n>In this paper, we introduce CRUISE, a comprehensive reconstruction-and-synthesis framework designed for V2X driving environments.
 - Score: 48.883658011726915
 - License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
 - Abstract:   Vehicle-to-everything (V2X) communication plays a crucial role in autonomous driving, enabling cooperation between vehicles and infrastructure. While simulation has significantly contributed to various autonomous driving tasks, its potential for data generation and augmentation in V2X scenarios remains underexplored. In this paper, we introduce CRUISE, a comprehensive reconstruction-and-synthesis framework designed for V2X driving environments. CRUISE employs decomposed Gaussian Splatting to accurately reconstruct real-world scenes while supporting flexible editing. By decomposing dynamic traffic participants into editable Gaussian representations, CRUISE allows for seamless modification and augmentation of driving scenes. Furthermore, the framework renders images from both ego-vehicle and infrastructure views, enabling large-scale V2X dataset augmentation for training and evaluation. Our experimental results demonstrate that: 1) CRUISE reconstructs real-world V2X driving scenes with high fidelity; 2) using CRUISE improves 3D detection across ego-vehicle, infrastructure, and cooperative views, as well as cooperative 3D tracking on the V2X-Seq benchmark; and 3) CRUISE effectively generates challenging corner cases. 
 
       
      
        Related papers
        - RoCo-Sim: Enhancing Roadside Collaborative Perception through Foreground   Simulation [30.744548212616007]
We present the first simulation framework RoCo-Sim for road-side collaborative perception.<n>RoCo-Sim is capable of generating diverse, multi-view consistent simulated roadside data.<n>Code and pre-trained models will be released soon.
arXiv  Detail & Related papers  (2025-03-13T14:33:42Z) - Vid2Sim: Realistic and Interactive Simulation from Video for Urban   Navigation [62.5805866419814]
Vid2Sim is a novel framework that bridges the sim2real gap through a scalable and cost-efficient real2sim pipeline for neural 3D scene reconstruction and simulation.<n>Experiments demonstrate that Vid2Sim significantly improves the performance of urban navigation in the digital twins and real world by 31.2% and 68.3% in success rate.
arXiv  Detail & Related papers  (2025-01-12T03:01:15Z) - CooPre: Cooperative Pretraining for V2X Cooperative Perception [47.00472259100765]
CooPre is a self-supervised learning framwork for V2X cooperative perception.<n>We develop a V2X bird-eye-view (BEV) guided masking strategy which effectively allows the model to pay attention to 3D features across heterogeneous V2X agents.<n>CooPre achieves a 4% mAP improvement on V2X-Real dataset and surpasses baseline performance using only 50% of the training data.
arXiv  Detail & Related papers  (2024-08-20T23:39:26Z) - V2V4Real: A Real-world Large-scale Dataset for Vehicle-to-Vehicle
  Cooperative Perception [49.7212681947463]
Vehicle-to-Vehicle (V2V) cooperative perception system has great potential to revolutionize the autonomous driving industry.
We present V2V4Real, the first large-scale real-world multi-modal dataset for V2V perception.
Our dataset covers a driving area of 410 km, comprising 20K LiDAR frames, 40K RGB frames, 240K annotated 3D bounding boxes for 5 classes, and HDMaps.
arXiv  Detail & Related papers  (2023-03-14T02:49:20Z) - V2XP-ASG: Generating Adversarial Scenes for Vehicle-to-Everything
  Perception [37.41995438002604]
V2X perception systems will soon be deployed at scale.
How can we evaluate and improve its performance under challenging traffic scenarios before the real-world deployment?
We propose the first open adversarial scene generator V2XP-ASG.
arXiv  Detail & Related papers  (2022-09-27T20:34:41Z) - V2X-ViT: Vehicle-to-Everything Cooperative Perception with Vision
  Transformer [58.71845618090022]
We build a holistic attention model, namely V2X-ViT, to fuse information across on-road agents.
V2X-ViT consists of alternating layers of heterogeneous multi-agent self-attention and multi-scale window self-attention.
To validate our approach, we create a large-scale V2X perception dataset.
arXiv  Detail & Related papers  (2022-03-20T20:18:25Z) - V2X-Sim: A Virtual Collaborative Perception Dataset for Autonomous
  Driving [26.961213523096948]
Vehicle-to-everything (V2X) denotes the collaboration between a vehicle and any entity in its surrounding.
We present the V2X-Sim dataset, the first public large-scale collaborative perception dataset in autonomous driving.
arXiv  Detail & Related papers  (2022-02-17T05:14:02Z) - LiDAR-based Panoptic Segmentation via Dynamic Shifting Network [56.71765153629892]
LiDAR-based panoptic segmentation aims to parse both objects and scenes in a unified manner.
We propose the Dynamic Shifting Network (DS-Net), which serves as an effective panoptic segmentation framework in the point cloud realm.
Our proposed DS-Net achieves superior accuracies over current state-of-the-art methods.
arXiv  Detail & Related papers  (2020-11-24T08:44:46Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.