A Multi-Modal Simulation Framework to Enable Digital Twin-based V2X Communications in Dynamic Environments
- URL: http://arxiv.org/abs/2303.06947v3
- Date: Wed, 17 Jul 2024 13:24:22 GMT
- Title: A Multi-Modal Simulation Framework to Enable Digital Twin-based V2X Communications in Dynamic Environments
- Authors: Lorenzo Cazzella, Francesco Linsalata, Maurizio Magarini, Matteo Matteucci, Umberto Spagnolini,
- Abstract summary: Digital Twins (DTs) for physical wireless environments have been recently proposed as accurate virtual representations of the propagation environment.
We propose a novel data-driven workflow for the creation of the DT of a Vehicle-to-Everything (V2X) communication scenario.
We showcase the proposed framework on the DT-aided blockage handover task for V2X link restoration.
- Score: 10.652127049174883
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Digital Twins (DTs) for physical wireless environments have been recently proposed as accurate virtual representations of the propagation environment that can enable multi-layer decisions at the physical communication equipment. At high-frequency bands, DTs can help to overcome the challenges emerging in high mobility conditions featuring vehicular environments. In this paper, we propose a novel data-driven workflow for the creation of the DT of a Vehicle-to-Everything (V2X) communication scenario and a multi-modal simulation framework for the generation of realistic sensor data and accurate mmWave/sub-THz wireless channels. The proposed method leverages an automotive simulation and testing framework and an accurate ray-tracing channel simulator. Simulations over an urban scenario show the achievable realistic sensor and channel modelling both at the infrastructure and at ego-vehicles. We showcase the proposed framework on the DT-aided blockage handover task for V2X link restoration, leveraging the framework's dynamic channel generation capabilities for realistic vehicular blockage simulation.
Related papers
- DrivingSphere: Building a High-fidelity 4D World for Closed-loop Simulation [54.02069690134526]
We propose DrivingSphere, a realistic and closed-loop simulation framework.
Its core idea is to build 4D world representation and generate real-life and controllable driving scenarios.
By providing a dynamic and realistic simulation environment, DrivingSphere enables comprehensive testing and validation of autonomous driving algorithms.
arXiv Detail & Related papers (2024-11-18T03:00:33Z) - GenDDS: Generating Diverse Driving Video Scenarios with Prompt-to-Video Generative Model [6.144680854063938]
GenDDS is a novel approach for generating driving scenarios for autonomous driving systems.
We employ the KITTI dataset, which includes real-world driving videos, to train the model.
We demonstrate that our model can generate high-quality driving videos that closely replicate the complexity and variability of real-world driving scenarios.
arXiv Detail & Related papers (2024-08-28T15:37:44Z) - S-NeRF++: Autonomous Driving Simulation via Neural Reconstruction and Generation [21.501865765631123]
S-NeRF++ is an innovative autonomous driving simulation system based on neural reconstruction.
S-NeRF++ is trained on widely-used self-driving datasets such as nuScenes and radiance.
System effectively utilizes noisy and sparse LiDAR data to refine training and address depth outliers.
arXiv Detail & Related papers (2024-02-03T10:35:42Z) - Vision-Assisted Digital Twin Creation for mmWave Beam Management [9.608394183713717]
We propose a practical Digital Twin creation pipeline and a channel simulator, that relies only on a single mounted camera and position information.
We demonstrate the performance benefits compared to methods that do not explicitly model the 3D environment, on downstream sub-tasks in beam acquisition.
arXiv Detail & Related papers (2024-01-31T12:23:55Z) - Reinforcement Learning with Human Feedback for Realistic Traffic
Simulation [53.85002640149283]
Key element of effective simulation is the incorporation of realistic traffic models that align with human knowledge.
This study identifies two main challenges: capturing the nuances of human preferences on realism and the unification of diverse traffic simulation models.
arXiv Detail & Related papers (2023-09-01T19:29:53Z) - TrafficBots: Towards World Models for Autonomous Driving Simulation and
Motion Prediction [149.5716746789134]
We show data-driven traffic simulation can be formulated as a world model.
We present TrafficBots, a multi-agent policy built upon motion prediction and end-to-end driving.
Experiments on the open motion dataset show TrafficBots can simulate realistic multi-agent behaviors.
arXiv Detail & Related papers (2023-03-07T18:28:41Z) - TrainSim: A Railway Simulation Framework for LiDAR and Camera Dataset
Generation [1.2165229201148093]
This paper presents a visual simulation framework able to generate realistic railway scenarios in a virtual environment.
It automatically produces inertial data and labeled datasets from emulated LiDARs and cameras.
arXiv Detail & Related papers (2023-02-28T11:00:13Z) - Generative AI-empowered Simulation for Autonomous Driving in Vehicular
Mixed Reality Metaverses [130.15554653948897]
In vehicular mixed reality (MR) Metaverse, distance between physical and virtual entities can be overcome.
Large-scale traffic and driving simulation via realistic data collection and fusion from the physical world is difficult and costly.
We propose an autonomous driving architecture, where generative AI is leveraged to synthesize unlimited conditioned traffic and driving data in simulations.
arXiv Detail & Related papers (2023-02-16T16:54:10Z) - COOPERNAUT: End-to-End Driving with Cooperative Perception for Networked
Vehicles [54.61668577827041]
We introduce COOPERNAUT, an end-to-end learning model that uses cross-vehicle perception for vision-based cooperative driving.
Our experiments on AutoCastSim suggest that our cooperative perception driving models lead to a 40% improvement in average success rate.
arXiv Detail & Related papers (2022-05-04T17:55:12Z) - VISTA 2.0: An Open, Data-driven Simulator for Multimodal Sensing and
Policy Learning for Autonomous Vehicles [131.2240621036954]
We present VISTA, an open source, data-driven simulator that integrates multiple types of sensors for autonomous vehicles.
Using high fidelity, real-world datasets, VISTA represents and simulates RGB cameras, 3D LiDAR, and event-based cameras.
We demonstrate the ability to train and test perception-to-control policies across each of the sensor types and showcase the power of this approach via deployment on a full scale autonomous vehicle.
arXiv Detail & Related papers (2021-11-23T18:58:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.