VPC-Net: Completion of 3D Vehicles from MLS Point Clouds
- URL: http://arxiv.org/abs/2008.03404v2
- Date: Mon, 1 Feb 2021 16:32:40 GMT
- Title: VPC-Net: Completion of 3D Vehicles from MLS Point Clouds
- Authors: Yan Xia, Yusheng Xu, Cheng Wang, Uwe Stilla
- Abstract summary: Vehicle Points CompletionNet (VPCNet)
We introduce a new encoder module to extract global features from the input instance, consisting of a transformer network and point feature enhancement layer.
A new refiner module is also presented to preserve the vehicle details from inputs and refine the complete outputs with fine-grained information.
- Score: 16.041815273682317
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a dynamic and essential component in the road environment of urban
scenarios, vehicles are the most popular investigation targets. To monitor
their behavior and extract their geometric characteristics, an accurate and
instant measurement of vehicles plays a vital role in traffic and
transportation fields. Point clouds acquired from the mobile laser scanning
(MLS) system deliver 3D information of road scenes with unprecedented detail.
They have proven to be an adequate data source in the fields of intelligent
transportation and autonomous driving, especially for extracting vehicles.
However, acquired 3D point clouds of vehicles from MLS systems are inevitably
incomplete due to object occlusion or self-occlusion. To tackle this problem,
we proposed a neural network to synthesize complete, dense, and uniform point
clouds for vehicles from MLS data, named Vehicle Points Completion-Net
(VPC-Net). In this network, we introduce a new encoder module to extract global
features from the input instance, consisting of a spatial transformer network
and point feature enhancement layer. Moreover, a new refiner module is also
presented to preserve the vehicle details from inputs and refine the complete
outputs with fine-grained information. Given sparse and partial point clouds as
inputs, the network can generate complete and realistic vehicle structures and
keep the fine-grained details from the partial inputs. We evaluated the
proposed VPC-Net in different experiments using synthetic and real-scan
datasets and applied the results to 3D vehicle monitoring tasks. Quantitative
and qualitative experiments demonstrate the promising performance of the
proposed VPC-Net and show state-of-the-art results.
Related papers
- Empowering Urban Traffic Management: Elevated 3D LiDAR for Data Collection and Advanced Object Detection Analysis [4.831084635928491]
This paper presents a novel framework that transforms the detection and analysis of 3D objects in traffic scenarios by utilizing the power of elevated LiDAR sensors.
Due to the limitation in obtaining real-world traffic datasets, we utilize the simulator to generate 3D point cloud for specific scenarios.
arXiv Detail & Related papers (2024-05-21T21:12:09Z) - AutoSynth: Learning to Generate 3D Training Data for Object Point Cloud
Registration [69.21282992341007]
Auto Synth automatically generates 3D training data for point cloud registration.
We replace the point cloud registration network with a much smaller surrogate network, leading to a $4056.43$ speedup.
Our results on TUD-L, LINEMOD and Occluded-LINEMOD evidence that a neural network trained on our searched dataset yields consistently better performance than the same one trained on the widely used ModelNet40 dataset.
arXiv Detail & Related papers (2023-09-20T09:29:44Z) - SKoPe3D: A Synthetic Dataset for Vehicle Keypoint Perception in 3D from
Traffic Monitoring Cameras [26.457695296042903]
We propose SKoPe3D, a unique synthetic vehicle keypoint dataset from a roadside perspective.
SKoPe3D contains over 150k vehicle instances and 4.9 million keypoints.
Our experiments highlight the dataset's applicability and the potential for knowledge transfer between synthetic and real-world data.
arXiv Detail & Related papers (2023-09-04T02:57:30Z) - HUM3DIL: Semi-supervised Multi-modal 3D Human Pose Estimation for
Autonomous Driving [95.42203932627102]
3D human pose estimation is an emerging technology, which can enable the autonomous vehicle to perceive and understand the subtle and complex behaviors of pedestrians.
Our method efficiently makes use of these complementary signals, in a semi-supervised fashion and outperforms existing methods with a large margin.
Specifically, we embed LiDAR points into pixel-aligned multi-modal features, which we pass through a sequence of Transformer refinement stages.
arXiv Detail & Related papers (2022-12-15T11:15:14Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - Collaborative 3D Object Detection for Automatic Vehicle Systems via
Learnable Communications [8.633120731620307]
We propose a novel collaborative 3D object detection framework that consists of three components.
Experiment results and bandwidth usage analysis demonstrate that our approach can save communication and computation costs.
arXiv Detail & Related papers (2022-05-24T07:17:32Z) - Self-supervised Point Cloud Completion on Real Traffic Scenes via
Scene-concerned Bottom-up Mechanism [14.255659581428333]
Point cloud completion aims to refer the complete shapes for incomplete 3D scans of objects.
Current deep learning-based approaches rely on large-scale complete shapes in the training process.
We propose a self-supervised point cloud completion method (TraPCC) for vehicles in real traffic scenes without any complete data.
arXiv Detail & Related papers (2022-03-20T14:42:37Z) - Cyber Mobility Mirror: Deep Learning-based Real-time 3D Object
Perception and Reconstruction Using Roadside LiDAR [14.566471856473813]
Cyber Mobility Mirror is a next-generation real-time traffic surveillance system for 3D object detection, classification, tracking, and reconstruction.
Results from field tests demonstrate that our prototype system can provide satisfactory perception performance with 96.99% precision and 83.62% recall.
High-fidelity real-time traffic conditions can be displayed on the GUI of the equipped vehicle with a frequency of 3-4 Hz.
arXiv Detail & Related papers (2022-02-28T01:58:24Z) - PC-DAN: Point Cloud based Deep Affinity Network for 3D Multi-Object
Tracking (Accepted as an extended abstract in JRDB-ACT Workshop at CVPR21) [68.12101204123422]
A point cloud is a dense compilation of spatial data in 3D coordinates.
We propose a PointNet-based approach for 3D Multi-Object Tracking (MOT)
arXiv Detail & Related papers (2021-06-03T05:36:39Z) - Multi-Modal Fusion Transformer for End-to-End Autonomous Driving [59.60483620730437]
We propose TransFuser, a novel Multi-Modal Fusion Transformer, to integrate image and LiDAR representations using attention.
Our approach achieves state-of-the-art driving performance while reducing collisions by 76% compared to geometry-based fusion.
arXiv Detail & Related papers (2021-04-19T11:48:13Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.