Monocular Vision based Crowdsourced 3D Traffic Sign Positioning with
Unknown Camera Intrinsics and Distortion Coefficients
- URL: http://arxiv.org/abs/2007.04592v1
- Date: Thu, 9 Jul 2020 07:03:17 GMT
- Title: Monocular Vision based Crowdsourced 3D Traffic Sign Positioning with
Unknown Camera Intrinsics and Distortion Coefficients
- Authors: Hemang Chawla, Matti Jukola, Elahe Arani, and Bahram Zonooz
- Abstract summary: We demonstrate an approach to computing 3D traffic sign positions without knowing the camera focal lengths, principal point, and distortion coefficients a priori.
We achieve an average single journey relative and absolute positioning accuracy of 0.26 m and 1.38 m, respectively.
- Score: 11.38332845467423
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autonomous vehicles and driver assistance systems utilize maps of 3D semantic
landmarks for improved decision making. However, scaling the mapping process as
well as regularly updating such maps come with a huge cost. Crowdsourced
mapping of these landmarks such as traffic sign positions provides an appealing
alternative. The state-of-the-art approaches to crowdsourced mapping use ground
truth camera parameters, which may not always be known or may change over time.
In this work, we demonstrate an approach to computing 3D traffic sign positions
without knowing the camera focal lengths, principal point, and distortion
coefficients a priori. We validate our proposed approach on a public dataset of
traffic signs in KITTI. Using only a monocular color camera and GPS, we achieve
an average single journey relative and absolute positioning accuracy of 0.26 m
and 1.38 m, respectively.
Related papers
- Neural Semantic Map-Learning for Autonomous Vehicles [85.8425492858912]
We present a mapping system that fuses local submaps gathered from a fleet of vehicles at a central instance to produce a coherent map of the road environment.
Our method jointly aligns and merges the noisy and incomplete local submaps using a scene-specific Neural Signed Distance Field.
We leverage memory-efficient sparse feature-grids to scale to large areas and introduce a confidence score to model uncertainty in scene reconstruction.
arXiv Detail & Related papers (2024-10-10T10:10:03Z) - Prior Based Online Lane Graph Extraction from Single Onboard Camera
Image [133.68032636906133]
We tackle online estimation of the lane graph from a single onboard camera image.
The prior is extracted from the dataset through a transformer based Wasserstein Autoencoder.
The autoencoder is then used to enhance the initial lane graph estimates.
arXiv Detail & Related papers (2023-07-25T08:58:26Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - Sparse Semantic Map-Based Monocular Localization in Traffic Scenes Using
Learned 2D-3D Point-Line Correspondences [29.419138863851526]
Given a query image, the goal is to estimate the camera pose corresponding to the prior map.
Existing approaches rely heavily on dense point descriptors at the feature level to solve the registration problem.
We propose a sparse semantic map-based monocular localization method, which solves 2D-3D registration via a well-designed deep neural network.
arXiv Detail & Related papers (2022-10-10T10:29:07Z) - Satellite Image Based Cross-view Localization for Autonomous Vehicle [59.72040418584396]
This paper shows that by using an off-the-shelf high-definition satellite image as a ready-to-use map, we are able to achieve cross-view vehicle localization up to a satisfactory accuracy.
Our method is validated on KITTI and Ford Multi-AV Seasonal datasets as ground view and Google Maps as the satellite view.
arXiv Detail & Related papers (2022-07-27T13:16:39Z) - Real-time Full-stack Traffic Scene Perception for Autonomous Driving
with Roadside Cameras [20.527834125706526]
We propose a novel framework for traffic scene perception with roadside cameras.
The proposed framework covers a full-stack of roadside perception, including object detection, object localization, object tracking, and multi-camera information fusion.
Our framework is deployed at a two-lane roundabout located at Ellsworth Rd. and State St., Ann Arbor, MI, USA, providing 7x24 real-time traffic flow monitoring and high-precision vehicle trajectory extraction.
arXiv Detail & Related papers (2022-06-20T13:33:52Z) - Semantic Image Alignment for Vehicle Localization [111.59616433224662]
We present a novel approach to vehicle localization in dense semantic maps using semantic segmentation from a monocular camera.
In contrast to existing visual localization approaches, the system does not require additional keypoint features, handcrafted localization landmark extractors or expensive LiDAR sensors.
arXiv Detail & Related papers (2021-10-08T14:40:15Z) - Automatic Map Update Using Dashcam Videos [1.6911482053867475]
We propose an SfM-based solution for automatic map update, with a focus on real-time change detection and localization.
Our system can locate the objects detected from 2D images in a 3D space, utilizing sparse SfM point clouds.
arXiv Detail & Related papers (2021-09-24T18:00:57Z) - Crowdsourced 3D Mapping: A Combined Multi-View Geometry and
Self-Supervised Learning Approach [10.610403488989428]
We propose a framework that estimates the 3D positions of semantically meaningful landmarks without assuming known camera intrinsics.
We utilize multi-view geometry as well as deep learning based self-calibration, depth, and ego-motion estimation for traffic sign positioning.
We achieve an average single-journey relative and absolute positioning accuracy of 39cm and 1.26m respectively.
arXiv Detail & Related papers (2020-07-25T12:10:16Z) - Road Curb Detection and Localization with Monocular Forward-view Vehicle
Camera [74.45649274085447]
We propose a robust method for estimating road curb 3D parameters using a calibrated monocular camera equipped with a fisheye lens.
Our approach is able to estimate the vehicle to curb distance in real time with mean accuracy of more than 90%.
arXiv Detail & Related papers (2020-02-28T00:24:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.