Monocular Camera Localization for Automated Vehicles Using Image
Retrieval
- URL: http://arxiv.org/abs/2109.06296v3
- Date: Thu, 30 Nov 2023 14:15:06 GMT
- Title: Monocular Camera Localization for Automated Vehicles Using Image
Retrieval
- Authors: Eunhyek Joa, Yibo Sun, and Francesco Borrelli
- Abstract summary: We address the problem of finding the current position and heading angle of an autonomous vehicle in real-time using a single camera.
Compared to methods which require LiDARs and high definition (HD) 3D maps in real-time, the proposed approach is easily scalable and computationally efficient.
- Score: 8.594652891734288
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address the problem of finding the current position and heading angle of
an autonomous vehicle in real-time using a single camera. Compared to methods
which require LiDARs and high definition (HD) 3D maps in real-time, the
proposed approach is easily scalable and computationally efficient, at the
price of lower precision.
The new method combines and adapts existing algorithms in three different
fields: image retrieval, mapping database, and particle filtering. The result
is a simple, real-time localization method using an image retrieval method
whose performance is comparable to other monocular camera localization methods
which use a map built with LiDARs.
We evaluate the proposed method using the KITTI odometry dataset and via
closed-loop experiments with an indoor 1:10 autonomous vehicle. The tests
demonstrate real-time capability and a 10cm level accuracy. Also, experimental
results of the closed-loop indoor tests show the presence of a positive
feedback loop between the localization error and the control error. Such
phenomena is analysed in details at the end of the article.
Related papers
- Improving Online Lane Graph Extraction by Object-Lane Clustering [106.71926896061686]
We propose an architecture and loss formulation to improve the accuracy of local lane graph estimates.
The proposed method learns to assign the objects to centerlines by considering the centerlines as cluster centers.
We show that our method can achieve significant performance improvements by using the outputs of existing 3D object detection methods.
arXiv Detail & Related papers (2023-07-20T15:21:28Z) - A comparison of uncertainty estimation approaches for DNN-based camera
localization [6.053739577423792]
This work compares the performances of three uncertainty estimation methods: Monte Carlo Dropout (MCD), Deep Ensemble (DE), and Deep Evidential Regression (DER)
We achieve accurate camera localization and a calibrated uncertainty, to the point that some method can be used for detecting localization failures.
arXiv Detail & Related papers (2022-11-02T16:15:28Z) - Multi-View Object Pose Refinement With Differentiable Renderer [22.040014384283378]
This paper introduces a novel multi-view 6 DoF object pose refinement approach focusing on improving methods trained on synthetic data.
It is based on the DPOD detector, which produces dense 2D-3D correspondences between the model vertices and the image pixels in each frame.
We report excellent performance in comparison to the state-of-the-art methods trained on the synthetic and real data.
arXiv Detail & Related papers (2022-07-06T17:02:22Z) - VPIT: Real-time Embedded Single Object 3D Tracking Using Voxel Pseudo Images [90.60881721134656]
We propose a novel voxel-based 3D single object tracking (3D SOT) method called Voxel Pseudo Image Tracking (VPIT)
Experiments on KITTI Tracking dataset show that VPIT is the fastest 3D SOT method and maintains competitive Success and Precision values.
arXiv Detail & Related papers (2022-06-06T14:02:06Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Learning Condition Invariant Features for Retrieval-Based Localization
from 1M Images [85.81073893916414]
We develop a novel method for learning more accurate and better generalizing localization features.
On the challenging Oxford RobotCar night condition, our method outperforms the well-known triplet loss by 24.4% in localization accuracy within 5m.
arXiv Detail & Related papers (2020-08-27T14:46:22Z) - 3D Lidar Mapping Relative Accuracy Automatic Evaluation Algorithm [7.4977759910523964]
HD (High Definition) map based on 3D lidar plays a vital role in autonomous vehicle localization, planning, decision-making, perception, etc.
Many 3D lidar mapping technologies related to SLAM are used in HD map construction to ensure its high accuracy.
To evaluate the accuracy of 3D lidar mapping, the most common methods use ground truth of poses to calculate the error between estimated poses and ground truth.
In this paper, we proposed a relative accuracy evaluation algorithm that can automatically evaluate the accuracy of HD map built by 3D lidar mapping without ground truth.
arXiv Detail & Related papers (2020-06-01T11:30:31Z) - Multi-View Optimization of Local Feature Geometry [70.18863787469805]
We address the problem of refining the geometry of local image features from multiple views without known scene or camera geometry.
Our proposed method naturally complements the traditional feature extraction and matching paradigm.
We show that our method consistently improves the triangulation and camera localization performance for both hand-crafted and learned local features.
arXiv Detail & Related papers (2020-03-18T17:22:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.