DA4AD: End-to-End Deep Attention-based Visual Localization for
Autonomous Driving
- URL: http://arxiv.org/abs/2003.03026v2
- Date: Mon, 13 Jul 2020 17:31:33 GMT
- Title: DA4AD: End-to-End Deep Attention-based Visual Localization for
Autonomous Driving
- Authors: Yao Zhou, Guowei Wan, Shenhua Hou, Li Yu, Gang Wang, Xiaofei Rui,
Shiyu Song
- Abstract summary: We present a visual localization framework based on novel deep attention aware features for autonomous driving.
Our method achieves a competitive localization accuracy when compared to the LiDAR-based localization solutions.
- Score: 19.02445537167235
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a visual localization framework based on novel deep attention
aware features for autonomous driving that achieves centimeter level
localization accuracy. Conventional approaches to the visual localization
problem rely on handcrafted features or human-made objects on the road. They
are known to be either prone to unstable matching caused by severe appearance
or lighting changes, or too scarce to deliver constant and robust localization
results in challenging scenarios. In this work, we seek to exploit the deep
attention mechanism to search for salient, distinctive and stable features that
are good for long-term matching in the scene through a novel end-to-end deep
neural network. Furthermore, our learned feature descriptors are demonstrated
to be competent to establish robust matches and therefore successfully estimate
the optimal camera poses with high precision. We comprehensively validate the
effectiveness of our method using a freshly collected dataset with high-quality
ground truth trajectories and hardware synchronization between sensors. Results
demonstrate that our method achieves a competitive localization accuracy when
compared to the LiDAR-based localization solutions under various challenging
circumstances, leading to a potential low-cost localization solution for
autonomous driving.
Related papers
- Learning Where to Look: Self-supervised Viewpoint Selection for Active Localization using Geometrical Information [68.10033984296247]
This paper explores the domain of active localization, emphasizing the importance of viewpoint selection to enhance localization accuracy.
Our contributions involve using a data-driven approach with a simple architecture designed for real-time operation, a self-supervised data training method, and the capability to consistently integrate our map into a planning framework tailored for real-world robotics applications.
arXiv Detail & Related papers (2024-07-22T12:32:09Z) - MapLocNet: Coarse-to-Fine Feature Registration for Visual Re-Localization in Navigation Maps [8.373285397029884]
Traditional localization approaches rely on high-definition (HD) maps, which consist of precisely annotated landmarks.
We propose a novel transformer-based neural re-localization method, inspired by image registration.
Our method significantly outperforms the current state-of-the-art OrienterNet on both the nuScenes and Argoverse datasets.
arXiv Detail & Related papers (2024-07-11T14:51:18Z) - Accurate Cooperative Localization Utilizing LiDAR-equipped Roadside Infrastructure for Autonomous Driving [2.0499240875882]
LiDAR now facilitates vehicle localization with centimeter-level accuracy.
These high-precision techniques often face reliability challenges in environments devoid of identifiable map features.
We propose a novel approach that utilizes road side units (RSU) with vehicle-to-infrastructure (V2I) communications to assist vehicle self-localization.
arXiv Detail & Related papers (2024-07-11T10:44:42Z) - Monocular Localization with Semantics Map for Autonomous Vehicles [8.242967098897408]
We propose a novel visual semantic localization algorithm that employs stable semantic features instead of low-level texture features.
First, semantic maps are constructed offline by detecting semantic objects, such as ground markers, lane lines, and poles, using cameras or LiDAR sensors.
Online visual localization is performed through data association of semantic features and map objects.
arXiv Detail & Related papers (2024-06-06T08:12:38Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - Exploring Contextual Representation and Multi-Modality for End-to-End
Autonomous Driving [58.879758550901364]
Recent perception systems enhance spatial understanding with sensor fusion but often lack full environmental context.
We introduce a framework that integrates three cameras to emulate the human field of view, coupled with top-down bird-eye-view semantic data to enhance contextual representation.
Our method achieves displacement error by 0.67m in open-loop settings, surpassing current methods by 6.9% on the nuScenes dataset.
arXiv Detail & Related papers (2022-10-13T05:56:20Z) - Deep Multi-Task Learning for Joint Localization, Perception, and
Prediction [68.50217234419922]
This paper investigates the issues that arise in state-of-the-art autonomy stacks under localization error.
We design a system that jointly performs perception, prediction, and localization.
Our architecture is able to reuse computation between both tasks, and is thus able to correct localization errors efficiently.
arXiv Detail & Related papers (2021-01-17T17:20:31Z) - Active Visual Localization in Partially Calibrated Environments [35.48595012305253]
Humans can robustly localize themselves without a map after they get lost following prominent visual cues or landmarks.
In this work, we aim at endowing autonomous agents the same ability. Such ability is important in robotics applications yet very challenging when an agent is exposed to partially calibrated environments.
We propose an indoor scene dataset ACR-6, which consists of both synthetic and real data and simulates challenging scenarios for active visual localization.
arXiv Detail & Related papers (2020-12-08T08:00:55Z) - Unsupervised Metric Relocalization Using Transform Consistency Loss [66.19479868638925]
Training networks to perform metric relocalization traditionally requires accurate image correspondences.
We propose a self-supervised solution, which exploits a key insight: localizing a query image within a map should yield the same absolute pose, regardless of the reference image used for registration.
We evaluate our framework on synthetic and real-world data, showing our approach outperforms other supervised methods when a limited amount of ground-truth information is available.
arXiv Detail & Related papers (2020-11-01T19:24:27Z) - Visual Localization for Autonomous Driving: Mapping the Accurate
Location in the City Maze [16.824901952766446]
We propose a novel feature voting technique for visual localization.
In our work, we craft the proposed feature voting method into three state-of-the-art visual localization networks.
Our approach can predict location robustly even in challenging inner-city settings.
arXiv Detail & Related papers (2020-08-13T03:59:34Z) - Object-based Illumination Estimation with Rendering-aware Neural
Networks [56.01734918693844]
We present a scheme for fast environment light estimation from the RGBD appearance of individual objects and their local image areas.
With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene.
arXiv Detail & Related papers (2020-08-06T08:23:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.