Appearance-Invariant 6-DoF Visual Localization using Generative
Adversarial Networks
- URL: http://arxiv.org/abs/2012.13191v1
- Date: Thu, 24 Dec 2020 10:43:43 GMT
- Title: Appearance-Invariant 6-DoF Visual Localization using Generative
Adversarial Networks
- Authors: Yimin Lin, Jianfeng Huang, Shiguo Lian
- Abstract summary: We propose a novel visual localization network when outside environment has changed such as different illumination, weather and season.
The visual localization network is composed of a feature extraction network and pose regression network.
Results show that our method outperforms state-of-the-art methods in the scenarios with various environment changes.
- Score: 7.04719493717788
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel visual localization network when outside environment has
changed such as different illumination, weather and season. The visual
localization network is composed of a feature extraction network and pose
regression network. The feature extraction network is made up of an encoder
network based on the Generative Adversarial Network CycleGAN, which can capture
intrinsic appearance-invariant feature maps from unpaired samples of different
weathers and seasons. With such an invariant feature, we use a 6-DoF pose
regression network to tackle long-term visual localization in the presence of
outdoor illumination, weather and season changes. A variety of challenging
datasets for place recognition and localization are used to prove our visual
localization network, and the results show that our method outperforms
state-of-the-art methods in the scenarios with various environment changes.
Related papers
- Long-Term Invariant Local Features via Implicit Cross-Domain
Correspondences [79.21515035128832]
We conduct a thorough analysis of the performance of current state-of-the-art feature extraction networks under various domain changes.
We propose a novel data-centric method, Implicit Cross-Domain Correspondences (iCDC)
iCDC represents the same environment with multiple Neural Radiance Fields, each fitting the scene under individual visual domains.
arXiv Detail & Related papers (2023-11-06T18:53:01Z) - Self-Supervised Feature Learning for Long-Term Metric Visual
Localization [16.987148593917905]
We present a novel self-supervised feature learning framework for metric visual localization.
We use a sequence-based image matching algorithm to generate image correspondences without ground-truth labels.
We can then sample image pairs to train a deep neural network that learns sparse features with associated descriptors and scores without ground-truth pose supervision.
arXiv Detail & Related papers (2022-11-30T21:15:05Z) - Real-time Local Feature with Global Visual Information Enhancement [6.640269424085467]
Current deep learning-based local feature algorithms always utilize convolution neural network (CNN) architecture with limited receptive field.
The proposed method introduces a global enhancement module to fuse global visual clues in a light-weight network.
Experiments on the public benchmarks demonstrate that the proposal can achieve considerable robustness against visual interference and meanwhile run in real time.
arXiv Detail & Related papers (2022-11-20T13:44:20Z) - Peripheral Vision Transformer [52.55309200601883]
We take a biologically inspired approach and explore to model peripheral vision in deep neural networks for visual recognition.
We propose to incorporate peripheral position encoding to the multi-head self-attention layers to let the network learn to partition the visual field into diverse peripheral regions given training data.
We evaluate the proposed network, dubbed PerViT, on the large-scale ImageNet dataset and systematically investigate the inner workings of the model for machine perception.
arXiv Detail & Related papers (2022-06-14T12:47:47Z) - Multiple-environment Self-adaptive Network for Aerial-view
Geo-localization [85.52750931345287]
Aerial-view geo-localization tends to determine an unknown position through matching the drone-view image with the geo-tagged satellite-view image.
We propose a Multiple-environment Self-adaptive Network (MuSe-Net) to adjust the domain shift caused by environmental changing.
In particular, MuSe-Net employs a two-branch neural network containing one multiple-environment style extraction network and one self-adaptive feature extraction network.
arXiv Detail & Related papers (2022-04-18T16:04:29Z) - Location-Free Camouflage Generation Network [82.74353843283407]
Camouflage is a common visual phenomenon, which refers to hiding the foreground objects into the background images, making them briefly invisible to the human eye.
This paper proposes a novel Location-free Camouflage Generation Network (LCG-Net) that fuse high-level features of foreground and background image, and generate result by one inference.
Experiments show that our method has results as satisfactory as state-of-the-art in the single-appearance regions and are less likely to be completely invisible, but far exceed the quality of the state-of-the-art in the multi-appearance regions.
arXiv Detail & Related papers (2022-03-18T10:33:40Z) - Wide and Narrow: Video Prediction from Context and Motion [54.21624227408727]
We propose a new framework to integrate these complementary attributes to predict complex pixel dynamics through deep networks.
We present global context propagation networks that aggregate the non-local neighboring representations to preserve the contextual information over the past frames.
We also devise local filter memory networks that generate adaptive filter kernels by storing the motion of moving objects in the memory.
arXiv Detail & Related papers (2021-10-22T04:35:58Z) - Sill-Net: Feature Augmentation with Separated Illumination
Representation [35.25230715669166]
We propose a novel neural network architecture called Separating-Illumination Network (Sill-Net)
Sill-Net learns to separate illumination features from images, and then during training we augment training samples with these separated illumination features in the feature space.
Experimental results demonstrate that our approach outperforms current state-of-the-art methods in several object classification benchmarks.
arXiv Detail & Related papers (2021-02-06T09:00:10Z) - Domain Adaptation of Learned Features for Visual Localization [60.6817896667435]
We tackle the problem of visual localization under changing conditions, such as time of day, weather, and seasons.
Recent learned local features based on deep neural networks have shown superior performance over classical hand-crafted local features.
We present a novel and practical approach, where only a few examples are needed to reduce the domain gap.
arXiv Detail & Related papers (2020-08-21T05:17:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.