Evaluation of Cross-View Matching to Improve Ground Vehicle Localization
with Aerial Perception
- URL: http://arxiv.org/abs/2003.06515v4
- Date: Sun, 15 Nov 2020 22:20:51 GMT
- Title: Evaluation of Cross-View Matching to Improve Ground Vehicle Localization
with Aerial Perception
- Authors: Deeksha Dixit, Surabhi Verma, Pratap Tokekar
- Abstract summary: Cross-view matching refers to the problem of finding the closest match for a given query ground view image to one from a database of aerial images.
In this paper, we evaluate cross-view matching for the task of localizing a ground vehicle over a longer trajectory.
- Score: 17.349420462716886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cross-view matching refers to the problem of finding the closest match for a
given query ground view image to one from a database of aerial images. If the
aerial images are geotagged, then the closest matching aerial image can be used
to localize the query ground view image. Due to the recent success of deep
learning methods, several cross-view matching techniques have been proposed.
These approaches perform well for the matching of isolated query images.
However, their evaluation over a trajectory is limited. In this paper, we
evaluate cross-view matching for the task of localizing a ground vehicle over a
longer trajectory. We treat these cross-view matches as sensor measurements
that are fused using a particle filter. We evaluate the performance of this
method using a city-wide dataset collected in a photorealistic simulation by
varying four parameters: height of aerial images, the pitch of the aerial
camera mount, FOV of the ground camera, and the methodology of fusing
cross-view measurements in the particle filter. We also report the results
obtained using our pipeline on a real-world dataset collected using Google
Street View and satellite view APIs.
Related papers
- Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Learning Dense Flow Field for Highly-accurate Cross-view Camera
Localization [15.89357790711828]
This paper addresses the problem of estimating the 3-DoF camera pose for a ground-level image with respect to a satellite image.
We propose a novel end-to-end approach that leverages the learning of dense pixel-wise flow fields in pairs of ground and satellite images.
Our approach reduces the median localization error by 89%, 19%, 80% and 35% on the KITTI, Ford multi-AV, VIGOR and Oxford RobotCar datasets.
arXiv Detail & Related papers (2023-09-27T10:26:26Z) - Render-and-Compare: Cross-View 6 DoF Localization from Noisy Prior [17.08552155321949]
In this work, we propose to go beyond the traditional ground-level setting and exploit the cross-view localization from aerial to ground.
As no public dataset exists for the studied problem, we collect a new dataset that provides a variety of cross-view images from smartphones and drones.
We develop a semi-automatic system to acquire ground-truth poses for query images.
arXiv Detail & Related papers (2023-02-13T11:43:47Z) - Uncertainty-aware Vision-based Metric Cross-view Geolocalization [25.87104194833264]
We present an end-to-end differentiable model that uses the ground and aerial images to predict a probability distribution over possible vehicle poses.
We improve the previous state-of-the-art by a large margin even without ground or aerial data from the test region.
arXiv Detail & Related papers (2022-11-22T10:23:20Z) - Cross-View Image Sequence Geo-localization [6.555961698070275]
Cross-view geo-localization aims to estimate the GPS location of a query ground-view image.
Recent approaches use panoramic ground-view images to increase the range of visibility.
We present the first cross-view geo-localization method that works on a sequence of limited Field-Of-View images.
arXiv Detail & Related papers (2022-10-25T19:46:18Z) - CVLNet: Cross-View Semantic Correspondence Learning for Video-based
Camera Localization [89.69214577915959]
This paper tackles the problem of Cross-view Video-based camera localization.
We propose estimating the query camera's relative displacement to a satellite image before similarity matching.
Experiments have demonstrated the effectiveness of video-based localization over single image-based localization.
arXiv Detail & Related papers (2022-08-07T07:35:17Z) - Satellite Image Based Cross-view Localization for Autonomous Vehicle [59.72040418584396]
This paper shows that by using an off-the-shelf high-definition satellite image as a ready-to-use map, we are able to achieve cross-view vehicle localization up to a satisfactory accuracy.
Our method is validated on KITTI and Ford Multi-AV Seasonal datasets as ground view and Google Maps as the satellite view.
arXiv Detail & Related papers (2022-07-27T13:16:39Z) - Cross-Camera Trajectories Help Person Retrieval in a Camera Network [124.65912458467643]
Existing methods often rely on purely visual matching or consider temporal constraints but ignore the spatial information of the camera network.
We propose a pedestrian retrieval framework based on cross-camera generation, which integrates both temporal and spatial information.
To verify the effectiveness of our method, we construct the first cross-camera pedestrian trajectory dataset.
arXiv Detail & Related papers (2022-04-27T13:10:48Z) - CroCo: Cross-Modal Contrastive learning for localization of Earth
Observation data [62.96337162094726]
It is of interest to localize a ground-based LiDAR point cloud on remote sensing imagery.
We propose a contrastive learning-based method that trains on DEM and high-resolution optical imagery.
In the best scenario, the Top-1 score of 0.71 and Top-5 score of 0.81 are obtained.
arXiv Detail & Related papers (2022-04-14T15:55:00Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - Leveraging Photogrammetric Mesh Models for Aerial-Ground Feature Point
Matching Toward Integrated 3D Reconstruction [19.551088857830944]
Integration of aerial and ground images has been proved as an efficient approach to enhance the surface reconstruction in urban environments.
Previous studies based on geometry-aware image rectification have alleviated this problem.
We propose a novel approach: leveraging photogrammetric mesh models for aerial-ground image matching.
arXiv Detail & Related papers (2020-02-21T01:47:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.