DeepTracks: Geopositioning Maritime Vehicles in Video Acquired from a
Moving Platform
- URL: http://arxiv.org/abs/2109.01235v1
- Date: Thu, 2 Sep 2021 22:36:16 GMT
- Title: DeepTracks: Geopositioning Maritime Vehicles in Video Acquired from a
Moving Platform
- Authors: Jianli Wei, Guanyu Xu, Alper Yilmaz
- Abstract summary: Given imagery from a camera mounted on a moving platform, we predict the geoposition of a target boat visible in images.
Our solution uses recent ML algorithms, the camera-scene geometry and Bayesian filtering.
We tested the performance of our approach using GPS ground truth and show the accuracy and speed of the estimated geopositions.
- Score: 0.6445605125467573
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Geopositioning and tracking a moving boat at sea is a very challenging
problem, requiring boat detection, matching and estimating its GPS location
from imagery with no common features. The problem can be stated as follows:
given imagery from a camera mounted on a moving platform with known GPS
location as the only valid sensor, we predict the geoposition of a target boat
visible in images. Our solution uses recent ML algorithms, the camera-scene
geometry and Bayesian filtering. The proposed pipeline first detects and tracks
the target boat's location in the image with the strategy of tracking by
detection. This image location is then converted to geoposition to the local
sea coordinates referenced to the camera GPS location using plane projective
geometry. Finally, target boat local coordinates are transformed to global GPS
coordinates to estimate the geoposition. To achieve a smooth geotrajectory, we
apply unscented Kalman filter (UKF) which implicitly overcomes small detection
errors in the early stages of the pipeline. We tested the performance of our
approach using GPS ground truth and show the accuracy and speed of the
estimated geopositions. Our code is publicly available at
https://github.com/JianliWei1995/AI-Track-at-Sea.
Related papers
- GAReT: Cross-view Video Geolocalization with Adapters and Auto-Regressive Transformers [53.80009458891537]
Cross-view video geo-localization aims to derive GPS trajectories from street-view videos by aligning them with aerial-view images.
Current CVGL methods use camera and odometry data, typically absent in real-world scenarios.
We propose GAReT, a fully transformer-based method for CVGL that does not require camera and odometry data.
arXiv Detail & Related papers (2024-08-05T21:29:33Z) - GeoCLIP: Clip-Inspired Alignment between Locations and Images for
Effective Worldwide Geo-localization [61.10806364001535]
Worldwide Geo-localization aims to pinpoint the precise location of images taken anywhere on Earth.
Existing approaches divide the globe into discrete geographic cells, transforming the problem into a classification task.
We propose GeoCLIP, a novel CLIP-inspired Image-to-GPS retrieval approach that enforces alignment between the image and its corresponding GPS locations.
arXiv Detail & Related papers (2023-09-27T20:54:56Z) - A Gis Aided Approach for Geolocalizing an Unmanned Aerial System Using
Deep Learning [0.4297070083645048]
We propose an alternative approach to geolocalize a UAS when GPS signal is degraded or denied.
Considering UAS has a downward-looking camera on its platform that can acquire real-time images as the platform flies, we apply modern deep learning techniques to achieve geolocalization.
We extract GIS information from OpenStreetMap (OSM) to semantically segment matched features into building and terrain classes.
arXiv Detail & Related papers (2022-08-25T17:51:15Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - Accurate 3-DoF Camera Geo-Localization via Ground-to-Satellite Image
Matching [102.39635336450262]
We address the problem of ground-to-satellite image geo-localization by matching a query image captured at the ground level against a large-scale database with geotagged satellite images.
Our new method is able to achieve the fine-grained location of a query image, up to pixel size precision of the satellite image.
arXiv Detail & Related papers (2022-03-26T20:10:38Z) - Continuous Self-Localization on Aerial Images Using Visual and Lidar
Sensors [25.87104194833264]
We propose a novel method for geo-tracking in outdoor environments by registering a vehicle's sensor information with aerial imagery of an unseen target region.
We train a model in a metric learning setting to extract visual features from ground and aerial images.
Our method is the first to utilize on-board cameras in an end-to-end differentiable model for metric self-localization on unseen orthophotos.
arXiv Detail & Related papers (2022-03-07T12:25:44Z) - ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints [94.60414567852536]
Long-range navigation requires both planning and reasoning about local traversability.
We propose a learning-based approach that integrates learning and planning.
ViKiNG can leverage its image-based learned controller and goal-directed to navigate to goals up to 3 kilometers away.
arXiv Detail & Related papers (2022-02-23T02:14:23Z) - Augmenting Depth Estimation with Geospatial Context [20.79700110152325]
We propose an end-to-end architecture for depth estimation.
We infer a synthetic ground-level depth map from a co-located overhead image, then fuses it inside of an encoder/decoder style segmentation network.
Results demonstrate that integrating geospatial context significantly reduces error compared to baselines.
arXiv Detail & Related papers (2021-09-20T23:24:17Z) - EagerMOT: 3D Multi-Object Tracking via Sensor Fusion [68.8204255655161]
Multi-object tracking (MOT) enables mobile robots to perform well-informed motion planning and navigation by localizing surrounding objects in 3D space and time.
Existing methods rely on depth sensors (e.g., LiDAR) to detect and track targets in 3D space, but only up to a limited sensing range due to the sparsity of the signal.
We propose EagerMOT, a simple tracking formulation that integrates all available object observations from both sensor modalities to obtain a well-informed interpretation of the scene dynamics.
arXiv Detail & Related papers (2021-04-29T22:30:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.