A Deep Learning Ensemble Framework for Off-Nadir Geocentric Pose
Prediction
- URL: http://arxiv.org/abs/2205.11230v3
- Date: Sat, 6 Aug 2022 08:16:21 GMT
- Title: A Deep Learning Ensemble Framework for Off-Nadir Geocentric Pose
Prediction
- Authors: Christopher Sun, Jai Sharma, Milind Maiti
- Abstract summary: Current software functions optimally only on near-nadir images, though off-nadir images are often the first sources of information following a natural disaster.
This study proposes a deep learning ensemble framework to predict geocentric pose using 5,923 near-nadir and off-nadir RGB satellite images of cities worldwide.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Computational methods to accelerate natural disaster response include change
detection, map alignment, and vision-aided navigation. Current software
functions optimally only on near-nadir images, though off-nadir images are
often the first sources of information following a natural disaster. The use of
off-nadir images for the aforementioned tasks requires the computation of
geocentric pose, which is an aerial vehicle's spatial orientation with respect
to gravity. This study proposes a deep learning ensemble framework to predict
geocentric pose using 5,923 near-nadir and off-nadir RGB satellite images of
cities worldwide. First, a U-Net Fully Convolutional Neural Network predicts
the pixel-wise above-ground elevation mask of the RGB images. Then, the
elevation masks are concatenated with the RGB images to form four-channel
inputs fed into a second convolutional model, which predicts orientation angle
and magnification scale. A performance accuracy of R2=0.917 significantly
outperforms previous methodologies. In addition, outlier removal is performed
through supervised interpolation, and a sensitivity analysis of elevation masks
is conducted to gauge the usefulness of data features, motivating future
avenues of feature engineering. The high-accuracy software built in this study
contributes to mapping and navigation procedures for effective disaster
response to save lives.
Related papers
- FlightScope: A Deep Comprehensive Review of Aircraft Detection Algorithms in Satellite Imagery [2.9687381456164004]
This paper critically evaluates and compares a suite of advanced object detection algorithms customized for the task of identifying aircraft within satellite imagery.
This research encompasses an array of methodologies including YOLO versions 5 and 8, Faster RCNN, CenterNet, RetinaNet, RTMDet, and DETR, all trained from scratch.
YOLOv5 emerges as a robust solution for aerial object detection, underlining its importance through superior mean average precision, Recall, and Intersection over Union scores.
arXiv Detail & Related papers (2024-04-03T17:24:27Z) - RFTrans: Leveraging Refractive Flow of Transparent Objects for Surface
Normal Estimation and Manipulation [50.10282876199739]
This paper introduces RFTrans, an RGB-D-based method for surface normal estimation and manipulation of transparent objects.
It integrates the RFNet, which predicts refractive flow, object mask, and boundaries, followed by the F2Net, which estimates surface normal from the refractive flow.
A real-world robot grasping task witnesses an 83% success rate, proving that refractive flow can help enable direct sim-to-real transfer.
arXiv Detail & Related papers (2023-11-21T07:19:47Z) - Combining geolocation and height estimation of objects from street level
imagery [5.887281983256354]
We propose a pipeline for combined multi-class object geolocation and height estimation from street level RGB imagery.
The proposed technique uses image metadata along with coordinates of objects detected in the image plane as found by a custom-trained Convolutional Neural Network.
arXiv Detail & Related papers (2023-05-14T19:40:02Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - GoToNet: Fast Monocular Scene Exposure and Exploration [0.6204265638103346]
We present a novel method for real-time environment exploration.
Our method requires only one look (image) to make a good tactical decision.
Two direction predictions, characterized by pixels dubbed the Goto and Lookat pixels, comprise the core of our method.
arXiv Detail & Related papers (2022-06-13T08:28:31Z) - DeepRM: Deep Recurrent Matching for 6D Pose Refinement [77.34726150561087]
DeepRM is a novel recurrent network architecture for 6D pose refinement.
The architecture incorporates LSTM units to propagate information through each refinement step.
DeepRM achieves state-of-the-art performance on two widely accepted challenging datasets.
arXiv Detail & Related papers (2022-05-28T16:18:08Z) - Sparse Auxiliary Networks for Unified Monocular Depth Prediction and
Completion [56.85837052421469]
Estimating scene geometry from data obtained with cost-effective sensors is key for robots and self-driving cars.
In this paper, we study the problem of predicting dense depth from a single RGB image with optional sparse measurements from low-cost active depth sensors.
We introduce Sparse Networks (SANs), a new module enabling monodepth networks to perform both the tasks of depth prediction and completion.
arXiv Detail & Related papers (2021-03-30T21:22:26Z) - Memory-Augmented Reinforcement Learning for Image-Goal Navigation [67.3963444878746]
We present a novel method that leverages a cross-episode memory to learn to navigate.
In order to avoid overfitting, we propose to use data augmentation on the RGB input during training.
We obtain this competitive performance from RGB input only, without access to additional sensors such as position or depth.
arXiv Detail & Related papers (2021-01-13T16:30:20Z) - Aerial Height Prediction and Refinement Neural Networks with Semantic
and Geometric Guidance [0.0]
This letter proposes a two-stage approach, where first a multi-task neural network is used to predict the height map resulting from a single RGB aerial input image.
Experiments on two publicly available datasets show that our method is capable of producing state-of-the-art results.
arXiv Detail & Related papers (2020-11-21T01:39:37Z) - Height estimation from single aerial images using a deep ordinal
regression network [12.991266182762597]
We deal with the ambiguous and unsolved problem of height estimation from a single aerial image.
Driven by the success of deep learning, especially deep convolution neural networks (CNNs), some researches have proposed to estimate height information from a single aerial image.
In this paper, we proposed to divide height values into spacing-increasing intervals and transform the regression problem into an ordinal regression problem.
arXiv Detail & Related papers (2020-06-04T12:03:51Z) - Where am I looking at? Joint Location and Orientation Estimation by
Cross-View Matching [95.64702426906466]
Cross-view geo-localization is a problem given a large-scale database of geo-tagged aerial images.
Knowing orientation between ground and aerial images can significantly reduce matching ambiguity between these two views.
We design a Dynamic Similarity Matching network to estimate cross-view orientation alignment during localization.
arXiv Detail & Related papers (2020-05-08T05:21:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.