Precise Aerial Image Matching based on Deep Homography Estimation
- URL: http://arxiv.org/abs/2107.08768v1
- Date: Mon, 19 Jul 2021 11:52:52 GMT
- Title: Precise Aerial Image Matching based on Deep Homography Estimation
- Authors: Myeong-Seok Oh, Yong-Ju Lee, Seong-Whan Lee
- Abstract summary: We propose a deep homography alignment network to precisely match two aerial images.
The proposed network is possible to train the matching network with a higher degree of freedom.
We introduce a method that can effectively learn the difficult-to-learn homography estimation network.
- Score: 21.948001630564363
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Aerial image registration or matching is a geometric process of aligning two
aerial images captured in different environments. Estimating the precise
transformation parameters is hindered by various environments such as time,
weather, and viewpoints. The characteristics of the aerial images are mainly
composed of a straight line owing to building and road. Therefore, the straight
lines are distorted when estimating homography parameters directly between two
images. In this paper, we propose a deep homography alignment network to
precisely match two aerial images by progressively estimating the various
transformation parameters. The proposed network is possible to train the
matching network with a higher degree of freedom by progressively analyzing the
transformation parameters. The precision matching performances have been
increased by applying homography transformation. In addition, we introduce a
method that can effectively learn the difficult-to-learn homography estimation
network. Since there is no published learning data for aerial image
registration, in this paper, a pair of images to which random homography
transformation is applied within a certain range is used for learning. Hence,
we could confirm that the deep homography alignment network shows high
precision matching performance compared with conventional works.
Related papers
- MsMorph: An Unsupervised pyramid learning network for brain image registration [4.000367245594772]
MsMorph is an image registration framework aimed at mimicking the manual process of registering image pairs.
It decodes semantic information at different scales and continuously compen-sates for the predicted deformation field.
The proposed method simulates the manual approach to registration, focusing on different regions of the image pairs and their neighborhoods.
arXiv Detail & Related papers (2024-10-23T19:20:57Z) - Predicting ptychography probe positions using single-shot phase retrieval neural network [5.889405057118457]
We propose a fundamentally new approach for ptychography probe position prediction for data with large position errors.
A neural network is used to make single-shot phase retrieval on individual diffraction patterns, yielding the object image at each scan point.
We show that our method can achieve good position prediction accuracy for data with large and accumulating errors on the order of $102$ pixels.
arXiv Detail & Related papers (2024-05-31T15:21:06Z) - Experimental Validation of Ultrasound Beamforming with End-to-End Deep Learning for Single Plane Wave Imaging [0.810120481608724]
Ultrafast ultrasound imaging insonifies a medium with one or a combination of a few plane waves at different beam-steered angles instead of many focused waves.
Deep learning approaches have been proposed to mitigate this disadvantage, in particular for single plane wave imaging.
We consider data-to-image networks which incorporate a conventional image formation techniques as differentiable layers in the network architecture.
arXiv Detail & Related papers (2024-04-22T13:58:36Z) - Cross-domain and Cross-dimension Learning for Image-to-Graph
Transformers [50.576354045312115]
Direct image-to-graph transformation is a challenging task that solves object detection and relationship prediction in a single model.
We introduce a set of methods enabling cross-domain and cross-dimension transfer learning for image-to-graph transformers.
We demonstrate our method's utility in cross-domain and cross-dimension experiments, where we pretrain our models on 2D satellite images before applying them to vastly different target domains in 2D and 3D.
arXiv Detail & Related papers (2024-03-11T10:48:56Z) - Dual-Flow Transformation Network for Deformable Image Registration with
Region Consistency Constraint [95.30864269428808]
Current deep learning (DL)-based image registration approaches learn the spatial transformation from one image to another by leveraging a convolutional neural network.
We present a novel dual-flow transformation network with region consistency constraint which maximizes the similarity of ROIs within a pair of images.
Experiments on four public 3D MRI datasets show that the proposed method achieves the best registration performance in accuracy and generalization.
arXiv Detail & Related papers (2021-12-04T05:30:44Z) - LocalTrans: A Multiscale Local Transformer Network for Cross-Resolution
Homography Estimation [52.63874513999119]
Cross-resolution image alignment is a key problem in multiscale giga photography.
Existing deep homography methods neglecting the explicit formulation of correspondences between them, which leads to degraded accuracy in cross-resolution challenges.
We propose a local transformer network embedded within a multiscale structure to explicitly learn correspondences between the multimodal inputs.
arXiv Detail & Related papers (2021-06-08T02:51:45Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Coarse-to-Fine Gaze Redirection with Numerical and Pictorial Guidance [74.27389895574422]
We propose a novel gaze redirection framework which exploits both a numerical and a pictorial direction guidance.
The proposed method outperforms the state-of-the-art approaches in terms of both image quality and redirection precision.
arXiv Detail & Related papers (2020-04-07T01:17:27Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z) - Leveraging Photogrammetric Mesh Models for Aerial-Ground Feature Point
Matching Toward Integrated 3D Reconstruction [19.551088857830944]
Integration of aerial and ground images has been proved as an efficient approach to enhance the surface reconstruction in urban environments.
Previous studies based on geometry-aware image rectification have alleviated this problem.
We propose a novel approach: leveraging photogrammetric mesh models for aerial-ground image matching.
arXiv Detail & Related papers (2020-02-21T01:47:59Z) - A Two-Stream Symmetric Network with Bidirectional Ensemble for Aerial
Image Matching [24.089374888914143]
We propose a novel method to precisely match two aerial images that were obtained in different environments via a two-stream deep network.
By internally augmenting the target image, the network considers the two-stream with the three input images and reflects the additional augmented pair in the training.
arXiv Detail & Related papers (2020-02-04T14:38:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.