Depth-Aware Multi-Grid Deep Homography Estimation with Contextual
Correlation
- URL: http://arxiv.org/abs/2107.02524v1
- Date: Tue, 6 Jul 2021 10:33:12 GMT
- Title: Depth-Aware Multi-Grid Deep Homography Estimation with Contextual
Correlation
- Authors: Lang Nie, Chunyu Lin, Kang Liao, Shuaicheng Liu, Yao Zhao
- Abstract summary: Homography estimation is an important task in computer vision, such as image stitching, video stabilization, and camera calibration.
Traditional homography estimation methods depend on the quantity and distribution of feature points, leading to poor robustness in textureless scenes.
We propose a contextual correlation layer, which can capture the long-range correlation on feature maps and flexibly be bridged in a learning framework.
We equip our network with depth perception capability, by introducing a novel depth-aware shape-preserved loss.
- Score: 38.95610086309832
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Homography estimation is an important task in computer vision, such as image
stitching, video stabilization, and camera calibration. Traditional homography
estimation methods heavily depend on the quantity and distribution of feature
points, leading to poor robustness in textureless scenes. The learning
solutions, on the contrary, try to learn robust deep features but demonstrate
unsatisfying performance in the scenes of low overlap rates. In this paper, we
address the two problems simultaneously, by designing a contextual correlation
layer, which can capture the long-range correlation on feature maps and
flexibly be bridged in a learning framework. In addition, considering that a
single homography can not represent the complex spatial transformation in
depth-varying images with parallax, we propose to predict multi-grid homography
from global to local. Moreover, we equip our network with depth perception
capability, by introducing a novel depth-aware shape-preserved loss. Extensive
experiments demonstrate the superiority of our method over other
state-of-the-art solutions in the synthetic benchmark dataset and real-world
dataset. The codes and models will be available at
https://github.com/nie-lang/Multi-Grid-Deep-Homogarphy.
Related papers
- Unveiling the Depths: A Multi-Modal Fusion Framework for Challenging
Scenarios [103.72094710263656]
This paper presents a novel approach that identifies and integrates dominant cross-modality depth features with a learning-based framework.
We propose a novel confidence loss steering a confidence predictor network to yield a confidence map specifying latent potential depth areas.
With the resulting confidence map, we propose a multi-modal fusion network that fuses the final depth in an end-to-end manner.
arXiv Detail & Related papers (2024-02-19T04:39:16Z) - Temporally Consistent Online Depth Estimation Using Point-Based Fusion [6.5514240555359455]
We aim to estimate temporally consistent depth maps of video streams in an online setting.
This is a difficult problem as future frames are not available and the method must choose between enforcing consistency and correcting errors from previous estimations.
We propose to address these challenges by using a global point cloud that is dynamically updated each frame, along with a learned fusion approach in image space.
arXiv Detail & Related papers (2023-04-15T00:04:18Z) - Parallax-Tolerant Unsupervised Deep Image Stitching [57.76737888499145]
We propose UDIS++, a parallax-tolerant unsupervised deep image stitching technique.
First, we propose a robust and flexible warp to model the image registration from global homography to local thin-plate spline motion.
To further eliminate the parallax artifacts, we propose to composite the stitched image seamlessly by unsupervised learning for seam-driven composition masks.
arXiv Detail & Related papers (2023-02-16T10:40:55Z) - Deep Convolutional Pooling Transformer for Deepfake Detection [54.10864860009834]
We propose a deep convolutional Transformer to incorporate decisive image features both locally and globally.
Specifically, we apply convolutional pooling and re-attention to enrich the extracted features and enhance efficacy.
The proposed solution consistently outperforms several state-of-the-art baselines on both within- and cross-dataset experiments.
arXiv Detail & Related papers (2022-09-12T15:05:41Z) - Joint Learning of Deep Texture and High-Frequency Features for
Computer-Generated Image Detection [24.098604827919203]
We propose a joint learning strategy with deep texture and high-frequency features for CG image detection.
A semantic segmentation map is generated to guide the affine transformation operation.
The combination of the original image and the high-frequency components of the original and rendered images are fed into a multi-branch neural network equipped with attention mechanisms.
arXiv Detail & Related papers (2022-09-07T17:30:40Z) - LocalTrans: A Multiscale Local Transformer Network for Cross-Resolution
Homography Estimation [52.63874513999119]
Cross-resolution image alignment is a key problem in multiscale giga photography.
Existing deep homography methods neglecting the explicit formulation of correspondences between them, which leads to degraded accuracy in cross-resolution challenges.
We propose a local transformer network embedded within a multiscale structure to explicitly learn correspondences between the multimodal inputs.
arXiv Detail & Related papers (2021-06-08T02:51:45Z) - Boosting Monocular Depth Estimation Models to High-Resolution via
Content-Adaptive Multi-Resolution Merging [14.279471205248534]
We show how a consistent scene structure and high-frequency details affect depth estimation performance.
We present a double estimation method that improves the whole-image depth estimation and a patch selection method that adds local details.
We demonstrate that by merging estimations at different resolutions with changing context, we can generate multi-megapixel depth maps with a high level of detail.
arXiv Detail & Related papers (2021-05-28T17:55:15Z) - Towards Unpaired Depth Enhancement and Super-Resolution in the Wild [121.96527719530305]
State-of-the-art data-driven methods of depth map super-resolution rely on registered pairs of low- and high-resolution depth maps of the same scenes.
We consider an approach to depth map enhancement based on learning from unpaired data.
arXiv Detail & Related papers (2021-05-25T16:19:16Z) - Monocular Depth Parameterizing Networks [15.791732557395552]
We propose a network structure that provides a parameterization of a set of depth maps with feasible shapes.
This allows us to search the shapes for a photo consistent solution with respect to other images.
Our experimental evaluation shows that our method generates more accurate depth maps and generalizes better than competing state-of-the-art approaches.
arXiv Detail & Related papers (2020-12-21T13:02:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.