Pixel-wise Deep Image Stitching
- URL: http://arxiv.org/abs/2112.06171v1
- Date: Sun, 12 Dec 2021 07:28:48 GMT
- Title: Pixel-wise Deep Image Stitching
- Authors: Hyeokjun Kweon, Hyeonseong Kim, Yoonsu Kang, Youngho Yoon, Wooseong
Jeong and Kuk-Jin Yoon
- Abstract summary: Image stitching aims at stitching the images taken from different viewpoints into an image with a wider field of view.
Existing methods warp the target image to the reference image using the estimated warp function.
We propose a novel deep image stitching framework exploiting the pixel-wise warp field to handle the large-parallax problem.
- Score: 21.824319551526294
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image stitching aims at stitching the images taken from different viewpoints
into an image with a wider field of view. Existing methods warp the target
image to the reference image using the estimated warp function, and a
homography is one of the most commonly used warping functions. However, when
images have large parallax due to non-planar scenes and translational motion of
a camera, the homography cannot fully describe the mapping between two images.
Existing approaches based on global or local homography estimation are not free
from this problem and suffer from undesired artifacts due to parallax. In this
paper, instead of relying on the homography-based warp, we propose a novel deep
image stitching framework exploiting the pixel-wise warp field to handle the
large-parallax problem. The proposed deep image stitching framework consists of
two modules: Pixel-wise Warping Module (PWM) and Stitched Image Generating
Module (SIGMo). PWM employs an optical flow estimation model to obtain
pixel-wise warp of the whole image, and relocates the pixels of the target
image with the obtained warp field. SIGMo blends the warped target image and
the reference image while eliminating unwanted artifacts such as misalignments,
seams, and holes that harm the plausibility of the stitched result. For
training and evaluating the proposed framework, we build a large-scale dataset
that includes image pairs with corresponding pixel-wise ground truth warp and
sample stitched result images. We show that the results of the proposed
framework are qualitatively superior to those of the conventional methods,
especially when the images have large parallax. The code and the proposed
dataset will be publicly available soon.
Related papers
- Single-image camera calibration with model-free distortion correction [0.0]
This paper proposes a method for estimating the complete set of calibration parameters from a single image of a planar speckle pattern covering the entire sensor.
The correspondence between image points and physical points on the calibration target is obtained using Digital Image Correlation.
At the end of the procedure, a dense and uniform model-free distortion map is obtained over the entire image.
arXiv Detail & Related papers (2024-03-02T16:51:35Z) - Image Deblurring using GAN [0.0]
This project focuses on the application of Generative Adversarial Network (GAN) in image deblurring.
The project defines a GAN model inflow and trains it with GoPRO dataset.
The network can obtain sharper pixels in image, achieving an average of 29.3 Peak Signal-to-Noise Ratio (PSNR) and 0.72 Structural Similarity Assessment (SSIM)
arXiv Detail & Related papers (2023-12-15T02:43:30Z) - Pixel-Inconsistency Modeling for Image Manipulation Localization [59.968362815126326]
Digital image forensics plays a crucial role in image authentication and manipulation localization.
This paper presents a generalized and robust manipulation localization model through the analysis of pixel inconsistency artifacts.
Experiments show that our method successfully extracts inherent pixel-inconsistency forgery fingerprints.
arXiv Detail & Related papers (2023-09-30T02:54:51Z) - Parallax-Tolerant Unsupervised Deep Image Stitching [57.76737888499145]
We propose UDIS++, a parallax-tolerant unsupervised deep image stitching technique.
First, we propose a robust and flexible warp to model the image registration from global homography to local thin-plate spline motion.
To further eliminate the parallax artifacts, we propose to composite the stitched image seamlessly by unsupervised learning for seam-driven composition masks.
arXiv Detail & Related papers (2023-02-16T10:40:55Z) - CoupAlign: Coupling Word-Pixel with Sentence-Mask Alignments for
Referring Image Segmentation [104.5033800500497]
Referring image segmentation aims at localizing all pixels of the visual objects described by a natural language sentence.
Previous works learn to straightforwardly align the sentence embedding and pixel-level embedding for highlighting the referred objects.
We propose CoupAlign, a simple yet effective multi-level visual-semantic alignment method.
arXiv Detail & Related papers (2022-12-04T08:53:42Z) - MC-Blur: A Comprehensive Benchmark for Image Deblurring [127.6301230023318]
In most real-world images, blur is caused by different factors, e.g., motion and defocus.
We construct a new large-scale multi-cause image deblurring dataset (called MC-Blur)
Based on the MC-Blur dataset, we conduct extensive benchmarking studies to compare SOTA methods in different scenarios.
arXiv Detail & Related papers (2021-12-01T02:10:42Z) - Single image deep defocus estimation and its applications [82.93345261434943]
We train a deep neural network to classify image patches into one of the 20 levels of blurriness.
The trained model is used to determine the patch blurriness which is then refined by applying an iterative weighted guided filter.
The result is a defocus map that carries the information of the degree of blurriness for each pixel.
arXiv Detail & Related papers (2021-07-30T06:18:16Z) - Image Inpainting with Edge-guided Learnable Bidirectional Attention Maps [85.67745220834718]
We present an edge-guided learnable bidirectional attention map (Edge-LBAM) for improving image inpainting of irregular holes.
Our Edge-LBAM method contains dual procedures,including structure-aware mask-updating guided by predict edges.
Extensive experiments show that our Edge-LBAM is effective in generating coherent image structures and preventing color discrepancy and blurriness.
arXiv Detail & Related papers (2021-04-25T07:25:16Z) - Deep Contrastive Patch-Based Subspace Learning for Camera Image Signal
Processing [5.678834480723395]
We present a specific patch-based, local subspace deep neural network that improves Camera ISP to be robust to heterogeneous artifacts.
We call our three-fold deep-trained model the Patch Subspace Learning Autoencoder (PSL-AE)
PSL-AE encodes patches extracted from noisy a nd clean image pairs, with different artifact types or distortion levels, by contrastive learning.
arXiv Detail & Related papers (2021-04-01T04:40:22Z) - Learning Edge-Preserved Image Stitching from Large-Baseline Deep
Homography [32.28310831466225]
We propose an image stitching learning framework, which consists of a large-baseline deep homography module and an edge-preserved deformation module.
Our method is superior to the existing learning method and shows competitive performance with state-of-the-art traditional methods.
arXiv Detail & Related papers (2020-12-11T08:43:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.