Robust Multi-Modal Image Stitching for Improved Scene Understanding
- URL: http://arxiv.org/abs/2312.17010v1
- Date: Thu, 28 Dec 2023 13:24:48 GMT
- Title: Robust Multi-Modal Image Stitching for Improved Scene Understanding
- Authors: Aritra Dutta, Dr. G Suseela, Asmita Sood
- Abstract summary: We've devised a unique and comprehensive image-stitching pipeline that taps into OpenCV's stitching module.
Our approach integrates feature-based matching, transformation estimation, and blending techniques to bring about panoramic views that are of top-tier quality.
- Score: 2.0476854378186102
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-modal image stitching can be a difficult feat. That's why, in this
paper, we've devised a unique and comprehensive image-stitching pipeline that
taps into OpenCV's stitching module. Our approach integrates feature-based
matching, transformation estimation, and blending techniques to bring about
panoramic views that are of top-tier quality - irrespective of lighting, scale
or orientation differences between images. We've put our pipeline to the test
with a varied dataset and found that it's very effective in enhancing scene
understanding and finding real-world applications.
Related papers
- ConDL: Detector-Free Dense Image Matching [2.7582789611575897]
We introduce a deep-learning framework designed for estimating dense image correspondences.
Our fully convolutional model generates dense feature maps for images, where each pixel is associated with a descriptor that can be matched across multiple images.
arXiv Detail & Related papers (2024-08-05T18:34:15Z) - Generative Powers of Ten [60.6740997942711]
We present a method that uses a text-to-image model to generate consistent content across multiple image scales.
We achieve this through a joint multi-scale diffusion sampling approach.
Our method enables deeper levels of zoom than traditional super-resolution methods.
arXiv Detail & Related papers (2023-12-04T18:59:25Z) - Patch-Wise Self-Supervised Visual Representation Learning: A Fine-Grained Approach [4.9204263448542465]
This study introduces an innovative, fine-grained dimension by integrating patch-level discrimination into self-supervised visual representation learning.
We employ a distinctive photometric patch-level augmentation, where each patch is individually augmented, independent from other patches within the same view.
We present a simple yet effective patch-matching algorithm to find the corresponding patches across the augmented views.
arXiv Detail & Related papers (2023-10-28T09:35:30Z) - Multi-Spectral Image Stitching via Spatial Graph Reasoning [52.27796682972484]
We propose a spatial graph reasoning based multi-spectral image stitching method.
We embed multi-scale complementary features from the same view position into a set of nodes.
By introducing long-range coherence along spatial and channel dimensions, the complementarity of pixel relations and channel interdependencies aids in the reconstruction of aligned multi-view features.
arXiv Detail & Related papers (2023-07-31T15:04:52Z) - Parallax-Tolerant Unsupervised Deep Image Stitching [57.76737888499145]
We propose UDIS++, a parallax-tolerant unsupervised deep image stitching technique.
First, we propose a robust and flexible warp to model the image registration from global homography to local thin-plate spline motion.
To further eliminate the parallax artifacts, we propose to composite the stitched image seamlessly by unsupervised learning for seam-driven composition masks.
arXiv Detail & Related papers (2023-02-16T10:40:55Z) - Single Stage Virtual Try-on via Deformable Attention Flows [51.70606454288168]
Virtual try-on aims to generate a photo-realistic fitting result given an in-shop garment and a reference person image.
We develop a novel Deformable Attention Flow (DAFlow) which applies the deformable attention scheme to multi-flow estimation.
Our proposed method achieves state-of-the-art performance both qualitatively and quantitatively.
arXiv Detail & Related papers (2022-07-19T10:01:31Z) - Bridging the Visual Gap: Wide-Range Image Blending [16.464837892640812]
We introduce an effective deep-learning model to realize wide-range image blending.
We experimentally demonstrate that our proposed method is able to produce visually appealing results.
arXiv Detail & Related papers (2021-03-28T15:07:45Z) - Learning Edge-Preserved Image Stitching from Large-Baseline Deep
Homography [32.28310831466225]
We propose an image stitching learning framework, which consists of a large-baseline deep homography module and an edge-preserved deformation module.
Our method is superior to the existing learning method and shows competitive performance with state-of-the-art traditional methods.
arXiv Detail & Related papers (2020-12-11T08:43:30Z) - TSIT: A Simple and Versatile Framework for Image-to-Image Translation [103.92203013154403]
We introduce a simple and versatile framework for image-to-image translation.
We provide a carefully designed two-stream generative model with newly proposed feature transformations.
This allows multi-scale semantic structure information and style representation to be effectively captured and fused by the network.
A systematic study compares the proposed method with several state-of-the-art task-specific baselines, verifying its effectiveness in both perceptual quality and quantitative evaluations.
arXiv Detail & Related papers (2020-07-23T15:34:06Z) - RetrieveGAN: Image Synthesis via Differentiable Patch Retrieval [76.87013602243053]
We propose a differentiable retrieval module to synthesize images from scene description with retrieved patches as reference.
We conduct extensive quantitative and qualitative experiments to demonstrate that the proposed method can generate realistic and diverse images.
arXiv Detail & Related papers (2020-07-16T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.