Guided Co-Modulated GAN for 360{\deg} Field of View Extrapolation
- URL: http://arxiv.org/abs/2204.07286v1
- Date: Fri, 15 Apr 2022 01:48:35 GMT
- Title: Guided Co-Modulated GAN for 360{\deg} Field of View Extrapolation
- Authors: Mohammad Reza Karimi Dastjerdi, Yannick Hold-Geoffroy, Jonathan
Eisenmann, Siavash Khodadadeh, and Jean-Fran\c{c}ois Lalonde
- Abstract summary: We propose a method to extrapolate a 360deg field of view from a single image.
Our method obtains state-of-the-art results and outperforms previous methods on standard image quality metrics.
- Score: 15.850166450573756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a method to extrapolate a 360{\deg} field of view from a single
image that allows for user-controlled synthesis of the out-painted content. To
do so, we propose improvements to an existing GAN-based in-painting
architecture for out-painting panoramic image representation. Our method
obtains state-of-the-art results and outperforms previous methods on standard
image quality metrics. To allow controlled synthesis of out-painting, we
introduce a novel guided co-modulation framework, which drives the image
generation process with a common pretrained discriminative model. Doing so
maintains the high visual quality of generated panoramas while enabling
user-controlled semantic content in the extrapolated field of view. We
demonstrate the state-of-the-art results of our method on field of view
extrapolation both qualitatively and quantitatively, providing thorough
analysis of our novel editing capabilities. Finally, we demonstrate that our
approach benefits the photorealistic virtual insertion of highly glossy objects
in photographs.
Related papers
- PrefPaint: Aligning Image Inpainting Diffusion Model with Human Preference [62.72779589895124]
We make the first attempt to align diffusion models for image inpainting with human aesthetic standards via a reinforcement learning framework.
We train a reward model with a dataset we construct, consisting of nearly 51,000 images annotated with human preferences.
Experiments on inpainting comparison and downstream tasks, such as image extension and 3D reconstruction, demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-10-29T11:49:39Z) - 360U-Former: HDR Illumination Estimation with Panoramic Adapted Vision Transformers [21.393389135740712]
We train 360U-Former as a GAN to generate HDRI from a limited field of view low dynamic range image (LDRI)
We evaluate our method using current illumination estimation protocols and datasets.
arXiv Detail & Related papers (2024-10-17T14:03:53Z) - Layered Rendering Diffusion Model for Zero-Shot Guided Image Synthesis [60.260724486834164]
This paper introduces innovative solutions to enhance spatial controllability in diffusion models reliant on text queries.
We present two key innovations: Vision Guidance and the Layered Rendering Diffusion framework.
We apply our method to three practical applications: bounding box-to-image, semantic mask-to-image and image editing.
arXiv Detail & Related papers (2023-11-30T10:36:19Z) - HawkI: Homography & Mutual Information Guidance for 3D-free Single Image to Aerial View [67.8213192993001]
We present HawkI, for synthesizing aerial-view images from text and an exemplar image.
HawkI blends the visual features from the input image within a pretrained text-to-2Dimage stable diffusion model.
At inference, HawkI employs a unique mutual information guidance formulation to steer the generated image towards faithfully replicating the semantic details of the input-image.
arXiv Detail & Related papers (2023-11-27T01:41:25Z) - Integrating View Conditions for Image Synthesis [14.738884513493227]
This paper introduces a pioneering framework that integrates viewpoint information to enhance the control of image editing tasks.
We distill three essential criteria -- consistency, controllability, and harmony -- that should be met for an image editing method.
arXiv Detail & Related papers (2023-10-24T16:55:07Z) - Reference-based Image Composition with Sketch via Structure-aware
Diffusion Model [38.1193912666578]
We introduce a multi-input-conditioned image composition model that incorporates a sketch as a novel modal, alongside a reference image.
Thanks to the edge-level controllability using sketches, our method enables a user to edit or complete an image sub-part.
Our framework fine-tunes a pre-trained diffusion model to complete missing regions using the reference image while maintaining sketch guidance.
arXiv Detail & Related papers (2023-03-31T06:12:58Z) - HORIZON: High-Resolution Semantically Controlled Panorama Synthesis [105.55531244750019]
Panorama synthesis endeavors to craft captivating 360-degree visual landscapes, immersing users in the heart of virtual worlds.
Recent breakthroughs in visual synthesis have unlocked the potential for semantic control in 2D flat images, but a direct application of these methods to panorama synthesis yields distorted content.
We unveil an innovative framework for generating high-resolution panoramas, adeptly addressing the issues of spherical distortion and edge discontinuity through sophisticated spherical modeling.
arXiv Detail & Related papers (2022-10-10T09:43:26Z) - ViCE: Self-Supervised Visual Concept Embeddings as Contextual and Pixel
Appearance Invariant Semantic Representations [77.3590853897664]
This work presents a self-supervised method to learn dense semantically rich visual embeddings for images inspired by methods for learning word embeddings in NLP.
arXiv Detail & Related papers (2021-11-24T12:27:30Z) - IMAGINE: Image Synthesis by Image-Guided Model Inversion [79.4691654458141]
We introduce an inversion based method, denoted as IMAge-Guided model INvErsion (IMAGINE), to generate high-quality and diverse images.
We leverage the knowledge of image semantics from a pre-trained classifier to achieve plausible generations.
IMAGINE enables the synthesis procedure to simultaneously 1) enforce semantic specificity constraints during the synthesis, 2) produce realistic images without generator training, and 3) give users intuitive control over the generation process.
arXiv Detail & Related papers (2021-04-13T02:00:24Z) - Image Completion via Inference in Deep Generative Models [16.99337751292915]
We consider image completion from the perspective of amortized inference in an image generative model.
We demonstrate superior sample quality and diversity compared to prior art on the CIFAR-10 and FFHQ-256 datasets.
arXiv Detail & Related papers (2021-02-24T02:59:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.