Image Completion with Heterogeneously Filtered Spectral Hints
- URL: http://arxiv.org/abs/2211.03700v1
- Date: Mon, 7 Nov 2022 17:15:16 GMT
- Title: Image Completion with Heterogeneously Filtered Spectral Hints
- Authors: Xingqian Xu, Shant Navasardyan, Vahram Tadevosyan, Andranik Sargsyan,
Yadong Mu, Humphrey Shi
- Abstract summary: We propose a new StyleGAN-based image completion network, Spectral Hint GAN (SH-GAN), inside which a spectral processing module, Spectral Hint Unit, is introduced.
From our inclusive experiments, we demonstrate that our model can reach FID scores of 3.4134 and 7.0277 on the benchmark datasets FFHQ and Places2.
- Score: 29.26481807829418
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Image completion with large-scale free-form missing regions is one of the
most challenging tasks for the computer vision community. While researchers
pursue better solutions, drawbacks such as pattern unawareness, blurry
textures, and structure distortion remain noticeable, and thus leave space for
improvement. To overcome these challenges, we propose a new StyleGAN-based
image completion network, Spectral Hint GAN (SH-GAN), inside which a carefully
designed spectral processing module, Spectral Hint Unit, is introduced. We also
propose two novel 2D spectral processing strategies, Heterogeneous Filtering
and Gaussian Split that well-fit modern deep learning models and may further be
extended to other tasks. From our inclusive experiments, we demonstrate that
our model can reach FID scores of 3.4134 and 7.0277 on the benchmark datasets
FFHQ and Places2, and therefore outperforms prior works and reaches a new
state-of-the-art. We also prove the effectiveness of our design via ablation
studies, from which one may notice that the aforementioned challenges, i.e.
pattern unawareness, blurry textures, and structure distortion, can be
noticeably resolved. Our code will be open-sourced at:
https://github.com/SHI-Labs/SH-GAN.
Related papers
- Coarse-Fine Spectral-Aware Deformable Convolution For Hyperspectral Image Reconstruction [15.537910100051866]
We study the inverse problem of Coded Aperture Snapshot Spectral Imaging (CASSI)
We propose Coarse-Fine Spectral-Aware Deformable Convolution Network (CFSDCN)
Our CFSDCN significantly outperforms previous state-of-the-art (SOTA) methods on both simulated and real HSI datasets.
arXiv Detail & Related papers (2024-06-18T15:15:12Z) - GS-IR: 3D Gaussian Splatting for Inverse Rendering [71.14234327414086]
We propose GS-IR, a novel inverse rendering approach based on 3D Gaussian Splatting (GS)
We extend GS, a top-performance representation for novel view synthesis, to estimate scene geometry, surface material, and environment illumination from multi-view images captured under unknown lighting conditions.
The flexible and expressive GS representation allows us to achieve fast and compact geometry reconstruction, photorealistic novel view synthesis, and effective physically-based rendering.
arXiv Detail & Related papers (2023-11-26T02:35:09Z) - Robust GAN inversion [5.1359892878090845]
We propose an approach which works in native latent space $W$ and tunes the generator network to restore missing image details.
We demonstrate the effectiveness of our approach on two complex datasets: Flickr-Faces-HQ and LSUN Church.
arXiv Detail & Related papers (2023-08-31T07:47:11Z) - PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with
Dual-Discriminators [120.06891448820447]
How to obtain clear and visually pleasant images has become a common concern of people.
The task of underwater image enhancement (UIE) has also emerged as the times require.
In this paper, we propose a physical model-guided GAN model for UIE, referred to as PUGAN.
Our PUGAN outperforms state-of-the-art methods in both qualitative and quantitative metrics.
arXiv Detail & Related papers (2023-06-15T07:41:12Z) - DPFNet: A Dual-branch Dilated Network with Phase-aware Fourier
Convolution for Low-light Image Enhancement [1.2645663389012574]
Low-light image enhancement is a classical computer vision problem aiming to recover normal-exposure images from low-light images.
convolutional neural networks commonly used in this field are good at sampling low-frequency local structural features in the spatial domain.
We propose a novel module using the Fourier coefficients, which can recover high-quality texture details under the constraint of semantics in the frequency phase.
arXiv Detail & Related papers (2022-09-16T13:56:09Z) - RSINet: Inpainting Remotely Sensed Images Using Triple GAN Framework [13.613245876782367]
We propose a novel inpainting method that individually focuses on each aspect of an image such as edges, colour and texture.
Each individual GAN also incorporates the attention mechanism that explicitly extracts the spectral and spatial features.
We evaluate our model, alongwith previous state of the art models, on the two well known remote sensing datasets, Open Cities AI and Earth on Canvas.
arXiv Detail & Related papers (2022-02-12T05:19:37Z) - Aggregated Contextual Transformations for High-Resolution Image
Inpainting [57.241749273816374]
We propose an enhanced GAN-based model, named Aggregated COntextual-Transformation GAN (AOT-GAN) for high-resolution image inpainting.
To enhance context reasoning, we construct the generator of AOT-GAN by stacking multiple layers of a proposed AOT block.
For improving texture synthesis, we enhance the discriminator of AOT-GAN by training it with a tailored mask-prediction task.
arXiv Detail & Related papers (2021-04-03T15:50:17Z) - UltraSR: Spatial Encoding is a Missing Key for Implicit Image
Function-based Arbitrary-Scale Super-Resolution [74.82282301089994]
In this work, we propose UltraSR, a simple yet effective new network design based on implicit image functions.
We show that spatial encoding is indeed a missing key towards the next-stage high-accuracy implicit image function.
Our UltraSR sets new state-of-the-art performance on the DIV2K benchmark under all super-resolution scales.
arXiv Detail & Related papers (2021-03-23T17:36:42Z) - Spatial-Spectral Feedback Network for Super-Resolution of Hyperspectral
Imagery [11.76638109321532]
High-dimensional and complex spectral patterns in hyperspectral image make it difficult to explore spatial information and spectral information among bands simultaneously.
The number of available hyperspectral training samples is extremely small, which can easily lead to overfitting when training a deep neural network.
We propose a novel Spatial-Spectral Feedback Network (SSFN) to refine low-level representations among local spectral bands with high-level information from global spectral bands.
arXiv Detail & Related papers (2021-03-07T13:28:48Z) - Semi-Supervised StyleGAN for Disentanglement Learning [79.01988132442064]
Current disentanglement methods face several inherent limitations.
We design new architectures and loss functions based on StyleGAN for semi-supervised high-resolution disentanglement learning.
arXiv Detail & Related papers (2020-03-06T22:54:46Z) - Spatial-Spectral Residual Network for Hyperspectral Image
Super-Resolution [82.1739023587565]
We propose a novel spectral-spatial residual network for hyperspectral image super-resolution (SSRNet)
Our method can effectively explore spatial-spectral information by using 3D convolution instead of 2D convolution, which enables the network to better extract potential information.
In each unit, we employ spatial and temporal separable 3D convolution to extract spatial and spectral information, which not only reduces unaffordable memory usage and high computational cost, but also makes the network easier to train.
arXiv Detail & Related papers (2020-01-14T03:34:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.