L-WaveBlock: A Novel Feature Extractor Leveraging Wavelets for
Generative Adversarial Networks
- URL: http://arxiv.org/abs/2311.05548v1
- Date: Thu, 9 Nov 2023 17:47:32 GMT
- Title: L-WaveBlock: A Novel Feature Extractor Leveraging Wavelets for
Generative Adversarial Networks
- Authors: Mirat Shah, Vansh Jain, Anmol Chokshi, Guruprasad Parasnis, Pramod
Bide
- Abstract summary: This paper introduces L-WaveBlock, a novel and robust feature extractor that leverages the capabilities of the Discrete Wavelet Transform (DWT) with deep learning methodologies.
L-WaveBlock is catered to quicken the convergence of GAN generators while simultaneously enhancing their performance.
The paper demonstrates the remarkable utility of L-WaveBlock across three datasets, a road satellite imagery dataset, the CelebA dataset and the GoPro dataset.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative Adversarial Networks (GANs) have risen to prominence in the field
of deep learning, facilitating the generation of realistic data from random
noise. The effectiveness of GANs often depends on the quality of feature
extraction, a critical aspect of their architecture. This paper introduces
L-WaveBlock, a novel and robust feature extractor that leverages the
capabilities of the Discrete Wavelet Transform (DWT) with deep learning
methodologies. L-WaveBlock is catered to quicken the convergence of GAN
generators while simultaneously enhancing their performance. The paper
demonstrates the remarkable utility of L-WaveBlock across three datasets, a
road satellite imagery dataset, the CelebA dataset and the GoPro dataset,
showcasing its ability to ease feature extraction and make it more efficient.
By utilizing DWT, L-WaveBlock efficiently captures the intricate details of
both structural and textural details, and further partitions feature maps into
orthogonal subbands across multiple scales while preserving essential
information at the same time. Not only does it lead to faster convergence, but
also gives competent results on every dataset by employing the L-WaveBlock. The
proposed method achieves an Inception Score of 3.6959 and a Structural
Similarity Index of 0.4261 on the maps dataset, a Peak Signal-to-Noise Ratio of
29.05 and a Structural Similarity Index of 0.874 on the CelebA dataset. The
proposed method performs competently to the state-of-the-art for the image
denoising dataset, albeit not better, but still leads to faster convergence
than conventional methods. With this, L-WaveBlock emerges as a robust and
efficient tool for enhancing GAN-based image generation, demonstrating superior
convergence speed and competitive performance across multiple datasets for
image resolution, image generation and image denoising.
Related papers
- WTDUN: Wavelet Tree-Structured Sampling and Deep Unfolding Network for Image Compressed Sensing [51.94493817128006]
We propose a novel wavelet-domain deep unfolding framework named WTDUN, which operates directly on the multi-scale wavelet subbands.
Our method utilizes the intrinsic sparsity and multi-scale structure of wavelet coefficients to achieve a tree-structured sampling and reconstruction.
arXiv Detail & Related papers (2024-11-25T12:31:03Z) - High-Precision Dichotomous Image Segmentation via Probing Diffusion Capacity [69.32473738284374]
We propose DiffDIS, a diffusion-driven segmentation model that taps into the potential of the pre-trained U-Net within diffusion models.
By leveraging the robust generalization capabilities and rich, versatile image representation prior to the SD models, we significantly reduce the inference time while preserving high-fidelity, detailed generation.
Experiments on the DIS5K dataset demonstrate the superiority of DiffDIS, achieving state-of-the-art results through a streamlined inference process.
arXiv Detail & Related papers (2024-10-14T02:49:23Z) - Wavelet-based Bi-dimensional Aggregation Network for SAR Image Change Detection [53.842568573251214]
Experimental results on three SAR datasets demonstrate that our WBANet significantly outperforms contemporary state-of-the-art methods.
Our WBANet achieves 98.33%, 96.65%, and 96.62% of percentage of correct classification (PCC) on the respective datasets.
arXiv Detail & Related papers (2024-07-18T04:36:10Z) - LeRF: Learning Resampling Function for Adaptive and Efficient Image Interpolation [64.34935748707673]
Recent deep neural networks (DNNs) have made impressive progress in performance by introducing learned data priors.
We propose a novel method of Learning Resampling (termed LeRF) which takes advantage of both the structural priors learned by DNNs and the locally continuous assumption.
LeRF assigns spatially varying resampling functions to input image pixels and learns to predict the shapes of these resampling functions with a neural network.
arXiv Detail & Related papers (2024-07-13T16:09:45Z) - Learning Image Deraining Transformer Network with Dynamic Dual
Self-Attention [46.11162082219387]
This paper proposes an effective image deraining Transformer with dynamic dual self-attention (DDSA)
Specifically, we only select the most useful similarity values based on top-k approximate calculation to achieve sparse attention.
In addition, we also develop a novel spatial-enhanced feed-forward network (SEFN) to further obtain a more accurate representation for achieving high-quality derained results.
arXiv Detail & Related papers (2023-08-15T13:59:47Z) - Toward Data-Driven STAP Radar [23.333816677794115]
We characterize our data-driven approach to space-time adaptive processing (STAP) radar.
We generate a rich example dataset of received radar signals by randomly placing targets of variable strengths in a predetermined region.
For each data sample within this region, we generate heatmap tensors in range, azimuth, and elevation of the output power of a beamformer.
In an airborne scenario, the moving radar creates a sequence of these time-indexed image stacks, resembling a video.
arXiv Detail & Related papers (2022-01-26T02:28:13Z) - ChiNet: Deep Recurrent Convolutional Learning for Multimodal Spacecraft
Pose Estimation [3.964047152162558]
This paper presents an innovative deep learning pipeline which estimates the relative pose of a spacecraft by incorporating the temporal information from a rendezvous sequence.
It leverages the performance of long short-term memory (LSTM) units in modelling sequences of data for the processing of features extracted by a convolutional neural network (CNN) backbone.
Three distinct training strategies, which follow a coarse-to-fine funnelled approach, are combined to facilitate feature learning and improve end-to-end pose estimation by regression.
arXiv Detail & Related papers (2021-08-23T16:48:58Z) - Real-Time High-Performance Semantic Image Segmentation of Urban Street
Scenes [98.65457534223539]
We propose a real-time high-performance DCNN-based method for robust semantic segmentation of urban street scenes.
The proposed method achieves the accuracy of 73.6% and 68.0% mean Intersection over Union (mIoU) with the inference speed of 51.0 fps and 39.3 fps.
arXiv Detail & Related papers (2020-03-11T08:45:53Z) - Learning-Based Human Segmentation and Velocity Estimation Using
Automatic Labeled LiDAR Sequence for Training [15.19884183320726]
We propose an automatic labeled sequential data generation pipeline for human recognition with point clouds.
Our approach uses a precise human model and reproduces a precise motion to generate realistic artificial data.
arXiv Detail & Related papers (2020-03-11T03:14:52Z) - Spatial-Spectral Residual Network for Hyperspectral Image
Super-Resolution [82.1739023587565]
We propose a novel spectral-spatial residual network for hyperspectral image super-resolution (SSRNet)
Our method can effectively explore spatial-spectral information by using 3D convolution instead of 2D convolution, which enables the network to better extract potential information.
In each unit, we employ spatial and temporal separable 3D convolution to extract spatial and spectral information, which not only reduces unaffordable memory usage and high computational cost, but also makes the network easier to train.
arXiv Detail & Related papers (2020-01-14T03:34:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.