UMLE: Unsupervised Multi-discriminator Network for Low Light Enhancement
- URL: http://arxiv.org/abs/2012.13177v2
- Date: Fri, 25 Dec 2020 02:36:11 GMT
- Title: UMLE: Unsupervised Multi-discriminator Network for Low Light Enhancement
- Authors: Yangyang Qu, Kai Chen, Chao Liu, Yongsheng Ou
- Abstract summary: Low-light scenarios will have serious implications for vision-based applications.
We propose a real-time unsupervised generative adversarial network (GAN) containing multiple discriminators.
Experiments indicate that our method is superior to the state-of-the-art methods in qualitative and quantitative evaluations.
- Score: 8.887169648516844
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Low-light image enhancement, such as recovering color and texture details
from low-light images, is a complex and vital task. For automated driving,
low-light scenarios will have serious implications for vision-based
applications. To address this problem, we propose a real-time unsupervised
generative adversarial network (GAN) containing multiple discriminators, i.e. a
multi-scale discriminator, a texture discriminator, and a color discriminator.
These distinct discriminators allow the evaluation of images from different
perspectives. Further, considering that different channel features contain
different information and the illumination is uneven in the image, we propose a
feature fusion attention module. This module combines channel attention with
pixel attention mechanisms to extract image features. Additionally, to reduce
training time, we adopt a shared encoder for the generator and the
discriminator. This makes the structure of the model more compact and the
training more stable. Experiments indicate that our method is superior to the
state-of-the-art methods in qualitative and quantitative evaluations, and
significant improvements are achieved for both autopilot positioning and
detection results.
Related papers
- A Non-Uniform Low-Light Image Enhancement Method with Multi-Scale
Attention Transformer and Luminance Consistency Loss [11.585269110131659]
Low-light image enhancement aims to improve the perception of images collected in dim environments.
Existing methods cannot adaptively extract the differentiated luminance information, which will easily cause over-exposure and under-exposure.
We propose a multi-scale attention Transformer named MSATr, which sufficiently extracts local and global features for light balance to improve the visual quality.
arXiv Detail & Related papers (2023-12-27T10:07:11Z) - LCPR: A Multi-Scale Attention-Based LiDAR-Camera Fusion Network for
Place Recognition [11.206532393178385]
We present a novel neural network named LCPR for robust multimodal place recognition.
Our method can effectively utilize multi-view camera and LiDAR data to improve the place recognition performance.
arXiv Detail & Related papers (2023-11-06T15:39:48Z) - SCANet: Self-Paced Semi-Curricular Attention Network for Non-Homogeneous
Image Dehazing [56.900964135228435]
Existing homogeneous dehazing methods struggle to handle the non-uniform distribution of haze in a robust manner.
We propose a novel self-paced semi-curricular attention network, called SCANet, for non-homogeneous image dehazing.
Our approach consists of an attention generator network and a scene reconstruction network.
arXiv Detail & Related papers (2023-04-17T17:05:29Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - A-ESRGAN: Training Real-World Blind Super-Resolution with Attention
U-Net Discriminators [0.0]
Blind image super-resolution(SR) is a long-standing task in CV that aims to restore low-resolution images suffering from unknown and complex distortions.
We present A-ESRGAN, a GAN model for blind SR tasks featuring an attention U-Net based, multi-scale discriminator.
arXiv Detail & Related papers (2021-12-19T02:50:23Z) - Learning Discriminative Representations for Multi-Label Image
Recognition [13.13795708478267]
We propose a unified deep network to learn discriminative features for the multi-label task.
By regularizing the whole network with the proposed loss, the performance of applying the wellknown ResNet-101 is improved significantly.
arXiv Detail & Related papers (2021-07-23T12:10:46Z) - Multi-view Contrastive Coding of Remote Sensing Images at Pixel-level [5.64497799927668]
A pixel-wise contrastive approach based on an unlabeled multi-view setting is proposed to overcome this limitation.
A pseudo-Siamese ResUnet is trained to learn a representation that aims to align features from the shifted positive pairs.
Results demonstrate both improvements in efficiency and accuracy over the state-of-the-art multi-view contrastive methods.
arXiv Detail & Related papers (2021-05-18T13:28:46Z) - Dual Contrastive Loss and Attention for GANs [82.713118646294]
We propose a novel dual contrastive loss and show that, with this loss, discriminator learns more generalized and distinguishable representations to incentivize generation.
We find attention to be still an important module for successful image generation even though it was not used in the recent state-of-the-art models.
By combining the strengths of these remedies, we improve the compelling state-of-the-art Fr'echet Inception Distance (FID) by at least 17.5% on several benchmark datasets.
arXiv Detail & Related papers (2021-03-31T01:10:26Z) - Attention Model Enhanced Network for Classification of Breast Cancer
Image [54.83246945407568]
AMEN is formulated in a multi-branch fashion with pixel-wised attention model and classification submodular.
To focus more on subtle detail information, the sample image is enhanced by the pixel-wised attention map generated from former branch.
Experiments conducted on three benchmark datasets demonstrate the superiority of the proposed method under various scenarios.
arXiv Detail & Related papers (2020-10-07T08:44:21Z) - A U-Net Based Discriminator for Generative Adversarial Networks [86.67102929147592]
We propose an alternative U-Net based discriminator architecture for generative adversarial networks (GANs)
The proposed architecture allows to provide detailed per-pixel feedback to the generator while maintaining the global coherence of synthesized images.
The novel discriminator improves over the state of the art in terms of the standard distribution and image quality metrics.
arXiv Detail & Related papers (2020-02-28T11:16:54Z) - Interpreting Galaxy Deblender GAN from the Discriminator's Perspective [50.12901802952574]
This research focuses on behaviors of one of the network's major components, the Discriminator, which plays a vital role but is often overlooked.
We demonstrate that our method clearly reveals attention areas of the Discriminator when differentiating generated galaxy images from ground truth images.
arXiv Detail & Related papers (2020-01-17T04:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.