Learning to See Low-Light Images via Feature Domain Adaptation
- URL: http://arxiv.org/abs/2312.06723v3
- Date: Wed, 20 Dec 2023 02:28:18 GMT
- Title: Learning to See Low-Light Images via Feature Domain Adaptation
- Authors: Qirui Yang, Qihua Cheng, Huanjing Yue, Le Zhang, Yihao Liu, Jingyu
Yang
- Abstract summary: We propose a single-stage network empowered by Feature Domain Adaptation (FDA) to decouple the denoising and color mapping tasks in raw LLIE.
FDA can explore the global and local correlations with fewer line buffers.
Our method achieves state-of-the-art performance with fewer computing costs.
- Score: 17.033219611079165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Raw low light image enhancement (LLIE) has achieved much better performance
than the sRGB domain enhancement methods due to the merits of raw data.
However, the ambiguity between noisy to clean and raw to sRGB mappings may
mislead the single-stage enhancement networks. The two-stage networks avoid
ambiguity by decoupling the two mappings but usually have large computing
complexity. To solve this problem, we propose a single-stage network empowered
by Feature Domain Adaptation (FDA) to decouple the denoising and color mapping
tasks in raw LLIE. The denoising encoder is supervised by the clean raw image,
and then the denoised features are adapted for the color mapping task by an FDA
module. We propose a Lineformer to serve as the FDA, which can well explore the
global and local correlations with fewer line buffers (friendly to the
line-based imaging process). During inference, the raw supervision branch is
removed. In this way, our network combines the advantage of a two-stage
enhancement process with the efficiency of single-stage inference. Experiments
on four benchmark datasets demonstrate that our method achieves
state-of-the-art performance with fewer computing costs (60% FLOPs of the
two-stage method DNF). Our codes will be released after the acceptance of this
work.
Related papers
- Retinex-RAWMamba: Bridging Demosaicing and Denoising for Low-Light RAW Image Enhancement [71.13353154514418]
Low-light image enhancement, particularly in cross-domain tasks such as mapping from the raw domain to the sRGB domain, remains a significant challenge.
We present a novel Mamba scanning mechanism, called RAWMamba, to effectively handle raw images with different CFAs.
We also present a Retinex Decomposition Module (RDM) grounded in Retinex prior, which decouples illumination from reflectance to facilitate more effective denoising and automatic non-linear exposure correction.
arXiv Detail & Related papers (2024-09-11T06:12:03Z) - A Modular System for Enhanced Robustness of Multimedia Understanding
Networks via Deep Parametric Estimation [30.904034138920057]
In multimedia understanding tasks, corrupted samples pose a critical challenge, because when fed to machine learning models they lead to performance degradation.
We propose SyMPIE to enhance input data for robust downstream multimedia understanding with minimal computational cost.
Our key insight is that most input corruptions can be modeled through global operations on color channels of images or spatial filters with small kernels.
arXiv Detail & Related papers (2024-02-28T15:24:58Z) - You Only Need One Color Space: An Efficient Network for Low-light Image Enhancement [50.37253008333166]
Low-Light Image Enhancement (LLIE) task tends to restore the details and visual information from corrupted low-light images.
We propose a novel trainable color space, named Horizontal/Vertical-Intensity (HVI)
It not only decouples brightness and color from RGB channels to mitigate the instability during enhancement but also adapts to low-light images in different illumination ranges due to the trainable parameters.
arXiv Detail & Related papers (2024-02-08T16:47:43Z) - Deep Linear Array Pushbroom Image Restoration: A Degradation Pipeline
and Jitter-Aware Restoration Network [26.86292926584254]
Linear Array Pushbroom (LAP) imaging technology is widely used in the realm of remote sensing.
Traditional methods for restoring LAP images, such as algorithms estimating the point spread function (PSF), exhibit limited performance.
We propose a Jitter-Aware Restoration Network (JARNet) to remove the distortion and blur in two stages.
arXiv Detail & Related papers (2024-01-16T07:26:26Z) - DiffIR: Efficient Diffusion Model for Image Restoration [108.82579440308267]
Diffusion model (DM) has achieved SOTA performance by modeling the image synthesis process into a sequential application of a denoising network.
Traditional DMs running massive iterations on a large model to estimate whole images or feature maps is inefficient for image restoration.
We propose DiffIR, which consists of a compact IR prior extraction network (CPEN), dynamic IR transformer (DIRformer), and denoising network.
arXiv Detail & Related papers (2023-03-16T16:47:14Z) - Overexposure Mask Fusion: Generalizable Reverse ISP Multi-Step
Refinement [10.186389326668305]
This paper presents a state-of-the-art solution to the task of RAW reconstruction.
Instead of from RGB to bayer, the pipeline trains from RGB to demosaiced RAW allowing use of perceptual loss functions.
arXiv Detail & Related papers (2022-10-20T18:21:41Z) - Degrade is Upgrade: Learning Degradation for Low-light Image Enhancement [52.49231695707198]
We investigate the intrinsic degradation and relight the low-light image while refining the details and color in two steps.
Inspired by the color image formulation, we first estimate the degradation from low-light inputs to simulate the distortion of environment illumination color, and then refine the content to recover the loss of diffuse illumination color.
Our proposed method has surpassed the SOTA by 0.95dB in PSNR on LOL1000 dataset and 3.18% in mAP on ExDark dataset.
arXiv Detail & Related papers (2021-03-19T04:00:27Z) - SFANet: A Spectrum-aware Feature Augmentation Network for
Visible-Infrared Person Re-Identification [12.566284647658053]
We propose a novel spectrum-aware feature augementation network named SFANet for cross-modality matching problem.
Learning with grayscale-spectrum images, our model can apparently reduce modality discrepancy and detect inner structure relations.
In feature-level, we improve the conventional two-stream network through balancing the number of specific and sharable convolutional blocks.
arXiv Detail & Related papers (2021-02-24T08:57:32Z) - Self-Supervised Representation Learning for RGB-D Salient Object
Detection [93.17479956795862]
We use Self-Supervised Representation Learning to design two pretext tasks: the cross-modal auto-encoder and the depth-contour estimation.
Our pretext tasks require only a few and un RGB-D datasets to perform pre-training, which make the network capture rich semantic contexts.
For the inherent problem of cross-modal fusion in RGB-D SOD, we propose a multi-path fusion module.
arXiv Detail & Related papers (2021-01-29T09:16:06Z) - Bi-directional Cross-Modality Feature Propagation with
Separation-and-Aggregation Gate for RGB-D Semantic Segmentation [59.94819184452694]
Depth information has proven to be a useful cue in the semantic segmentation of RGBD images for providing a geometric counterpart to the RGB representation.
Most existing works simply assume that depth measurements are accurate and well-aligned with the RGB pixels and models the problem as a cross-modal feature fusion.
In this paper, we propose a unified and efficient Crossmodality Guided to not only effectively recalibrate RGB feature responses, but also to distill accurate depth information via multiple stages and aggregate the two recalibrated representations alternatively.
arXiv Detail & Related papers (2020-07-17T18:35:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.