Revisiting Lightweight Low-Light Image Enhancement: From a YUV Color Space Perspective
- URL: http://arxiv.org/abs/2601.17349v1
- Date: Sat, 24 Jan 2026 07:27:54 GMT
- Title: Revisiting Lightweight Low-Light Image Enhancement: From a YUV Color Space Perspective
- Authors: Hailong Yan, Shice Liu, Xiangtao Zhang, Lujian Yao, Fengxiang Yang, Jinwei Chen, Bo Li,
- Abstract summary: We propose a novel YUV-based paradigm that strategically restores channels using a Dual-Stream Global-Local Attention module for the Y channel, a Y-guided Local-Aware Frequency Attention module for the UV channels, and a Guided Interaction module for final feature fusion.<n>Our model establishes a new state-of-the-art on multiple benchmarks, delivering superior visual quality with a significantly lower parameter count.
- Score: 17.507319835166406
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the current era of mobile internet, Lightweight Low-Light Image Enhancement (L3IE) is critical for mobile devices, which faces a persistent trade-off between visual quality and model compactness. While recent methods employ disentangling strategies to simplify lightweight architectural design, such as Retinex theory and YUV color space transformations, their performance is fundamentally limited by overlooking channel-specific degradation patterns and cross-channel interactions. To address this gap, we perform a frequency-domain analysis that confirms the superiority of the YUV color space for L3IE. We identify a key insight: the Y channel primarily loses low-frequency content, while the UV channels are corrupted by high-frequency noise. Leveraging this finding, we propose a novel YUV-based paradigm that strategically restores channels using a Dual-Stream Global-Local Attention module for the Y channel, a Y-guided Local-Aware Frequency Attention module for the UV channels, and a Guided Interaction module for final feature fusion. Extensive experiments validate that our model establishes a new state-of-the-art on multiple benchmarks, delivering superior visual quality with a significantly lower parameter count.
Related papers
- One-Shot Refiner: Boosting Feed-forward Novel View Synthesis via One-Step Diffusion [57.824020826432815]
We present a novel framework for high-fidelity novel view synthesis (NVS) from sparse images.<n>We design a Dual-Domain Detail Perception Module, which enables handling high-resolution images without being limited by the ViT backbone.<n>We develop a feature-guided diffusion network, which can preserve high-frequency details during the restoration process.
arXiv Detail & Related papers (2026-01-20T17:11:55Z) - IrisNet: Infrared Image Status Awareness Meta Decoder for Infrared Small Targets Detection [92.56025546608699]
IrisNet is a novel meta-learned framework that adapts detection strategies to the input infrared image status.<n>Our approach establishes a dynamic mapping between infrared image features and entire decoder parameters.<n> Experiments on NUDT-SIRST, NUAA-SIRST, and IRSTD-1K datasets demonstrate the superiority of our IrisNet.
arXiv Detail & Related papers (2025-11-25T13:53:54Z) - Rethinking Diffusion Model-Based Video Super-Resolution: Leveraging Dense Guidance from Aligned Features [51.5076190312734]
Video Super-Resolution approaches suffer from error accumulation, spatial artifacts, and a trade-off between perceptual quality and fidelity.<n>We propose a novelly Guided diffusion model with Aligned Features for Video Super-Resolution (DGAF-VSR)<n>Experiments on synthetic and real-world datasets demonstrate that DGAF-VSR surpasses state-of-the-art methods in key aspects of VSR.
arXiv Detail & Related papers (2025-11-21T03:40:45Z) - UHDRes: Ultra-High-Definition Image Restoration via Dual-Domain Decoupled Spectral Modulation [0.07352098890194292]
Ultra-high-definition (UHD) images often suffer from severe degradations such as blur, haze, rain, or low-light conditions.<n>We propose UHDRes, a novel lightweight dual-domain decoupled spectral modulation framework for UHD image restoration.
arXiv Detail & Related papers (2025-11-07T06:28:30Z) - FRBNet: Revisiting Low-Light Vision through Frequency-Domain Radial Basis Network [7.386546521017689]
We revisit low-light image formation and extend the classical Lambertian model to better characterize low-light conditions.<n>We propose a novel and end-to-end trainable module named textbfFrequency-domain textbfRadial textbfBasis textbfNetwork.<n>As a plug-and-play module, FRBNet can be integrated into existing networks for low-light downstream tasks without modifying loss functions.
arXiv Detail & Related papers (2025-10-27T15:46:07Z) - FreSca: Scaling in Frequency Space Enhances Diffusion Models [55.75504192166779]
This paper explores frequency-based control within latent diffusion models.<n>We introduce FreSca, a novel framework that decomposes noise difference into low- and high-frequency components.<n>FreSca operates without any model retraining or architectural change, offering model- and task-agnostic control.
arXiv Detail & Related papers (2025-04-02T22:03:11Z) - Triple-domain Feature Learning with Frequency-aware Memory Enhancement for Moving Infrared Small Target Detection [12.641645684148136]
Infrared small target detection presents significant challenges due to target sizes and low contrast against backgrounds.
We propose a new Triple-domain Strategy (Tridos) with frequency-aware memory enhancement on-temporal domain for infrared small target detection.
Inspired by human visual system, our memory enhancement is designed to capture the spatial relations of infrared targets among video frames.
arXiv Detail & Related papers (2024-06-11T05:21:30Z) - Learning Spatial Adaptation and Temporal Coherence in Diffusion Models for Video Super-Resolution [151.1255837803585]
We propose a novel approach, pursuing Spatial Adaptation and Temporal Coherence (SATeCo) for video super-resolution.
SATeCo pivots on learning spatial-temporal guidance from low-resolution videos to calibrate both latent-space high-resolution video denoising and pixel-space video reconstruction.
Experiments conducted on the REDS4 and Vid4 datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-03-25T17:59:26Z) - CVT-xRF: Contrastive In-Voxel Transformer for 3D Consistent Radiance Fields from Sparse Inputs [65.80187860906115]
We propose a novel approach to improve NeRF's performance with sparse inputs.
We first adopt a voxel-based ray sampling strategy to ensure that the sampled rays intersect with a certain voxel in 3D space.
We then randomly sample additional points within the voxel and apply a Transformer to infer the properties of other points on each ray, which are then incorporated into the volume rendering.
arXiv Detail & Related papers (2024-03-25T15:56:17Z) - LYT-NET: Lightweight YUV Transformer-based Network for Low-light Image Enhancement [0.0]
LYT-Net is a novel lightweight transformer-based model for low-light image enhancement (LLIE)<n>In our method we adopt a dual-path approach, treating chrominance channels U and V and luminance channel Y as separate entities to help the model better handle illumination adjustment and corruption restoration.<n>Our comprehensive evaluation on established LLIE datasets demonstrates that, despite its low complexity, our model outperforms recent LLIE methods.
arXiv Detail & Related papers (2024-01-26T21:02:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.