HazeCLIP: Towards Language Guided Real-World Image Dehazing
- URL: http://arxiv.org/abs/2407.13719v1
- Date: Thu, 18 Jul 2024 17:18:25 GMT
- Title: HazeCLIP: Towards Language Guided Real-World Image Dehazing
- Authors: Ruiyi Wang, Wenhao Li, Xiaohong Liu, Chunyi Li, Zicheng Zhang, Xiongkuo Min, Guangtao Zhai,
- Abstract summary: Existing methods have achieved remarkable performance in single image dehazing, particularly on synthetic datasets.
This paper introduces HazeCLIP, a language-guided adaptation framework designed to enhance the real-world performance of pre-trained dehazing networks.
- Score: 62.4454483961341
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing methods have achieved remarkable performance in single image dehazing, particularly on synthetic datasets. However, they often struggle with real-world hazy images due to domain shift, limiting their practical applicability. This paper introduces HazeCLIP, a language-guided adaptation framework designed to enhance the real-world performance of pre-trained dehazing networks. Inspired by the Contrastive Language-Image Pre-training (CLIP) model's ability to distinguish between hazy and clean images, we utilize it to evaluate dehazing results. Combined with a region-specific dehazing technique and tailored prompt sets, CLIP model accurately identifies hazy areas, providing a high-quality, human-like prior that guides the fine-tuning process of pre-trained networks. Extensive experiments demonstrate that HazeCLIP achieves the state-of-the-art performance in real-word image dehazing, evaluated through both visual quality and no-reference quality assessments. The code is available: https://github.com/Troivyn/HazeCLIP .
Related papers
- Contrasting Deepfakes Diffusion via Contrastive Learning and Global-Local Similarities [88.398085358514]
Contrastive Deepfake Embeddings (CoDE) is a novel embedding space specifically designed for deepfake detection.
CoDE is trained via contrastive learning by additionally enforcing global-local similarities.
arXiv Detail & Related papers (2024-07-29T18:00:10Z) - Prompt-based test-time real image dehazing: a novel pipeline [9.229160145438469]
We present Prompt-based Test-Time Dehazing (PTTD) to help generate visually pleasing results of real-captured hazy images.
We experimentally observe that given a dehazing model trained on synthetic data, fine-tuning the statistics (ie, mean and standard deviation) of encoding features is able to narrow the domain gap.
PTTD is model-agnostic and can be equipped with various state-of-the-art dehazing models trained on synthetic hazy-clean pairs to tackle the real image dehazing task.
arXiv Detail & Related papers (2023-09-29T16:50:38Z) - Non-Contrastive Learning Meets Language-Image Pre-Training [145.6671909437841]
We study the validity of non-contrastive language-image pre-training (nCLIP)
We introduce xCLIP, a multi-tasking framework combining CLIP and nCLIP, and show that nCLIP aids CLIP in enhancing feature semantics.
arXiv Detail & Related papers (2022-10-17T17:57:46Z) - UCL-Dehaze: Towards Real-world Image Dehazing via Unsupervised
Contrastive Learning [57.40713083410888]
This paper explores contrastive learning with an adversarial training effort to leverage unpaired real-world hazy and clean images.
We propose an effective unsupervised contrastive learning paradigm for image dehazing, dubbed UCL-Dehaze.
We conduct comprehensive experiments to evaluate our UCL-Dehaze and demonstrate its superiority over the state-of-the-arts.
arXiv Detail & Related papers (2022-05-04T03:25:13Z) - Toward Fast, Flexible, and Robust Low-Light Image Enhancement [87.27326390675155]
We develop a new Self-Calibrated Illumination (SCI) learning framework for fast, flexible, and robust brightening images in real-world low-light scenarios.
Considering the computational burden of the cascaded pattern, we construct the self-calibrated module which realizes the convergence between results of each stage.
We make comprehensive explorations to SCI's inherent properties including operation-insensitive adaptability and model-irrelevant generality.
arXiv Detail & Related papers (2022-04-21T14:40:32Z) - Mutual Learning for Domain Adaptation: Self-distillation Image Dehazing
Network with Sample-cycle [7.452382358080454]
We propose a mutual learning dehazing framework for domain adaption.
Specifically, we first devise two siamese networks: a teacher network in the synthetic domain and a student network in the real domain.
We show that the framework outperforms state-of-the-art dehazing techniques in terms of subjective and objective evaluation.
arXiv Detail & Related papers (2022-03-17T16:32:14Z) - DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting [91.56988987393483]
We present a new framework for dense prediction by implicitly and explicitly leveraging the pre-trained knowledge from CLIP.
Specifically, we convert the original image-text matching problem in CLIP to a pixel-text matching problem and use the pixel-text score maps to guide the learning of dense prediction models.
Our method is model-agnostic, which can be applied to arbitrary dense prediction systems and various pre-trained visual backbones.
arXiv Detail & Related papers (2021-12-02T18:59:32Z) - FD-GAN: Generative Adversarial Networks with Fusion-discriminator for
Single Image Dehazing [48.65974971543703]
We propose a fully end-to-end Generative Adversarial Networks with Fusion-discriminator (FD-GAN) for image dehazing.
Our model can generator more natural and realistic dehazed images with less color distortion and fewer artifacts.
Experiments have shown that our method reaches state-of-the-art performance on both public synthetic datasets and real-world images.
arXiv Detail & Related papers (2020-01-20T04:36:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.