VBLC: Visibility Boosting and Logit-Constraint Learning for Domain
Adaptive Semantic Segmentation under Adverse Conditions
- URL: http://arxiv.org/abs/2211.12256v1
- Date: Tue, 22 Nov 2022 13:16:41 GMT
- Title: VBLC: Visibility Boosting and Logit-Constraint Learning for Domain
Adaptive Semantic Segmentation under Adverse Conditions
- Authors: Mingjia Li, Binhui Xie, Shuang Li, Chi Harold Liu, Xinjing Cheng
- Abstract summary: Generalizing models trained on normal visual conditions to target domains under adverse conditions is demanding in the practical systems.
We propose Visibility Boosting and Logit-Constraint learning (VBLC), tailored for superior normal-to-adverse adaptation.
VBLC explores the potential of getting rid of reference images and resolving the mixture of adverse conditions simultaneously.
- Score: 31.992504022101215
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generalizing models trained on normal visual conditions to target domains
under adverse conditions is demanding in the practical systems. One prevalent
solution is to bridge the domain gap between clear- and adverse-condition
images to make satisfactory prediction on the target. However, previous methods
often reckon on additional reference images of the same scenes taken from
normal conditions, which are quite tough to collect in reality. Furthermore,
most of them mainly focus on individual adverse condition such as nighttime or
foggy, weakening the model versatility when encountering other adverse
weathers. To overcome the above limitations, we propose a novel framework,
Visibility Boosting and Logit-Constraint learning (VBLC), tailored for superior
normal-to-adverse adaptation. VBLC explores the potential of getting rid of
reference images and resolving the mixture of adverse conditions
simultaneously. In detail, we first propose the visibility boost module to
dynamically improve target images via certain priors in the image level. Then,
we figure out the overconfident drawback in the conventional cross-entropy loss
for self-training method and devise the logit-constraint learning, which
enforces a constraint on logit outputs during training to mitigate this pain
point. To the best of our knowledge, this is a new perspective for tackling
such a challenging task. Extensive experiments on two normal-to-adverse domain
adaptation benchmarks, i.e., Cityscapes -> ACDC and Cityscapes ->
FoggyCityscapes + RainCityscapes, verify the effectiveness of VBLC, where it
establishes the new state of the art. Code is available at
https://github.com/BIT-DA/VBLC.
Related papers
- Locality Alignment Improves Vision-Language Models [55.275235524659905]
Vision language models (VLMs) have seen growing adoption in recent years, but many still struggle with basic spatial reasoning errors.
We propose a new efficient post-training stage for ViTs called locality alignment.
We show that locality-aligned backbones improve performance across a range of benchmarks.
arXiv Detail & Related papers (2024-10-14T21:01:01Z) - Test-Time Training for Semantic Segmentation with Output Contrastive
Loss [12.535720010867538]
Deep learning-based segmentation models have achieved impressive performance on public benchmarks, but generalizing well to unseen environments remains a major challenge.
This paper introduces Contrastive Loss (OCL), known for its capability to learn robust and generalized representations, to stabilize the adaptation process.
Our method excels even when applied to models initially pre-trained using domain adaptation methods on test domain data, showcasing its resilience and adaptability.
arXiv Detail & Related papers (2023-11-14T03:13:47Z) - Robust Monocular Depth Estimation under Challenging Conditions [81.57697198031975]
State-of-the-art monocular depth estimation approaches are highly unreliable under challenging illumination and weather conditions.
We tackle these safety-critical issues with md4all: a simple and effective solution that works reliably under both adverse and ideal conditions.
arXiv Detail & Related papers (2023-08-18T17:59:01Z) - Similarity Min-Max: Zero-Shot Day-Night Domain Adaptation [52.923298434948606]
Low-light conditions not only hamper human visual experience but also degrade the model's performance on downstream vision tasks.
This paper challenges a more complicated scenario with border applicability, i.e., zero-shot day-night domain adaptation.
We propose a similarity min-max paradigm that considers them under a unified framework.
arXiv Detail & Related papers (2023-07-17T18:50:15Z) - Condition-Invariant Semantic Segmentation [77.10045325743644]
We implement Condition-Invariant Semantic (CISS) on the current state-of-the-art domain adaptation architecture.
Our method achieves the second-best performance on the normal-to-adverse Cityscapes$to$ACDC benchmark.
CISS is shown to generalize well to domains unseen during training, such as BDD100K-night and ACDC-night.
arXiv Detail & Related papers (2023-05-27T03:05:07Z) - Refign: Align and Refine for Adaptation of Semantic Segmentation to
Adverse Conditions [78.71745819446176]
Refign is a generic extension to self-training-based UDA methods which leverages cross-domain correspondences.
Refign consists of two steps: (1) aligning the normal-condition image to the corresponding adverse-condition image using an uncertainty-aware dense matching network, and (2) refining the adverse prediction with the normal prediction using an adaptive label correction mechanism.
The approach introduces no extra training parameters, minimal computational overhead -- during training only -- and can be used as a drop-in extension to improve any given self-training-based UDA method.
arXiv Detail & Related papers (2022-07-14T11:30:38Z) - Toward Fast, Flexible, and Robust Low-Light Image Enhancement [87.27326390675155]
We develop a new Self-Calibrated Illumination (SCI) learning framework for fast, flexible, and robust brightening images in real-world low-light scenarios.
Considering the computational burden of the cascaded pattern, we construct the self-calibrated module which realizes the convergence between results of each stage.
We make comprehensive explorations to SCI's inherent properties including operation-insensitive adaptability and model-irrelevant generality.
arXiv Detail & Related papers (2022-04-21T14:40:32Z) - PixMatch: Unsupervised Domain Adaptation via Pixelwise Consistency
Training [4.336877104987131]
Unsupervised domain adaptation is a promising technique for semantic segmentation.
We present a novel framework for unsupervised domain adaptation based on the notion of target-domain consistency training.
Our approach is simpler, easier to implement, and more memory-efficient during training.
arXiv Detail & Related papers (2021-05-17T19:36:28Z) - Domain-invariant Similarity Activation Map Contrastive Learning for
Retrieval-based Long-term Visual Localization [30.203072945001136]
In this work, a general architecture is first formulated probabilistically to extract domain invariant feature through multi-domain image translation.
And then a novel gradient-weighted similarity activation mapping loss (Grad-SAM) is incorporated for finer localization with high accuracy.
Extensive experiments have been conducted to validate the effectiveness of the proposed approach on the CMUSeasons dataset.
Our performance is on par with or even outperforms the state-of-the-art image-based localization baselines in medium or high precision.
arXiv Detail & Related papers (2020-09-16T14:43:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.