Competitive Simplicity for Multi-Task Learning for Real-Time Foggy Scene
Understanding via Domain Adaptation
- URL: http://arxiv.org/abs/2012.05304v1
- Date: Wed, 9 Dec 2020 20:38:34 GMT
- Title: Competitive Simplicity for Multi-Task Learning for Real-Time Foggy Scene
Understanding via Domain Adaptation
- Authors: Naif Alshammari, Samet Akcay, and Toby P. Breckon
- Abstract summary: We propose a multi-task learning approach capable of performing in real-time semantic scene understanding and monocular depth estimation under foggy weather conditions.
Our model incorporates RGB colour, depth, and luminance images via distinct encoders with dense connectivity and features fusing.
- Score: 17.530091734327296
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automotive scene understanding under adverse weather conditions raises a
realistic and challenging problem attributable to poor outdoor scene visibility
(e.g. foggy weather). However, because most contemporary scene understanding
approaches are applied under ideal-weather conditions, such approaches may not
provide genuinely optimal performance when compared to established a priori
insights on extreme-weather understanding. In this paper, we propose a complex
but competitive multi-task learning approach capable of performing in real-time
semantic scene understanding and monocular depth estimation under foggy weather
conditions by leveraging both recent advances in adversarial training and
domain adaptation. As an end-to-end pipeline, our model provides a novel
solution to surpass degraded visibility in foggy weather conditions by
transferring scenes from foggy to normal using a GAN-based model. For optimal
performance in semantic segmentation, our model generates depth to be used as
complementary source information with RGB in the segmentation network. We
provide a robust method for foggy scene understanding by training two models
(normal and foggy) simultaneously with shared weights (each model is trained on
each weather condition independently). Our model incorporates RGB colour,
depth, and luminance images via distinct encoders with dense connectivity and
features fusing, and leverages skip connections to produce consistent depth and
segmentation predictions. Using this architectural formulation with light
computational complexity at inference time, we are able to achieve comparable
performance to contemporary approaches at a fraction of the overall model
complexity.
Related papers
- DepthFM: Fast Monocular Depth Estimation with Flow Matching [22.206355073676082]
Current discriminative approaches to this problem are limited due to blurry artifacts.
State-of-the-art generative methods suffer from slow sampling due to their SDE nature.
We observe that this can be effectively framed using flow matching, since its straight trajectories through solution space offer efficiency and high quality.
arXiv Detail & Related papers (2024-03-20T17:51:53Z) - Physical Perception Network and an All-weather Multi-modality Benchmark
for Adverse Weather Image Fusion [4.3773535988950725]
Multi-modality image fusion (MMIF) integrates the complementary information from different modal images to provide comprehensive and objective interpretation of a scenes.
Existing MMIF methods lack the ability to resist different weather interferences in real-life scenarios.
We have established a benchmark for MMIF research under extreme weather conditions.
arXiv Detail & Related papers (2024-02-03T09:02:46Z) - WeatherDepth: Curriculum Contrastive Learning for Self-Supervised Depth Estimation under Adverse Weather Conditions [42.99525455786019]
We propose WeatherDepth, a self-supervised robust depth estimation model with curriculum contrastive learning.
The proposed solution is proven to be easily incorporated into various architectures and demonstrates state-of-the-art (SoTA) performance on both synthetic and real weather datasets.
arXiv Detail & Related papers (2023-10-09T09:26:27Z) - Robust Monocular Depth Estimation under Challenging Conditions [81.57697198031975]
State-of-the-art monocular depth estimation approaches are highly unreliable under challenging illumination and weather conditions.
We tackle these safety-critical issues with md4all: a simple and effective solution that works reliably under both adverse and ideal conditions.
arXiv Detail & Related papers (2023-08-18T17:59:01Z) - Similarity Min-Max: Zero-Shot Day-Night Domain Adaptation [52.923298434948606]
Low-light conditions not only hamper human visual experience but also degrade the model's performance on downstream vision tasks.
This paper challenges a more complicated scenario with border applicability, i.e., zero-shot day-night domain adaptation.
We propose a similarity min-max paradigm that considers them under a unified framework.
arXiv Detail & Related papers (2023-07-17T18:50:15Z) - Exploring the Application of Large-scale Pre-trained Models on Adverse
Weather Removal [97.53040662243768]
We propose a CLIP embedding module to make the network handle different weather conditions adaptively.
This module integrates the sample specific weather prior extracted by CLIP image encoder together with the distribution specific information learned by a set of parameters.
arXiv Detail & Related papers (2023-06-15T10:06:13Z) - Toward Fast, Flexible, and Robust Low-Light Image Enhancement [87.27326390675155]
We develop a new Self-Calibrated Illumination (SCI) learning framework for fast, flexible, and robust brightening images in real-world low-light scenarios.
Considering the computational burden of the cascaded pattern, we construct the self-calibrated module which realizes the convergence between results of each stage.
We make comprehensive explorations to SCI's inherent properties including operation-insensitive adaptability and model-irrelevant generality.
arXiv Detail & Related papers (2022-04-21T14:40:32Z) - Multi-Model Learning for Real-Time Automotive Semantic Foggy Scene
Understanding via Domain Adaptation [17.530091734327296]
We propose an efficient end-to-end automotive semantic scene understanding approach that is robust to foggy weather conditions.
Our approach incorporates RGB colour, depth and luminance images via distinct encoders with dense connectivity.
Our model achieves comparable performance to contemporary approaches at a fraction of the overall model complexity.
arXiv Detail & Related papers (2020-12-09T21:04:05Z) - Object-based Illumination Estimation with Rendering-aware Neural
Networks [56.01734918693844]
We present a scheme for fast environment light estimation from the RGBD appearance of individual objects and their local image areas.
With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene.
arXiv Detail & Related papers (2020-08-06T08:23:19Z) - DeFeat-Net: General Monocular Depth via Simultaneous Unsupervised
Representation Learning [65.94499390875046]
DeFeat-Net is an approach to simultaneously learn a cross-domain dense feature representation.
Our technique is able to outperform the current state-of-the-art with around 10% reduction in all error measures.
arXiv Detail & Related papers (2020-03-30T13:10:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.