Competitive Simplicity for Multi-Task Learning for Real-Time Foggy Scene
Understanding via Domain Adaptation
- URL: http://arxiv.org/abs/2012.05304v1
- Date: Wed, 9 Dec 2020 20:38:34 GMT
- Title: Competitive Simplicity for Multi-Task Learning for Real-Time Foggy Scene
Understanding via Domain Adaptation
- Authors: Naif Alshammari, Samet Akcay, and Toby P. Breckon
- Abstract summary: We propose a multi-task learning approach capable of performing in real-time semantic scene understanding and monocular depth estimation under foggy weather conditions.
Our model incorporates RGB colour, depth, and luminance images via distinct encoders with dense connectivity and features fusing.
- Score: 17.530091734327296
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automotive scene understanding under adverse weather conditions raises a
realistic and challenging problem attributable to poor outdoor scene visibility
(e.g. foggy weather). However, because most contemporary scene understanding
approaches are applied under ideal-weather conditions, such approaches may not
provide genuinely optimal performance when compared to established a priori
insights on extreme-weather understanding. In this paper, we propose a complex
but competitive multi-task learning approach capable of performing in real-time
semantic scene understanding and monocular depth estimation under foggy weather
conditions by leveraging both recent advances in adversarial training and
domain adaptation. As an end-to-end pipeline, our model provides a novel
solution to surpass degraded visibility in foggy weather conditions by
transferring scenes from foggy to normal using a GAN-based model. For optimal
performance in semantic segmentation, our model generates depth to be used as
complementary source information with RGB in the segmentation network. We
provide a robust method for foggy scene understanding by training two models
(normal and foggy) simultaneously with shared weights (each model is trained on
each weather condition independently). Our model incorporates RGB colour,
depth, and luminance images via distinct encoders with dense connectivity and
features fusing, and leverages skip connections to produce consistent depth and
segmentation predictions. Using this architectural formulation with light
computational complexity at inference time, we are able to achieve comparable
performance to contemporary approaches at a fraction of the overall model
complexity.
Related papers
- WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning [69.82211470647349]
We introduce the first generalist weather foundation model (WeatherGFM)
It addresses a wide spectrum of weather understanding tasks in a unified manner.
Our model can effectively handle up to ten weather understanding tasks, including weather forecasting, super-resolution, weather image translation, and post-processing.
arXiv Detail & Related papers (2024-11-08T09:14:19Z) - DepthFM: Fast Monocular Depth Estimation with Flow Matching [22.206355073676082]
Current discriminative approaches to this problem are limited due to blurry artifacts.
State-of-the-art generative methods suffer from slow sampling due to their SDE nature.
We observe that this can be effectively framed using flow matching, since its straight trajectories through solution space offer efficiency and high quality.
arXiv Detail & Related papers (2024-03-20T17:51:53Z) - WeatherDepth: Curriculum Contrastive Learning for Self-Supervised Depth Estimation under Adverse Weather Conditions [42.99525455786019]
We propose WeatherDepth, a self-supervised robust depth estimation model with curriculum contrastive learning.
The proposed solution is proven to be easily incorporated into various architectures and demonstrates state-of-the-art (SoTA) performance on both synthetic and real weather datasets.
arXiv Detail & Related papers (2023-10-09T09:26:27Z) - Robust Monocular Depth Estimation under Challenging Conditions [81.57697198031975]
State-of-the-art monocular depth estimation approaches are highly unreliable under challenging illumination and weather conditions.
We tackle these safety-critical issues with md4all: a simple and effective solution that works reliably under both adverse and ideal conditions.
arXiv Detail & Related papers (2023-08-18T17:59:01Z) - Exploring the Application of Large-scale Pre-trained Models on Adverse
Weather Removal [97.53040662243768]
We propose a CLIP embedding module to make the network handle different weather conditions adaptively.
This module integrates the sample specific weather prior extracted by CLIP image encoder together with the distribution specific information learned by a set of parameters.
arXiv Detail & Related papers (2023-06-15T10:06:13Z) - Toward Fast, Flexible, and Robust Low-Light Image Enhancement [87.27326390675155]
We develop a new Self-Calibrated Illumination (SCI) learning framework for fast, flexible, and robust brightening images in real-world low-light scenarios.
Considering the computational burden of the cascaded pattern, we construct the self-calibrated module which realizes the convergence between results of each stage.
We make comprehensive explorations to SCI's inherent properties including operation-insensitive adaptability and model-irrelevant generality.
arXiv Detail & Related papers (2022-04-21T14:40:32Z) - Numerical Weather Forecasting using Convolutional-LSTM with Attention
and Context Matcher Mechanisms [10.759556555869798]
We introduce a novel deep learning architecture for forecasting high-resolution weather data.
Our Weather Model achieves significant performance improvements compared to baseline deep learning models.
arXiv Detail & Related papers (2021-02-01T08:30:42Z) - Multi-Model Learning for Real-Time Automotive Semantic Foggy Scene
Understanding via Domain Adaptation [17.530091734327296]
We propose an efficient end-to-end automotive semantic scene understanding approach that is robust to foggy weather conditions.
Our approach incorporates RGB colour, depth and luminance images via distinct encoders with dense connectivity.
Our model achieves comparable performance to contemporary approaches at a fraction of the overall model complexity.
arXiv Detail & Related papers (2020-12-09T21:04:05Z) - Object-based Illumination Estimation with Rendering-aware Neural
Networks [56.01734918693844]
We present a scheme for fast environment light estimation from the RGBD appearance of individual objects and their local image areas.
With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene.
arXiv Detail & Related papers (2020-08-06T08:23:19Z) - DeFeat-Net: General Monocular Depth via Simultaneous Unsupervised
Representation Learning [65.94499390875046]
DeFeat-Net is an approach to simultaneously learn a cross-domain dense feature representation.
Our technique is able to outperform the current state-of-the-art with around 10% reduction in all error measures.
arXiv Detail & Related papers (2020-03-30T13:10:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.