Competitive Simplicity for Multi-Task Learning for Real-Time Foggy Scene
Understanding via Domain Adaptation
- URL: http://arxiv.org/abs/2012.05304v1
- Date: Wed, 9 Dec 2020 20:38:34 GMT
- Title: Competitive Simplicity for Multi-Task Learning for Real-Time Foggy Scene
Understanding via Domain Adaptation
- Authors: Naif Alshammari, Samet Akcay, and Toby P. Breckon
- Abstract summary: We propose a multi-task learning approach capable of performing in real-time semantic scene understanding and monocular depth estimation under foggy weather conditions.
Our model incorporates RGB colour, depth, and luminance images via distinct encoders with dense connectivity and features fusing.
- Score: 17.530091734327296
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automotive scene understanding under adverse weather conditions raises a
realistic and challenging problem attributable to poor outdoor scene visibility
(e.g. foggy weather). However, because most contemporary scene understanding
approaches are applied under ideal-weather conditions, such approaches may not
provide genuinely optimal performance when compared to established a priori
insights on extreme-weather understanding. In this paper, we propose a complex
but competitive multi-task learning approach capable of performing in real-time
semantic scene understanding and monocular depth estimation under foggy weather
conditions by leveraging both recent advances in adversarial training and
domain adaptation. As an end-to-end pipeline, our model provides a novel
solution to surpass degraded visibility in foggy weather conditions by
transferring scenes from foggy to normal using a GAN-based model. For optimal
performance in semantic segmentation, our model generates depth to be used as
complementary source information with RGB in the segmentation network. We
provide a robust method for foggy scene understanding by training two models
(normal and foggy) simultaneously with shared weights (each model is trained on
each weather condition independently). Our model incorporates RGB colour,
depth, and luminance images via distinct encoders with dense connectivity and
features fusing, and leverages skip connections to produce consistent depth and
segmentation predictions. Using this architectural formulation with light
computational complexity at inference time, we are able to achieve comparable
performance to contemporary approaches at a fraction of the overall model
complexity.
Related papers
- Leveraging Stable Diffusion for Monocular Depth Estimation via Image Semantic Encoding [1.0445560141983634]
We propose a novel image-based semantic embedding that extracts contextual information directly from visual features.
Our method achieves performance comparable to state-of-the-art models while addressing the shortcomings of CLIP embeddings in handling outdoor scenes.
arXiv Detail & Related papers (2025-02-01T15:37:22Z) - Multi-Modality Driven LoRA for Adverse Condition Depth Estimation [61.525312117638116]
We propose Multi-Modality Driven LoRA (MMD-LoRA) for Adverse Condition Depth Estimation.
It consists of two core components: Prompt Driven Domain Alignment (PDDA) and Visual-Text Consistent Contrastive Learning (VTCCL)
It achieves state-of-the-art performance on the nuScenes and Oxford RobotCar datasets.
arXiv Detail & Related papers (2024-12-28T14:23:58Z) - WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning [69.82211470647349]
We introduce the first generalist weather foundation model (WeatherGFM)
It addresses a wide spectrum of weather understanding tasks in a unified manner.
Our model can effectively handle up to ten weather understanding tasks, including weather forecasting, super-resolution, weather image translation, and post-processing.
arXiv Detail & Related papers (2024-11-08T09:14:19Z) - WeatherDepth: Curriculum Contrastive Learning for Self-Supervised Depth Estimation under Adverse Weather Conditions [42.99525455786019]
We propose WeatherDepth, a self-supervised robust depth estimation model with curriculum contrastive learning.
The proposed solution is proven to be easily incorporated into various architectures and demonstrates state-of-the-art (SoTA) performance on both synthetic and real weather datasets.
arXiv Detail & Related papers (2023-10-09T09:26:27Z) - Robust Monocular Depth Estimation under Challenging Conditions [81.57697198031975]
State-of-the-art monocular depth estimation approaches are highly unreliable under challenging illumination and weather conditions.
We tackle these safety-critical issues with md4all: a simple and effective solution that works reliably under both adverse and ideal conditions.
arXiv Detail & Related papers (2023-08-18T17:59:01Z) - Exploring the Application of Large-scale Pre-trained Models on Adverse
Weather Removal [97.53040662243768]
We propose a CLIP embedding module to make the network handle different weather conditions adaptively.
This module integrates the sample specific weather prior extracted by CLIP image encoder together with the distribution specific information learned by a set of parameters.
arXiv Detail & Related papers (2023-06-15T10:06:13Z) - Multi-Model Learning for Real-Time Automotive Semantic Foggy Scene
Understanding via Domain Adaptation [17.530091734327296]
We propose an efficient end-to-end automotive semantic scene understanding approach that is robust to foggy weather conditions.
Our approach incorporates RGB colour, depth and luminance images via distinct encoders with dense connectivity.
Our model achieves comparable performance to contemporary approaches at a fraction of the overall model complexity.
arXiv Detail & Related papers (2020-12-09T21:04:05Z) - Object-based Illumination Estimation with Rendering-aware Neural
Networks [56.01734918693844]
We present a scheme for fast environment light estimation from the RGBD appearance of individual objects and their local image areas.
With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene.
arXiv Detail & Related papers (2020-08-06T08:23:19Z) - DeFeat-Net: General Monocular Depth via Simultaneous Unsupervised
Representation Learning [65.94499390875046]
DeFeat-Net is an approach to simultaneously learn a cross-domain dense feature representation.
Our technique is able to outperform the current state-of-the-art with around 10% reduction in all error measures.
arXiv Detail & Related papers (2020-03-30T13:10:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.