Mutual-GAN: Towards Unsupervised Cross-Weather Adaptation with Mutual
Information Constraint
- URL: http://arxiv.org/abs/2106.16000v1
- Date: Wed, 30 Jun 2021 11:44:22 GMT
- Title: Mutual-GAN: Towards Unsupervised Cross-Weather Adaptation with Mutual
Information Constraint
- Authors: Jiawei Chen and Yuexiang Li and Kai Ma and Yefeng Zheng
- Abstract summary: Convolutional neural network (CNN) have proven its success for semantic segmentation, which is a core task of emerging industrial applications such as autonomous driving.
In practical applications, the outdoor weather and illumination are changeable, e.g., cloudy and nighttime, which results in a significant drop of semantic segmentation accuracy of CNN only trained with daytime data.
We propose a novel generative adversarial network (namely Mutual-GAN) to alleviate the accuracy decline when daytime-trained neural network is applied to videos captured under adverse weather conditions.
- Score: 32.67453558911618
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Convolutional neural network (CNN) have proven its success for semantic
segmentation, which is a core task of emerging industrial applications such as
autonomous driving. However, most progress in semantic segmentation of urban
scenes is reported on standard scenarios, i.e., daytime scenes with favorable
illumination conditions. In practical applications, the outdoor weather and
illumination are changeable, e.g., cloudy and nighttime, which results in a
significant drop of semantic segmentation accuracy of CNN only trained with
daytime data. In this paper, we propose a novel generative adversarial network
(namely Mutual-GAN) to alleviate the accuracy decline when daytime-trained
neural network is applied to videos captured under adverse weather conditions.
The proposed Mutual-GAN adopts mutual information constraint to preserve
image-objects during cross-weather adaptation, which is an unsolved problem for
most unsupervised image-to-image translation approaches (e.g., CycleGAN). The
proposed Mutual-GAN is evaluated on two publicly available driving video
datasets (i.e., CamVid and SYNTHIA). The experimental results demonstrate that
our Mutual-GAN can yield visually plausible translated images and significantly
improve the semantic segmentation accuracy of daytime-trained deep learning
network while processing videos under challenging weathers.
Related papers
- SIGMA:Sinkhorn-Guided Masked Video Modeling [69.31715194419091]
Sinkhorn-guided Masked Video Modelling ( SIGMA) is a novel video pretraining method.
We distribute features of space-time tubes evenly across a limited number of learnable clusters.
Experimental results on ten datasets validate the effectiveness of SIGMA in learning more performant, temporally-aware, and robust video representations.
arXiv Detail & Related papers (2024-07-22T08:04:09Z) - Genuine Knowledge from Practice: Diffusion Test-Time Adaptation for
Video Adverse Weather Removal [53.15046196592023]
We introduce test-time adaptation into adverse weather removal in videos.
We propose the first framework that integrates test-time adaptation into the iterative diffusion reverse process.
arXiv Detail & Related papers (2024-03-12T14:21:30Z) - Transform-Equivariant Consistency Learning for Temporal Sentence
Grounding [66.10949751429781]
We introduce a novel Equivariant Consistency Regulation Learning framework to learn more discriminative representations for each video.
Our motivation comes from that the temporal boundary of the query-guided activity should be consistently predicted.
In particular, we devise a self-supervised consistency loss module to enhance the completeness and smoothness of the augmented video.
arXiv Detail & Related papers (2023-05-06T19:29:28Z) - Unsupervised Video Domain Adaptation for Action Recognition: A
Disentanglement Perspective [37.45565756522847]
We consider the generation of cross-domain videos from two sets of latent factors.
TranSVAE framework is then developed to model such generation.
Experiments on the UCF-HMDB, Jester, and Epic-Kitchens datasets verify the effectiveness and superiority of TranSVAE.
arXiv Detail & Related papers (2022-08-15T17:59:31Z) - Unsupervised Foggy Scene Understanding via Self Spatial-Temporal Label
Diffusion [51.11295961195151]
We exploit the characteristics of the foggy image sequence of driving scenes to densify the confident pseudo labels.
Based on the two discoveries of local spatial similarity and adjacent temporal correspondence of the sequential image data, we propose a novel Target-Domain driven pseudo label Diffusion scheme.
Our scheme helps the adaptive model achieve 51.92% and 53.84% mean intersection-over-union (mIoU) on two publicly available natural foggy datasets.
arXiv Detail & Related papers (2022-06-10T05:16:50Z) - Adaptive graph convolutional networks for weakly supervised anomaly
detection in videos [42.3118758940767]
We propose a weakly supervised adaptive graph convolutional network (WAGCN) to model the contextual relationships among video segments.
We fully consider the influence of other video segments on the current segment when generating the anomaly probability score for each segment.
arXiv Detail & Related papers (2022-02-14T06:31:34Z) - AuxAdapt: Stable and Efficient Test-Time Adaptation for Temporally
Consistent Video Semantic Segmentation [81.87943324048756]
In video segmentation, generating temporally consistent results across frames is as important as achieving frame-wise accuracy.
Existing methods rely on optical flow regularization or fine-tuning with test data to attain temporal consistency.
This paper presents an efficient, intuitive, and unsupervised online adaptation method, AuxAdapt, for improving the temporal consistency of most neural network models.
arXiv Detail & Related papers (2021-10-24T07:07:41Z) - Let There be Light: Improved Traffic Surveillance via Detail Preserving
Night-to-Day Transfer [19.33490492872067]
We propose a framework to alleviate the accuracy decline when object detection is taken to adverse conditions by using image translation method.
To alleviate the detail corruptions caused by Generative Adversarial Networks (GANs), we propose to utilize Kernel Prediction Network (KPN) based method to refine the nighttime to daytime image translation.
arXiv Detail & Related papers (2021-05-11T13:18:50Z) - Semantic Segmentation of Medium-Resolution Satellite Imagery using
Conditional Generative Adversarial Networks [3.4797121357690153]
We propose Conditional Generative Adversarial Networks (CGAN) based approach of image-to-image translation for high-resolution satellite imagery.
We find that the CGAN model outperforms the CNN model of similar complexity by a significant margin on an unseen imbalanced test dataset.
arXiv Detail & Related papers (2020-12-05T18:18:45Z) - A Prospective Study on Sequence-Driven Temporal Sampling and Ego-Motion
Compensation for Action Recognition in the EPIC-Kitchens Dataset [68.8204255655161]
Action recognition is one of the top-challenging research fields in computer vision.
ego-motion recorded sequences have become of important relevance.
The proposed method aims to cope with it by estimating this ego-motion or camera motion.
arXiv Detail & Related papers (2020-08-26T14:44:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.