Robust Semantic Segmentation in Adverse Weather Conditions by means of
Fast Video-Sequence Segmentation
- URL: http://arxiv.org/abs/2007.00290v1
- Date: Wed, 1 Jul 2020 07:29:35 GMT
- Title: Robust Semantic Segmentation in Adverse Weather Conditions by means of
Fast Video-Sequence Segmentation
- Authors: Andreas Pfeuffer and Klaus Dietmayer
- Abstract summary: Video-segmentation approaches capture temporal information of the previous video-frames in addition to current image information.
Video-segmentation approaches, which are often based on recurrent neural networks, cannot be applied in real-time applications anymore.
In this work, the LSTM-ICNet is sped up by modifying the recurrent units of the network so that it becomes real-time capable again.
- Score: 12.788867618508688
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Computer vision tasks such as semantic segmentation perform very well in good
weather conditions, but if the weather turns bad, they have problems to achieve
this performance in these conditions. One possibility to obtain more robust and
reliable results in adverse weather conditions is to use video-segmentation
approaches instead of commonly used single-image segmentation methods.
Video-segmentation approaches capture temporal information of the previous
video-frames in addition to current image information, and hence, they are more
robust against disturbances, especially if they occur in only a few frames of
the video-sequence. However, video-segmentation approaches, which are often
based on recurrent neural networks, cannot be applied in real-time applications
anymore, since their recurrent structures in the network are computational
expensive. For instance, the inference time of the LSTM-ICNet, in which
recurrent units are placed at proper positions in the single-segmentation
approach ICNet, increases up to 61 percent compared to the basic ICNet. Hence,
in this work, the LSTM-ICNet is sped up by modifying the recurrent units of the
network so that it becomes real-time capable again. Experiments on different
datasets and various weather conditions show that the inference time can be
decreased by about 23 percent by these modifications, while they achieve
similar performance than the LSTM-ICNet and outperform the single-segmentation
approach enormously in adverse weather conditions.
Related papers
- Genuine Knowledge from Practice: Diffusion Test-Time Adaptation for
Video Adverse Weather Removal [53.15046196592023]
We introduce test-time adaptation into adverse weather removal in videos.
We propose the first framework that integrates test-time adaptation into the iterative diffusion reverse process.
arXiv Detail & Related papers (2024-03-12T14:21:30Z) - Distortion-Aware Network Pruning and Feature Reuse for Real-time Video
Segmentation [49.17930380106643]
We propose a novel framework to speed up any architecture with skip-connections for real-time vision tasks.
Specifically, at the arrival of each frame, we transform the features from the previous frame to reuse them at specific spatial bins.
We then perform partial computation of the backbone network on the regions of the current frame that captures temporal differences between the current and previous frame.
arXiv Detail & Related papers (2022-06-20T07:20:02Z) - AuxAdapt: Stable and Efficient Test-Time Adaptation for Temporally
Consistent Video Semantic Segmentation [81.87943324048756]
In video segmentation, generating temporally consistent results across frames is as important as achieving frame-wise accuracy.
Existing methods rely on optical flow regularization or fine-tuning with test data to attain temporal consistency.
This paper presents an efficient, intuitive, and unsupervised online adaptation method, AuxAdapt, for improving the temporal consistency of most neural network models.
arXiv Detail & Related papers (2021-10-24T07:07:41Z) - RiWNet: A moving object instance segmentation Network being Robust in
adverse Weather conditions [13.272209740926156]
We focus on a new possibility, that is, to improve its resilience to weather interference through the network's structural design.
We propose a novel FPN structure called RiWFPN with a progressive top-down interaction and attention refinement module.
We extend SOLOV2 to capture temporal information in video to learn motion information, and propose a moving object instance segmentation network with RiWFPN called RiWNet.
arXiv Detail & Related papers (2021-09-04T08:55:36Z) - Mutual-GAN: Towards Unsupervised Cross-Weather Adaptation with Mutual
Information Constraint [32.67453558911618]
Convolutional neural network (CNN) have proven its success for semantic segmentation, which is a core task of emerging industrial applications such as autonomous driving.
In practical applications, the outdoor weather and illumination are changeable, e.g., cloudy and nighttime, which results in a significant drop of semantic segmentation accuracy of CNN only trained with daytime data.
We propose a novel generative adversarial network (namely Mutual-GAN) to alleviate the accuracy decline when daytime-trained neural network is applied to videos captured under adverse weather conditions.
arXiv Detail & Related papers (2021-06-30T11:44:22Z) - Energy-Efficient Model Compression and Splitting for Collaborative
Inference Over Time-Varying Channels [52.60092598312894]
We propose a technique to reduce the total energy bill at the edge device by utilizing model compression and time-varying model split between the edge and remote nodes.
Our proposed solution results in minimal energy consumption and $CO$ emission compared to the considered baselines.
arXiv Detail & Related papers (2021-06-02T07:36:27Z) - Temporal Modulation Network for Controllable Space-Time Video
Super-Resolution [66.06549492893947]
Space-time video super-resolution aims to increase the spatial and temporal resolutions of low-resolution and low-frame-rate videos.
Deformable convolution based methods have achieved promising STVSR performance, but they could only infer the intermediate frame pre-defined in the training stage.
We propose a Temporal Modulation Network (TMNet) to interpolate arbitrary intermediate frame(s) with accurate high-resolution reconstruction.
arXiv Detail & Related papers (2021-04-21T17:10:53Z) - Intrinsic Temporal Regularization for High-resolution Human Video
Synthesis [59.54483950973432]
temporal consistency is crucial for extending image processing pipelines to the video domain.
We propose an effective intrinsic temporal regularization scheme, where an intrinsic confidence map is estimated via the frame generator to regulate motion estimation.
We apply our intrinsic temporal regulation to single-image generator, leading to a powerful " INTERnet" capable of generating $512times512$ resolution human action videos.
arXiv Detail & Related papers (2020-12-11T05:29:45Z) - Noisy-LSTM: Improving Temporal Awareness for Video Semantic Segmentation [29.00635219317848]
This paper presents a new model named Noisy-LSTM, which is trainable in an end-to-end manner.
We also present a simple yet effective training strategy, which replaces a frame in video sequence with noises.
arXiv Detail & Related papers (2020-10-19T13:08:15Z) - iSeeBetter: Spatio-temporal video super-resolution using recurrent
generative back-projection networks [0.0]
We present iSeeBetter, a novel GAN-based structural-temporal approach to video super-resolution (VSR)
iSeeBetter extracts spatial and temporal information from the current and neighboring frames using the concept of recurrent back-projection networks as its generator.
Our results demonstrate that iSeeBetter offers superior VSR fidelity and surpasses state-of-the-art performance.
arXiv Detail & Related papers (2020-06-13T01:36:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.