Local Memory Attention for Fast Video Semantic Segmentation
- URL: http://arxiv.org/abs/2101.01715v1
- Date: Tue, 5 Jan 2021 18:57:09 GMT
- Title: Local Memory Attention for Fast Video Semantic Segmentation
- Authors: Matthieu Paul, Martin Danelljan, Luc Van Gool, Radu Timofte
- Abstract summary: We propose a novel neural network module that transforms an existing single-frame semantic segmentation model into a video semantic segmentation pipeline.
Our approach aggregates a rich representation of the semantic information in past frames into a memory module.
We observe an improvement in segmentation performance on Cityscapes by 1.7% and 2.1% in mIoU respectively, while increasing inference time of ERFNet by only 1.5ms.
- Score: 157.7618884769969
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel neural network module that transforms an existing
single-frame semantic segmentation model into a video semantic segmentation
pipeline. In contrast to prior works, we strive towards a simple and general
module that can be integrated into virtually any single-frame architecture. Our
approach aggregates a rich representation of the semantic information in past
frames into a memory module. Information stored in the memory is then accessed
through an attention mechanism. This provides temporal appearance cues from
prior frames, which are then fused with an encoding of the current frame
through a second attention-based module. The segmentation decoder processes the
fused representation to predict the final semantic segmentation. We integrate
our approach into two popular semantic segmentation networks: ERFNet and
PSPNet. We observe an improvement in segmentation performance on Cityscapes by
1.7% and 2.1% in mIoU respectively, while increasing inference time of ERFNet
by only 1.5ms.
Related papers
- Temporally Consistent Referring Video Object Segmentation with Hybrid Memory [98.80249255577304]
We propose an end-to-end R-VOS paradigm that explicitly models temporal consistency alongside the referring segmentation.
Features of frames with automatically generated high-quality reference masks are propagated to segment remaining frames.
Extensive experiments demonstrate that our approach enhances temporal consistency by a significant margin.
arXiv Detail & Related papers (2024-03-28T13:32:49Z) - Multi-grained Temporal Prototype Learning for Few-shot Video Object
Segmentation [156.4142424784322]
Few-Shot Video Object (FSVOS) aims to segment objects in a query video with the same category defined by a few annotated support images.
We propose to leverage multi-grained temporal guidance information for handling the temporal correlation nature of video data.
Our proposed video IPMT model significantly outperforms previous models on two benchmark datasets.
arXiv Detail & Related papers (2023-09-20T09:16:34Z) - DPANET:Dual Pooling Attention Network for Semantic Segmentation [0.0]
We propose a lightweight and flexible neural network named Dual Pool Attention Network(DPANet)
The first component is spatial pool attention module, we formulate an easy and powerful method densely to extract contextual characteristics.
The second component is channel pool attention module. So, the aim of this module is stripping them out, in order to construct relationship of all channels and heighten different channels semantic information selectively.
arXiv Detail & Related papers (2022-10-11T13:29:33Z) - Adaptive Intermediate Representations for Video Understanding [50.64187463941215]
We introduce a new way to leverage semantic segmentation as an intermediate representation for video understanding.
We propose a general framework which learns the intermediate representations (optical flow and semantic segmentation) jointly with the final video understanding task.
We obtain more powerful visual representations for videos which lead to performance gains over the state-of-the-art.
arXiv Detail & Related papers (2021-04-14T21:37:23Z) - LSMVOS: Long-Short-Term Similarity Matching for Video Object [3.3518869877513895]
Semi-supervised video object segmentation refers to segmenting the object in subsequent frames given the object label in the first frame.
This paper explores a new propagation method, uses short-term matching modules to extract the information of the previous frame and apply it in propagation.
By combining the long-term matching module with the short-term matching module, the whole network can achieve efficient video object segmentation without online fine tuning.
arXiv Detail & Related papers (2020-09-02T01:32:05Z) - Interactive Video Object Segmentation Using Global and Local Transfer
Modules [51.93009196085043]
We develop a deep neural network, which consists of the annotation network (A-Net) and the transfer network (T-Net)
Given user scribbles on a frame, A-Net yields a segmentation result based on the encoder-decoder architecture.
We train the entire network in two stages, by emulating user scribbles and employing an auxiliary loss.
arXiv Detail & Related papers (2020-07-16T06:49:07Z) - Video Object Segmentation with Episodic Graph Memory Networks [198.74780033475724]
A graph memory network is developed to address the novel idea of "learning to update the segmentation model"
We exploit an episodic memory network, organized as a fully connected graph, to store frames as nodes and capture cross-frame correlations by edges.
The proposed graph memory network yields a neat yet principled framework, which can generalize well both one-shot and zero-shot video object segmentation tasks.
arXiv Detail & Related papers (2020-07-14T13:19:19Z) - Dual Temporal Memory Network for Efficient Video Object Segmentation [42.05305410986511]
One of the fundamental challenges in Video Object (VOS) is how to make the most use of the temporal information to boost the performance.
We present an end-to-end network which stores short- and long-term video sequence information preceding the current frame as the temporal memories.
Our network consists of two temporal sub-networks including a short-term memory sub-network and a long-term memory sub-network.
arXiv Detail & Related papers (2020-03-13T06:07:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.