Towards Efficient Scene Understanding via Squeeze Reasoning
- URL: http://arxiv.org/abs/2011.03308v3
- Date: Tue, 20 Jul 2021 06:29:08 GMT
- Title: Towards Efficient Scene Understanding via Squeeze Reasoning
- Authors: Xiangtai Li, Xia Li, Ansheng You, Li Zhang, Guangliang Cheng, Kuiyuan
Yang, Yunhai Tong, Zhouchen Lin
- Abstract summary: We propose a novel framework called Squeeze Reasoning.
Instead of propagating information on the spatial map, we first learn to squeeze the input feature into a channel-wise global vector.
We show that our approach can be modularized as an end-to-end trained block and can be easily plugged into existing networks.
- Score: 71.1139549949694
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph-based convolutional model such as non-local block has shown to be
effective for strengthening the context modeling ability in convolutional
neural networks (CNNs). However, its pixel-wise computational overhead is
prohibitive which renders it unsuitable for high resolution imagery. In this
paper, we explore the efficiency of context graph reasoning and propose a novel
framework called Squeeze Reasoning. Instead of propagating information on the
spatial map, we first learn to squeeze the input feature into a channel-wise
global vector and perform reasoning within the single vector where the
computation cost can be significantly reduced. Specifically, we build the node
graph in the vector where each node represents an abstract semantic concept.
The refined feature within the same semantic category results to be consistent,
which is thus beneficial for downstream tasks. We show that our approach can be
modularized as an end-to-end trained block and can be easily plugged into
existing networks. {Despite its simplicity and being lightweight, the proposed
strategy allows us to establish the considerable results on different semantic
segmentation datasets and shows significant improvements with respect to strong
baselines on various other scene understanding tasks including object
detection, instance segmentation and panoptic segmentation.} Code is available
at \url{https://github.com/lxtGH/SFSegNets}.
Related papers
- Self-Supervised Node Representation Learning via Node-to-Neighbourhood
Alignment [10.879056662671802]
Self-supervised node representation learning aims to learn node representations from unlabelled graphs that rival the supervised counterparts.
In this work, we present simple-yet-effective self-supervised node representation learning via aligning the hidden representations of nodes and their neighbourhood.
We learn node representations that achieve promising node classification performance on a set of graph-structured datasets from small- to large-scale.
arXiv Detail & Related papers (2023-02-09T13:21:18Z) - Dynamic Graph Message Passing Networks for Visual Recognition [112.49513303433606]
Modelling long-range dependencies is critical for scene understanding tasks in computer vision.
A fully-connected graph is beneficial for such modelling, but its computational overhead is prohibitive.
We propose a dynamic graph message passing network, that significantly reduces the computational complexity.
arXiv Detail & Related papers (2022-09-20T14:41:37Z) - Local Permutation Equivariance For Graph Neural Networks [2.208242292882514]
We develop a new method, named locally permutation-equivariant graph neural networks.
It provides a framework for building graph neural networks that operate on local node neighbourhoods.
We experimentally validate the method on a range of graph benchmark classification tasks.
arXiv Detail & Related papers (2021-11-23T13:10:34Z) - RICE: Refining Instance Masks in Cluttered Environments with Graph
Neural Networks [53.15260967235835]
We propose a novel framework that refines the output of such methods by utilizing a graph-based representation of instance masks.
We train deep networks capable of sampling smart perturbations to the segmentations, and a graph neural network, which can encode relations between objects, to evaluate the segmentations.
We demonstrate an application that uses uncertainty estimates generated by our method to guide a manipulator, leading to efficient understanding of cluttered scenes.
arXiv Detail & Related papers (2021-06-29T20:29:29Z) - Inferring community characteristics in labelled networks [6.85316573653194]
We introduce a new generative model, the feature-first block model (FFBM)
We present a method to efficiently sample from the posterior distribution of the FFBM parameters.
The main advantages of the proposed approach are that the whole feature-space is used automatically, and features can be rank-ordered implicitly according to impact.
arXiv Detail & Related papers (2021-05-28T12:07:10Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z) - Quiver Signal Processing (QSP) [145.6921439353007]
We state the basics for a signal processing framework on quiver representations.
We propose a signal processing framework that allows us to handle heterogeneous multidimensional information in networks.
arXiv Detail & Related papers (2020-10-22T08:40:15Z) - Spatial Pyramid Based Graph Reasoning for Semantic Segmentation [67.47159595239798]
We apply graph convolution into the semantic segmentation task and propose an improved Laplacian.
The graph reasoning is directly performed in the original feature space organized as a spatial pyramid.
We achieve comparable performance with advantages in computational and memory overhead.
arXiv Detail & Related papers (2020-03-23T12:28:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.