Multi-Scale Semantics-Guided Neural Networks for Efficient
Skeleton-Based Human Action Recognition
- URL: http://arxiv.org/abs/2111.03993v1
- Date: Sun, 7 Nov 2021 03:50:50 GMT
- Title: Multi-Scale Semantics-Guided Neural Networks for Efficient
Skeleton-Based Human Action Recognition
- Authors: Pengfei Zhang and Cuiling Lan and Wenjun Zeng and Junliang Xing and
Jianru Xue and Nanning Zheng
- Abstract summary: A simple yet effective multi-scale semantics-guided neural network is proposed for skeleton-based action recognition.
MS-SGN achieves the state-of-the-art performance on the NTU60, NTU120, and SYSU datasets.
- Score: 140.18376685167857
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Skeleton data is of low dimension. However, there is a trend of using very
deep and complicated feedforward neural networks to model the skeleton sequence
without considering the complexity in recent year. In this paper, a simple yet
effective multi-scale semantics-guided neural network (MS-SGN) is proposed for
skeleton-based action recognition. We explicitly introduce the high level
semantics of joints (joint type and frame index) into the network to enhance
the feature representation capability of joints. Moreover, a multi-scale
strategy is proposed to be robust to the temporal scale variations. In
addition, we exploit the relationship of joints hierarchically through two
modules, i.e., a joint-level module for modeling the correlations of joints in
the same frame and a frame-level module for modeling the temporal dependencies
of frames. With an order of magnitude smaller model size than most previous
methods, MSSGN achieves the state-of-the-art performance on the NTU60, NTU120,
and SYSU datasets.
Related papers
- Multi-Scale Spatial-Temporal Self-Attention Graph Convolutional Networks for Skeleton-based Action Recognition [0.0]
In this paper, we propose self-attention GCN hybrid model, Multi-Scale Spatial-Temporal self-attention (MSST)-GCN.
We utilize spatial self-attention module with adaptive topology to understand intra-frame interactions within a frame among different body parts, and temporal self-attention module to examine correlations between frames of a node.
arXiv Detail & Related papers (2024-04-03T10:25:45Z) - SkeletonMAE: Graph-based Masked Autoencoder for Skeleton Sequence
Pre-training [110.55093254677638]
We propose an efficient skeleton sequence learning framework, named Skeleton Sequence Learning (SSL)
In this paper, we build an asymmetric graph-based encoder-decoder pre-training architecture named SkeletonMAE.
Our SSL generalizes well across different datasets and outperforms the state-of-the-art self-supervised skeleton-based action recognition methods.
arXiv Detail & Related papers (2023-07-17T13:33:11Z) - Neural Attentive Circuits [93.95502541529115]
We introduce a general purpose, yet modular neural architecture called Neural Attentive Circuits (NACs)
NACs learn the parameterization and a sparse connectivity of neural modules without using domain knowledge.
NACs achieve an 8x speedup at inference time while losing less than 3% performance.
arXiv Detail & Related papers (2022-10-14T18:00:07Z) - DG-STGCN: Dynamic Spatial-Temporal Modeling for Skeleton-based Action
Recognition [77.87404524458809]
We propose a new framework for skeleton-based action recognition, namely Dynamic Group Spatio-Temporal GCN (DG-STGCN)
It consists of two modules, DG-GCN and DG-TCN, respectively, for spatial and temporal modeling.
DG-STGCN consistently outperforms state-of-the-art methods, often by a notable margin.
arXiv Detail & Related papers (2022-10-12T03:17:37Z) - Pose-Guided Graph Convolutional Networks for Skeleton-Based Action
Recognition [32.07659338674024]
Graph convolutional networks (GCNs) can model the human body skeletons as spatial and temporal graphs.
In this work, we propose pose-guided GCN (PG-GCN), a multi-modal framework for high-performance human action recognition.
The core idea of this module is to utilize a trainable graph to aggregate features from the skeleton stream with that of the pose stream, which leads to a network with more robust feature representation ability.
arXiv Detail & Related papers (2022-10-10T02:08:49Z) - Deep Image Clustering with Contrastive Learning and Multi-scale Graph
Convolutional Networks [58.868899595936476]
This paper presents a new deep clustering approach termed image clustering with contrastive learning and multi-scale graph convolutional networks (IcicleGCN)
Experiments on multiple image datasets demonstrate the superior clustering performance of IcicleGCN over the state-of-the-art.
arXiv Detail & Related papers (2022-07-14T19:16:56Z) - SpatioTemporal Focus for Skeleton-based Action Recognition [66.8571926307011]
Graph convolutional networks (GCNs) are widely adopted in skeleton-based action recognition.
We argue that the performance of recent proposed skeleton-based action recognition methods is limited by the following factors.
Inspired by the recent attention mechanism, we propose a multi-grain contextual focus module, termed MCF, to capture the action associated relation information.
arXiv Detail & Related papers (2022-03-31T02:45:24Z) - Self-attention based anchor proposal for skeleton-based action
recognition [3.611872164105537]
Skeleton sequences are widely used for action recognition task due to its lightweight and compact characteristics.
Recent graph convolutional network (GCN) approaches have achieved great success for skeleton-based action recognition.
We propose a novel self-attention based skeleton-anchor proposal (SAP) module to comprehensively model the internal relations of a human body.
arXiv Detail & Related papers (2021-12-17T10:05:57Z) - Relational State-Space Model for Stochastic Multi-Object Systems [24.234120525358456]
This paper introduces the relational state-space model (R-SSM), a sequential hierarchical latent variable model.
R-SSM makes use of graph neural networks (GNNs) to simulate the joint state transitions of multiple correlated objects.
The utility of R-SSM is empirically evaluated on synthetic and real time-series datasets.
arXiv Detail & Related papers (2020-01-13T03:45:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.