Extended multi-stream temporal-attention module for skeleton-based human action recognition (HAR)
- URL: http://arxiv.org/abs/2411.06553v1
- Date: Sun, 10 Nov 2024 18:28:52 GMT
- Title: Extended multi-stream temporal-attention module for skeleton-based human action recognition (HAR)
- Authors: Faisal Mehmood, Xin Guo, Enqing Chen, Muhammad Azeem Akbar, Arif Ali Khan, Sami Ullah,
- Abstract summary: Graph convolutional networks (GCNs) are an effective skeleton-based human action recognition (HAR) technique.
GCNs enable the specification of CNNs to a non-Euclidean frame that is more flexible.
- Score: 4.822426770727152
- License:
- Abstract: Graph convolutional networks (GCNs) are an effective skeleton-based human action recognition (HAR) technique. GCNs enable the specification of CNNs to a non-Euclidean frame that is more flexible. The previous GCN-based models still have a lot of issues: (I) The graph structure is the same for all model layers and input data.
Related papers
- Temporal-Channel Topology Enhanced Network for Skeleton-Based Action
Recognition [26.609509266693077]
We propose a novel CNN architecture, Temporal-Channel Topology Enhanced Network (TCTE-Net), to learn spatial and temporal topologies for skeleton-based action recognition.
TCTE-Net shows state-of-the-art performance compared to CNN-based methods and achieves superior performance compared to GCN-based methods.
arXiv Detail & Related papers (2023-02-25T03:09:07Z) - DG-STGCN: Dynamic Spatial-Temporal Modeling for Skeleton-based Action
Recognition [77.87404524458809]
We propose a new framework for skeleton-based action recognition, namely Dynamic Group Spatio-Temporal GCN (DG-STGCN)
It consists of two modules, DG-GCN and DG-TCN, respectively, for spatial and temporal modeling.
DG-STGCN consistently outperforms state-of-the-art methods, often by a notable margin.
arXiv Detail & Related papers (2022-10-12T03:17:37Z) - Pose-Guided Graph Convolutional Networks for Skeleton-Based Action
Recognition [32.07659338674024]
Graph convolutional networks (GCNs) can model the human body skeletons as spatial and temporal graphs.
In this work, we propose pose-guided GCN (PG-GCN), a multi-modal framework for high-performance human action recognition.
The core idea of this module is to utilize a trainable graph to aggregate features from the skeleton stream with that of the pose stream, which leads to a network with more robust feature representation ability.
arXiv Detail & Related papers (2022-10-10T02:08:49Z) - Skeleton-based Action Recognition via Adaptive Cross-Form Learning [75.92422282666767]
Skeleton-based action recognition aims to project skeleton sequences to action categories, where sequences are derived from multiple forms of pre-detected points.
Existing methods tend to improve GCNs by leveraging multi-form skeletons due to their complementary cues.
We present Adaptive Cross-Form Learning (ACFL), which empowers well-designed GCNs to generate complementary representation from single-form skeletons.
arXiv Detail & Related papers (2022-06-30T07:40:03Z) - Topology-aware Convolutional Neural Network for Efficient Skeleton-based
Action Recognition [15.93566875893684]
We propose a pure CNN architecture named Topology-aware CNN (Ta-CNN) in this paper.
We develop a novel cross-channel feature augmentation module, which is a combo of map-attend-group-map operations.
In particular, we develop a novel cross-channel feature augmentation module, which is a combo of map-attend-group-map operations.
arXiv Detail & Related papers (2021-12-08T09:02:50Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Dynamic GCN: Context-enriched Topology Learning for Skeleton-based
Action Recognition [40.467040910143616]
We propose Dynamic GCN, in which a novel convolutional neural network named Contextencoding Network (CeN) is introduced to learn skeleton topology automatically.
CeN is extremely lightweight yet effective, and can be embedded into a graph convolutional layer.
Dynamic GCN achieves better performance with $2times$$4times$ fewer FLOPs than existing methods.
arXiv Detail & Related papers (2020-07-29T09:12:06Z) - Simple and Deep Graph Convolutional Networks [63.76221532439285]
Graph convolutional networks (GCNs) are a powerful deep learning approach for graph-structured data.
Despite their success, most of the current GCN models are shallow, due to the em over-smoothing problem.
We propose the GCNII, an extension of the vanilla GCN model with two simple yet effective techniques.
arXiv Detail & Related papers (2020-07-04T16:18:06Z) - Structure-Aware Human-Action Generation [126.05874420893092]
Graph convolutional networks (GCNs) are promising way to leverage structure information to learn structure representations.
We propose a variant of GCNs to leverage the powerful self-attention mechanism to adaptively sparsify a complete action graph in the temporal space.
Our method could dynamically attend to important past frames and construct a sparse graph to apply in the GCN framework, well-capturing the structure information in action sequences.
arXiv Detail & Related papers (2020-07-04T00:18:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.