ANUBIS: Review and Benchmark Skeleton-Based Action Recognition Methods
with a New Dataset
- URL: http://arxiv.org/abs/2205.02071v2
- Date: Thu, 5 May 2022 01:06:52 GMT
- Title: ANUBIS: Review and Benchmark Skeleton-Based Action Recognition Methods
with a New Dataset
- Authors: Zhenyue Qin, Yang Liu, Madhawa Perera, Saeed Anwar, Tom Gedeon, Pan
Ji, Dongwoo Kim
- Abstract summary: We present a review in the form of a taxonomy on existing works of skeleton-based action recognition.
To promote more fair and comprehensive evaluation, we collect ANUBIS, a large-scale human skeleton dataset.
- Score: 26.581495230711198
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Skeleton-based action recognition, as a subarea of action recognition, is
swiftly accumulating attention and popularity. The task is to recognize actions
performed by human articulation points. Compared with other data modalities, 3D
human skeleton representations have extensive unique desirable characteristics,
including succinctness, robustness, racial-impartiality, and many more. We aim
to provide a roadmap for new and existing researchers a on the landscapes of
skeleton-based action recognition for new and existing researchers. To this
end, we present a review in the form of a taxonomy on existing works of
skeleton-based action recognition. We partition them into four major
categories: (1) datasets; (2) extracting spatial features; (3) capturing
temporal patterns; (4) improving signal quality. For each method, we provide
concise yet informatively-sufficient descriptions. To promote more fair and
comprehensive evaluation on existing approaches of skeleton-based action
recognition, we collect ANUBIS, a large-scale human skeleton dataset. Compared
with previously collected dataset, ANUBIS are advantageous in the following
four aspects: (1) employing more recently released sensors; (2) containing
novel back view; (3) encouraging high enthusiasm of subjects; (4) including
actions of the COVID pandemic era. Using ANUBIS, we comparably benchmark
performance of current skeleton-based action recognizers. At the end of this
paper, we outlook future development of skeleton-based action recognition by
listing several new technical problems. We believe they are valuable to solve
in order to commercialize skeleton-based action recognition in the near future.
The dataset of ANUBIS is available at:
http://hcc-workshop.anu.edu.au/webs/anu101/home.
Related papers
- SkeleTR: Towrads Skeleton-based Action Recognition in the Wild [86.03082891242698]
SkeleTR is a new framework for skeleton-based action recognition.
It first models the intra-person skeleton dynamics for each skeleton sequence with graph convolutions.
It then uses stacked Transformer encoders to capture person interactions that are important for action recognition in general scenarios.
arXiv Detail & Related papers (2023-09-20T16:22:33Z) - One-Shot Action Recognition via Multi-Scale Spatial-Temporal Skeleton
Matching [77.6989219290789]
One-shot skeleton action recognition aims to learn a skeleton action recognition model with a single training sample.
This paper presents a novel one-shot skeleton action recognition technique that handles skeleton action recognition via multi-scale spatial-temporal feature matching.
arXiv Detail & Related papers (2023-07-14T11:52:10Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - PYSKL: Towards Good Practices for Skeleton Action Recognition [77.87404524458809]
PYSKL is an open-source toolbox for skeleton-based action recognition based on PyTorch.
It implements six different algorithms under a unified framework to ease the comparison of efficacy and efficiency.
PYSKL supports the training and testing of nine skeleton-based action recognition benchmarks and achieves state-of-the-art recognition performance on eight of them.
arXiv Detail & Related papers (2022-05-19T09:58:32Z) - Towards a Deeper Understanding of Skeleton-based Gait Recognition [4.812321790984493]
In recent years, most gait recognition methods used the person's silhouette to extract the gait features.
Model-based methods do not suffer from these problems and are able to represent the temporal motion of body joints.
In this work, we propose an approach based on Graph Convolutional Networks (GCNs) that combines higher-order inputs, and residual networks.
arXiv Detail & Related papers (2022-04-16T18:23:37Z) - Skeleton-Based Mutually Assisted Interacted Object Localization and
Human Action Recognition [111.87412719773889]
We propose a joint learning framework for "interacted object localization" and "human action recognition" based on skeleton data.
Our method achieves the best or competitive performance with the state-of-the-art methods for human action recognition.
arXiv Detail & Related papers (2021-10-28T10:09:34Z) - UNIK: A Unified Framework for Real-world Skeleton-based Action
Recognition [11.81043814295441]
We introduce UNIK, a novel skeleton-based action recognition method that is able to generalize across datasets.
To study the cross-domain generalizability of action recognition in real-world videos, we re-evaluate state-of-the-art approaches as well as the proposed UNIK.
Results show that the proposed UNIK, with pre-training on Posetics, generalizes well and outperforms state-of-the-art when transferred onto four target action classification datasets.
arXiv Detail & Related papers (2021-07-19T02:00:28Z) - Revisiting Skeleton-based Action Recognition [107.08112310075114]
PoseC3D is a new approach to skeleton-based action recognition, which relies on a 3D heatmap instead stack a graph sequence as the base representation of human skeletons.
On four challenging datasets, PoseC3D consistently obtains superior performance, when used alone on skeletons and in combination with the RGB modality.
arXiv Detail & Related papers (2021-04-28T06:32:17Z) - Unifying Graph Embedding Features with Graph Convolutional Networks for
Skeleton-based Action Recognition [18.001693718043292]
We propose a novel framework, which unifies 15 graph embedding features into the graph convolutional network for human action recognition.
Our model is validated by three large-scale datasets, namely NTU-RGB+D, Kinetics and SYSU-3D.
arXiv Detail & Related papers (2020-03-06T02:31:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.