PYSKL: Towards Good Practices for Skeleton Action Recognition
- URL: http://arxiv.org/abs/2205.09443v1
- Date: Thu, 19 May 2022 09:58:32 GMT
- Title: PYSKL: Towards Good Practices for Skeleton Action Recognition
- Authors: Haodong Duan, Jiaqi Wang, Kai Chen, Dahua Lin
- Abstract summary: PYSKL is an open-source toolbox for skeleton-based action recognition based on PyTorch.
It implements six different algorithms under a unified framework to ease the comparison of efficacy and efficiency.
PYSKL supports the training and testing of nine skeleton-based action recognition benchmarks and achieves state-of-the-art recognition performance on eight of them.
- Score: 77.87404524458809
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present PYSKL: an open-source toolbox for skeleton-based action
recognition based on PyTorch. The toolbox supports a wide variety of skeleton
action recognition algorithms, including approaches based on GCN and CNN. In
contrast to existing open-source skeleton action recognition projects that
include only one or two algorithms, PYSKL implements six different algorithms
under a unified framework with both the latest and original good practices to
ease the comparison of efficacy and efficiency. We also provide an original
GCN-based skeleton action recognition model named ST-GCN++, which achieves
competitive recognition performance without any complicated attention schemes,
serving as a strong baseline. Meanwhile, PYSKL supports the training and
testing of nine skeleton-based action recognition benchmarks and achieves
state-of-the-art recognition performance on eight of them. To facilitate future
research on skeleton action recognition, we also provide a large number of
trained models and detailed benchmark results to give some insights. PYSKL is
released at https://github.com/kennymckormick/pyskl and is actively maintained.
We will update this report when we add new features or benchmarks. The current
version corresponds to PYSKL v0.2.
Related papers
- Mask and Compress: Efficient Skeleton-based Action Recognition in Continual Learning [29.561972935011877]
We introduce CHARON (Continual Human Action Recognition On skeletoNs), which maintains consistent performance while operating within an efficient framework.
Our experiments on Split NTU-60 and the proposed Split NTU-120 datasets demonstrate that CHARON sets a new benchmark in this domain.
arXiv Detail & Related papers (2024-07-01T15:48:49Z) - Expressive Keypoints for Skeleton-based Action Recognition via Skeleton Transformation [14.033701085783177]
We propose Expressive Keypoints that incorporates hand and foot details to form a fine-grained skeletal representation, improving the discriminative ability for existing models in discerning intricate actions.
A plug-and-play Instance Pooling module is exploited to extend our approach to multi-person scenarios without surging computation costs.
arXiv Detail & Related papers (2024-06-26T01:48:56Z) - SkeleTR: Towrads Skeleton-based Action Recognition in the Wild [86.03082891242698]
SkeleTR is a new framework for skeleton-based action recognition.
It first models the intra-person skeleton dynamics for each skeleton sequence with graph convolutions.
It then uses stacked Transformer encoders to capture person interactions that are important for action recognition in general scenarios.
arXiv Detail & Related papers (2023-09-20T16:22:33Z) - FastPoseGait: A Toolbox and Benchmark for Efficient Pose-based Gait
Recognition [11.985433662623036]
FastPoseGait is an open-source toolbox for pose-based gait recognition based on PyTorch.
Our toolbox supports a set of cutting-edge pose-based gait recognition algorithms and a variety of related benchmarks.
arXiv Detail & Related papers (2023-09-02T02:05:58Z) - One-Shot Action Recognition via Multi-Scale Spatial-Temporal Skeleton
Matching [77.6989219290789]
One-shot skeleton action recognition aims to learn a skeleton action recognition model with a single training sample.
This paper presents a novel one-shot skeleton action recognition technique that handles skeleton action recognition via multi-scale spatial-temporal feature matching.
arXiv Detail & Related papers (2023-07-14T11:52:10Z) - Generative Action Description Prompts for Skeleton-based Action
Recognition [15.38417530693649]
We propose a Generative Action-description Prompts (GAP) approach for skeleton-based action recognition.
We employ a pre-trained large-scale language model as the knowledge engine to automatically generate text descriptions for body parts movements of actions.
Our proposed GAP method achieves noticeable improvements over various baseline models without extra cost at inference.
arXiv Detail & Related papers (2022-08-10T12:55:56Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - ANUBIS: Review and Benchmark Skeleton-Based Action Recognition Methods
with a New Dataset [26.581495230711198]
We present a review in the form of a taxonomy on existing works of skeleton-based action recognition.
To promote more fair and comprehensive evaluation, we collect ANUBIS, a large-scale human skeleton dataset.
arXiv Detail & Related papers (2022-05-04T14:03:43Z) - Joint-bone Fusion Graph Convolutional Network for Semi-supervised
Skeleton Action Recognition [65.78703941973183]
We propose a novel correlation-driven joint-bone fusion graph convolutional network (CD-JBF-GCN) as an encoder and use a pose prediction head as a decoder.
Specifically, the CD-JBF-GC can explore the motion transmission between the joint stream and the bone stream.
The pose prediction based auto-encoder in the self-supervised training stage allows the network to learn motion representation from unlabeled data.
arXiv Detail & Related papers (2022-02-08T16:03:15Z) - A Self-Supervised Gait Encoding Approach with Locality-Awareness for 3D
Skeleton Based Person Re-Identification [65.18004601366066]
Person re-identification (Re-ID) via gait features within 3D skeleton sequences is a newly-emerging topic with several advantages.
This paper proposes a self-supervised gait encoding approach that can leverage unlabeled skeleton data to learn gait representations for person Re-ID.
arXiv Detail & Related papers (2020-09-05T16:06:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.