Skeleton Cloud Colorization for Unsupervised 3D Action Representation
Learning
- URL: http://arxiv.org/abs/2108.01959v2
- Date: Thu, 5 Aug 2021 04:48:58 GMT
- Title: Skeleton Cloud Colorization for Unsupervised 3D Action Representation
Learning
- Authors: Siyuan Yang, Jun Liu, Shijian Lu, Meng Hwa Er, Alex C. Kot
- Abstract summary: Skeleton-based human action recognition has attracted increasing attention in recent years.
We design a novel skeleton cloud colorization technique that is capable of learning skeleton representations from unlabeled skeleton sequence data.
We show that the proposed method outperforms existing unsupervised and semi-supervised 3D action recognition methods by large margins.
- Score: 65.88887113157627
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Skeleton-based human action recognition has attracted increasing attention in
recent years. However, most of the existing works focus on supervised learning
which requiring a large number of annotated action sequences that are often
expensive to collect. We investigate unsupervised representation learning for
skeleton action recognition, and design a novel skeleton cloud colorization
technique that is capable of learning skeleton representations from unlabeled
skeleton sequence data. Specifically, we represent a skeleton action sequence
as a 3D skeleton cloud and colorize each point in the cloud according to its
temporal and spatial orders in the original (unannotated) skeleton sequence.
Leveraging the colorized skeleton point cloud, we design an auto-encoder
framework that can learn spatial-temporal features from the artificial color
labels of skeleton joints effectively. We evaluate our skeleton cloud
colorization approach with action classifiers trained under different
configurations, including unsupervised, semi-supervised and fully-supervised
settings. Extensive experiments on NTU RGB+D and NW-UCLA datasets show that the
proposed method outperforms existing unsupervised and semi-supervised 3D action
recognition methods by large margins, and it achieves competitive performance
in supervised 3D action recognition as well.
Related papers
- SkeletonMAE: Graph-based Masked Autoencoder for Skeleton Sequence
Pre-training [110.55093254677638]
We propose an efficient skeleton sequence learning framework, named Skeleton Sequence Learning (SSL)
In this paper, we build an asymmetric graph-based encoder-decoder pre-training architecture named SkeletonMAE.
Our SSL generalizes well across different datasets and outperforms the state-of-the-art self-supervised skeleton-based action recognition methods.
arXiv Detail & Related papers (2023-07-17T13:33:11Z) - One-Shot Action Recognition via Multi-Scale Spatial-Temporal Skeleton
Matching [77.6989219290789]
One-shot skeleton action recognition aims to learn a skeleton action recognition model with a single training sample.
This paper presents a novel one-shot skeleton action recognition technique that handles skeleton action recognition via multi-scale spatial-temporal feature matching.
arXiv Detail & Related papers (2023-07-14T11:52:10Z) - Self-Supervised 3D Action Representation Learning with Skeleton Cloud
Colorization [75.0912240667375]
3D Skeleton-based human action recognition has attracted increasing attention in recent years.
Most of the existing work focuses on supervised learning which requires a large number of labeled action sequences.
In this paper, we address self-supervised 3D action representation learning for skeleton-based action recognition.
arXiv Detail & Related papers (2023-04-18T08:03:26Z) - Contrastive Self-Supervised Learning for Skeleton Representations [2.528877542605869]
We use a contrastive self-supervised learning method, SimCLR, to learn representations that capture the semantics of skeleton point clouds.
To pre-train the representations, we normalise six existing datasets to obtain more than 40 million skeleton frames.
We evaluate the quality of the learned representations with three downstream tasks: skeleton reconstruction, motion prediction, and activity classification.
arXiv Detail & Related papers (2022-11-10T02:45:36Z) - SkeletonMAE: Spatial-Temporal Masked Autoencoders for Self-supervised
Skeleton Action Recognition [13.283178393519234]
Self-supervised skeleton-based action recognition has attracted more attention.
With utilizing the unlabeled data, more generalizable features can be learned to alleviate the overfitting problem.
We propose a spatial-temporal masked autoencoder framework for self-supervised 3D skeleton-based action recognition.
arXiv Detail & Related papers (2022-09-01T20:54:27Z) - Skeleton-Contrastive 3D Action Representation Learning [35.06361753065124]
This paper strives for self-supervised learning of a feature space suitable for skeleton-based action recognition.
Our approach achieves state-of-the-art performance for self-supervised learning from skeleton data on the challenging PKU and NTU datasets.
arXiv Detail & Related papers (2021-08-08T14:44:59Z) - Revisiting Skeleton-based Action Recognition [107.08112310075114]
PoseC3D is a new approach to skeleton-based action recognition, which relies on a 3D heatmap instead stack a graph sequence as the base representation of human skeletons.
On four challenging datasets, PoseC3D consistently obtains superior performance, when used alone on skeletons and in combination with the RGB modality.
arXiv Detail & Related papers (2021-04-28T06:32:17Z) - A Self-Supervised Gait Encoding Approach with Locality-Awareness for 3D
Skeleton Based Person Re-Identification [65.18004601366066]
Person re-identification (Re-ID) via gait features within 3D skeleton sequences is a newly-emerging topic with several advantages.
This paper proposes a self-supervised gait encoding approach that can leverage unlabeled skeleton data to learn gait representations for person Re-ID.
arXiv Detail & Related papers (2020-09-05T16:06:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.