Balanced Representation Learning for Long-tailed Skeleton-based Action
Recognition
- URL: http://arxiv.org/abs/2308.14024v1
- Date: Sun, 27 Aug 2023 07:25:51 GMT
- Title: Balanced Representation Learning for Long-tailed Skeleton-based Action
Recognition
- Authors: Hongda Liu, Yunlong Wang, Min Ren, Junxing Hu, Zhengquan Luo, Guangqi
Hou, Zhenan Sun
- Abstract summary: We propose a novel balanced representation learning method to address the long-tailed problem in action recognition.
We design a detached action-aware learning schedule to further mitigate the bias in the representation space.
The proposed method is validated on four skeleton datasets.
- Score: 33.212205680007756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Skeleton-based action recognition has recently made significant progress.
However, data imbalance is still a great challenge in real-world scenarios. The
performance of current action recognition algorithms declines sharply when
training data suffers from heavy class imbalance. The imbalanced data actually
degrades the representations learned by these methods and becomes the
bottleneck for action recognition. How to learn unbiased representations from
imbalanced action data is the key to long-tailed action recognition. In this
paper, we propose a novel balanced representation learning method to address
the long-tailed problem in action recognition. Firstly, a spatial-temporal
action exploration strategy is presented to expand the sample space
effectively, generating more valuable samples in a rebalanced manner. Secondly,
we design a detached action-aware learning schedule to further mitigate the
bias in the representation space. The schedule detaches the representation
learning of tail classes from training and proposes an action-aware loss to
impose more effective constraints. Additionally, a skip-modal representation is
proposed to provide complementary structural information. The proposed method
is validated on four skeleton datasets, NTU RGB+D 60, NTU RGB+D 120, NW-UCLA,
and Kinetics. It not only achieves consistently large improvement compared to
the state-of-the-art (SOTA) methods, but also demonstrates a superior
generalization capacity through extensive experiments. Our code is available at
https://github.com/firework8/BRL.
Related papers
- ReL-SAR: Representation Learning for Skeleton Action Recognition with Convolutional Transformers and BYOL [6.603505460200282]
Unsupervised representation learning is of prime importance to leverage unlabeled skeleton data.
We design a lightweight convolutional transformer framework, named ReL-SAR, for jointly modeling spatial and temporal cues in skeleton sequences.
We capitalize on Bootstrap Your Own Latent (BYOL) to learn robust representations from unlabeled skeleton sequence data.
arXiv Detail & Related papers (2024-09-09T16:03:26Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Scaling Laws Beyond Backpropagation [64.0476282000118]
We study the ability of Direct Feedback Alignment to train causal decoder-only Transformers efficiently.
We find that DFA fails to offer more efficient scaling than backpropagation.
arXiv Detail & Related papers (2022-10-26T10:09:14Z) - Moderately-Balanced Representation Learning for Treatment Effects with
Orthogonality Information [14.040918087553177]
Estimating the average treatment effect (ATE) from observational data is challenging due to selection bias.
We propose a moderately-balanced representation learning framework.
This framework protects the representation from being over-balanced via multi-task learning.
arXiv Detail & Related papers (2022-09-05T13:20:12Z) - Constructing Balance from Imbalance for Long-tailed Image Recognition [50.6210415377178]
The imbalance between majority (head) classes and minority (tail) classes severely skews the data-driven deep neural networks.
Previous methods tackle with data imbalance from the viewpoints of data distribution, feature space, and model design.
We propose a concise paradigm by progressively adjusting label space and dividing the head classes and tail classes.
Our proposed model also provides a feature evaluation method and paves the way for long-tailed feature learning.
arXiv Detail & Related papers (2022-08-04T10:22:24Z) - Efficient Augmentation for Imbalanced Deep Learning [8.38844520504124]
We study a convolutional neural network's internal representation of imbalanced image data.
We measure the generalization gap between a model's feature embeddings in the training and test sets, showing that the gap is wider for minority classes.
This insight enables us to design an efficient three-phase CNN training framework for imbalanced data.
arXiv Detail & Related papers (2022-07-13T09:43:17Z) - Value-Consistent Representation Learning for Data-Efficient
Reinforcement Learning [105.70602423944148]
We propose a novel method, called value-consistent representation learning (VCR), to learn representations that are directly related to decision-making.
Instead of aligning this imagined state with a real state returned by the environment, VCR applies a $Q$-value head on both states and obtains two distributions of action values.
It has been demonstrated that our methods achieve new state-of-the-art performance for search-free RL algorithms.
arXiv Detail & Related papers (2022-06-25T03:02:25Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Exploratory State Representation Learning [63.942632088208505]
We propose a new approach called XSRL (eXploratory State Representation Learning) to solve the problems of exploration and SRL in parallel.
On one hand, it jointly learns compact state representations and a state transition estimator which is used to remove unexploitable information from the representations.
On the other hand, it continuously trains an inverse model, and adds to the prediction error of this model a $k$-step learning progress bonus to form the objective of a discovery policy.
arXiv Detail & Related papers (2021-09-28T10:11:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.