SMA-STN: Segmented Movement-Attending Spatiotemporal Network
forMicro-Expression Recognition
- URL: http://arxiv.org/abs/2010.09342v1
- Date: Mon, 19 Oct 2020 09:23:24 GMT
- Title: SMA-STN: Segmented Movement-Attending Spatiotemporal Network
forMicro-Expression Recognition
- Authors: Jiateng Liu, Wenming Zheng, Yuan Zong
- Abstract summary: This paper proposes a segmented movement-attending network (SMA-STN) to reveal subtle movement changes visually in an efficient way.
Extensive experiments on three widely used benchmarks, i.e., CALoss II, SAMM, and SHIC, show that the proposed SMA-STN achieves better MER performance than other state-of-the-art methods.
- Score: 20.166205708651194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Correctly perceiving micro-expression is difficult since micro-expression is
an involuntary, repressed, and subtle facial expression, and efficiently
revealing the subtle movement changes and capturing the significant segments in
a micro-expression sequence is the key to micro-expression recognition (MER).
To handle the crucial issue, in this paper, we firstly propose a dynamic
segmented sparse imaging module (DSSI) to compute dynamic images as
local-global spatiotemporal descriptors under a unique sampling protocol, which
reveals the subtle movement changes visually in an efficient way. Secondly, a
segmented movement-attending spatiotemporal network (SMA-STN) is proposed to
further unveil imperceptible small movement changes, which utilizes a
spatiotemporal movement-attending module (STMA) to capture long-distance
spatial relation for facial expression and weigh temporal segments. Besides, a
deviation enhancement loss (DE-Loss) is embedded in the SMA-STN to enhance the
robustness of SMA-STN to subtle movement changes in feature level. Extensive
experiments on three widely used benchmarks, i.e., CASME II, SAMM, and SHIC,
show that the proposed SMA-STN achieves better MER performance than other
state-of-the-art methods, which proves that the proposed method is effective to
handle the challenging MER problem.
Related papers
- Unsupervised Representation Learning of Complex Time Series for Maneuverability State Identification in Smart Mobility [0.0]
In smart mobility, MTS plays a crucial role in providing temporal dynamics of behaviors such as maneuver patterns.
In this work, we aim to address challenges associated with modeling MTS data collected from a vehicle using sensors.
Our goal is to investigate the effectiveness of two distinct unsupervised representation learning approaches in identifying maneuvering states in smart mobility.
arXiv Detail & Related papers (2024-08-26T15:16:18Z) - Three-Stream Temporal-Shift Attention Network Based on Self-Knowledge Distillation for Micro-Expression Recognition [21.675660978188617]
Micro-expression recognition is crucial in many fields, including criminal analysis and psychotherapy.
A three-stream temporal-shift attention network based on self-knowledge distillation called SKD-TSTSAN is proposed in this paper.
arXiv Detail & Related papers (2024-06-25T13:22:22Z) - Adaptive Temporal Motion Guided Graph Convolution Network for Micro-expression Recognition [48.21696443824074]
We propose a novel framework for micro-expression recognition, named the Adaptive Temporal Motion Guided Graph Convolution Network (ATM-GCN)
Our framework excels at capturing temporal dependencies between frames across the entire clip, thereby enhancing micro-expression recognition at the clip level.
arXiv Detail & Related papers (2024-06-13T10:57:24Z) - Multi-Scale Spatio-Temporal Graph Convolutional Network for Facial Expression Spotting [11.978551396144532]
We propose a Multi-Scale Spatio-Temporal Graph Conal Network (SpoT-CN) for facial expression spotting.
We track both short- and long-term motion of facial muscles in compact sliding windows whose window length adapts to the temporal receptive field of the network.
This network learns both local and global features from multiple scales of facial graph structures using our proposed facial localvolution graph pooling (FLGP)
arXiv Detail & Related papers (2024-03-24T03:10:39Z) - Unleashing Network Potentials for Semantic Scene Completion [50.95486458217653]
This paper proposes a novel SSC framework - Adrial Modality Modulation Network (AMMNet)
AMMNet introduces two core modules: a cross-modal modulation enabling the interdependence of gradient flows between modalities, and a customized adversarial training scheme leveraging dynamic gradient competition.
Extensive experimental results demonstrate that AMMNet outperforms state-of-the-art SSC methods by a large margin.
arXiv Detail & Related papers (2024-03-12T11:48:49Z) - Spatio-Temporal Branching for Motion Prediction using Motion Increments [55.68088298632865]
Human motion prediction (HMP) has emerged as a popular research topic due to its diverse applications.
Traditional methods rely on hand-crafted features and machine learning techniques.
We propose a noveltemporal-temporal branching network using incremental information for HMP.
arXiv Detail & Related papers (2023-08-02T12:04:28Z) - Motion Gait: Gait Recognition via Motion Excitation [5.559482051571756]
We propose Motion Excitation Module (MEM) to guide-temporal features to focus on human parts with large dynamic changes.
MEM learns the difference information between frames and intervals, so as to obtain the representation of changes temporal motion changes.
We present the Fine Feature Extractor (EFF), which independently learns according to the spatial-temporal representations of human horizontal parts of individuals.
arXiv Detail & Related papers (2022-06-22T13:47:14Z) - ProgressiveMotionSeg: Mutually Reinforced Framework for Event-Based
Motion Segmentation [101.19290845597918]
This paper presents a Motion Estimation (ME) module and an Event Denoising (ED) module jointly optimized in a mutually reinforced manner.
Taking temporal correlation as guidance, ED module calculates the confidence that each event belongs to real activity events, and transmits it to ME module to update energy function of motion segmentation for noise suppression.
arXiv Detail & Related papers (2022-03-22T13:40:26Z) - Short and Long Range Relation Based Spatio-Temporal Transformer for
Micro-Expression Recognition [61.374467942519374]
We propose a novel a-temporal transformer architecture -- to the best of our knowledge, the first purely transformer based approach for micro-expression recognition.
The architecture comprises a spatial encoder which learns spatial patterns, a temporal dimension classification for temporal analysis, and a head.
A comprehensive evaluation on three widely used spontaneous micro-expression data sets, shows that the proposed approach consistently outperforms the state of the art.
arXiv Detail & Related papers (2021-12-10T22:10:31Z) - Unsupervised Motion Representation Learning with Capsule Autoencoders [54.81628825371412]
Motion Capsule Autoencoder (MCAE) models motion in a two-level hierarchy.
MCAE is evaluated on a novel Trajectory20 motion dataset and various real-world skeleton-based human action datasets.
arXiv Detail & Related papers (2021-10-01T16:52:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.