MMA: A Momentum Mamba Architecture for Human Activity Recognition with Inertial Sensors
- URL: http://arxiv.org/abs/2511.21550v1
- Date: Wed, 26 Nov 2025 16:21:36 GMT
- Title: MMA: A Momentum Mamba Architecture for Human Activity Recognition with Inertial Sensors
- Authors: Thai-Khanh Nguyen, Uyen Vo, Tan M. Nguyen, Thieu N. Vo, Trung-Hieu Le, Cuong Pham,
- Abstract summary: Human activity recognition (HAR) from inertial sensors is essential for ubiquitous computing, mobile health, and ambient intelligence.<n>We introduce Momentum Mamba, a momentum-augmented SSM that incorporates second-order dynamics to improve stability of information flow across time steps.<n> Experiments on multiple HAR benchmarks demonstrate consistent gains over vanilla Mamba and Transformer baselines in accuracy, robustness, and convergence speed.
- Score: 6.7394387525196215
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Human activity recognition (HAR) from inertial sensors is essential for ubiquitous computing, mobile health, and ambient intelligence. Conventional deep models such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and transformers have advanced HAR but remain limited by vanishing or exloding gradients, high computational cost, and difficulty in capturing long-range dependencies. Structured state-space models (SSMs) like Mamba address these challenges with linear complexity and effective temporal modeling, yet they are restricted to first-order dynamics without stable longterm memory mechanisms. We introduce Momentum Mamba, a momentum-augmented SSM that incorporates second-order dynamics to improve stability of information flow across time steps, robustness, and long-sequence modeling. Two extensions further expand its capacity: Complex Momentum Mamba for frequency-selective memory scaling. Experiments on multiple HAR benchmarks demonstrate consistent gains over vanilla Mamba and Transformer baselines in accuracy, robustness, and convergence speed. With only moderate increases in training cost, momentum-augmented SSMs offer a favorable accuracy-efficiency balance, establishing them as a scalable paradigm for HAR and a promising principal framework for broader sequence modeling applications.
Related papers
- Scalable Spatio-Temporal SE(3) Diffusion for Long-Horizon Protein Dynamics [51.85385061275941]
Molecular dynamics (MD) simulations remain the gold standard for studying protein dynamics.<n>Recent generative models have shown promise in accelerating simulations, yet they struggle with long-horizon generation.<n>We present STAR-MD, a scalable diffusion model that generates physically plausible protein trajectories over micro-scale timescales.
arXiv Detail & Related papers (2026-02-02T14:13:28Z) - DeMa: Dual-Path Delay-Aware Mamba for Efficient Multivariate Time Series Analysis [22.768341734517815]
Transformer-based models suffer from computational complexity and high memory overhead.<n>Mamba has emerged as a promising linear-time alternative with high expressiveness.<n>DeMa is a dual-path delay-aware Mamba backbone.
arXiv Detail & Related papers (2026-01-09T04:54:56Z) - MODE: Efficient Time Series Prediction with Mamba Enhanced by Low-Rank Neural ODEs [41.50535363508025]
Time series prediction plays a pivotal role across diverse domains such as finance, healthcare, energy systems, and environmental modeling.<n>We propose MODE, a unified framework that integrates Low-Rank Neural Ordinary Differential Equations (Neural ODEs) with an Enhanced Mamba architecture.<n>Our contributions include: (1) a unified and efficient architecture for long-term time series modeling, (2) integration of Mamba's selective scanning with low-rank Neural ODEs for enhanced temporal representation, and (3) substantial improvements in efficiency and scalability enabled by low-rank approximation and dynamic selective scanning.
arXiv Detail & Related papers (2026-01-01T11:23:20Z) - MS-SSM: A Multi-Scale State Space Model for Efficient Sequence Modeling [60.648359990090846]
State-space models (SSMs) have recently attention as an efficient alternative to computationally expensive attention-based models for sequence modeling.<n>This paper introduces a multi-scale SSM framework that represents sequence dynamics across multiple resolution and processing each resolution with specialized state-space dynamics.
arXiv Detail & Related papers (2025-12-29T19:36:28Z) - TSkel-Mamba: Temporal Dynamic Modeling via State Space Model for Human Skeleton-based Action Recognition [59.99922360648663]
TSkel-Mamba is a hybrid Transformer-Mamba framework that effectively captures both spatial and temporal dynamics.<n>The MTI module employs multi-scale Cycle operators to capture cross-channel temporal interactions, a critical factor in action recognition.
arXiv Detail & Related papers (2025-12-12T11:55:16Z) - SAMBA: Toward a Long-Context EEG Foundation Model via Spatial Embedding and Differential Mamba [2.1232375739287006]
Long-sequence electroencephalogram (EEG) modeling is essential for developing generalizable EEG representation models.<n>We propose textitSAMBA, a self-supervised learning framework with a Mamba-based U-shaped encoder-decoder architecture.<n>We show that SAMBA consistently outperforms state-of-the-art methods while maintaining low memory consumption and inference time.
arXiv Detail & Related papers (2025-11-23T18:31:18Z) - The Curious Case of In-Training Compression of State Space Models [49.819321766705514]
State Space Models (SSMs) tackle long sequence modeling tasks efficiently, offer both parallelizable training and fast inference.<n>Key design challenge is striking the right balance between maximizing expressivity and limiting this computational burden.<n>Our approach, textscCompreSSM, applies to Linear Time-Invariant SSMs such as Linear Recurrent Units, but is also extendable to selective models.
arXiv Detail & Related papers (2025-10-03T09:02:33Z) - Quantum-Optimized Selective State Space Model for Efficient Time Series Prediction [39.146761527401424]
We propose a hybrid quantum-optimized approach that integrates state space dynamics with a variational quantum gate.<n>We empirically validate Q-SSM on three widely used benchmarks, i.e., ETT, Traffic, and Exchange Rate.<n>Results show that Q-SSM consistently improves over strong baselines.
arXiv Detail & Related papers (2025-08-29T22:00:48Z) - Routing Mamba: Scaling State Space Models with Mixture-of-Experts Projection [88.47928738482719]
Linear State Space Models (SSMs) offer remarkable performance gains in sequence modeling.<n>Recent advances, such as Mamba, further enhance SSMs with input-dependent gating and hardware-aware implementations.<n>We introduce Routing Mamba (RoM), a novel approach that scales SSM parameters using sparse mixtures of linear projection experts.
arXiv Detail & Related papers (2025-06-22T19:26:55Z) - Structured State Space Model Dynamics and Parametrization for Spiking Neural Networks [0.8321953606016751]
Multi-state spiking neurons offer compelling alternatives to conventional deep learning models.<n>State space models (SSMs) excel in long sequence processing using linear state-intrinsic recurrence resembling spiking neurons' subthreshold regime.<n>Here, we establish a mathematical bridge between SSMs and second-order spiking neuron models.
arXiv Detail & Related papers (2025-06-04T13:54:02Z) - Learning to Dissipate Energy in Oscillatory State-Space Models [51.98491034847041]
State-space models (SSMs) are a class of networks for sequence learning.<n>We show that D-LinOSS consistently outperforms previous LinOSS methods on long-range learning tasks.
arXiv Detail & Related papers (2025-05-17T23:15:17Z) - Physics-informed Multiple-Input Operators for efficient dynamic response prediction of structures [0.07916635054977067]
MIONet predicts structural responses continuously over both space and time.<n>Model is validated on both a simple beam and the KW-51 bridge, achieving FEM level accuracy within seconds.
arXiv Detail & Related papers (2025-05-11T18:45:58Z) - MambaVT: Spatio-Temporal Contextual Modeling for robust RGB-T Tracking [51.28485682954006]
We propose a pure Mamba-based framework (MambaVT) to fully exploit intrinsic-temporal contextual modeling for robust visible-thermal tracking.
Specifically, we devise the long-range cross-frame integration component to globally adapt to target appearance variations.
Experiments show the significant potential of vision Mamba for RGB-T tracking, with MambaVT achieving state-of-the-art performance on four mainstream benchmarks.
arXiv Detail & Related papers (2024-08-15T02:29:00Z) - Longhorn: State Space Models are Amortized Online Learners [51.10124201221601]
State-space models (SSMs) offer linear decoding efficiency while maintaining parallelism during training.
In this work, we explore SSM design through the lens of online learning, conceptualizing SSMs as meta-modules for specific online learning problems.
We introduce a novel deep SSM architecture, Longhorn, whose update resembles the closed-form solution for solving the online associative recall problem.
arXiv Detail & Related papers (2024-07-19T11:12:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.