Damba-ST: Domain-Adaptive Mamba for Efficient Urban Spatio-Temporal Prediction
- URL: http://arxiv.org/abs/2506.18939v2
- Date: Thu, 10 Jul 2025 07:42:46 GMT
- Title: Damba-ST: Domain-Adaptive Mamba for Efficient Urban Spatio-Temporal Prediction
- Authors: Rui An, Yifeng Zhang, Ziran Liang, Wenqi Fan, Yuxuan Liang, Xuequn Shang, Qing Li,
- Abstract summary: We propose Damba-ST, a domain-adaptive Mamba-based model for efficient urban-temporal prediction.<n>Damba-ST retains Mamba's linear complexity while significantly enhancing complexity to heterogeneous domains.<n>It achieves state-of-the-art performance on prediction tasks and demonstrates strong zero-shot generalization.
- Score: 27.924276998605816
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Training urban spatio-temporal foundation models that generalize well across diverse regions and cities is critical for deploying urban services in unseen or data-scarce regions. Recent studies have typically focused on fusing cross-domain spatio-temporal data to train unified Transformer-based models. However, these models suffer from quadratic computational complexity and high memory overhead, limiting their scalability and practical deployment. Inspired by the efficiency of Mamba, a state space model with linear time complexity, we explore its potential for efficient urban spatio-temporal prediction. However, directly applying Mamba as a spatio-temporal backbone leads to negative transfer and severe performance degradation. This is primarily due to spatio-temporal heterogeneity and the recursive mechanism of Mamba's hidden state updates, which limit cross-domain generalization. To overcome these challenges, we propose Damba-ST, a novel domain-adaptive Mamba-based model for efficient urban spatio-temporal prediction. Damba-ST retains Mamba's linear complexity advantage while significantly enhancing its adaptability to heterogeneous domains. Specifically, we introduce two core innovations: (1) a domain-adaptive state space model that partitions the latent representation space into a shared subspace for learning cross-domain commonalities and independent, domain-specific subspaces for capturing intra-domain discriminative features; (2) three distinct Domain Adapters, which serve as domain-aware proxies to bridge disparate domain distributions and facilitate the alignment of cross-domain commonalities. Extensive experiments demonstrate the generalization and efficiency of Damba-ST. It achieves state-of-the-art performance on prediction tasks and demonstrates strong zero-shot generalization, enabling seamless deployment in new urban environments without extensive retraining or fine-tuning.
Related papers
- InceptionMamba: An Efficient Hybrid Network with Large Band Convolution and Bottleneck Mamba [21.47782205082816]
InceptionNeXt has shown excellent competitiveness in image classification and a number of downstream tasks.<n>Built on parallel one-dimensional strip convolutions, InceptionNeXt suffers from limited ability of capturing spatial dependencies along different dimensions.<n>We propose a novel backbone architecture termed InceptionMamba to overcome these limitations.
arXiv Detail & Related papers (2025-06-10T12:31:05Z) - HMamba: Hyperbolic Mamba for Sequential Recommendation [39.60869234694072]
Hyperbolic Mamba is a novel architecture that unifies the efficiency of Mamba's selective state space mechanism with hyperbolic geometry's hierarchical representational power.<n>We show that Hyperbolic Mamba achieves 3-11% improvement while retaining Mamba's linear-time efficiency, enabling real-world deployment.
arXiv Detail & Related papers (2025-05-14T07:34:36Z) - EventMamba: Enhancing Spatio-Temporal Locality with State Space Models for Event-Based Video Reconstruction [66.84997711357101]
EventMamba is a specialized model designed for event-based video reconstruction tasks.<n>We show that EventMamba markedly improves speed while delivering superior visual quality compared to Transformer-based methods.
arXiv Detail & Related papers (2025-03-25T14:46:45Z) - DA-Mamba: Domain Adaptive Hybrid Mamba-Transformer Based One-Stage Object Detection [0.3683202928838613]
We present the first domain-adaptive Mamba-based one-stage object detection model, DA-Mamba.<n>Inspired by the global modeling and linear complexity of the Mamba architecture, we present the first domain-adaptive Mamba-based one-stage object detection model, DA-Mamba.
arXiv Detail & Related papers (2025-02-16T15:58:54Z) - OccMamba: Semantic Occupancy Prediction with State Space Models [24.697645636701797]
OccMamba is the first Mamba-based network for semantic occupancy prediction.<n>Inspired by the global modeling and linear complexity of the Mamba architecture, we present the first OccMamba network for semantic occupancy prediction.
arXiv Detail & Related papers (2024-08-19T10:07:00Z) - MambaVT: Spatio-Temporal Contextual Modeling for robust RGB-T Tracking [51.28485682954006]
We propose a pure Mamba-based framework (MambaVT) to fully exploit intrinsic-temporal contextual modeling for robust visible-thermal tracking.
Specifically, we devise the long-range cross-frame integration component to globally adapt to target appearance variations.
Experiments show the significant potential of vision Mamba for RGB-T tracking, with MambaVT achieving state-of-the-art performance on four mainstream benchmarks.
arXiv Detail & Related papers (2024-08-15T02:29:00Z) - DGMamba: Domain Generalization via Generalized State Space Model [80.82253601531164]
Domain generalization(DG) aims at solving distribution shift problems in various scenes.
Mamba, as an emerging state space model (SSM), possesses superior linear complexity and global receptive fields.
We propose a novel framework for DG, named DGMamba, that excels in strong generalizability toward unseen domains.
arXiv Detail & Related papers (2024-04-11T14:35:59Z) - Memory-Efficient Prompt Tuning for Incremental Histopathology
Classification [69.46798702300042]
We present a memory-efficient prompt tuning framework to cultivate model generalization potential in economical memory cost.
We have extensively evaluated our framework with two histopathology tasks, i.e., breast cancer metastasis classification and epithelium-stroma tissue classification.
arXiv Detail & Related papers (2024-01-22T03:24:45Z) - Domain-incremental Cardiac Image Segmentation with Style-oriented Replay
and Domain-sensitive Feature Whitening [67.6394526631557]
M&Ms should incrementally learn from each incoming dataset and progressively update with improved functionality as time goes by.
In medical scenarios, this is particularly challenging as accessing or storing past data is commonly not allowed due to data privacy.
We propose a novel domain-incremental learning framework to recover past domain inputs first and then regularly replay them during model optimization.
arXiv Detail & Related papers (2022-11-09T13:07:36Z) - Normalization Perturbation: A Simple Domain Generalization Method for
Real-World Domain Shifts [133.99270341855728]
Real-world domain styles can vary substantially due to environment changes and sensor noises.
Deep models only know the training domain style.
We propose Normalization Perturbation to overcome this domain style overfitting problem.
arXiv Detail & Related papers (2022-11-08T17:36:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.