ms-Mamba: Multi-scale Mamba for Time-Series Forecasting

πŸ“… 2025-04-10
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing time-series forecasting models are largely constrained to a single temporal resolution, limiting their ability to capture cross-scale dynamic patterns. To address this, we propose MS-Mambaβ€”the first Mamba-based architecture explicitly designed for multi-scale forecasting. Our approach innovatively integrates multi-scale modeling into the Mamba framework via heterogeneous Ξ”-sampling-rate parallel Mamba blocks that separately capture fast- and slow-varying dynamics; multi-branch temporal downsampling; cross-scale feature fusion; and an end-to-end learnable scheduling mechanism. Evaluated on multiple standard time-series forecasting benchmarks, MS-Mamba consistently outperforms state-of-the-art Transformer-based models and single-scale Mamba variants, achieving an average 7.2% reduction in MAE. These results empirically validate the effectiveness and superiority of multi-scale state-space modeling for time-series forecasting.

Technology Category

Application Category

πŸ“ Abstract
The problem of Time-series Forecasting is generally addressed by recurrent, Transformer-based and the recently proposed Mamba-based architectures. However, existing architectures generally process their input at a single temporal scale, which may be sub-optimal for many tasks where information changes over multiple time scales. In this paper, we introduce a novel architecture called Multi-scale Mamba (ms-Mamba) to address this gap. ms-Mamba incorporates multiple temporal scales by using multiple Mamba blocks with different sampling rates ($Delta$s). Our experiments on many benchmarks demonstrate that ms-Mamba outperforms state-of-the-art approaches, including the recently proposed Transformer-based and Mamba-based models.
Problem

Research questions and friction points this paper is trying to address.

Addresses single-scale limitation in time-series forecasting models
Proposes multi-scale Mamba for varying temporal information
Outperforms existing Transformer and Mamba-based approaches
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-scale Mamba with varying sampling rates
Multiple Mamba blocks for different temporal scales
Outperforms Transformer and Mamba-based models
πŸ”Ž Similar Papers
No similar papers found.
Y
Yusuf Meric Karadag
Dept. of Computer Eng. and ROMER Robotics Center, Middle East Technical University
Sinan Kalkan
Sinan Kalkan
Dept. of Computer Eng., Middle East Technical University
Computer VisionDeep LearningRobotics
I
I. Dino
Dept. of Computer Eng. and ROMER Robotics Center, Middle East Technical University