DC-Mamber: A Dual Channel Prediction Model based on Mamba and Linear Transformer for Multivariate Time Series Forecasting

📅 2025-07-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of jointly modeling local temporal patterns and global inter-variable dependencies in multivariate time series forecasting (MTSF), this paper proposes DC-Mamba, a dual-channel forecasting framework. It innovatively integrates channel-independent Mamba—capable of capturing long-range univariate temporal dynamics—with channel-mixed linear Transformer—efficiently modeling cross-variable global dependencies—via a dual-encoder architecture and a collaborative embedding-fusion mechanism. Unlike standard Transformers (high computational complexity, weak local sensitivity) and vanilla Mamba (inherently limited in parallel global context modeling), DC-Mamba retains linear time complexity while substantially enhancing both local pattern sensitivity and global contextual aggregation. Extensive experiments on eight public benchmarks demonstrate that DC-Mamba achieves state-of-the-art prediction accuracy, outperforming leading methods across diverse domains. Moreover, it exhibits superior computational efficiency and strong generalization capability.

Technology Category

Application Category

📝 Abstract
In multivariate time series forecasting (MTSF), existing strategies for processing sequences are typically categorized as channel-independent and channel-mixing. The former treats all temporal information of each variable as a token, focusing on capturing local temporal features of individual variables, while the latter constructs a token from the multivariate information at each time step, emphasizing the modeling of global temporal dependencies. Current mainstream models are mostly based on Transformer and the emerging Mamba. Transformers excel at modeling global dependencies through self-attention mechanisms but exhibit limited sensitivity to local temporal patterns and suffer from quadratic computational complexity, restricting their efficiency in long-sequence processing. In contrast, Mamba, based on state space models (SSMs), achieves linear complexity and efficient long-range modeling but struggles to aggregate global contextual information in parallel. To overcome the limitations of both models, we propose DC-Mamber, a dual-channel forecasting model based on Mamba and linear Transformer for time series forecasting. Specifically, the Mamba-based channel employs a channel-independent strategy to extract intra-variable features, while the Transformer-based channel adopts a channel-mixing strategy to model cross-timestep global dependencies. DC-Mamber first maps the raw input into two distinct feature representations via separate embedding layers. These representations are then processed by a variable encoder (built on Mamba) and a temporal encoder (built on linear Transformer), respectively. Finally, a fusion layer integrates the dual-channel features for prediction. Extensive experiments on eight public datasets confirm DC-Mamber's superior accuracy over existing models.
Problem

Research questions and friction points this paper is trying to address.

Overcoming limitations in multivariate time series forecasting models
Balancing local and global temporal feature extraction
Improving efficiency and accuracy in long-sequence processing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-channel model combines Mamba and Transformer
Mamba extracts intra-variable temporal features
Linear Transformer models cross-timestep dependencies
🔎 Similar Papers
No similar papers found.
B
Bing Fan
School of Computer Science and Technology, University of Science and Technology of China, Hefei, Anhui, China
S
Shusen Ma
Institute of Advanced Technology, University of Science and Technology of China, Hefei, Anhui, China
Yun-Bo Zhao
Yun-Bo Zhao
University of Science and Technology of China
Human-Machine SystemsSmart ManufacturingNetworked Control Systems
Y
Yu Kang
Institute of Advanced Technology, University of Science and Technology of China, Hefei, Anhui, China; Department of Automation, University of Science and Technology of China, Hefei, Anhui, China; Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Anhui, China