🤖 AI Summary
This work addresses multivariate time series forecasting and missing value imputation. We propose TSRM, a lightweight temporal representation model featuring a novel hybrid architecture that synergistically integrates a multi-branch CNN encoder with Transformer-style self-attention mechanisms. TSRM employs a shared spatiotemporal representation layer for joint modeling, enabling unified support for both forecasting and imputation tasks. Its design emphasizes efficient feature aggregation and end-to-end multi-task training to balance expressive power and computational efficiency. Evaluated on seven mainstream benchmarks, TSRM consistently outperforms state-of-the-art methods: it achieves significant improvements in both forecasting accuracy and imputation quality, reduces average parameter count by 37%, and accelerates inference speed by 2.1×. These results validate the effectiveness and generalizability of TSRM’s lightweight architecture.
📝 Abstract
We introduce a temporal feature encoding architecture called Time Series Representation Model (TSRM) for multivariate time series forecasting and imputation. The architecture is structured around CNN-based representation layers, each dedicated to an independent representation learning task and designed to capture diverse temporal patterns, followed by an attention-based feature extraction layer and a merge layer, designed to aggregate extracted features. The architecture is fundamentally based on a configuration that is inspired by a Transformer encoder, with self-attention mechanisms at its core. The TSRM architecture outperforms state-of-the-art approaches on most of the seven established benchmark datasets considered in our empirical evaluation for both forecasting and imputation tasks. At the same time, it significantly reduces complexity in the form of learnable parameters. The source code is available at https://github.com/RobertLeppich/TSRM.