Quantized symbolic time series approximation

📅 2024-11-20
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high storage overhead and poor scalability of symbolic time-series approximation methods—particularly on high-performance hardware and with large language models (LLMs)—this paper proposes QABBA, the first symbolic representation framework integrating low-bit quantization into the ABBA (Adaptive Brownian Bridge Approximation) paradigm. Leveraging theoretical analysis, QABBA derives a tight upper bound on quantization error, enabling high-fidelity symbolic reconstruction without retraining downstream embeddings. The resulting discrete symbol sequences are directly consumable by LLMs for end-to-end time-series regression fine-tuning. Evaluated on the Monash Time Series Regression Benchmark, QABBA achieves new state-of-the-art performance while reducing storage requirements by orders of magnitude. Crucially, it preserves ABBA’s original reconstruction accuracy and computational efficiency. Extensive experiments across diverse domains confirm QABBA’s robustness and generalizability.

Technology Category

Application Category

📝 Abstract
Time series are ubiquitous in numerous science and engineering domains, e.g., signal processing, bioinformatics, and astronomy. Previous work has verified the efficacy of symbolic time series representation in a variety of engineering applications due to its storage efficiency and numerosity reduction. The most recent symbolic aggregate approximation technique, ABBA, has been shown to preserve essential shape information of time series and improve downstream applications, e.g., neural network inference regarding prediction and anomaly detection in time series. Motivated by the emergence of high-performance hardware which enables efficient computation for low bit-width representations, we present a new quantization-based ABBA symbolic approximation technique, QABBA, which exhibits improved storage efficiency while retaining the original speed and accuracy of symbolic reconstruction. We prove an upper bound for the error arising from quantization and discuss how the number of bits should be chosen to balance this with other errors. An application of QABBA with large language models (LLMs) for time series regression is also presented, and its utility is investigated. By representing the symbolic chain of patterns on time series, QABBA not only avoids the training of embedding from scratch, but also achieves a new state-of-the-art on Monash regression dataset. The symbolic approximation to the time series offers a more efficient way to fine-tune LLMs on the time series regression task which contains various application domains. We further present a set of extensive experiments performed across various well-established datasets to demonstrate the advantages of the QABBA method for symbolic approximation.
Problem

Research questions and friction points this paper is trying to address.

Develops QABBA for efficient symbolic time series approximation
Balances quantization error and storage efficiency in QABBA
Enhances LLM performance for time series regression tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Quantization-based ABBA symbolic approximation technique
Improved storage efficiency with original speed
Enhances LLMs for time series regression
🔎 Similar Papers
No similar papers found.
E
Erin Carson
Department of Numerical Mathematics, Charles University, Prague, Czech Republic
Xinye Chen
Xinye Chen
Sorbonne Université, CNRS, LIP6
Scientific ComputingMachine LearningAlgorithmHigh Performance ComputingMathematical Software
C
Cheng Kang
Department of Cybernetics, Czech Technical University in Prague, Prague, Czech Republic