MSDformer: Multi-scale Discrete Transformer For Time Series Generation
Chen, Zhicheng, Feng, Shibo, Xiao, Xi, Zhang, Zhong, Li, Qing, Gao, Xingyu, Zhao, Peilin
–arXiv.org Artificial Intelligence
Discrete Token Modeling (DTM), which employs vector quantization techniques, has demonstrated remarkable success in modeling non-natural language modalities, particularly in time series generation. While our prior work SDformer established the first DTM-based framework to achieve state-of-the-art performance in this domain, two critical limitations persist in existing DTM approaches: 1) their inability to capture multi-scale temporal patterns inherent to complex time series data, and 2) the absence of theoretical foundations to guide model optimization. To address these challenges, we proposes a novel multi-scale DTM-based time series generation method, called Multi-Scale Discrete Transformer (MSDformer). MSDformer employs a multi-scale time series tokenizer to learn discrete token representations at multiple scales, which jointly characterize the complex nature of time series data. Subsequently, MSDformer applies a multi-scale autoregressive token modeling technique to capture the multi-scale patterns of time series within the discrete latent space. Theoretically, we validate the effectiveness of the DTM method and the rationality of MSDformer through the rate-distortion theorem. Comprehensive experiments demonstrate that MSDformer significantly outperforms state-of-the-art methods. Both theoretical analysis and experimental results demonstrate that incorporating multi-scale information and modeling multi-scale patterns can substantially enhance the quality of generated time series in DTM-based approaches. The code will be released upon acceptance.
arXiv.org Artificial Intelligence
May-21-2025
- Country:
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Health & Medicine (0.93)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (1.00)
- Statistical Learning (0.93)
- Natural Language > Large Language Model (0.93)
- Representation & Reasoning (1.00)
- Vision (0.94)
- Machine Learning
- Data Science > Data Mining (1.00)
- Artificial Intelligence
- Information Technology