Probabilistic Transformer for Time Series Analysis David S. Matteson Department of Statistics and Data Science Department of Statistics and Data Science Cornell University
–Neural Information Processing Systems
Generative modeling of multivariate time series has remained challenging partly due to the complex, non-deterministic dynamics across long-distance time steps. In this paper, we propose deep probabilistic methods that combine state-space models (SSMs) with transformer architectures. In contrast to previously proposed SSMs, our approaches use attention mechanism to model non-Markovian dynamics in the latent space and avoid recurrent neural networks entirely. We also extend our models to include several layers of stochastic variables organized in a hierarchy for further expressiveness. Compared to transformer models, ours are probabilistic, non-autoregressive, and capable of generating diverse long-term forecasts with accounted uncertainty. Extensive experiments show that our models consistently outperform competitive baselines on various tasks and datasets, including time series forecasting and human motion prediction.
Neural Information Processing Systems
Feb-10-2025, 16:44:12 GMT
- Genre:
- Research Report > New Finding (0.68)