Moirai 2.0: When Less Is More for Time Series Forecasting
Liu, Chenghao, Aksu, Taha, Liu, Juncheng, Liu, Xu, Yan, Hanshu, Pham, Quang, Savarese, Silvio, Sahoo, Doyen, Xiong, Caiming, Li, Junnan
–arXiv.org Artificial Intelligence
We introduce Moirai 2.0, a decoder-only time-series foundation model trained on a new corpus of 36M series. The model adopts quantile forecasting and multi-token prediction, improving both probabilistic accuracy and inference efficiency. On the Gift-Eval benchmark, it ranks among the top pretrained models while achieving a strong trade-off between accuracy, speed, and model size. Compared to Moirai 1.0, Moirai 2.0 replaces masked-encoder training, multi-patch inputs, and mixture-distribution outputs with a simpler decoder-only architecture, single patch, and quantile loss. Ablation studies isolate these changes -- showing that the decoder-only backbone along with recursive multi-quantile decoding contribute most to the gains. Additional experiments show that Moirai 2.0 outperforms larger models from the same family and exhibits robust domain-level results. In terms of efficiency and model size, Moirai 2.0 is twice as fast and thirty times smaller than its prior best version, Moirai 1.0-Large, while also performing better. Model performance plateaus with increasing parameter count and declines at longer horizons, motivating future work on data scaling and long-horizon modeling. We release code and evaluation details to support further research.
arXiv.org Artificial Intelligence
Nov-25-2025
- Country:
- North America (0.28)
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Information Technology (0.48)
- Technology: