UniCast: A Unified Multimodal Prompting Framework for Time Series Forecasting
Park, Sehyuk, Han, Soyeon Caren, Hovy, Eduard
–arXiv.org Artificial Intelligence
Time series forecasting is a foundational task across domains, such as finance, healthcare, and environmental monitoring. While recent advances in Time Series Foundation Models (TSFMs) have demonstrated strong generalisation through large-scale pretraining, existing models operate predominantly in a unimodal setting, ignoring the rich multimodal context, such as visual and textual signals, that often accompanies time series data in real-world scenarios. This paper introduces a novel parameter-efficient multimodal framework, UniCast, that extends TSFMs to jointly leverage time series, vision, and text modalities for enhanced forecasting performance. Our method integrates modality-specific embed-dings from pretrained Vision and Text Encoders with a frozen TSFM via soft prompt tuning, enabling efficient adaptation with minimal parameter updates. This design not only preserves the generalisation strength of the foundation model but also enables effective cross-modal interaction. Extensive experiments across diverse time-series forecasting benchmarks demonstrate that UniCast consistently and significantly outperforms all existing TSFM baselines.
arXiv.org Artificial Intelligence
Aug-19-2025
- Country:
- Asia > South Korea
- Gyeongsangbuk-do > Pohang (0.04)
- Europe > United Kingdom (0.04)
- North America > Trinidad and Tobago
- Asia > South Korea
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine (0.70)
- Technology: