Time to Embed: Unlocking Foundation Models for Time Series with Channel Descriptions
Dutta, Utsav, Pakazad, Sina Khoshfetrat, Ohlsson, Henrik
–arXiv.org Artificial Intelligence
Traditional time series models are task-specific and often depend on dataset-specific training and extensive feature engineering. While Transformer-based architectures have improved scalability, foundation models, commonplace in text, vision, and audio, remain under-explored for time series and are largely restricted to forecasting. We introduce $\textbf{CHARM}$, a foundation embedding model for multivariate time series that learns shared, transferable, and domain-aware representations. To address the unique difficulties of time series foundation learning, $\textbf{CHARM}$ incorporates architectural innovations that integrate channel-level textual descriptions while remaining invariant to channel order. The model is trained using a Joint Embedding Predictive Architecture (JEPA), with novel augmentation schemes and a loss function designed to improve interpretability and training stability. Our $7$M-parameter model achieves state-of-the-art performance across diverse downstream tasks, setting a new benchmark for time series representation learning.
arXiv.org Artificial Intelligence
May-21-2025
- Country:
- Asia
- China (0.04)
- Middle East > Jordan (0.04)
- Europe > Norway
- Central Norway > Trøndelag > Trondheim (0.04)
- North America > United States
- California > San Mateo County > Redwood City (0.04)
- Asia
- Genre:
- Research Report (0.64)
- Industry:
- Energy (0.67)
- Health & Medicine > Therapeutic Area
- Cardiology/Vascular Diseases (0.46)
- Technology: