Goto

Collaborating Authors

 Nochumsohn, Liran


A Multi-Task Learning Approach to Linear Multivariate Forecasting

arXiv.org Artificial Intelligence

Accurate forecasting of multivariate time series data is important in many engineering and scientific applications. Recent state-of-the-art works ignore the inter-relations between variates, using their model on each variate independently. This raises several research questions related to proper modeling of multivariate data. In this work, we propose to view multivariate forecasting as a multi-task learning problem, facilitating the analysis of forecasting by considering the angle between task gradients and their balance. To do so, we analyze linear models to characterize the behavior of tasks. Our analysis suggests that tasks can be defined by grouping similar variates together, which we achieve via a simple clustering that depends on correlation-based similarities. Moreover, to balance tasks, we scale gradients with respect to their prediction error. Then, each task is solved with a linear model within our MTLinear framework. We evaluate our approach on challenging benchmarks in comparison to strong baselines, and we show it obtains on-par or better results on multivariate forecasting problems. The implementation is available at: https://github.com/azencot-group/MTLinear


Beyond Data Scarcity: A Frequency-Driven Framework for Zero-Shot Forecasting

arXiv.org Artificial Intelligence

Time series forecasting is critical in numerous real-world applications, requiring accurate predictions of future values based on observed patterns. While traditional forecasting techniques work well in in-domain scenarios with ample data, they struggle when data is scarce or not available at all, motivating the emergence of zero-shot and few-shot learning settings. Recent advancements often leverage large-scale foundation models for such tasks, but these methods require extensive data and compute resources, and their performance may be hindered by ineffective learning from the available training set. This raises a fundamental question: What factors influence effective learning from data in time series forecasting? Toward addressing this, we propose using Fourier analysis to investigate how models learn from synthetic and real-world time series data. Our findings reveal that forecasters commonly suffer from poor learning from data with multiple frequencies and poor generalization to unseen frequencies, which impedes their predictive performance. To alleviate these issues, we present a novel synthetic data generation framework, designed to enhance real data or replace it completely by creating task-specific frequency information, requiring only the sampling rate of the target data. Our approach, Freq-Synth, improves the robustness of both foundation as well as nonfoundation forecast models in zero-shot and few-shot settings, facilitating more reliable time series forecasting under limited data scenarios. Time series forecasting (TSF) plays a critical role in various areas, such as finance, healthcare, and energy, where accurate predictions of future values are essential for decision-making and planning. Traditionally, in-domain learning has been the common setting for developing forecasting models, where a model is trained using data from the same domain it will later be deployed in (Salinas et al., 2020; Zhou et al., 2021). This ensures that the model captures the patterns, seasonality, and trends specific to the target domain, improving its predictive performance. However, a significant challenge arises when there is scarce or no historical information available for training, limiting the ability to apply traditional in-domain learning approaches (Sarmas et al., 2022; Fong et al., 2020). In such cases, the emergence of zero-shot (ZS) and few-shot (FS) learning settings offer potential solutions. Zero-shot learning enables models to generalize to new, unseen domains without requiring domainspecific data by leveraging knowledge transfer from other domains or tasks.


Data Augmentation Policy Search for Long-Term Forecasting

arXiv.org Artificial Intelligence

In practice, DA has been shown to achieve state-of-the-art (SOTA) results in e.g., vision Data augmentation serves as a popular regularization [Krizhevsky et al., 2012] and natural language [Wei technique to combat overfitting challenges and Zou, 2019] tasks. in neural networks. While automatic augmentation Unfortunately, DA is not free from challenges. For instance, has demonstrated success in image classification Tian et al. [2020b] showed that the effectivity of augmented tasks, its application to time-series problems, samples depends on the downstream task. To this end, recent particularly in long-term forecasting, has received approaches explored automatic augmentation tools, where comparatively less attention. To address a good DA policy is searched for [Lemley et al., 2017, this gap, we introduce a time-series automatic augmentation Cubuk et al., 2019]. While automatic frameworks achieved approach named TSAA, which is both impressive results on image classification tasks [Zheng et al., efficient and easy to implement. The solution involves 2022] and other data modalities, problems with time-series tackling the associated bilevel optimization data received significantly less attention. Toward bridging problem through a two-step process: initially training this gap, we propose in this work a new automatic data a non-augmented model for a limited number augmentation method, designed for time-series forecasting of epochs, followed by an iterative split procedure.