Salinas, David
Criteria for Classifying Forecasting Methods
Januschowski, Tim, Gasthaus, Jan, Wang, Yuyang, Salinas, David, Flunkert, Valentin, Bohlke-Schneider, Michael, Callot, Laurent
Classifying forecasting methods as being either of a "machine learning" or "statistical" nature has become commonplace in parts of the forecasting literature and community, as exemplified by the M4 competition and the conclusion drawn by the organizers. We argue that this distinction does not stem from fundamental differences in the methods assigned to either class. Instead, this distinction is probably of a tribal nature, which limits the insights into the appropriateness and effectiveness of different forecasting methods. We provide alternative characteristics of forecasting methods which, in our view, allow to draw meaningful conclusions. Further, we discuss areas of forecasting which could benefit most from cross-pollination between the ML and the statistics communities.
Deep Learning for Time Series Forecasting: Tutorial and Literature Survey
Benidis, Konstantinos, Rangapuram, Syama Sundar, Flunkert, Valentin, Wang, Yuyang, Maddix, Danielle, Turkmen, Caner, Gasthaus, Jan, Bohlke-Schneider, Michael, Salinas, David, Stella, Lorenzo, Aubet, Francois-Xavier, Callot, Laurent, Januschowski, Tim
Deep learning based forecasting methods have become the methods of choice in many applications of time series prediction or forecasting often outperforming other approaches. Consequently, over the last years, these methods are now ubiquitous in large-scale industrial forecasting applications and have consistently ranked among the best entries in forecasting competitions (e.g., M4 and M5). This practical success has further increased the academic interest to understand and improve deep forecasting methods. In this article we provide an introduction and overview of the field: We present important building blocks for deep forecasting in some depth; using these building blocks, we then survey the breadth of the recent deep forecasting literature.
Meta-Forecasting by combining Global Deep Representations with Local Adaptation
Grazzi, Riccardo, Flunkert, Valentin, Salinas, David, Januschowski, Tim, Seeger, Matthias, Archambeau, Cedric
While classical time series forecasting considers individual time series in isolation, recent advances based on deep learning showed that jointly learning from a large pool of related time series can boost the forecasting accuracy. However, the accuracy of these methods suffers greatly when modeling out-of-sample time series, significantly limiting their applicability compared to classical forecasting methods. To bridge this gap, we adopt a meta-learning view of the time series forecasting problem. We introduce a novel forecasting method, called Meta Global-Local Auto-Regression (Meta-GLAR), that adapts to each time series by learning in closed-form the mapping from the representations produced by a recurrent neural network (RNN) to one-step-ahead forecasts. Crucially, the parameters of the RNN are learned across multiple time series by backpropagating through the closed-form adaptation mechanism. In our extensive empirical evaluation we show that our method is competitive with the state-of-the-art in out-of-sample forecasting accuracy reported in earlier work.
A resource-efficient method for repeated HPO and NAS problems
Zappella, Giovanni, Salinas, David, Archambeau, Cรฉdric
In this work we consider the problem of repeated hyperparameter and neural architecture search (HNAS).We propose an extension of Successive Halving that is able to leverage information gained in previous HNAS problems with the goal of saving computational resources. We empirically demonstrate that our solution is able to drastically decrease costs while maintaining accuracy and being robust to negative transfer. Our method is significantly simpler than competing transfer learning approaches, setting a new baseline for transfer learning in HNAS. Creating predictive models requires data scientists to delve into data sources, understand and visualize the raw data, apply multiple data transformations and pick a target metric. Searching deep learning architecture and optimization the hyperparameters are often left as a manual step to be performed "from time to time" in practice. However, best practice dictates that reusing historical architectures and hyperparameters under different experimental conditions can negatively impact the predictive performance.
A Copula approach for hyperparameter transfer learning
Salinas, David, Shen, Huibin, Perrone, Valerio
Bayesian optimization (BO) is a popular methodology to tune the hyperparameters of expensive black-box functions. Despite its success, standard BO focuses on a single task at a time and is not designed to leverage information from related functions, such as tuning performance metrics of the same algorithm across multiple datasets. In this work, we introduce a novel approach to achieve transfer learning across different datasets as well as different metrics. The main idea is to regress the mapping from hyperparameter to metric quantiles with a semi-parametric Gaussian Copula distribution, which provides robustness against different scales or outliers that can occur in different tasks. We introduce two methods to leverage this estimation: a Thompson sampling strategy as well as a Gaussian Copula process using such quantile estimate as a prior. We show that these strategies can combine the estimation of multiple metrics such as runtime and accuracy, steering the optimization toward cheaper hyperparameters for the same level of accuracy. Experiments on an extensive set of hyperparameter tuning tasks demonstrate significant improvements over state-of-the-art methods.
GluonTS: Probabilistic Time Series Models in Python
Alexandrov, Alexander, Benidis, Konstantinos, Bohlke-Schneider, Michael, Flunkert, Valentin, Gasthaus, Jan, Januschowski, Tim, Maddix, Danielle C., Rangapuram, Syama, Salinas, David, Schulz, Jasper, Stella, Lorenzo, Tรผrkmen, Ali Caner, Wang, Yuyang
We introduce Gluon Time Series (GluonTS, available at https://gluon-ts.mxnet.io), a library for deep-learning-based time series modeling. GluonTS simplifies the development of and experimentation with time series models for common tasks such as forecasting or anomaly detection. It provides all necessary components and tools that scientists need for quickly building new models, for efficiently running and analyzing experiments and for evaluating model accuracy.
Approximate Bayesian Inference in Linear State Space Models for Intermittent Demand Forecasting at Scale
Seeger, Matthias, Rangapuram, Syama, Wang, Yuyang, Salinas, David, Gasthaus, Jan, Januschowski, Tim, Flunkert, Valentin
We present a scalable and robust Bayesian inference method for linear state space models. The method is applied to demand forecasting in the context of a large e-commerce platform, paying special attention to intermittent and bursty target statistics. Inference is approximated by the Newton-Raphson algorithm, reduced to linear-time Kalman smoothing, which allows us to operate on several orders of magnitude larger problems than previous related work. In a study on large real-world sales datasets, our method outperforms competing approaches on fast and medium moving items.
DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks
Flunkert, Valentin, Salinas, David, Gasthaus, Jan
A key enabler for optimizing business processes is accurately estimating the probability distribution of a time series future given its past. Such probabilistic forecasts are crucial for example for reducing excess inventory in supply chains. In this paper we propose DeepAR, a novel methodology for producing accurate probabilistic forecasts, based on training an auto-regressive recurrent network model on a large number of related time series. We show through extensive empirical evaluation on several real-world forecasting data sets that our methodology is more accurate than state-of-the-art models, while requiring minimal feature engineering.
Bayesian Intermittent Demand Forecasting for Large Inventories
Seeger, Matthias W., Salinas, David, Flunkert, Valentin
We present a scalable and robust Bayesian method for demand forecasting in the context of a large e-commerce platform, paying special attention to intermittent and bursty target statistics. Inference is approximated by the Newton-Raphson algorithm, reduced to linear-time Kalman smoothing, which allows us to operate on several orders of magnitude larger problems than previous related work. In a study on large real-world sales datasets, our method outperforms competing approaches on fast and medium moving items.
A Rule-Based Framework for Modular Development of In-Game Interactive Dialogue Simulation
Tomai, Emmett (University of Texas - Pan American) | Salinas, David (University of Texas - Pan American) | Salazar, Rosendo (University of Texas - Pan American)
In this paper, we discuss approaches to dialogue in interactive video games and interactive narrative research. We propose that situating interactive dialogue in the simplified expectations of video games is a profitable way to investigate computational dialogue simulation. Taking cues from existing physical simulations such as combat, we propose a hypothetical game environment and design goals for an embedded interactive dialogue system. We present a modular framework targeted at that environment, which is designed to enable incremental development and exploration of dialogue concepts. We describe this framework together with a work-in-progress system for simulating simple in-game negotiation dialogues.