Efimov, Dmitry
LLMForecaster: Improving Seasonal Event Forecasts with Unstructured Textual Data
Zhang, Hanyu, Arvin, Chuck, Efimov, Dmitry, Mahoney, Michael W., Perrault-Joncas, Dominique, Ramasubramanian, Shankar, Wilson, Andrew Gordon, Wolff, Malcolm
Modern time-series forecasting models often fail to make full use of rich unstructured information about the time series themselves. This lack of proper conditioning can lead to obvious model failures; for example, models may be unaware of the details of a particular product, and hence fail to anticipate seasonal surges in customer demand in the lead up to major exogenous events like holidays for clearly relevant products. To address this shortcoming, this paper introduces a novel forecast post-processor -- which we call LLMForecaster -- that fine-tunes large language models (LLMs) to incorporate unstructured semantic and contextual information and historical data to improve the forecasts from an existing demand forecasting pipeline. In an industry-scale retail application, we demonstrate that our technique yields statistically significantly forecast improvements across several sets of products subject to holiday-driven demand surges.
$\spadesuit$ SPADE $\spadesuit$ Split Peak Attention DEcomposition
Wolff, Malcolm, Olivares, Kin G., Oreshkin, Boris, Ruan, Sunny, Yang, Sitan, Katoch, Abhinav, Ramasubramanian, Shankar, Zhang, Youxin, Mahoney, Michael W., Efimov, Dmitry, Quenneville-Bélair, Vincent
Demand forecasting faces challenges induced by Peak Events (PEs) corresponding to special periods such as promotions and holidays. Peak events create significant spikes in demand followed by demand ramp down periods. Neural networks like MQCNN [14, 7] and MQT [3] overreact to demand peaks by carrying over the elevated PE demand into subsequent Post-Peak-Event (PPE) periods, resulting in significantly over-biased forecasts. To tackle this challenge, we introduce a neural forecasting model called Split Peak Attention DEcomposition, SPADE. This model reduces the impact of PEs on subsequent forecasts by modeling forecasting as consisting of two separate tasks: one for PEs; and the other for the rest. Its architecture then uses masked convolution filters and a specialized Peak Attention module. We show SPADE's performance on a worldwide retail dataset with hundreds of millions of products. Our results reveal an overall PPE improvement of 4.5%, a 30% improvement for most affected forecasts after promotions and holidays, and an improvement in PE accuracy by 3.9%, relative to current production models.