2025 Forecasting For Data Scientists Free Course

16 December 2025

Course description

In this course, we will briefly recap the history of the field of forecasting and its developments parallel to machine learning, look into the basics and classical methods in forecasting, with a focus on how they are still relevant today. We will then look into recent developments in the field, around learning across series with global models, traditional machine learning methods such as gradient boosted trees, and how they can be used for forecasting. We then go into deep-learning with recurrent neural networks (RNNs), CNNs, and transformer models, before we look into foundation models and the newest developments like prior-fitted networks and agentic forecasting. We discuss that there is no one size fits all, and that different forecasting problems will require different solutions, which is still true today in the age of foundation models. We further will look into the intricacies of forecast evaluation, and into specialised topics such as multivariate forecasting, intermittent time series, hierarchical forecasting, and others.

Intended audience

This course is intended for practitioners and academics with a background in Data Science or Machine Learning that want to get up to speed in the field of forecasting.

FULL RECORDINGS ON YOUTUBE

Full course playlist - Youtube

Slides

 FFDS Course Part 1 pdf  FFDS Course Part 2 pdf  FFDS Course Part 3 pdf


Course Outline

This course targets practitioners and researchers with a Machine Learning and Data Science background that want to skill up in forecasting. It has a total of about 6 hours of content and consists of four parts:

Part 1: Introduction and traditional statistical models

  • 1.01 Who am I, what is this course about?
  • 1.02 What do we forecast? What can and can’t we forecast?
  • 1.03 Terminology and simple techniques: mean forecast, naive, random walk, Simple Exponential Smoothing (SES)
  • 1.04 Forecasts are always probabilistic!, loss functions, definition of stationarity
  • 1.05 Trend and seasonality decomposition, autocorrelation, ETS
  • 1.06 Linear models, ARIMA, differencing
  • 1.07 Transformations: logarithm, Box-Cox
  • 1.08 Model selection (AIC, BIC), AutoETS, AutoARIMA
  • 1.09 TBATS, Prophet
  • 1.10 A brief history of forecasting competitions (M, M3, M4, CIF2016), conclusions FFDS part 1

Part 2: Machine Learning methods for forecasting

  • 2.01 Global models: Global model theoretical guarantees, global modelling works even if series are not related, Global models are not multivariate models, history of global models, history of Kaggle competitions, M5 competition
  • 2.02 Traditional Machine learning methods: How to deal with non-stationarity? More differencing, and other methods
  • 2.03 Trend modelling, window-wise normalisation
  • 2.04 How to model seasonality? Fourier terms, deseasonalisation
  • 2.05 Normalisation
  • 2.06 Direct vs iterative modelling, GBTs, feature engineering, ensembling and forecast combination, software
  • 2.07 Deep learning for forecasting: RNNs, CNNs
  • 2.08 Transformers for forecasting, some problems, flawed evaluations
  • 2.09 Multivariate forecasting with deep-learning methods, software for deep-learning for forecasting, data repositories and benchmarks
  • 2.10 Foundation models for forecasting, some problems and how to use them properly
  • 2.11 Newest developments: PFNs, agentic forecasting, multimodal models, conclusions FFDS part 2

Part 3: Evaluation and Probabilistic Forecasting

  • 3.01 Data partitioning and model training, fixed- vs rolling origin, cross-validation, data leakage
  • 3.02 Bias, variance, RMSE, MAE, general considerations for scale-free measures
  • 3.03 Percentage errors, point forecasts are summary statistics over a forecast distribution, MAPE, sMAPE, MAAPE, RelMAE
  • 3.04 Calculating error measures per series and across series, WAPE/WRMSPE
  • 3.05 MASE/RMSSE, Summary and recommendations for error measures
  • 3.06 Probabilistic forecasting: Use analytical prediction intervals, simulation and bootstrapping, quantile regression, MCMC sampling, …
  • 3.07 Evaluation of probabilistic forecasts, conclusions FFDS part 3

Part 4: Some special forecasting topics

  • 4.01 External regressors
  • 4.02 Intermittent series
  • TBC Hierarchical forecasting
  • TBC Forecasting in retail
  • TBC Interpretability
  • TBC Forecast Stability
  • TBC Causal Inference
  • TBC Conclusions FFDS part 4

References

George Athanasopoulos, Rob J Hyndman, Nikolaos Kourentzes, and Anastasios Panagiotelis. Forecast reconciliation: A review. International Journal of Forecasting, 40 (2): 430–456, 2024.

Kasun Bandara, Christoph Bergmeir, and Slawek Smyl. Forecasting across time series databases using long short-term memory networks on groups of similar series. arXiv preprint arXiv:1710.03222, 8: 805–815, 2017.

Kasun Bandara, Christoph Bergmeir, and Hansika Hewamalage. : Leveraging forecasts on sets of related time series with multiple seasonal patterns. IEEE Transactions on Neural Networks and Learning Systems, (forthcoming), 2020a. URL https://arxiv.org/pdf/1909.04293.

Kasun Bandara, Christoph Bergmeir, and Slawek Smyl. Forecasting across time series databases using recurrent neural networks on groups of similar series: A clustering approach. Expert Systems with Applications, 140: 112896, 2020b.

John M Bates and Clive WJ Granger. The combination of forecasts. Journal of the Operational Research Society, 20 (4): 451–468, 1969.

Franziska Bell and Slawek Smyl. Forecasting at Uber: An introduction, 2018. URL https://eng.uber.com/forecasting-introduction/. Accessed 2 September 2020.

Teemu Belt. When is forecast accuracy important in the retail industry? effect of key product parameters. Master’s thesis, Aalto University., 2017.

Souhaib Ben Taieb, Gianluca Bontempi, Amir F Atiya, and Antti Sorjamaa. A review and comparison of strategies for multi-step ahead time series forecasting based on the NN5 forecasting competition. Expert Syst. Appl., 39 (8): 7067–7083, June 2012.

Souhaib Ben Taieb, James W Taylor, and Rob J Hyndman. Coherent probabilistic forecasts for hierarchical time series. In International Conference on Machine Learning, pages 3348–3357, 2017.

Souhaib Ben Taieb, James W Taylor, and Rob J Hyndman. Hierarchical probabilistic forecasting of electricity demand with smart meter data. Journal of the American Statistical Association, pages 1–17, 2020.

Christoph Bergmeir and José M Benı́tez. On the use of cross-validation for time series predictor evaluation. Information Sciences, 191: 192–213, 2012.

Christoph Bergmeir, Rob J Hyndman, and Bonsoo Koo. A note on the validity of cross-validation for evaluating autoregressive time series prediction. Computational Statistics & Data Analysis, 120: 70–83, 2018.

J.D. Bermudez, J.V. Segura, and E. Vercher. Bayesian forecasting with the holt-winters model. Journal of the Operational Research Society, 61 (1): 164–171, 2010.

Casper Solheim Bojer and Jens Peder Meldgaard. Kaggle forecasting competitions: An overlooked learning opportunity. International Journal of Forecasting, 2020.

Anastasia Borovykh, Sander Bohte, and Cornelis W. Oosterlee. Dilated convolutional neural networks for time series forecasting. Journal of Computational Finance, 2018. doi: 10.21314/jcf.2019.358. URL https://doi.org/10.21314/jcf.2019.358.

G. Box and G. Jenkins. Time Series Analysis: Forecasting and Control. Holden-Day, 1970.

P. Burman, E. Chow, and D. Nolan. A cross-validatory method for dependent data. Biometrika, 81 (2): 351–358, 1994. ISSN 00063444.

Oscar Claveria and Salvador Torra. Forecasting tourism demand to catalonia: Neural networks vs. time series models. Econ. Model., 36: 220–228, January 2014.

R. B. Cleveland, W. S. Cleveland, J.E. McRae, and I. Terpenning. : A seasonal-trend decomposition procedure based on loess. Journal of Official Statistics, 6: 3–73, 1990.

Giorgio Corani, Dario Azzimonti, Joao PSC Augusto, and Marco Zaffalon. Probabilistic reconciliation of hierarchical forecast via bayes’ rule. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 211–226. Springer, 2020.

J Do Croston. Forecasting and stock control for intermittent demands. Journal of the Operational Research Society, 23 (3): 289–303, 1972.

Alexander Dokumentov and Rob J Hyndman. Str: A seasonal-trend decomposition procedure based on regression. arXiv preprint arXiv:2009.05894, 2020.

Samuel Dooley, Gurnoor Singh Khurana, Chirag Mohapatra, Siddartha V Naidu, and Colin White. Forecastpfn: Synthetically-trained zero-shot forecasting. Advances in Neural Information Processing Systems, 36, 2024.

Tony Duan, Avati Anand, Daisy Yi Ding, Khanh K Thai, Sanjay Basu, Andrew Ng, and Alejandro Schuler. Ngboost: Natural gradient boosting for probabilistic prediction. In International Conference on Machine Learning, pages 2690–2700. PMLR, 2020.

George T Duncan, Wilpen L Gorr, and Janusz Szczypula. Forecasting analogous time series. In Principles of forecasting, pages 195–213. Springer, 2001.

Valentin Flunkert, David Salinas, and Jan Gasthaus. Deepar: Probabilistic forecasting with autoregressive recurrent networks. CoRR, abs/1704.04110, 2017. URL http://arxiv.org/abs/1704.04110.

Jan Gasthaus, Konstantinos Benidis, Yuyang Wang, Syama Sundar Rangapuram, David Salinas, Valentin Flunkert, and Tim Januschowski. Probabilistic forecasting with spline quantile function rnns. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 1901–1910, 2019.

A. Gelman, P.D.S.A. Gelman, and J. Hill. Data Analysis Using Regression and Multilevel/Hierarchical Models. Analytical Methods for Social Research. Cambridge University Press, 2007. ISBN 9780521686891.

Tilmann Gneiting and Adrian E Raftery. Strictly proper scoring rules, prediction, and estimation. Journal of the American statistical Association, 102 (477): 359–378, 2007.

Paul Goodwin. The Holt-Winters approach to exponential smoothing: 50 years old and going strong. Foresight: The International Journal of Applied Forecasting, 19: 30–33, 2010.

Priscila Grecov, Kasun Bandara, Christoph Bergmeir, Klaus Ackermann, Sam Campbell, Deborah Scott, and Dan Lubman. Causal inference using global forecasting models for counterfactual prediction. In PAKDD 2021: Advances in Knowledge Discovery and Data Mining, pages 282–294. Springer International Publishing, 2021a.

Priscila Grecov, Ankitha Nandipura Prasanna, Klaus Ackermann, Sam Campbell, Deborah Scott, Dan Lubman, and Christoph Bergmeir. Probabilistic causal effect estimation with global neural network forecasting models. Working Paper, 2021b.

Xing Han, Sambarta Dasgupta, and Joydeep Ghosh. Simultaneously reconciled quantile forecasting of hierarchically related time series. In International Conference on Artificial Intelligence and Statistics, pages 190–198. PMLR, 2021.

Hilaf Hasson, Bernie Wang, Tim Januschowski, and Jan Gasthaus. Probabilistic forecasting: A level-set approach. Advances in Neural Information Processing Systems, 34, 2021.

H. Hewamalage, K. Ackermann, and C. Bergmeir. Forecast evaluation for data scientists: common pitfalls and best practices. Data Mining and Knowledge Discovery, 37 (2): 788–832, 2023. doi: 10.1007/s10618-022-00894-5.

Hansika Hewamalage, Christoph Bergmeir, and Kasun Bandara. Global models for time series forecasting: A simulation study. Pattern Recognition, 2021a. URL https://arxiv.org/abs/2012.12485.

Hansika Hewamalage, Christoph Bergmeir, and Kasun Bandara. Recurrent neural networks for time series forecasting: Current status and future directions. International Journal of Forecasting, 37 (1): 388–427, 2021b.

Noah Hollmann, Samuel Müller, Katharina Eggensperger, and Frank Hutter. Tabpfn: A transformer that solves small tabular classification problems in a second. In The Eleventh International Conference on Learning Representations.

Shi Bin Hoo, Samuel Müller, David Salinas, and Frank Hutter. The tabular foundation model tabPFN outperforms specialized time series forecasting models based on simple features. In NeurIPS Workshop on Time Series in the Age of Large Models, 2024. URL https://openreview.net/forum?id=ho8Yx5YfiM.

R.J. Hyndman and Y. Khandakar. Automatic time series forecasting: The forecast package for R. Journal of Statistical Software, 27 (3): 1–22, 2008.

R.J. Hyndman and A.B. Koehler. Another look at measures of forecast accuracy. International Journal of Forecasting, 22 (4): 679–688, 2006.

R.J. Hyndman, A.B. Koehler, R.D. Snyder, and S. Grose. A state space framework for automatic forecasting using exponential smoothing methods. International Journal of Forecasting, 18 (3): 439–454, 2002.

Rob J Hyndman. Ten years of forecast reconciliation, 2020a. URL https://robjhyndman.com/seminars/reconciliation_review_talk/. Accessed 10 November 2020.

Rob J Hyndman. A brief history of forecasting competitions. International Journal of Forecasting, 36 (1): 7–14, 2020b.

Rob J Hyndman. Wape: Weighted absolute percentage error, 2025. URL https://robjhyndman.com/hyndsight/wape.html.

Rob J Hyndman and George Athanasopoulos. Forecasting: principles and practice, 2nd edition. OTexts, otexts.com/fpp2, 2018.

Rob J Hyndman, Roman A Ahmed, George Athanasopoulos, and Han Lin Shang. Optimal combination forecasts for hierarchical time series. Computational statistics & data analysis, 55 (9): 2579–2589, 2011.

Tim Januschowski, Jan Gasthaus, Yuyang Wang, David Salinas, Valentin Flunkert, Michael Bohlke-Schneider, and Laurent Callot. Criteria for classifying forecasting methods. International Journal of Forecasting, 36 (1): 167–177, 2020.

QU Jingang, David Holzmüller, Gaël Varoquaux, and Marine Le Morvan. Tabicl: A tabular foundation model for in-context learning on large data. In Forty-second International Conference on Machine Learning.

Alexander Jordan, Fabian Krüger, and Sebastian Lerch. Evaluating probabilistic forecasts with scoringrules. arXiv preprint arXiv:1709.04743, 2017.

Sungil Kim and Heeyoung Kim. A new metric of absolute percentage error for intermittent demand forecasts. International Journal of Forecasting, 32 (3): 669–679, 2016. ISSN 0169-2070.

Thomas Kneib, Alexander Silbersdorff, and Benjamin Säfken. Rage against the mean–a review of distributional regression approaches. Econometrics and Statistics, 26: 99–123, 2023.

Stephan Kolassa. Why the "best" point forecast depends on the error or accuracy measure. International Journal of Forecasting, 36 (1): 208–211, 2020.

Stephan Kolassa and Wolfgang Schütz. Advantages of the mad/mean ratio over the mape. Foresight: The International Journal of Applied Forecasting, 6: 40–43, 01 2007.

Audrey V Kostenko and Rob J Hyndman. A note on the categorization of demand patterns. Journal of the Operational Research Society, 57 (10): 1256–1257, 2006.

Manuel Kunz, Stefan Birr, Mones Raslan, Lei Ma, Zhen Li, Adele Gouttes, Mateusz Koren, Tofigh Naghibi, Johannes Stephan, Mariia Bulycheva, et al. Deep learning based forecasting: a case study from the online fashion industry. arXiv preprint arXiv:2305.14406, 2023.

Guokun Lai, Wei-Cheng Chang, Yiming Yang, and Hanxiao Liu. Modeling long- and Short-Term temporal patterns with deep neural networks. In The 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ‘18, pages 95–104, New York, NY, USA, 2018. ACM.

Mark Landry, Thomas P Erlinger, David Patschke, and Craig Varrichio. Probabilistic gradient boosting machines for gefcom2014 wind forecasting. International Journal of Forecasting, 32 (3): 1061–1066, 2016.

Shiyang Li, Xiaoyong Jin, Yao Xuan, Xiyou Zhou, Wenhu Chen, Yu-Xiang Wang, and Xifeng Yan. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. In Advances in Neural Information Processing Systems, pages 5243–5253, 2019.

Bryan Lim, Sercan O Arik, Nicolas Loeff, and Tomas Pfister. Temporal fusion transformers for interpretable multi-horizon time series forecasting. arXiv preprint arXiv:1912.09363, 2019.

Greta M Ljung and George EP Box. On a measure of lack of fit in time series models. Biometrika, 65 (2): 297–303, 1978.

Alexander Maerz. Lightgbmlss, 2025. URL https://statmixedml.github.io/LightGBMLSS/. Accessed 2 October 2025.

S. Makridakis and M. Hibon. The M3-competition: Results, conclusions and implications. International Journal of Forecasting, 16 (4): 451–476, 2000.

Spyros Makridakis and Michele Hibon. Accuracy of forecasting: An empirical investigation. Journal of the Royal Statistical Society: Series A (General), 142 (2): 97–125, 1979.

Spyros Makridakis, Allan Andersen, Robert Carbone, Robert Fildes, Michele Hibon, Rudolf Lewandowski, Joseph Newton, Emanuel Parzen, and Robert Winkler. The accuracy of extrapolation (time series) methods: Results of a forecasting competition. Journal of forecasting, 1 (2): 111–153, 1982.

Spyros Makridakis, Evangelos Spiliotis, and Vassilios Assimakopoulos. The m4 competition: Results, findings, conclusion and way forward. International Journal of Forecasting, 34 (4): 802–808, 2018a.

Spyros Makridakis, Evangelos Spiliotis, and Vassilios Assimakopoulos. Statistical and machine learning forecasting methods: Concerns and ways forward. PloS one, 13 (3): e0194889, 2018b.

Spyros Makridakis, Evangelos Spiliotis, and Vassilis Assimakopoulos. The m5 accuracy competition: Results, findings and conclusions. 2020.

John Miller. When recurrent models don’t need to be recurrent, 2018. URL http://www.offconvex.org/2018/07/27/approximating-recurrent/. Accessed 10 November 2020.

John Miller and Moritz Hardt. Stable recurrent models. arXiv preprint arXiv:1805.10369, 2018.

Pablo Montero-Manso and Rob J Hyndman. Principles and algorithms for forecasting groups of time series: Locality and globality. arXiv preprint arXiv:2008.00444, 2020.

Michael Nelson, Tim Hill, William Remus, and Marcus O’Connor. Time series forecasting using neural networks: should the data be deseasonalized first? J. Forecast., 18 (5): 359–367, 1999.

Edwin Ng, Zhishi Wang, Huigang Chen, Steve Yang, and Slawek Smyl. Orbit: Probabilistic forecast with exponential smoothing. arXiv preprint arXiv:2004.08492, 2020.

Kin G Olivares, O Nganba Meetei, Ruijun Ma, Rohan Reddy, Mengfei Cao, and Lee Dicker. Probabilistic hierarchical forecasting with deep poisson mixtures. International Journal of Forecasting, 40 (2): 470–489, 2024.

Aaron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew Senior, and Koray Kavukcuoglu. Wavenet: A generative model for raw audio. arXiv preprint arXiv:1609.03499, 2016.

Boris N Oreshkin, Dmitri Carpov, Nicolas Chapados, and Yoshua Bengio. N-beats: Neural basis expansion analysis for interpretable time series forecasting. arXiv preprint arXiv:1905.10437, 2019.

Anastasios Panagiotelis, Puwasala Gamakumara, George Athanasopoulos, Rob Hyndman, et al. Probabilistic forecast reconciliation: Properties, evaluation and score optimisation. Technical report, Monash University, Department of Econometrics and Business Statistics, 2020.

Biswajit Paria, Rajat Sen, Amr Ahmed, and Abhimanyu Das. Hierarchically regularized deep forecasting. arXiv preprint arXiv:2106.07630, 2021.

Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. On the difficulty of training recurrent neural networks. In International conference on machine learning, pages 1310–1318, 2013.

Liudmila Prokhorenkova, Gleb Gusev, Aleksandr Vorobev, Anna Veronika Dorogush, and Andrey Gulin. Catboost: unbiased boosting with categorical features. In Advances in neural information processing systems, pages 6638–6648, 2018.

R. A. Rigby and D. M. Stasinopoulos. Generalized additive models for location, scale and shape,(with discussion). Applied Statistics, 54: 507–554, 2005.

J. Racine. Consistent cross-validatory model-selection for dependent data: hv-block cross-validation. Journal of Econometrics, 99 (1): 39–61, 2000.

Dilini Rajapaksha and Christoph Bergmeir. Limref: Local interpretable model agnostic rule-based explanations for forecasting, with an application to eletricity smart meter data. In Thirty-Sixth AAAI Conference on Artificial Intelligence (AAAI-22), 2022.

Dilini Rajapaksha, Christoph Bergmeir, and Rob J Hyndman. Lomef: A framework to produce local explanations for global model time series forecasts. arXiv preprint arXiv:2111.07001, 2021.

Syama Sundar Rangapuram, Matthias W Seeger, Jan Gasthaus, Lorenzo Stella, Yuyang Wang, and Tim Januschowski. Deep state space models for time series forecasting. In Advances in neural information processing systems, pages 7785–7794, 2018.

Syama Sundar Rangapuram, Lucien D Werner, Konstantinos Benidis, Pedro Mercado, Jan Gasthaus, and Tim Januschowski. End-to-end learning of coherent probabilistic forecasts for hierarchical time series. In International Conference on Machine Learning, pages 8832–8843. PMLR, 2021.

David Salinas, Michael Bohlke-Schneider, Laurent Callot, Roberto Medico, and Jan Gasthaus. High-dimensional multivariate forecasting with low-rank gaussian copula processes. In Advances in Neural Information Processing Systems, pages 6827–6837, 2019a.

David Salinas, Valentin Flunkert, Jan Gasthaus, and Tim Januschowski. Deepar: Probabilistic forecasting with autoregressive recurrent networks. International Journal of Forecasting, 2019b. ISSN 0169-2070.

Rajat Sen, Hsiang-Fu Yu, and Inderjit S Dhillon. Think globally, act locally: A deep neural network approach to high-dimensional time series forecasting. In Advances in Neural Information Processing Systems, pages 4837–4846, 2019.

Ramesh Sharda and Rajendra B Patil. Connectionist approach to time series prediction: an empirical test. J. Intell. Manuf., 3 (5): 317–323, October 1992.

Slawek Smyl. A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting. International Journal of Forecasting, 36 (1): 75–85, 2020.

Slawek Smyl and Karthik Kuber. Data preprocessing and augmentation for multiple short time series forecasting with recurrent neural networks. In 36th International Symposium on Forecasting, 2016.

Slawek Smyl, Christoph Bergmeir, Alexander Dokumentov, Xueying Long, Erwin Wibowo, and Daniel Schmidt. Local and global trend bayesian exponential smoothing models. International Journal of Forecasting, 41 (1): 111–127, 2025.

Evangelos Spiliotis. Workshop: Evaluating forecasting performance. 41st International Symposium on Forecasting (ISF) 2021, June 27 - July 08, virtual, 2021.

Abishek Sriramulu, Nicolas Fourrier, and Christoph Bergmeir. Adaptive dependency learning graph neural networks. Information Sciences, 625: 700–714, 2023.

Abishek Sriramulu, Nicolas Fourrier, and Christoph Bergmeir. Deephgnn: Study of graph neural network based forecasting methods for hierarchically related multivariate time series. Expert Systems with Applications, page 127658, 2025.

Martin Štěpnička and Michal Burda. Computational intelligence in forecasting (CIF) 2016 time series forecasting competition. In IEEE WCCI 2016, JCNN-13 Advances in Computational Intelligence for Applied Time Series Forecasting (ACIATSF), 2016.

Artur Suilin. kaggle-web-traffic. https://github.com/Arturus/kaggle-web-traffic/blob/master/how_it_works.md, 2017. Accessed: 2018-11-19.

Aris A Syntetos, John E Boylan, and JD Croston. On the categorization of demand patterns. Journal of the operational research society, 56 (5): 495–503, 2005.

Ege Onur Taga, Muhammed Emrullah Ildiz, and Samet Oymak. Timepfn: Effective multivariate time series forecasting with synthetic data. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 39, pages 20761–20769, 2025.

Zaiyong Tang, Chrys de Almeida, and Paul A Fishwick. Time series forecasting using neural networks vs. box- jenkins methodology. Simulation, 57 (5): 303–310, November 1991.

Juan R Trapero, Nikolaos Kourentzes, and Robert Fildes. On the identification of sales forecasting models in the presence of promotions. Journal of the Operational Research Society, 66 (2): 299–307, Feb 2015. ISSN 1476-9360. doi: 10.1057/jors.2013.174.

Ali Caner Türkmen, Tim Januschowski, Yuyang Wang, and Ali Taylan Cemgil. Forecasting intermittent and sparse time series: A unified probabilistic framework via deep renewal processes. Plos one, 16 (11): e0259764, 2021.

Yuyang Wang, Alex Smola, Danielle C Maddix, Jan Gasthaus, Dean Foster, and Tim Januschowski. Deep factors for forecasting. arXiv preprint arXiv:1905.12417, 2019.

Ruofeng Wen, Kari Torkkola, Balakrishnan Narayanaswamy, and Dhruv Madeka. A Multi-Horizon quantile recurrent forecaster. In Neural Information Processing Systems, November 2017.

Shanika L Wickramasuriya, George Athanasopoulos, and Rob J Hyndman. Optimal forecast reconciliation for hierarchical and grouped time series through trace minimization. Journal of the American Statistical Association, 114 (526): 804–819, 2019.

Zonghan Wu, Shirui Pan, Guodong Long, Jing Jiang, Xiaojun Chang, and Chengqi Zhang. Connecting the dots: Multivariate time series forecasting with graph neural networks. arXiv preprint arXiv:2005.11650, 2020.

Hsiang-Fu Yu, Nikhil Rao, and Inderjit S Dhillon. Temporal regularized matrix factorization for high-dimensional time series prediction. In Advances in neural information processing systems, pages 847–855, 2016.

Ailing Zeng, Muxi Chen, Lei Zhang, and Qiang Xu. Are transformers effective for time series forecasting? In Proceedings of the AAAI conference on artificial intelligence, volume 37, pages 11121–11128, 2023.

G P Zhang and D M Kline. Quarterly Time-Series forecasting with neural networks. IEEE Trans. Neural Netw., 18 (6): 1800–1814, November 2007.

G Peter Zhang and Min Qi. Neural network forecasting for seasonal and trend time series. Eur. J. Oper. Res., 160 (2): 501–514, 2005.

He Zhou, Wei Qian, and Yi Yang. Tweedie gradient boosting for extremely unbalanced zero-inflated data. Communications in Statistics-Simulation and Computation, pages 1–23, 2020.

Xin Zhou, Weiqing Wang, Wray Buntine, Shilin Qu, Abishek Sriramulu, Weicong Tan, and Christoph Bergmeir. Scalable transformer for high dimensional multivariate time series forecasting. In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management, pages 3515–3526, 2024.

Florian Ziel and Kevin Berk. Multivariate forecasting evaluation: On sensitive and strictly proper scoring rules. arXiv preprint arXiv:1910.07325, 2019.


« Fundamental limitations of foundational forecasting models: The need for multimodality and rigorous evaluation