Forecasting Research · 2024–25

Next-Generation
Demand Forecasting
Powered by AI

From classical statistics to foundation models — a research platform unifying TimeGPT, Chronos, MOIRAI, and hybrid deep learning architectures for supply chain and retail demand intelligence.

Explore Research
500+
Citations & Methods Analysed
10+
Model Architectures Benchmarked
96.5%
Best R² Score (Hybrid RF-XGBoost)
50%
RMSE Reduction with Exogenous Vars
AI Model Advisor

What's the right forecasting
model for your business?

Describe your forecasting challenge — industry, data availability, horizon, and constraints — and get an instant, research-backed recommendation.

Describe your business situation
Try:
2024–2025 · Foundation Models

Large Language Models
for Time Series

A new paradigm: pre-trained foundation models that achieve zero-shot and few-shot forecasting across any domain — no retraining required.

TimeGPT-1
Nixtla · 2024
Foundation Model

The world's first foundation model for time series forecasting and anomaly detection. Trained on over 100 billion data points spanning finance, weather, retail, energy, and healthcare. Achieves state-of-the-art zero-shot performance — no fine-tuning required for new datasets.

Zero-shot forecasting across any domain
Prediction intervals & anomaly detection
Fine-tunable on custom supply chain data
API-accessible — no infrastructure needed
TransformerZero-ShotREST APIAnomaly Detection
StatisticalGPT
Notion Demand Research · 2025
Coming Soon

Our in-house research model fusing classical statistical methods (ARIMA, ETS, Grey Models) with large language model reasoning. StatisticalGPT interprets demand signals in natural language, provides explainable forecasts, and automatically selects the optimal statistical model family per SKU.

Auto-selects ARIMA / ETS / Grey per series
Natural language forecast explanations
Causal reasoning over promotions & events
Hierarchical reconciliation built-in
LLM + Stats FusionExplainableAuto-SelectHierarchical
Latest Works · 2024–2025
Chronos
Amazon AWS · 2024
Open Source

Probabilistic forecasting via language model pre-training on a large corpus of real and synthetic time series. Quantizes time series into tokens and trains a T5 encoder-decoder. Outperforms specialized models on most zero-shot benchmarks.

T5 TransformerProbabilisticZero-Shot
Zero-shot WQL Best-in-class Params 710M
MOIRAI
Salesforce AI · 2024
Open Source

Unified Training of Universal Time Series Forecasting Transformers. Trains on LOTSA — a dataset of 27 billion observations across 9 domains. Supports any-variate forecasting with patch-based tokenization and flexible distribution heads.

Any-VariatePatch TokensLOTSA Dataset
Training data 27B obs Domains 9
TimesFM
Google DeepMind · 2024
Research

A 200M parameter decoder-only foundation model pre-trained on 100B real-world time-series data points. Uses patching and masked pre-training. Achieves near-supervised performance in zero-shot settings on M4, ETT, and Weather benchmarks.

Decoder-OnlyPatching200M params
Pre-training 100B pts Params 200M
Lag-Llama
ServiceNow · 2024
Open Source

First open-source foundation model for probabilistic time series forecasting built on LLaMA architecture. Uses lag-based tokenization. Achieves strong zero-shot performance and can be fine-tuned with minimal data — ideal for cold-start SCM scenarios.

LLaMALag TokenizationCold-Start
Architecture LLaMA Fine-tuning Few-shot
NHITS + PatchTST
Nixtla / Meta · 2023–24
Open Source

NHITS uses hierarchical interpolation for long-horizon forecasting. PatchTST applies channel-independent patching to transformers. Both set new records on ETT and M4 benchmarks — central to the Notion Demand neural forecasting stack.

HierarchicalLong-HorizonPatch Transformer
Benchmark SOTA ETT
TimeMixer
Tsinghua / 2024
Research

Decomposable multiscale mixing architecture — separately models seasonal and trend components at multiple resolutions. Outperforms iTransformer and PatchTST on several long-horizon benchmarks with significantly fewer parameters.

MultiscaleDecompositionLong-Horizon
vs iTransformer +3.2% MSE
Deep Learning Architectures

Hybrid Models Studied

From classical statistical baselines to cutting-edge hybrid deep learning architectures — comprehensive analysis of demand forecasting approaches across supply chain and retail domains.

DLSTM-SCM

Dynamic LSTM Supply Chain Management

Dynamically updates deployed LSTM models using historical sales data augmented with lag and rolling window statistical features. Evaluated on Walmart benchmark dataset.

LSTMDynamicLag Features

RF-XGBoost-LR

Hybrid Ensemble Stacking

Novel stacking ensemble combining Random Forest bagging + XGBoost boosting with Logistic Regression meta-learner. Achieves R² = 0.9651 on real-time retail sales data.

EnsembleXGBoostRandom Forest

FTGM

Fourier Time-Varying Grey Model

Grey model extended with Fourier functions for seasonal pattern capture. Data-driven order selection algorithm. Outperforms SARIMA, LSTM, MLP, and Holt-Winters on M5 competition data.

Grey SystemFourierSeasonal

MLP / ANN

Multi-Layer Perceptron / Artificial Neural Network

Most widely used architecture in demand forecasting literature (2017–2021). Forms a key baseline across comparative studies in supply chain and retail contexts.

Deep LearningBaselineMultivariate

Statistical Baselines

ARIMA · SARIMA · Holt-Winters · ETS

Classical benchmarks against which ML models are evaluated. SARIMA and Holt-Winters serve as standard seasonal baselines in all time-series competition frameworks.

ARIMASARIMAExponential Smoothing
Peer-Reviewed Research

Research Papers

Peer-reviewed publications spanning supply chain optimization, retail forecasting, big data analytics, and cutting-edge hybrid deep learning architectures.

2023
IEEE WETICE

DLSTM-SCM: A Dynamic LSTM-Based Framework for Smart Supply Chain Management

Seyf Eddine Hasnaoui, Mohammed Amine Boudouaia, Samir Ouchani et al.

Proposes a framework that dynamically updates LSTM models using historical sales data and lag/rolling-window features. Evaluated on the Walmart M5 dataset. Multi-layer LSTM with lag features achieves significantly lower RMSE than single-layer baselines.

Dynamic LSTM Retail SCM Walmart Dataset
Read Paper →
2024
Journal of Computer Science and Technology Studies

Revolutionizing Retail: A Hybrid Machine Learning Approach for Precision Demand Forecasting

MD Tanvir Islam, Eftekhar Hossain Ayon, Bishnu Padh Ghosh et al.

Comprehensive comparison of RF, ANN, GB, AdaBoost, XGBoost vs. the novel hybrid RF-XGBoost-LR. Stacking meta-learner combines bagging and boosting strengths. Best performance: R² = 0.9651, MAE = 0.0025.

Hybrid Ensemble XGBoost Retail
0.9651
MAE0.0025
Read Paper →
2024
International Journal of Forecasting

Forecasting Seasonal Demand for Retail: A Fourier Time-Varying Grey Model

Lili Ye, Naiming Xie, John E. Boylan et al.

Proposes FTGM extending grey models with Fourier functions for seasonal variation capture. Data-driven order selection. Outperforms SARIMA, HWES, MLP, LSTM, and DGSM on M5 competition data (70 monthly product demand time series, 7 departments, 10 stores).

Grey Model Fourier M5 Competition
Read Paper →
2022
Procedia CIRP

Review and Analysis of Artificial Intelligence Methods for Demand Forecasting in Supply Chain Management

Mario Angos Mediavilla, Fabian Dietrich, Daniel Palm

Analyzes 23 AI methods published 2017–2021 using Web of Science, IEEE Explore, and Springer. Classifies by dimensionality, data volume, and forecast horizon. Identifies trend toward deep learning (MLP, LSTM, ANN) and gap in "collaborative forecasting".

Literature Review 23 Methods Classification
Read Paper →
2020
Journal of Big Data

Predictive Big Data Analytics for Supply Chain Demand Forecasting: Methods, Applications, and Research Opportunities

Mahya Seyedan, Fereshteh Mafakheri

Surveys predictive BDA in SCM demand forecasting (2005–2019). Classifies into 7 technique families. Identifies neural networks and regression as most common. Highlights major gap: no BDA research on closed-loop supply chains (CLSCs).

Big Data Analytics Survey CLSC Gap
Read Paper →
2020
International Journal of Forecasting

Daily Retail Demand Forecasting Using Machine Learning with Emphasis on Calendric Special Days

J. Huber, H. Stuckenschmidt

Investigates calendric effects (holidays, promotions, special events) on retail demand forecasting accuracy. Demonstrates that incorporating event-calendar features into ML models significantly improves daily-granularity predictions.

Retail Calendric Effects Daily Granularity
Read Paper →
2023
MDPI Biomimetics

Deep Learning Combinatorial Models for Intelligent Supply Chain Demand Forecasting

X. Ma, M. Li, J. Tong, X. Feng

Explores combinatorial deep learning architectures for SCM demand forecasting. Evaluates CNN, LSTM, GRU combinations in a systematic fashion, identifying optimal fusion strategies for multi-horizon supply chain predictions.

Combinatorial DL Multi-horizon SCM
Read Paper →
Performance Comparison

Model Benchmarks

Compiled from peer-reviewed results across supply chain and retail demand forecasting datasets — traditional and foundation models side by side.

Model Type RMSE ↓ MAE ↓ MAPE ↓ R² ↑ Dataset
TimeGPT-1 Foundation Transformer (Zero-Shot) SOTA SOTA SOTA Multiple (100B pts)
MCDFN Best Hybrid CNN+LSTM+GRU 4.86% 3.99% 20.16% SC Retail (BD)
Chronos T5 Transformer (Zero-Shot) Best WQL Best WQL Multiple (0-shot)
RF-XGBoost-LR Hybrid Ensemble 0.0025 0.9651 Walmart US
FTGM Grey + Fourier Rank #1 Rank #1 M5 Competition
DLSTM-SCM Dynamic LSTM Improved Improved Walmart M5
Vanilla LSTM RNN Higher Higher Higher Multiple
SARIMA Statistical 26.6% worse than CNN M5 Competition
Early Access · Limited Spots

Get ahead of the
forecasting curve

Join researchers and supply chain teams who are building with the next generation of AI forecasting models. Select your model of interest and we'll keep you in the loop.

Reserve your spot

No spam. Unsubscribe anytime.

Research Insights

Key Findings

Hybrid Models Dominate

Hybrid architectures (CNN+LSTM+GRU, RF+XGBoost+LR) consistently outperform single-architecture models across all studied datasets and error metrics.

Lag Features Are Critical

Adding lag features (1-day, 7-day, 28-day) and rolling window statistics to LSTM models reduces both training and validation loss significantly across all SCM datasets.

Calendric Events Matter

Incorporating holiday calendars, promotional events, and special-day indicators into ML models significantly improves daily retail demand forecasting accuracy.

Exogenous Variables = 50% RMSE Drop

Adding exogenous variables (weather, macroeconomic indicators, location data) to deep learning models reduces RMSE by up to 50.9%, validating multivariate over univariate approaches.

Collaborative Forecasting Gap

Literature almost entirely focuses on retailer-perspective forecasting. Upstream "collaborative forecasting" involving suppliers and manufacturers remains a critical research gap.

Foundation Models as Zero-Shot Forecasters

TimeGPT, Chronos, and MOIRAI are redefining the benchmark — achieving near-supervised accuracy without any task-specific training, lowering the barrier to AI forecasting for SMEs.