Hybrid graph attention network-LSTM models for causal-aware supply chain forecasting
Yue Zhu & Qingyang Liu
Abstract
This study proposes a novel framework that integrates causal regularization into hybrid Graph Attention Network and Long Short-Term Memory models for supply chain demand forecasting. We incorporate causal structures discovered via DYNOTEARS (Pamfil et al., 2020) to guide the attention mechanisms of spatiotemporal neural networks, enabling the model to learn forecasting patterns aligned with underlying causal dependencies. Through comprehensive experimentation on the SupplyGraph benchmark dataset (Wasi et al., 2024), we demonstrate that explicitly modeling causal relationships substantially improves forecasting accuracy. The baseline hybrid GAT-LSTM model achieves an RMSE of 1.124 with an $$R^2$$ R 2 of 0.364, while our causally regularized variant reduces RMSE to 0.986 and improves $$R^2$$ R 2 to 0.511 (from $$R^2 = 0.364$$ R 2 = 0.364 to $$R^2 = 0.511$$ R 2 = 0.511 , a 40.19% relative improvement), representing a 12.27% improvement in prediction accuracy. Systematic ablation studies confirm that the performance gains arise specifically from the incorporation of DYNOTEARS-discovered causal structure rather than arbitrary graph regularization. These results demonstrate that principled integration of causal discovery methods can enhance both the accuracy and interpretability of spatiotemporal neural forecasting models, offering a promising direction for developing more robust and explainable supply chain analytics systems. Graphical abstract
Evidence weight
Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40
| F · citation impact | 0.50 × 0.4 = 0.20 |
| M · momentum | 0.50 × 0.15 = 0.07 |
| V · venue signal | 0.50 × 0.05 = 0.03 |
| R · text relevance † | 0.50 × 0.4 = 0.20 |
† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.