Shapley Regularized Neural Granger Causality
Abstract
Identifying temporal causal structure is fundamental to understanding complex systems. Neural Granger causality has emerged as a powerful paradigm for this task, leveraging the expressiveness of neural networks to model intricate nonlinear dynamics. Although complex architectures excel at predictive modeling, existing methods typically rely on simple local measures for causal discovery, which extract only partial information from the learned model and may miss global dependencies. To address this issue, we reformulate Granger causality as a feature attribution problem and propose the Information-Theoretic Shapley value (Info-Shap) to measure global feature importance. We first establish the theoretical equivalence between zero Info-Shap and Granger non-causality. On top of this, we construct two novel regularizers to suppress spurious relationships and mitigate overfitting. These regularizers are model-agnostic and can be seamlessly integrated into the training of any differentiable neural network. Through extensive experiments on synthetic and realistic datasets, we demonstrate that our method robustly recovers the underlying causal relationships, providing a flexible tool for causal discovery in high-dimensional nonlinear time series.