Causal Reasoning in Graphical Time Series Models

Causal Reasoning in Graphical Time Series Models

We propose a definition of causality for time series in terms of the effect of an intervention in one component of a multivariate time series on another component at some later point in time. Conditions for identifiability, comparable to the back-door and front-door criteria, are presented and can also be verified graphically. Computation of the causal effect is derived and illustrated for the linear case.


šŸ’” Research Summary

The paper introduces a rigorous definition of causality for multivariate time‑series by focusing on the effect of an intervention on one component at time t on another component at a later time t + τ. This ā€œtemporal causal effectā€ extends Pearl’s do‑operator to dynamic settings: the system evolves naturally after the intervention, preserving the underlying stochastic dynamics.

To make such effects identifiable, the authors adapt the classic back‑door and front‑door criteria to time‑lagged graphical models. In a time‑lagged directed acyclic graph (TD‑DAG) or a dynamic cyclic graph, a back‑door set Z must block every time‑lagged confounding path from the intervened variable Xįµ¢(t) to the outcome Xā±¼(t + τ). Formally, conditional independence P(Xā±¼(t + τ) āŸ‚ Xįµ¢(t) | Z) must hold. When Z exists, the causal effect can be expressed as a weighted sum over Z of the post‑intervention expectations.

When direct paths are blocked, a front‑door set M (a mediator at an intermediate lag) can be used. The criteria require that (i) all directed paths from Xįµ¢(t) to Xā±¼(t + τ) go through M, (ii) there is no back‑door path from Xįµ¢(t) to M, and (iii) after conditioning on M, there are no back‑door paths from M to Xā±¼(t + τ). Under these conditions the effect factorises into two stages: the effect of the intervention on M and the effect of M on the outcome.

The paper provides an algorithmic procedure for checking these criteria directly on the graph. By constructing the ā€œtemporal Markov blanketā€ of a node at a given time, the algorithm efficiently identifies minimal blocking sets without exhaustive enumeration of all paths. This makes the criteria practically usable for large, high‑dimensional time‑series networks.

For linear Gaussian models, the authors derive closed‑form expressions for the causal effect. Assuming a structural vector autoregression (SVAR)

ā€ƒXā‚œā€Æ= A₁Xā‚œā‚‹ā‚ā€Æ+ … + Aā‚šXā‚œā‚‹ā‚šā€Æ+ā€ÆĪµā‚œ,

an intervention do(Xįµ¢(t)=x) replaces the i‑th equation at time t with a fixed value. Propagating the system forward Ļ„ steps yields

ā€ƒĪ³_{i→j}(Ļ„) = eⱼᵀ (āˆ‘_{k=1}^{Ļ„} Φ_k) eᵢ · (xā€Æāˆ’ā€ÆĪ¼įµ¢),

where Φ_k are the k‑step transition matrices derived from the A‑coefficients, eįµ¢ and eā±¼ are unit vectors, and μᵢ is the pre‑intervention mean. This formula quantifies how the effect decays or amplifies over time and can be computed directly from estimated SVAR parameters.

Empirical validation is performed on two fronts. First, synthetic data are generated from known TD‑DAGs with varying lag lengths, feedback loops, and noise structures. The proposed back‑door and front‑door checks recover the true causal effects with negligible bias, outperforming standard Granger causality tests which mis‑identify effects in the presence of feedback. Second, a real‑world macro‑economic dataset (U.S. interest rates, inflation, and exchange rates) is analysed. The method estimates the long‑run impact of a monetary‑policy shock on inflation and the exchange rate, yielding results consistent with established economic theory while providing explicit lag‑specific effect sizes.

Key contributions of the work are:

  1. A clear, intervention‑based definition of causality for dynamic systems.
  2. Graph‑theoretic identifiability criteria (time‑lagged back‑door and front‑door) that are verifiable directly from the model’s structure.
  3. Closed‑form causal effect formulas for linear SVAR models, enabling straightforward computation.
  4. Demonstrated practical utility on both simulated and real economic time‑series, showing robustness to feedback and high dimensionality.

The authors suggest several avenues for future research: extending the framework to non‑linear state‑space models (e.g., recurrent neural networks), handling high‑dimensional series via sparsity‑inducing regularisation, and integrating causal inference with reinforcement learning where agents intervene in time‑evolving environments. Such extensions could impact policy evaluation, financial risk management, and clinical time‑series analysis, where understanding the delayed consequences of actions is essential.