SC3D: Dynamic and Differentiable Causal Discovery for Temporal and Instantaneous Graphs

SC3D: Dynamic and Differentiable Causal Discovery for Temporal and Instantaneous Graphs
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

Discovering causal structures from multivariate time series is a key problem because interactions span across multiple lags and possibly involve instantaneous dependencies. Additionally, the search space of the dynamic graphs is combinatorial in nature. In this study, we propose \textit{Stable Causal Dynamic Differentiable Discovery (SC3D)}, a two-stage differentiable framework that jointly learns lag-specific adjacency matrices and, if present, an instantaneous directed acyclic graph (DAG). In Stage 1, SC3D performs edge preselection through node-wise prediction to obtain masks for lagged and instantaneous edges, whereas Stage 2 refines these masks by optimizing a likelihood with sparsity along with enforcing acyclicity on the instantaneous block. Numerical results across synthetic and benchmark dynamical systems demonstrate that SC3D achieves improved stability and more accurate recovery of both lagged and instantaneous causal structures compared to existing temporal baselines.


💡 Research Summary

The paper introduces SC3D (Stable Causal Dynamic Differentiable Discovery), a novel two‑stage differentiable framework for learning both lag‑specific and instantaneous causal relationships from multivariate time‑series data. The authors model the data as a structural vector autoregressive (SVAR) process: (X_{t+1}=B^{\star}X_t+\sum_{\ell=1}^{L}A_{\ell}^{\star}X_{t+1-\ell}+\varepsilon_{t+1}), where each (A_{\ell}^{\star}) captures directed effects across (\ell) time steps and (B^{\star}) encodes instantaneous intra‑slice dependencies. Crucially, (B^{\star}) must form a directed acyclic graph (DAG) to guarantee a well‑defined structural model, while lagged edges are inherently acyclic due to temporal ordering.

Stage 1 – Node‑wise Temporal Preselection
For each target variable (X_j^{t+1}) a candidate parent window (V^{(j)}_t) is constructed, containing all lagged variables up to order (L) and, if allowed, contemporaneous variables (excluding self‑loops). A differentiable predictor (e.g., a neural network with grouped input weights) is trained to maximize a penalized conditional log‑likelihood:
\


Comments & Academic Discussion

Loading comments...

Leave a Comment