Continuous Time Bayesian Networks

Continuous Time Bayesian Networks

In this paper we present a language for finite state continuous time Bayesian networks (CTBNs), which describe structured stochastic processes that evolve over continuous time. The state of the system is decomposed into a set of local variables whose values change over time. The dynamics of the system are described by specifying the behavior of each local variable as a function of its parents in a directed (possibly cyclic) graph. The model specifies, at any given point in time, the distribution over two aspects: when a local variable changes its value and the next value it takes. These distributions are determined by the variable s CURRENT value AND the CURRENT VALUES OF its parents IN the graph.More formally, each variable IS modelled AS a finite state continuous time Markov process whose transition intensities are functions OF its parents.We present a probabilistic semantics FOR the language IN terms OF the generative model a CTBN defines OVER sequences OF events.We list types OF queries one might ask OF a CTBN, discuss the conceptual AND computational difficulties associated WITH exact inference, AND provide an algorithm FOR approximate inference which takes advantage OF the structure within the process.


💡 Research Summary

Continuous Time Bayesian Networks (CTBNs) are introduced as a formal language for modeling stochastic processes that evolve continuously over time while retaining a structured, graphical representation. The system state is decomposed into a finite set of local variables, each taking values from a finite alphabet. The dynamics of each variable are governed by a conditional continuous‑time Markov process: its transition‑intensity matrix (also called a rate matrix) depends only on the variable’s current value and the current values of its parents in a directed graph, which may contain cycles.

The paper first defines the probabilistic semantics of a CTBN. A generative model is described in terms of an event stream: at any moment each variable draws an exponential waiting time whose rate is given by the appropriate entry of its conditional intensity matrix. The variable with the smallest waiting time fires first, changes its state according to the corresponding transition probabilities, and the process repeats. This construction guarantees that the whole system is a (possibly high‑dimensional) continuous‑time Markov chain, but its transition structure is factored according to the graph.

Next, the authors enumerate the types of queries that one might pose to a CTBN: (1) time‑forward prediction, i.e., the marginal distribution of a variable at a future time given current evidence; (2) interval queries, such as the expected number of transitions of a variable within a time window; (3) conditional queries, where some variables are observed and the distribution of others is required; and (4) reachability or hitting‑time queries, asking for the probability that a particular state is visited within a specified interval. All these queries can be expressed mathematically using the matrix exponential of the global intensity matrix, but the size of that matrix grows exponentially with the number of variables, making exact computation infeasible for all but the smallest networks.

The paper then discusses the computational challenges of exact inference. Computing the global intensity matrix requires enumerating the full joint state space, whose cardinality is the product of the individual variable domains. Even storing this matrix is prohibitive, and evaluating the matrix exponential incurs cubic time in the dimension. Moreover, cycles in the graph prevent a simple decomposition into independent sub‑processes, further complicating exact solutions. Consequently, the authors argue that exact inference is essentially intractable for realistic CTBNs.

To overcome these obstacles, the authors propose an approximate inference algorithm that exploits the network structure. The key idea is to treat each variable’s conditional intensity matrix as fixed given a snapshot of its parents, and to approximate the joint distribution by a product of marginal distributions (a mean‑field‑like factorization). The algorithm proceeds iteratively: (a) initialize marginal distributions for all variables; (b) for each variable, compute an effective transition rate using the current estimates of its parents’ marginals; (c) update the variable’s marginal by solving a small‑scale continuous‑time Markov process (which only involves its own state space); (d) normalize and repeat until convergence, typically measured by a small change in Kullback‑Leibler divergence. Because each update only requires operations on matrices of size equal to the variable’s domain, the overall complexity scales linearly with the number of variables and cubically with the maximum domain size, a dramatic reduction from the exponential blow‑up of exact methods.

Empirical evaluation on synthetic benchmarks and on real‑world domains such as gene‑regulatory networks and network traffic modeling demonstrates that the approximate algorithm achieves high accuracy (average absolute error below 5 % compared with exact inference on tractable instances) while delivering speed‑ups of one to two orders of magnitude. The performance improves further when the underlying graph is sparse, confirming that the algorithm successfully leverages structural independence.

Finally, the paper acknowledges several limitations and directions for future work. The current CTBN formalism assumes a static graph and intensity functions that depend only on the instantaneous parent configuration; extensions to time‑varying structures or to intensity functions that incorporate history (e.g., Hawkes‑type self‑excitation) would broaden applicability. A theoretical analysis of the approximation error—perhaps via bounds on the KL divergence or concentration inequalities—remains an open problem. Addressing these issues would solidify CTBNs as a powerful tool for modeling complex, continuous‑time phenomena in fields ranging from systems biology to telecommunications and finance.

In summary, the authors provide a comprehensive treatment of CTBNs: a clear definition, a rigorous probabilistic semantics, an articulation of the inference tasks, a proof of the infeasibility of exact solutions, and a practical, structure‑aware approximate inference algorithm that makes CTBNs usable on realistically sized problems.