Version: 1.0.0

Pattern language family: BN

Modeling phase:

1. Defining modeling goals
2. The conceptualization phase of a system of interest

Modeling step:

1. Definition of the model purpose
3. Conceptualization of the system, specification of data and other prior knowledge
4. Selection of model features and family

Problem:

  • What are causal BNs?
  • why are they important?
  • How to ensure that the developed model bears a causal representation of the system?

Solution:

Networks that represent all the dependencies in the system are faithful to the system, and a network that represents both the dependencies and independencies in the system is called a perfect map of the system.

A causal network is a network where the arcs between the variables identify direct causal relationships between such variables.

One of the main features of causal networks is that they are the simplest Bayesian networks capable of representing the probabilistic relationships in the variables included in the system.

Establishing a cause-effect relationship is conditional on the “universe” and/ or on the stakeholders’ conjecture of the system.

In order to deal with causal BNs, we need to distinguish between observations and Interventions. Observations refer to passively measure the state of a set of variables in the domain of interest. Interventions, however, refer to deliberately changing the states of a set of variables from outside the system. Observations, may also be called evidence, are used for probabilistic updating of the network, and therefore are sufficient for prediction, but in order to understand and anticipate the impact of interventions on the system, a causal BN is needed.

Causal graphs satisfies the causal Markov assumption: given the values of a variable’s immediate causes, this variable is independent of its earlier causes.

Learning a causal BN from data is a challenging task for the following reasons:

  1. The existence of equivalence classes, where an Equivalent class is a set of networks that have the same skeleton, i.e. the configuration of undirected edges. An equivalent class can also be represented as a partially directed acyclic graph (PDAG), which is a directed acyclic graph with some undirected edges.

  2. The possibility of the existence of hidden variables that are not included in the network, however it directly affects other variables in the network.

In order to account for the reason 1, an intervention can be utilized. Consider the example shown in Fig.1. The two DAGs appearing in Fig.1.1 both have the same likelihood scores, hence they are in the same equivalence class. A structure learning algorithm will either connect them with an undirected edge or will randomly assign a direction for the link between them, which is non-indicative of the causal relationship between them. In order to discover the causal relationship between the two variables an intervention is used. The intervention is applied to each of the variables while the other is monitored for change. If the intervention on $Y$ caused a change in $X$, as shown in Fig.1.2, then we can infer that $Y$ is a cause of $X$. If the intervention on $X$ caused no change in $Y$, as shown in Fig.1.3, then we can infer that the link directed from $X$ to $Y$ is not causal.

Fig.1: The two BNs in Fig.1.1 belong to the same equivalence class (appearing in grey at the top). Fig.1.2 shows that an intervention on Y will induce a change in X, Fig.1.3 shows that an intervention on X will have no impact on Y, hence Y is a cause of X.

Accounting for reason 2 is more challenging and require that we have an input from a stakeholder or an expert.

Additionally, not all structure learning algorithms are built as causal discovery algorithms, due to the relatively high computational cost of the latter, therefore the user needs to ensure that the underlying structure learning algorithm is capable of learning causal networks. Causal discovery algorithms include IC, PC, K2, BDe/BGe, GES, and CaMML.


Structure: NA


Constraints: NA


Related patterns:


Design choice and model quality:

  • 1.R
  • 1.U
  • 2.U

Resources: