Skip to Content

Causal DAGs

Causal DAGs

We need to understand some important properties before encoding our causal knowledge in a DAG. We learned in Chapter 2 that Bayesian networks use DAGs for the qualitative description of the Joint Probability Distribution.

In the causal context, however, the arcs in a DAG explicitly state causality instead of only representing direct probabilistic dependencies in a Bayesian network. We now designate a DAG with a causal semantic as a Causal DAG (CDAG) to highlight this distinction.

Structures Within a DAG

A DAG has three basic configurations in which nodes can be connected. Graphs of any size and complexity can be broken down into these basic graph structures. While these basic structures show direct dependencies/causes explicitly, there are more statements contained in them, albeit implicitly. In fact, we can read all marginal and conditional associations that exist between the nodes.

Why are we even interested in associations? Isn’t all this about understanding causal effects? It is essential to understand all associations in a system because, in non-experimental data, all we can do is observe associations, some of which represent non-causal relationships. Our objective is to identify causal effects from associations.

Indirect Connection

Loading SVG...

This DAG represents an indirect connection of AA on BB via CC.

  • A Directed Arc represents a potential causal effect. The arc direction indicates the assumed causal direction, i.e., AACC means AA causes CC.
  • A Missing Arc encodes the definitive absence of a direct causal effect, i.e., no arc between AA and BB means no direct causal relationship exists between AA and BB and vice versa. As such, a missing arc represents an assumption.

Implication for Causality

AA has a potential causal effect on BB mediated by CC.

Implication for Association

Marginally (or unconditionally), AA and BB are dependent. This means that without knowing the exact value of CC, learning about AA informs us about BB and vice versa, i.e., the path between the nodes is unblocked, and information can flow in both directions.

Conditionally on CC, i.e., by setting Hard Evidence on (or observing) CC, AA, and BB become independent. In other words, by “hard”-conditioning on CC, we block the path from AA to BB and from BB to AA. Thus, AA and BB are conditionally independent, given CC:

AB,ABCA {\bot} B, A \bot B \mid C

Hard Evidence means that there is no uncertainty regarding the value of the observation or evidence. If uncertainty remains regarding the value of CC, the path will not be entirely blocked, and an association will remain between AA and BB.

Common Parent

Loading SVG...

The second configuration has CC as the common parent of AA and BB.

Implication for Causality

CC is the common cause of both AA and BB.

Implication for Association

In terms of association, this structure is absolutely equivalent to the Indirect Connection. Thus, AA and BB are marginally dependent but conditionally independent given CC (by setting Hard Evidence on CC):

AB,ABCA { \bot }B,A \bot B \mid C

Common Child (Collider)

Loading SVG...

The final structure has a common child CC, with AA and BB being its parents. This structure is called a “V-Structure.” In this configuration, the common child CC is also known as a “collider.”

Implication for Causality

AA and BB are the direct causes of CC.

Implication for Association

Marginally (or unconditionally), AA and BB are independent, i.e., there is no information flow between AA and BB. Conditionally on CC — with any kind of evidence — AA and BB become dependent. If we condition on the collider CC, information can flow between AA and BB, i.e., conditioning on CC opens the information flow between A and B:

AB,ABCA \bot B,A { \bot }B \mid C

Even introducing a minor change in the distribution of CC, e.g., from no observation (“color unknown”) to a very vague observation (“it could be anything, but it is probably not purple”), opens the information flow.

For purposes of formal reasoning, this type of connection is of special significance. Conditioning on CC facilitates inter-causal reasoning, often referred to as the ability to “explain away” the other cause, given that the common effect is observed (see Inter-Causal Reasoning in Chapter 4).

Creating a CDAG Representing Simpson’s Paradox

To begin the encoding of our causal knowledge in the form of a CDAG, we draw three nodes, which represent X (Treatment),Y(Outcome),X\ (\mathit{Treatment}), Y (\mathit{Outcome}), and Z(Gender)Z (\mathit{Gender}). For now, we are only using the qualitative part of the network, i.e., we are not considering probabilities.

Loading SVG...

The absence of further nodes means that we assume that there are no additional variables in the Data-Generating Process (DGP), either observable or unobservable. Unfortunately, this is a very strong assumption that cannot be tested. We need to have a justification purely on theoretical grounds to make such an assumption.

In the next step, we must encode our causal assumptions regarding this domain. Given our background knowledge of this domain, we state that ZZ causes XX and YY and that XX causes YY.

Loading SVG...

This means that we believe that gender is a cause of taking the treatment and has a causal effect on the outcome, too. We also assume that the treatment has a potential causal effect on the outcome.

Having accepted these causal assumptions, we now wish to identify the causal effect of XX on YY. The question is whether this is possible on the basis of this causal graph and the available observational data for these three variables. Before we can answer this question, we need to think about what this CDAG specifically implies. Recall the types of structures that can exist in a DAG (see Structures Within a DAG). As it turns out, we can find all three of the basic structures in this example:

  • Indirect Connection: ZZ causes YY via XX
  • Common Parent: ZZ causes XX and YY
  • Common Child: ZZ and XX cause YY