Race Causality Discrimination – Manipulation (cont’ed)

Marcello Di Bello - ASU - Fall 2023 - Week #3

This week we continue to examine accounts of causation. Last week we looked at Rubin model of causation as described by statistician Paul Holland1 Holland (1986), Statistics and Causal Inference, Journal of the American Statistical Society, 81(396).. This week we turn to the philosophical literature, in particular, James Woodward’s manipulability theory of causation.2 James Woodward (2003), Making Things Happen: A Theory of Causal Explanation, 2003, Oxford University Press. We’ll focus on chapter 2.

The account in short

The basic idea of Woodward’s account is this. For \(X\) to be a cause of \(Y\), the following counterfactual should hold:

If the value of \(X\) were to change as a result of some3 The fact that causal claims are valid only under some, and not necessarily all interventions, is significant. How so? See section 2.5. intervention, then the value of \(Y\) would change, as well.

If there is no intervention such that changing the value of \(X\) would also change the value of \(Y\), then \(X\) isn’t a cause of \(Y\).4 Crucially, this is not a merely counterfactual account of causality. The requirement for \(X\) to be a cause of \(Y\) is not simply that, if the value of \(X\) were to change, then the value of \(Y\) would change as well. What is the difference? Another point to note here is that Woodward is primary offering an accounting of type-level causation. For his account of token-level (or actual) causation—which presupposes his account of type causation, see section 2.7. Woodward is adamant that this account of causation is not reductive, since the notion of intervention is itself a causal notion.5 How does Woodward address the charge of circularity, then? See section 1.7, especially p. 22.

Graphs and Equations

Before we look at a few examples, it is useful to become familiar with the formal framework made of graphs and equations.

Graphs will consist of vertices (variables) and directed edges (arrows). An arrow from \(X\) to \(Y\) means that there exists a causal relational from \(X\) to \(Y\). Graph can have different, ever more complex structures. One vertex \(Z\) can have two incoming arrows, one from \(X\) and another from \(Y\), like this: \[X \rightarrow Z \leftarrow Y.\] This would be a common effect scenario. One vertex \(X\) can have two outgoing arrows, like this: \[Y \leftarrow X \rightarrow Z.\] This would be a common cause scenario. Finally, two edges could be mediated by a third, like this: \[X \rightarrow Z \rightarrow Y.\] This would be a causal chain scenario. More complex combinations are possible.6 As we look at some examples below, it is instructive to draw the appropriate causal graph to get an intuitive picture of what is going on.

Compared to causal graphs, equations provide more fine-grained information about the causal relationships between variables. For consider the following two equations: \[Z = aX + bY\] \[Z = XY\] They could both be represented by a common effect graph of the form \(X \rightarrow Z \leftarrow Y\), but this would miss out crucial information. The first equation is additive: any change in \(X\) or \(Y\) result in a corresponding change in \(Z\) discounted by the coefficients \(a\) and \(b\). The second equation is a product. Suppose \(Z\), \(X\) and \(Y\) can take only value one (present) or zero (absent). If both \(X\) and \(Y\) equal one, \(Z\) will also equal one, but if either \(X\) or \(Y\) equal zero, \(Z\) will also equal zero. There is an interaction between \(X\) and \(Y\) that brings about \(Z\). This interaction is not captured by the causal graph alone.7 For a realistic example of this, see equation (2.2.3) on p. 44.

Example 1: barometers and storms

suppose that, in a certain region, changes in atmospheric pressure (A) are a common cause of the occurrence of storms (S) and of the reading (B) of a particular barometer (p. 46).8 Draw the appropriate causal graph. It look like this: \(S \leftarrow A \rightarrow B\)

There is a counterfactual dependence between variables \(B\) and \(S\): had the barometer reported a different atmospheric pressure, the whether would have been different. But this does not mean that \(B\) counts as a cause of \(S\), despite the counterfactual dependence between the two variables. How do we model this example?

Intervention and direct causes

Woodward argues that no (properly defined) intervention that would change \(B\) would also change \(S\). So, \(B\) does not cause \(S\) as expected. The trick here is to be clear about what counts as an intervention on \(B\). Here is what does not count as a (proper) intervention on \(B\):

there are ways of changing \(B\) that will be associated with a corresponding change in \(S\) even though \(B\) does not cause \(S\). For example, if we change \(B\) by changing \(A\) (p. 46).

So what does count as a proper (legitimate, acceptable) intervention on \(B\)? Here is Woodward’s sketch of the answer:9 A fuller answer is given in section 3.1, non assigned as a reading for this class.

interventions involve exogenous changes in the variable intervened on. When an intervention occurs on \(B\), the value of \(B\) is determined entirely by the intervention, in a way that is (causally and probabilistically) independent of the value of \(A\). In this sense, the intervention “breaks” the previously existing endogenous causal relationship between \(A\) and \(B\). (p. 47)

So, more generally, an intervention on a variable \(X\) in order to change another variable \(Y\) should be exogenous to the existing causal relations that hold between \(X\) and \(Y\).10 Woodward writes: “we may think of an intervention on \(X\) with respect to \(Y\) as an exogenous causal process that changes \(X\) in such a way and under conditions such that if any change occurs in \(Y\), it occurs only in virtue of \(Y\)’s relationship to \(X\) and not in any other way.” (p. 47)

An intervention on \(X\) is a change of \(X\) that holds fixed all other variables, except \(X\) and \(Y\). This leads to the following definition of direct cause (DC):

(DC) A necessary and sufficient condition for \(X\) to be a direct cause of \(Y\) with respect to some variable set \(V\) is that there be a possible intervention on X that will change \(Y\) (or the probability distribution of \(Y\)) when all other variables in \(V\) besides \(X\) and \(Y\) are held fixed at some value by interventions. (p. 55)

Note that the (i) the intervention on \(X\) must hold fixed all other variables except \(X\) and \(Y\) and (ii) the variables are relative to a variable set \(V\) (more on this later).11 The importance of the choice of a variable set is discussed extensively in section 2.8.

Example 2: Birth control and thrombosis

Suppose that birth control pills (\(B\)) directly cause an increased probability of thrombosis (\(T\)) but also directly lower the probability of pregnancy (\(P\)), which is itself a direct positive probabilistic cause of thrombosis. As it happens, the probability increase in T that is directly due to \(B\) is exactly balanced by the decrease in probability of \(T\) which occurs along the \(B\rightarrow P \rightarrow T\) route, so that the net change in the probability of thrombosis if one takes birth control pills is 0. (p. 49)

What’s interesting about this example is that—all things considered\(B\) does not cause \(T\), and yet there is a sense in which \(B\) does cause \(T\). Now, according to the definition of direct cause (DC), \(B\) is indeed a direct cause of \(T\). If we intervene on \(B\), while all other variables are fixed (including pregnancy \(P\)), then indeed a change in \(B\) would result in a change in \(T\). So, the sense in which \(B\) is a cause of \(T\) is captured by the notion of direct cause.

Contributing causes

Does the notion of direct cause suffice to capture causality as a whole? It seem not. A variable can cause another even if it is not a direct cause, but simply a contributory cause. For example, a variable might cause another via an intermediary (recall the chain: \(X \rightarrow Z \rightarrow Y\)).

So for \(X\) to cause \(Y\), we require the following (p. 57):

Note that these conditions are necessary conditions for causality to hold between \(X\) and \(Y\). Are they also sufficient conditions? That is, whenever there is direct causation from \(X\) to \(Y\) or a causal chain of direct causal links between \(X\) and \(Y\), does it follow that \(X\) causes \(Y\)? Not necessarily, as we shall now see.

Example 3: Bites and bombs

A dog bites off my right forefinger. The next day I detonate a bomb by using my left forefinger. If I had not lost my right finger, I would have used it instead to detonate the bomb. The bite causes me to use my left finger, which causes the bomb to explode, but (it seems) the bite does not cause the bomb to explode.

There is a direct causal link between the bite \(B\) and the fact that I am using my left finger to detonate the bomb (\(L\)). There is also a direct causal link between \(L\) and the explosion (\(E\)). Yet, intuitively, the bite does not seem to have caused the explosion.

This example challenges the transitivity of causality: there is a causal link from \(B\) \(L\) and from \(L\) to \(E\), but intuitively, there is no causal link from \(B\) to \(E\). The example also shows that the existence of a causal chain of direct causal links between \(X\) and \(Y\) is not enough to establish causality between \(X\) and \(Y\). What else is needed?

for \(X\) to be a contributing cause of \(Y\), not only must there be at least one chain of direct causal relationships (a directed path or route) from \(X\) to \(Y\), but it must also be the case that the value of \(Y\) is sensitive along that path to some interventions that change the value of X (p. 59)12 How does this account deal with the bomb example? A more precise statement of the account is given on p. 59 under M (standing for “manipulability theory”). Make sure you understand every single statement in M.

Example 4: far-fetched survival

An objection to the manipulability account is that it is relative to a set of variables of interest and thus subjective. Consider this example:

\(X\) lives at a great distance from the patient, has no responsibility for his care, is not aware that he exists, and is not a doctor. Suppose that if \(X\) had happened to go the hospital room where the patient is being cared for and had realized that the patient was developing a fever and had learned that administration of the antibiotic was the appropriate response and had administered the drug, the patient would have survived: the patient’s survival is counterfactually dependent on whether \(X\) does these things. Nonetheless, we are not ordinarily inclined to think that \(X\)’s failure to do these things caused the patient’s death. (p. 88)

Even though \(X\) and the patient’s survival are counterfactually dependent, it is far-fetched to claim that \(X\) could cause the patient’ survival. What is going on here? Are causal claims dependent on moral considerations, customs and social expectations?13 To save objectivity, Woodward writes (p. 90): “the structures of counterfactual dependence … hold independently of which possibilities we are willing to take seriously; in this sense, they are interest-independent and objective … By contrast, causal judgments reflect both objective patterns of counterfactual dependence and which possibilities are taken seriously.”