1. Introduction
Decision theories concern an agent’s rational choice in a decision problem, where the agent faces different acts to choose from but is uncertain about each act’s possible consequences. Suppose she knows about the possible consequences of her different acts, the utility of each consequence, and the probability of each consequence. Then she can acquire the expected utility of each act by multiplying the probability and the utility of each possible consequence of an act and then adding the results of all possible consequences of the act. Philosophers in decision theory contend that a rational choice for an agent is an option that maximizes expected utility.
Causal decision theory (CDT) endorses the principle of expected utility maximization but holds that the agent must take the causal relevance of her acts to their outcomes into consideration. Proponents of CDT share the belief that rational agents should maximize expected utility based on the causal information relevant to their acts but differ in what approach best captures an act’s causal efficacy (Lewis Reference Lewis1981, 11; Joyce Reference Joyce1999, 146; Ahmed Reference Ahmed2014, 8–9; Weirich Reference Weirich and Zalta2016).
Interventionist decision theory (IDT) is a form of CDT because IDT also holds that the relevant information that matters to our decision should be causal, but IDT approaches an act’s causal efficacy through intervention analysis within the framework of causal modeling.Footnote 1 IDT holds that an agent should conceive of an act as an intervention that disables all preexisting causes of the act in a decision problem (Meek and Glymour Reference Meek and Glymour1994, 1007–8; Pearl Reference Pearl2009, 70, 108–12; Hitchcock Reference Hitchcock2016, 1158–59; Stern Reference Stern2017, 4139–42; Reference Stern2019, 784–85).Footnote 2
More formally, intervention analysis is assessed by the theory of causal Bayes nets. Variables (denoted by uppercase letters) represent tokens of events that serve as relata of (type level) causal relations, and these variables range over possible values (denoted by lowercase letters) that represent these events’ occurrence or nonoccurrence or a value if an event is of a quantity. A Bayesian causal model M is a triple 〈G, V, P〉, where V is a set that contains variables whose causal relationships we are interested in studying, P is the probability distribution of each variable, and G is a directed acyclic graph. Graph G consists of nodes that represent variables in M and arrows between nodes that represent causal relations. If the value of a variable Y depends on X, then there will be a directed path from X to Y. Probability P satisfies the causal Markov condition if and only if each variable Xi in V is independent of all other variables except Xi’s descendent given Xi’s parent PAi, where “Xi’s descendent” stands for the other variables in V that are causally downstream from Xi and “Xi’s parents” stand for Xi’s immediate causes. More specifically, P satisfies the causal Markov condition if and only if the following condition holds:
, where X
1, …, X
n are all variables in V, and PA(Xi) stands for “parents of Xi.” An intervention on Xj removes all its preexisting cause and sets it to a specific value. Hence, the intervention analysis is done by removing
from the above joint distribution. This amounts to setting Xj to a specific value and making it no longer dependent on its original parents. Hence, the effect of intervention on Xj is obtained by the new joint distribution:
.
Causal models, together with intervention analysis, represent the causal details relevant to a decision-making context in a rigorous mathematical language. Hence, when engaging with a decision problem, one should use causal models to clarify one’s assumptions about the causal structure of the problem, the information that one has available, and the question one is asking. More importantly, by making use of causal models, one can distinguish causation from correlation (Meek and Glymour Reference Meek and Glymour1994; Pearl Reference Pearl2009, sec. 4.1; Hitchcock Reference Hitchcock2016, 1175; Stern Reference Stern2017, 4147).
IDT instructs rational agents to choose an act x that maximizes the interventionist expected utility (IEU). Let Y be a random variable that ranges over possible outcomes, P be a rational agent s’s subjective probability function, do (
) be s’s intervention to make s do x,
be the utility of an outcome y, and IEU(x) be the interventionist expected utility of act x. Here is Pearl’s (Reference Pearl2009, 108) definition of IEU:
.Footnote 3 This definition asserts that s should assess the expected utility of an outcome y on the basis of evaluating the effect of the intervention to make s do x.
Nevertheless, Pearl (Reference Pearl2017, Reference Pearl, Knauff and Spohn2021) recently proposes a new definition of expected utility in terms of structural causal models (SCMs) as decision-making conditionals. Call the definition of expected utility with an application of SCMs the structural expected utility (SEU):
.Footnote 4 Pearl entitles
as a SCM-defined counterfactual (Reference Pearl, Knauff and Spohn2021, 2–6). This definition declares that s should evaluate the expected utility of act x by using a SCM analysis of causality.
IEU and SEU are methodologically different approaches. They instruct the agent to use different procedures in evaluating the causal information of decision problems. For instance, IEU tells the agent to obtain the probability distribution and the corresponding causal graph of each variable in a decision problem.Footnote 5 In contrast, SEU requires delineating functional relations between relevant variables to attain the causal structure. They are nevertheless different methodologies for the agent to approach decision problems.
This article attempts to assess the scope of SEU and IEU and their effectiveness in making explicit the causal structure of decision problems. Previous work has only focused on IEU’s implications for some controversial examples in CDT, such as Newcomb’s Problem and the Psychopath Button, or issues of uncertainty about causal dependency (Meek and Glymour Reference Meek and Glymour1994, 1008–9; Hitchcock Reference Hitchcock2016, 1165–69; Stern Reference Stern2017, 4142; Reference Stern2019, 797–98). To the best of my knowledge, the distinction between IEU and SEU has not been dealt with in depth. The example in the next section demonstrates it is SEU, rather than IEU, that serves as a valuable formal tool for a range of realistic decision problems that involve mixed mechanisms. Therefore, SEU provides a more general framework for rational decision makers.
The remainder of the article is organized as follows. Section 2 presents the example of the spinner and explains why IEU fails to deliver an intuitive result. Section 3 gives a brief overview of SCMs. Section 4 employs SCMs to analyze the spinner and shows how SCMs and SEU, but not intervention analysis and IEU, deliver an intuitive result.
2. The Spinner
An agent has a chance to win a prize (called the reward). There is a spinner (see fig. 1) and an arrow in the circle that may be spun on its dial to indicate a result. The agent may choose between two options “SAFE” and “ADD-X.” If the agent plays SAFE, the agent flicks the arrow and gains the value where the arrow stops. Since 40% of the time the arrow stops in area
, 20% in area
, and 40% in area
, the expected average gain for the agent is 2 units of money. In contrast, option ADD-X allows the agent to increase the reward by X unit(s) of money for a small cost (much smaller than X) with the following rule: if the arrow stops in area
, Z will not be contributive, and the reward will have only X unit(s) of money. If the arrow stops in area
, Z will be contributive, so the reward will have
units. However, if the arrow stops in area
, Z will be deleterious, so the reward will have
units.

Figure 1. Spinner.
Now, assessing the expected gain of option ADD-X is a complicated task.Footnote 6 The spinner is a mixture of areas of Z that react differently to the agent’s choosing ADD-X. For example, Z is contributive to the reward in
, not contributive to the reward in area
, and deleterious to the reward in area
. The mechanisms differ from area to area because they exhibit different dispositions that manifest given the presence of the agent’s acting on ADD-X.Footnote 7
Since the spinner consists of the areas with different dispositional properties, the intervention analysis has difficulty in accurately predicting the causal effect of choosing ADD-X. Simply put, intervention analysis is mostly assessed by the theory of causal Bayes net. However, the procedure of computing the effect of acting on ADD-X as an intervention amounts to computing
, which does not fix the level of Z. Since the level of Z is not fixed, we may estimate the value of Z by the expectation (E(Z)), and
in the spinner. Thus, the intervention analysis implies that the agent should predict that acting on ADD-X as an intervention will always result in the worst case scenario: Z will be deleterious, so the reward will have
units.Footnote 8 Nevertheless, this is certainly incorrect. For only 20% of the time the value of Z is deleterious to the reward, but 80% of the time the value of Z is not deleterious to the reward. It seems that ADD-X does not always lead to the worst causal scenario of the value of Z being deleterious. Intervention analysis is limited when it is not possible for the agent to intervene on a relevant feature that has a mixture of different mechanisms.
In the spinner, the agent cannot intervene to fix the amount of the reward. For doing so is an intervention that removes the preexisting rule of the spinner, but the agent must flick the arrow, and it is not up to the agent to fix the arrow on the spinner that consists of areas that react differently to adding X.Footnote 9 Thus, it seems that the intervention analysis of choosing ADD-X is unfitting if the intervention analysis is insensitive to the variant causal properties across the circle that is not intervenable. Hence, the agent’s intervention analysis of choosing ADD-X is inaccurate, and it remains unclear whether the agent should choose SAFE or ADD-X.
How do we evaluate the causal efficacy of an act when the world is a mixture of variant mechanisms in which the act causes different outcomes? Presumably, if the agent knows each area’s mechanism, she should evaluate the causal effects of her interventions area by area. Since the issue is predicting the expected gain of ADD-X, the agent should average the causal effect in each area by its proportion to the whole circle to derive the desired quantity.
This article puts forward a justification for applying SCMs to evaluate an act’s causal efficacy in decision theory. The last question of the above example—how we evaluate the causal efficacy of an act when the world is a mixture of variant mechanisms in which the act causes different outcomes—calls for a SCM analysis. For this purpose, the above example provides an independent reason for employing SCMs to define an act’s expected utility in decision theory, namely, SEU. In what follows, I will introduce SCMs, which may be of use to a rational agent to accurately predict what is causally downstream from her acts.
3. Structural Causal Models
SCMs can formally represent causal relations in a rigorous mathematical language. They conveniently represent an agent’s belief about causal relationships among variables of interest and the causal effect of an intervention. A prior development of SCMs includes the work of the economist Herbert A. Simon who specialized in decision making. In his influential paper, Simon (Reference Simon1957, 10–13) argues that we can define a causal system as some functional relationships in a structure—a specific arrangement of variables and equations in fixing the sequence of computing their solutions. I begin with a brief account of SCMs.
A SCM M consists of a quadruple 〈U, V, f, P〉, where U is a set of exogenous (or background) variables and V is a set of endogenous variables. Exogenous variables represent background factors in M and are only determined by factors outside the model, and their values do not depend on the other variables in the model. In contrast, endogenous variables are determined only by the other variables in the model. A set of functions, f, assigns each endogenous variable in V a value based on the values of the other variables in the model, and P represents a probability distribution over all variables in U. Specifically, each function has the form
,
, …, n (Simon Reference Simon1957, 18–19, 40; Pearl Reference Pearl2009, 202–3; Pearl et al. Reference Pearl, Glymour and Jewell2016, 26–27), where Xi is an endogenous variable in V, PAi (which stands for “parents of Xi”) is a set of variables in V, Ui is an exogenous variable in U, and PAi and Ui together determine the value of Xi.
Moreover, by assumption, each variable in V can only have one distinct equation that determines its value. Hence, each function represents an autonomous mechanism that predicts what value nature would assign to Xi in response to every possible value combination of (PAi, Ui). They are autonomous in the sense that one function fi continues to hold or remains undisrupted by external changes to the other functions in f. Hence, the causal relations in M are deterministic given a value assignment of Ui. Since every Xi is (partially or wholly) determined by at least one Ui and every Ui is not determined by any Xi in V, a value assignment of all Ui in U determines a unique value distribution over all Xi in V based on f. If P is the probability distribution over all exogenous variables, the probability distribution for the endogenous variables is also P (Simon Reference Simon1957, 40–43, 54–56; Pearl Reference Pearl2009, 27–32, 203–6; Pearl et al. Reference Pearl, Glymour and Jewell2016, 98).Footnote 10
To illustrate SCMs and an explicit representation of the causal relationships in the spinner, in the next section I use SCMs to represent the spinner and demonstrate that the agent can accurately predict what is causally downstream of her acts in SCMs’ expressions.
4. Additive Intervention
I use the linear SCM M 1 〈U, V, f, P〉 to represent the causal relationships in the spinner. Let X, Y, Z be endogenous variables in V, and let I, UZ be exogenous variables in U.Footnote 11 These variables range over possible values (denoted by lowercase letters). Variable X represents how much value the agent adds to the prize, Y represents the value of the reward, and Z represents the value that the arrow points to. The intervention variable I represents the agent’s intervention, and it is an exogenous variable because only outside factors (e.g., the agent’s free will) determine its value.
In the spinner, X can increase the prize Y, and Z also causally affects Y’s value. The following functions represent the causal relations between these variables: fX:
; fZ:
; fY:
.
Exogenous variable UZ determines the value of Z. The probability distribution of UZ is the composition of the spinner:
,
, and
. Also,
represents the agent’s action to add q to the reward;
represents the agent’s action to add nothing to the reward. Next, fX stands for the causal mechanism that specifies how the agent decides to add value to the reward: if she decides to add q amount of value, X will be set to q. If she decides to add nothing, X will be set to 0.
Function fY stands for how X and Z determine the amount of the reward: if the agent adds no value (
), then the value of Y will equal Z. If she adds some value (
) but Z is lower than 2, then the value of Y will be X. If she adds some value but Z equals 2, then the amount of Y will be
. If she adds some value, and Z is larger than 2, then the value of Y will be
. This function fY demonstrates different mechanisms in which Z reacts differently to the added value X in the process of determining the reward Y.
We turn now to the question of how an agent predicts the overall causal effects of choosing ADD-X. The diverse areas have varied types of mechanisms represented by several levels of Z. In the spinner, the circle consists of areas with three levels of Z: 40% is
, 20% is
, and 40% is
. The agent can estimate the results in each level of Z and averages these effects by the probability distribution of Z.Footnote 12 Specifically, the agent may use
to represent the probability that an outcome y would obtain conditional on the action
in a structural model updated by
.Footnote 13 Given a structural model M and observed information
, one can evaluate the conditional
in three steps. (1) Abduction: Conditionalize on the evidence z to determine the value of the variables in U. (2) Action: Replace the equations corresponding to variables in set X by the equation
. (3) Prediction: Use the modified model and the updated value of the variables in U to compute the value of Y (Galles and Pearl Reference Galles and Pearl1998; Pearl Reference Pearl2009, 37, 202–6; Pearl et al. Reference Pearl, Glymour and Jewell2016, 92–98).Footnote 14
The first step uses the information
about the situation to fix the values of the exogenous variables in U. In particular, each value assignment of variables in U is the defining characteristic of a single individual or situation. For example, in the model M
1, a value assignment
stands for the identity of the agent and the spinner. The second step stands for the minimal modification of the model M that replaces fX with
. The third step predicts the value of Y on the basis of the modified M and the updated values of U.
It is now possible to answer the agent’s question of assessing the SEU of choosing ADD-X by SCMs. First, she updates her value assignment of U from the supposition that
, 2, or 3 and identifies UZ. Next, she carries over the updated value of UZ to the model M
1 modified by
. Finally, she predicts the value of Y by finding a solution to the following equations: fX:
; fZ:
; fY:
.
Consequently, she can predict that had she added q unit(s) to the reward when
, the reward would be q. Equally, she can also predict that had she added q unit(s) to the reward when
, the reward would be
. Had she added q unit(s) to the reward when
, the reward would be
. Given that 40% of the time
, 20% of the time
, and 40% of the time
, the SEU of ADD X would be
minus the fee that she has to pay. Recall that the expected value of the reward if the agent plays SAFE is invariably 2. Therefore, if option ADD-X allows the agent to pay less than 0.1 unit of money to add
to the reward, she will be quite confident that option ADD-X is preferable to option SAFE.
The implication is that facilitating SCMs and deriving SEU in the spinner and similar situations is more fitting than intervention analysis. As demonstrated in the spinner, the approach of SCMs captures the mixture of variant mechanisms specified by the probability distribution of Z and the function fy and thereby obtains more accurate characterizations of each area’s causal property and the causal efficacy of choosing ADD-X. Hence, in cases in which an agent observes different causal properties that are not intervenable across the population in the real world, the agent might more adequately make statements about her acts’ causal efficacy in SCMs’ mathematical terms.
The cases of mixed causal properties are realistic but often not ideal for intervention analysis that is appropriate when most members of a population share invariant causal profiles. These cases are common when an act causally affects an extensive system. For example, a socioeconomic policy affects diverse citizens, an educational program affects numerous students, a business decision affects countless customers, and an approved drug affects various patients. These complicated situations are not rare in decision problems.
In this article, I have identified the example of the spinner that underlines the importance of SCMs and SEU. In that example, the characterization of the causal effect of the act delivered by IEU and the characterization delivered by SEU diverge and the latter—not the former—seems intuitively correct. Moreover, the language of SCMs and SEU is richer than intervention analysis and IEU because SCMs and SEU enable the agent to make necessary mathematical statements that relate directly to various causal dispositions in the real world.Footnote 15 The theoretical implication of the spinner is that SEU is recommended in similar situations and that SEU might be a foundation for a more general decision theory.Footnote 16