1. Introduction
Philosophers, psychologists and economists have known for a long time that there are a number of decision problems for which the preferences over the available options that seem rational to many people cannot be accommodated within orthodox decision theory in the natural way. In response to this observation, a number of alternatives to the orthodoxy have been proposed (Allais Reference Allais1953; Quiggin Reference Quiggin1982; Schmeidler Reference Schmeidler1989; Buchak Reference Buchak2013). In this paper, I offer an argument against those alternatives and in favour of the orthodoxy. This argument is very general: it is effective against any deviation from the orthodoxy. As a result, we need some account of the preferences that seem rational and yet which this orthodoxy cannot accommodate naturally: we need an error theory for our intuition that they are rational, or a way of making the orthodoxy accommodate them. I will focus here on those preferences that seem to encode sensitivity to risk. And I will focus on the alternative to the orthodoxy proposed by Lara Buchak’s risk-weighted expected utility theory, which is intended to accommodate these preferences (Buchak Reference Buchak2013). I will show that, in fact, the orthodoxy can be made to accommodate the preferences in question; and I will argue that this is in fact the correct way to accommodate them. Thus, the paper has two parts: the first is a general objection to any non-expected utility theory; the second is a specific account of how to accommodate within the orthodoxy the preferences that Buchak’s theory permits.
2. The argument for orthodox expected utility theory
2.1. Decision problems and the framework of decision theory
Here’s a decision problem. An agent is choosing between two options: on the first, which we will call Safe, she is guaranteed to receive £50; on the second, which we will call Risky, a fair coin will be flipped and she will receive £100 if the coin lands heads and £0 if the coin lands tails. There are three components of this decision problem. First: the states of the world, namely, the state in which the coin lands heads and the state in which the coin lands tails. Second: the outcomes, namely, £0, £50, £100. Third: the acts between which our agent is choosing, namely, Safe and Risky. In general, a decision problem consists of these three components:
S is the set of states (or possible worlds). Thus, in our example, S = {Heads, Tails}. Degrees of belief or credences will be assigned to a finite algebra F of subsets of the set of states S. These are propositions represented as sets of states or possible worlds.Footnote 1
χ is the set of outcomes. Thus, in our example, S = {£0,£50,£100}. We will take outcomes to be entire descriptions of the world, rather than merely changes in the agent’s wealth. Thus, £0 is really Status quo + £0; £50 is really Status quo + £50; and so on. But we will continue to denote them just by the change in the status quo that they represent. Utilities will be assigned to the elements of χ.
A is the set of acts. Thus, in our example, A = {Safe,Risky}. We represent acts as finite-valued functions from S to χ. That is, they take ways the world might be and return the outcome of the act if the world were that way. Thus, for our purposes, we can represent an act f in the set of acts A using the following notation:
, where
are the values that the function f might take – that is, the possible outcomes of the acts – and, for each
, the proposition
says that f will have outcome
. Thus, if we represent propositions as sets of possible worlds, as we did above,
. So
is the set of states of the world in which f has outcome
. Thus, in our example above,
and
. We assume that all such propositions
are in the algebra
. For each outcome x in
, there is an act in
– which we write
– that has outcome x regardless of the state of the world. That is, representing the act
as a function from states to outcomes,
for all states s in
. We call
the constant act on x. Let
be the set of constant acts. They will prove particularly important in Section 5.2 below.
2.2. The business of decision theory
With this framework in place, we can state the business of decision theory. It is concerned with the relationship between two sorts of attitudes, which I will call external attitudes and internal attitudes.Footnote 2 The external attitudes are typically taken to be represented by the agent’s preference ordering .
is an ordering on the set
of acts. If f and g are acts in
, we say that
if the agent weakly prefers act f to act g. The internal attitudes, on the other hand, are typically taken to be given by the agent’s credences and her utilities. As mentioned above, her credences are defined on propositions in a
-algebra
on the set of states
. They measure how strongly she believes those propositions. And her utilities are defined on the outcomes in
. They measure how strongly she desires or values those outcomes.Footnote 3 If you are a constructivist about the internal attitudes, then you will take only the external attitudes to be real: you will then take the business of decision theory to be the representation of the external attitudes by treating the agent as if she has internal attitudes and as if she combines those attitudes in a particular way to give her external attitudes. If, on the other hand, you are a realist about the internal attitudes, then you will take both sorts of attitudes to be real: you will then say that a rational agent’s internal and external attitudes ought to relate in a particular way; indeed, they ought to relate as if she obtains her external attitudes by combining her internal attitudes in a particular way. We will have more to say about the business of decision theory later (cf. Section 5.2 below).
2.3. The EU rule of combination
Expected utility theory posits only two types of internal attitudes: these are given by the agent’s credences and utilities. Her credences are given by a credence function , which we assume to be a probability function on
. Her utilities are given by a utility function
. As with most decision theories, expected utility theory posits one type of external attitude, namely, the agent’s preference ordering. Expected utility theory then employs the following rule of combination, which states how her internal and external attitudes ought to relate:
EU Rule of Combination Suppose is an act in
– that is, if
is true, the outcome of f is
. Then define
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0001.png?pub-status=live)
Then if the agent is rational, then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0002.png?pub-status=live)
That is, an agent’s preferences ought to order acts by their subjective expected utility.
A number of decision theorists wish to deny the EU Rule of Combination. Buchak is amongst them, but there are other so-called non-expected utility theorists (Allais Reference Allais1953; Quiggin Reference Quiggin1982; Schmeidler Reference Schmeidler1989). I disagree with them about the rule of combination; however, as we will see in the second half of this paper, I agree with them about the rationality of the preferences that they try to capture by reformulating the rule of combination. In Section 4, I try to effect a reconciliation between these two positions – the correctness of the EU Rule of Combination, on the one hand, and the rationality of risk-sensitive preferences, on the other. In this part of the paper, I wish to argue that we ought to combine our internal attitudes in exactly the way that expected utility theory suggests. That is, I want to argue for the EU Rule of Combination.
How can we tell between different rules of combination? It is commonly assumed that representation theorems help us to do this, but this is a mistake. A representation theorem presupposes a rule of combination. Relative to a particular rule of combination, it demonstrates that, for any agent whose preferences satisfy certain axioms, there are internal attitudes with certain properties (unique to some extent) such that these internal attitudes determine the preferences in line with that rule of combination. As many authors have emphasized, given a different rule of combination, there will often be different internal attitudes with different properties that determine the same preferences, but this time in line with this different rule of combination (Zynda Reference Zynda2000; Eriksson and Hájek Reference Eriksson and Hájek2007; Meacham and Weisberg Reference Meacham and Weisberg2011).
While both constructivists and realists must appeal to rules of combination, my argument for the EU Rule of Combination applies primarily to the realist. I attempt to show that, for an agent with a credence function of a certain sort and a utility function, they are irrational if they don’t combine those two functions in a particular way and set their preferences in line with that way of combining them. It is directed at an agent whose credence function and utility function have a psychological reality beyond her being represented as having them. Thus, it does not apply to the constructivist, who thinks of the credence function and utility function as merely convenient mathematical ways of representing the preference ordering.
2.4. Estimates and the EU rule of combination
Finally, we come to state our argument in favour of the EU Rule of Combination. It draws on a mathematical result due to Bruno de Finetti, which we present as Theorem1 below.
(EU1) A rational agent will weakly prefer one option to another if, and only if, her estimate of the utility of the first is at least her estimate of the utility of the second.
(EU2) A rational agent’s estimate of a quantity will be her subjective expectation of it.
Therefore,
(EUC) A rational agent’s preference numbering
will be determined by the EU Rule of Combination.
The first premise (EU1) is supposed to be intuitively plausible. Suppose I desire only chocolate – obtaining as much of it as possible is my only goal. And suppose my estimate of the quantity of chocolate in the wrapper on my left is greater than my estimate of the quantity of chocolate in the wrapper on my right. And suppose that, nonetheless, I weakly prefer choosing the chocolate in the wrapper on my right. Then I would seem irrational. Likewise, if I desire only utility – surely an analytic truth if there are any – then I would seem irrational if my estimate of the utility of an act g were higher than my estimate of the utility another act f and yet I were to weakly prefer f to g. This is the argument for premise (EU1).
The second premise (EU2) is based on a mathematical argument together with a plausible claim about the goodness of estimates. Estimates, so the plausible claim goes, are better the closer they are to the true quantity they estimate. Indeed, we might take this to be an analytic truth. That is, we might take it to be a necessary condition on something being an estimate of a quantity that it is valued for its proximity to the actual value of that quantity. Thus, if I estimate that the amount of chocolate remaining in my cupboard is 73 g and my friend estimates that it is 79 g and in fact it is 80 g, then her estimate is better than mine. The mathematical argument is a generalization of a result due to de Finetti, which says, very roughly, that if an agent has estimates that are not expectations of the quantities that they estimate, there are alternative estimates of those quantities that are guaranteed to be closer to the true values of the quantities; so estimates that aren’t expectations are irrational.
Let’s make all of this precise. Suppose X is a quantity. Mathematically, we might understand this as a random variable, which is a function that takes a state of the world s and returns X(s), which is the value that this quantity takes in this state of the world. Thus, if C is the quantity of chocolate in my cupboard in grams, and @ is the actual state of the world, where there is 80 g of chocolate in my cupboard, then . Now, given what we said above about the goodness of estimates, we can measure the badness or disvalue of an estimate e(X) of a quantity X given a state of the world s by the distance between e(X) and X(s). Now, I will focus on just one measure of distance here, for the sake of simplicity, but the result also holds of a wide range of distance measures that mathematicians call the Bregman divergences.Footnote 4 Having said that, in Section 2.5, I will offer a reason to prefer the measure of distance I use here to all other measures. The measure of distance between two numbers x and y that I will use here is the square of their difference
, which is itself a Bregman divergence. For obvious reasons, we call this the quadratic distance measure and we write it
. Thus, relative to this measure of distance, the badness of an estimate e(X) of a quantity X given a state of the world s is
.
Now, in the argument we wish to give for (EU2), we are interested in evaluating the goodness or badness not only of a single estimate of a single quantity, but also of a set of estimates of a set of quantities. So our next job is to say how we measure this. Suppose is a set of quantities for which our agent has estimates. One of these quantities might be the quantity of chocolate in my cupboard, for instance; another might be the quantity of chocolate in my hand; another still might be the distance between my house and the nearest chocolate shop; and so on. And suppose that e is her estimate function – that is, e takes each quantity X in
and returns our agent’s estimate e(X) of that quantity. Then we will measure the badness of an estimate function at a state of the world by adding together the badness of each of the individual estimates that comprise it. Thus, the badness of e at the state of the world s is the sum of each
for each X in
. So the badness of an estimate function e defined on the quantities in
at a state of the world s is
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0003.png?pub-status=live)
With these notions defined, we have nearly defined everything that we need for our argument for (EU2). But there is one final observation to make. Consider our credences. It seems natural to say that my credence in a true proposition is better, epistemically speaking, the closer it is to the maximal credence, which is 1. And it seems natural to say that my credence in a false proposition is better, epistemically speaking, the closer it is to 0. That is, our credence in a proposition can be seen as an estimate of a particular quantity, namely, the quantity that takes value 1 if the proposition is true and value 0 if the proposition is false (Jeffrey Reference Jeffrey1986; Joyce Reference Joyce1998). Given a proposition A, call this the indicator quantity for A: thus, A(s) = 0 if A is false at s; A(s) = 1 if A is true at s.
Now, suppose that our agent has credences in all propositions in a finite algebra . Let’s say that her credence function is the function that assigns to each of these propositions her credence in it. Then the observation that a credence in a proposition is, or at least should be evaluated as if it is, an estimate of the indicator quantity for that proposition suggests that the badness of a credence function c at a state of the world s should be given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0004.png?pub-status=live)
That is, it is the sum of the distance between each credence, c(A), and the indicator quantity, A, corresponding to the proposition to which the credence is assigned.
Finally, we can say that an agent with a credence function c defined on the propositions in the finite algebra and an estimate function e defined on a finite set of quantities
should be evaluated at a state of the world s as follows: her badness is given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0005.png?pub-status=live)
That completes our account of how badly an agent is doing who has credences in propositions in and certain estimates in quantities in
.
Next, we turn to what we might show using this account. Let’s say that our agent’s credence function c defined on finite algebra is probabilistic if
(i) (Range)
for all A in
.
(ii) (Normalization)
, where
is the tautologous proposition; that is, it is true at all states of the world.
(iii) (Additivity)
if A and B are mutually exclusive propositions; that is, A and B are not true together at any state of the world.
Now suppose that c is probabilistic. Then we say that the estimate function e defined on is expectational with respect to c if
(iv) (Expectation) For each X in
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0006.png?pub-status=live)
So an estimate function is expectational with respect to a probabilistic credence function if its estimate of each quantity is the weighted sum of the possible values of that quantity where the weights are given by the credence assigned to the relevant state of the world by the credence function. We say that a pair (c, e) is probabilistic and expectational if c is probabilistic and e is expectational with respect to c – that is, if they jointly satisfy (i)–(iv).
For instance, suppose there are just two states of the world, and
. And let C be the quantity of chocolate in my cupboard in grams. Let’s suppose that, in state
, there is a meagre 80 g of chocolate in my cupboard (so
), whereas in state
, there is veritable bounty, namely, 1000 g (so
). And suppose that, having resisted the temptation to indulge in wishful thinking, I have credence
in state
and
in state
. Then my credences are probabilistic (since they sum to 1), and my estimate of C is expectational with respect to my credences just in case it is
.
In number to establish (EU2), the second premise of the argument for the EU Rule of Combination, we need to show that it is a requirement of rationality that an agent have a probabilistic credence function and an estimate function that is expectational with respect to it. Our argument is based on the following mathematical theorem, which is due to de Finetti (Reference de Finetti1974, 136).
Theorem 1 (de Finetti) Suppose c is a credence function on and e is an estimate function on
.
(i) If (c,e) is not probabilistic and expectational, then there is another pair
that is probabilistic and expectational such that
(ii) If (c,e) is probabilistic and expectational, then there is no other pair
such that
Thus, if an agent either has a credence function that is not a probability function, or if her credence function is a probability function but her estimates of quantities are not all given by her expectations of those quantities relative to that credence function, then there are alternative credences and estimates that, taken together, will be less bad than her credences and estimates taken together; that is, the alternative credences and estimates will be closer to the quantities that they estimate however those quantities turn out to be. What’s more, if her credence function is a probability function, and if her estimates are given by her expectations, then there are no alternative credences and estimates that are guaranteed to be better than hers; indeed, there are no alternative credences and estimates that are guaranteed to do at least as well as hers. I provide a proof of this result in the Appendix.
This gives us an argument for having credences that are probabilities and estimates that are expectations. If we fail to do this, the theorem says, there are alternative credences and estimates we might have had that are guaranteed to do better than our credences; and there is nothing that is guaranteed to do better than those alternatives; indeed, there is nothing else that is even guaranteed to do just as well as them. Compare: I am offered two gambles on a fair coin toss. On the first, if the coin lands heads, I receive £5; if it lands tails, I lose £6. On the second, if the coin lands heads, I receive £10; if it lands tails, I lose £3. Now suppose I choose the first gamble. You would charge me with irrationality. After all, the second is guaranteed to be better than the first; whether the coin comes up heads or tails, I’ll end up with more money if I’ve taken the second gamble. We are using a similar piece of reasoning here to argue that an agent is irrational if she has credences that are not probabilities, or if she has credences that are probabilities, but her estimates are not expectations with respect to them. That is, we are appealing to the so-called Dominance Principle, which says that an option is irrational if there is an alternative that is guaranteed to be better than it, and if there is nothing that is guaranteed to be better than that alternative. This completes our justification of the second premise (EU2) of our argument for the EU Rule of Combination.
You might worry here that, in the preceding justification of (EU2), we appeal to one principle of rational choice in number to justify another: we are appealing to the Dominance Principle in number to establish the EU Rule of Combination. And of course we are. But that is permissible in this context. After all, the Dominance Principle is an uncontroversial principle of decision theory. It is agreed upon by all parties to the current debate. Buchak and all other non-expected utility theorists disagree with me and other expected utility theorists about how credences and utilities should combine to give preferences. But we all agree that if one option is guaranteed to be better than another, and there is nothing that is guaranteed to be better than the first, then the second is irrational. So the argument strategy is legitimate.
Having given our justification for (EU2), this completes our argument for the EU Rule of Combination. According to the first premise (EU1), our preference numbering over acts should match our estimates of the utility of those acts: that is, I should weakly prefer one act to another iff my estimate of the utility of the first is at least as great as my estimate of the utility of the second. According to the second premise (EU2), our estimate of a given quantity, whether it is the utility of an act or the mass of chocolate in my fridge, should be our expectation of that quantity; that is, it should be the weighted average of the possible values that that quantity might take where the weights are given by our credences in the relevant states of the world. Putting these together, we obtain the EU Rule of Combination.
2.5. Measuring the badness of estimates
Before we leave our argument for the EU Rule of Combination, it is worth noting two things about the distance measure that we used to give the badness of an estimate of a given quantity at a given state of the world, and the function
that we used to give the badness of a set of estimates in a set of quantities at a given world. First, as noted above, Theorem1 holds for a wide range of alternative measures of distance; indeed, for any of the so-called Bregman divergences. However, second, it is also true that Theorem1 fails for a wide range of alternative measures; indeed, it fails for the so-called absolute value measure
, which takes the distance between real numbers x and y to be the difference between them, that is,
. Thus, this argument will be compelling to the extent that we can justify using the quadratic measure
, or some other Bregman divergence, instead of the absolute value measure
. Arguments have been given for this assumption in the case where we are measuring only the badness of credences (D’Agostino and Sinigaglia Reference D’Agostino, Sinigaglia, Suárez, Dorato and Rédei2010; Leitgeb and Pettigrew Reference Leitgeb and Pettigrew2010; Pettigrew ta Reference Pettigrew2016). In this context, it looks most promising to extend the argument of D’Agostino and Sinigaglia (Reference D’Agostino, Sinigaglia, Suárez, Dorato and Rédei2010). The arguments of Leitgeb and Pettigrew (Reference Leitgeb and Pettigrew2010) assume too much, and the argument of Pettigrew (ta) is too closely bound to the case of credences.
Above, we assumed that we begin with a measure of the distance between a single estimate e(X) of a single quantity X and the true value X(s) of X at a state of the world s; and then we measure the distance between an entire estimate function e defined on a set
of quantities, on the one hand, and the true values of those quantities at a state of the world s, on the other hand, by summing the distances,
, between each e(X) and X(s) for X in
. If we adapt the argument given by D’Agostino and Sinigaglia (Reference D’Agostino, Sinigaglia, Suárez, Dorato and Rédei2010), we do not make this assumption: instead, we justify it. That is, we assume that the badness of an estimate function e defined on
at a state of the world s is given by some function
, and we lay down conditions on this function such that, if
satisfies all of the conditions, then there is a continuous and strictly increasing function
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0009.png?pub-status=live)
Since we appealed to the Dominance Principle in number to justify the EU Rule of Combination, and since the Dominance Principle pays attention only to the numbering of options at a world, rather than their cardinal utilities at that world, it does not matter whether we use the sum of the squared differences between the values or whether we use some strictly increasing transformation of that sum. Thus, this characterization of is sufficient for our purposes.
Here are the conditions that D’Agostino and Sinigaglia (Reference D’Agostino, Sinigaglia, Suárez, Dorato and Rédei2010) place on our measure of the badness of an estimate function.Footnote 5
Extensionality Let us say that the estimate profile of e at s is the multiset of all pairs (e(X),X(s)) for X in – that is,
.Footnote 6 Then, if e has the same estimate profile at s as
has at
, then
.
That is, the badness of your estimate function is a function only of its estimate profile. It does not depend on the particular quantities to which you assign estimates. If you and I assign estimates to very different quantities, but our estimate profiles match, then our estimates are exactly as bad as each other.
Accurate Extension Invariance If e is an estimate function on and
, then let
be the estimate function on
that agrees with e on all quantities in
–
is sometimes called the restriction of e to
. Then, if the estimates that e assigns to quantities not in
are equal to the true values of those quantities at s, then
.
That is, adding perfectly accurate estimates to your estimate function does not affect its badness.
Difference Supervenience If e assigns an estimate to just one quantity X, then for some continuous and strictly increasing function
.
That is, the badness of a single estimate is a continuous and strictly increasing function of the difference between that estimate and the true value of the quantity.
Separability If and
(i)
and
(ii)
,
then .
That is, if two estimate functions are equally bad on some subset of the quantities to which they assign estimates, then one is worse than the other if it is worse on the remaining quantities.
Taken together, these four properties entail that there are continuous and strictly increasing functions and
such that:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0010.png?pub-status=live)
Indeed, these four conditions are equivalent to the existence of two such functions H and f. Thus, what we need for our conclusion is a further property that ensures that . That is the job of the final condition on
. To state it, we need the notion of an number-reversing swap. Suppose e is an estimate function defined on a set of quantities
and s is a state of the world. And suppose that the estimates that e assigns to quantities X and Y are numbered correctly at s. That is,
(i)
and
or
(ii)
and
.
Then, if we define to be the estimate function that is obtained from e by swapping its estimates for X and Y – so
and
– then we say that
is an number-reversing swap of e, since the estimates that
assigns to quantities X and Y are numbered incorrectly at s. Our next condition says two things: first, it says that an number-reversing swap always increases the badness of the estimates; second, it says that if you compare two number-reversing swaps on the same estimate function and if (a) the two swaps involve swapping estimates that are themselves equally far apart and (b) the two swaps involve quantities whose true values are equally far apart, then the badness of the swaps is equal. The motivation for this condition is the following: The badness of a set of estimates is supposed to be determined by the extent to which they match the truth about the quantities that they estimate. As well as matching the quantitative facts about those quantities – such as their values – it also seems to be a good thing to match the qualitative facts about them – such as their numbering. Clearly e matches this qualitative fact for X and Y, whereas
does not. Thus, other things being equal,
is worse than e. And other things are equal, since all that has changed in the move from e to
is that the quantities to which the estimates e(X) and e(Y) are assigned have been switched. Moreover, if we consider two possible number-reversing swaps on the same estimate function where (a) and (b) hold, then the effect of each swap on the badness of the estimate function should be the same, since there is nothing to tell between them.
The Badness of Number-Reversing Swaps Suppose e is defined on
and suppose
are quantities in
. And suppose e(X),e(Y) are numbered as X(s), Y(s) are; and
are numbered as
are. And suppose
and
. Then
.
We now have the following theorem:
Theorem 2 (D’Agostino & Dardanoni) If satisfies Extensionality, Accurate Extension Invariance, Difference Supervenience, Separability and The Badness of Number-Reversing Swaps, then there is a continuous and strictly increasing functions
such that:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0011.png?pub-status=live)
This gives us what we need.
3. Expected utility and risk-weighted expected utility
In the previous section, we gave our defence of the EU Rule of Combination. In this section, we describe Lara Buchak’s proposed alternative. To do this, we’ll illustrate the difference between expected utility and risk-weighted expected utility using a particular act as an example. We’ll first describe the expected utility of that act, and then we’ll show how to define its risk-weighted expected utility. Our example is the following act: . The agent’s probabilistic credences over the events
and her utilities for the outcomes
are given as follows:
In Figure 1, we exploit a useful diagrammatic way of representing the expected utility of h, which is used by Quiggin (Reference Quiggin1993), Wakker (Reference Wakker2010) and Buchak (Reference Buchak2013).
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_f0001_oc.png?pub-status=live)
Figure 1. The expected utility of h is given by the grey area.Note: It is obtained by summing the areas of the four vertical rectangles: working from right to left, their areas are
.
Figure 1 suggests two ways in which we might reformulate . These will be very useful in understanding how expected utility theory relates to Buchak’s proposal.
First, it is clear that
depends only on the utilities of the outcomes to which the act f may give rise and the probabilities that f will produce outcomes with those utilities. Thus, given an act
and a utility function u, we might redescribe f as
where
are the utilities to which f might give rise numbered from least to greatest – that is,
. For instance, in our example act h:
,
,
.
is the proposition that f will give rise to
. Thus,
. For instance, in our example act h:
,
,
.
The second reformulation of
builds on this first and is illustrated in Figure 3. Suppose
is the numbered utility-based description of f relative to u. Then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_f0002_oc.png?pub-status=live)
Figure 2. Again, the expected utility of h is given by the grey area.Note: It is obtained by summing the areas of the three vertical rectangles (the middle two vertical rectangles from Figure 1 have been merged): working from right to left, their areas are
.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_f0003_oc.png?pub-status=live)
Figure 3. Once again, the expected utility of h is given by the grey area.Notes: It is obtained by summing the areas of the three horizontal rectangles. Working from bottom to top, their areas are
,
and
.
With this in hand, we’re ready to formulate Buchak’s alternative to expected utility theory. Buchak is motivated by the apparent rationality of risk-sensitive behaviour. Notoriously, some seemingly rational risk-sensitive behaviour cannot be captured by expected utility theory at all: for instance, Allais described seemingly rational preferences that cannot be generated by any rational credence function and utility function in the way prescribed by expected utility theory (Allais Reference Allais1953). Moreover, there are other seemingly rationally preferences that can be generated by a credence function and utility function in line with expected utility theory, but which seem to be rational even for agents who do not have credences and utilities that would generate them in this way. Thus, for instance, consider the two acts described in the introduction to this article: and
. Suppose that our agent strictly prefers Safe to Risky: that is,
. Can expected utility theory capture the rationality of this preference? Suppose that, since the coin is known to be fair, rationality requires that the agent assigns credences to the two states of the world as follows:
. Then it is still possible to describe a utility function on the outcomes
,
,
that generates these preferences in the way expected utility theory requires. Let
and
. That is, suppose the agent treats money as a dependent good: how much utility it gives depends on how much of it she has already; so, money has diminishing marginal utility for this agent. Then, for an agent with this credence function and utility function,
, as required. So expected utility theory can capture the rationality of these preferences. However, as Buchak rightly observes, those preferences – that is,
– seem rational not only for an agent for whom money has diminishing marginal utility; they seem rational even for an agent whose utility is linear in money. And this is something that expected utility cannot capture. Thus, Buchak is interested not only in saving the Allais preferences, but also in saving other risk-sensitive behaviour without attributing the risk-sensitive behaviour to the shape of the utility function (Buchak Reference Buchak2013,Chapter 1).
How does Buchak hope to capture these risk-sensitive preferences? Where expected utility theory countenances only two types of internal attitude as relevant to preferences, Buchak countenances a third as well: this third component is supposed to capture the agent’s attitude to risk, and it is given by a function , which Buchak assumes to be strictly increasing, continuous, and taking the following values,
and
(Buchak Reference Buchak2013,Section 2.2). Buchak’s risk-weighted expected utility theory then employs the following rule of combination, which states how an agent’s internal and external attitudes ought to relate, where the agent has credence function c, utility function u and risk function r:
REU Rule of Combination (Buchak Reference Buchak2013, 53) Suppose is the numbered utility-based description of act f relative to utility function u. Then let
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0014.png?pub-status=live)
Then if the agent is rational, then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0015.png?pub-status=live)
In Figure 4, we illustrate the risk-weighted expected utility of our example act h when the agent has the risk function . Notice that the formulation of
is exactly like the formulation of
that we gave above except that each probability weight is transformed by the agent’s risk function. Thus, if
(for all
), then, as Figure 4 illustrates, the lowest utility to which the act can give rise – namely,
– contributes just as much to
as it does to
– it contributes
to both. But further increases in utility – such as the increase from getting at least utility
to getting at least
– make less of a contribution since their probability –
– is acted on by the risk function, and it is this reduced value –
– that weights the possible increases in utility. Thus, such an agent displays risk-averse behaviour.
is such a risk function.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_f0004_oc.png?pub-status=live)
Figure 4. The risk-weighted expected utility of h is given by the grey area, where
.
Similarly, if (for all
), then the lowest utility to which the act can give rise contributes just as much to
as it does to
, but further increases in utility make more of a contribution since their probability is acted on by the risk function and it is this increased value that weights the possible increases in utility. This is illustrated in Figure 5. Such an agent displays risk-seeking behaviour.
is such a risk function.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_f0005_oc.png?pub-status=live)
Figure 5. The risk-weighted expected utility of h is given by the grey area, where
.
It’s also easy to see that, if (for
), then
. Thus, expected utility theory is the special case of risk-weighted expected utility theory given by a linear risk function. In such a situation, we say that the agent is risk-neutral. This means that Buchak’s theory permits any preferences that expected utility theory permits. But it also permits a whole lot more. For instance, one can easily recover the Allais preferences or the preference
described above by attributing to an agent a certain sort of risk function – in both cases, a risk-averse risk function.
This, then, is Buchak’s proposal.
4. Redescribing the outcomes
Moving from expected utility theory to risk-weighted expected utility theory involves an agent evaluating an act in the way illustrated in Figure 3 to evaluating it in the way illustrated in Figure 4. In number to begin to see how we can redescribe the REU rule of combination as an instance of the EU rule of combination, we reformulate the REU rule in the way illustrated in Figure 6.Footnote 7 Thus, we can reformulate as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0016.png?pub-status=live)
And we can reformulate this as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0017.png?pub-status=live)
since .
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_f0006_oc.png?pub-status=live)
Figure 6. Again, the risk-weighted expected utility of f is given by the grey area, where
.
Now, suppose we let
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0018.png?pub-status=live)
Then we have:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0019.png?pub-status=live)
Reformulating Buchak’s rule of combination in this way suggests two accounts of it. On the first, utilities attach ultimately to outcomes , and they are weighted not by an agent’s credences but rather by a function of those credences that encodes the agent’s attitude to risk (given by a risk function). On this account, we group
together to give this weighting. Thus, we assume that this weighting has a particular form: it is obtained from a credence function c and a risk function r to give
; this weighting then attaches to
to give
. This is the account that Buchak favours.
On the second account, credences do provide the weightings for utility, as in the EU rule of combination, but utilities attach ultimately to outcome-act pairs . On this account, we group
together to give this utility; this utility is then weighted by
to give
. That is, we say that an agent’s utility function is defined on a new outcome space: it is not defined on a set of outcomes
, but on a particular subset of
, which we will call
.
is the set of outcome-act pairs
such that
is a possible outcome of f: that is,
. Now, just as the first account assumed that the weightings of the utilities have a certain form – namely, they are generated by a risk function and probability function in a certain way – so this account assumes something about the form of the new utility function
on
: we assume that a certain relation holds between the utility that
assigns to outcome-act pairs in which the act is the constant act over the outcome and the utility
to outcome-act pairs in which this is not the case. We assume that the following holds:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn10001.png?pub-status=live)
If a utility function on satisfies this property, we say that it encodes attitudes to risk relative to risk function r. Thus, on this account an agent evaluates an act as follows:
She begins with a risk function r and a probability function c.
She then assigns utilities to all constant outcome-act pairs
, defining
on
, where
.
Finally, she extends
to cover all outcome-act pairs in
in the unique way required in number to make
a utility function that encodes attitudes to risk relative to r. That is, she obtains
by weighting
in a certain way that is determined by her probability function and her attitudes to risk.
Let’s see this in action in our example act h; we’ll consider h from the point of view of two risk functions, and
. Recall:
is a risk-averse risk function;
is risk seeking. We begin by assigning utility to all constant outcome-act pairs
:
Then we do the same trick as above and amalgamate the outcome-act pairs with the same utility: thus, again, is the event in which the act gives outcome-act pair
,
is the event in which it gives
or
, and
the event in which it gives
. Next, we assign utilities to
,
,
and
in such a way as to make
encode attitudes to risk relative to the risk function r.
Let’s start by considering the utility of , the lowest outcome of h. Suppose our risk function is
; then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0020.png?pub-status=live)
And now suppose our risk function is ; then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0021.png?pub-status=live)
Thus, the risk-averse agent – that is, the agent with risk function – values this lowest outcome
as the result of h more than she values the same outcome as the result of a certain gift of
, whereas the risk-seeking agent – with risk function
– values it less. And this is true in general: if
for all x, the utility of the lowest outcome as a result of h will be more valuable than the same outcome as a result of the constant act on that outcome; if
it will be less valuable.
Next, let us consider the utility of , the highest outcome of h. Suppose her risk function is
; then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0022.png?pub-status=live)
And now suppose her risk function is ; then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0023.png?pub-status=live)
Thus, the risk-averse agent – that is, the agent with risk function – values this highest outcome
as the result of h less than she values the same outcome as the result of a certain gift of
, whereas the risk-seeking agent – with risk function
– values it more. And, again, this is true in general: if
for all x, the utility of the highest outcome as a result of h will be less valuable than the same outcome as a result of the constant act on that outcome; if
it will be more valuable.
This seems right. The risk-averse agent wants the highest utility, but also cares about how sure she was to obtain it. Thus, if she obtains from h, she knows she was guaranteed to obtain at least this much utility from h or from
(since
is the lowest possible outcome of each act). But she also knows that h gave her some chance of getting more utility. So she values
more than
. But if she obtains
from h, she knows she was pretty lucky to get this much utility, while she knows that she would have been guaranteed that much if she had obtained
from
. So she values
less than
. And similarly, but in reverse, for the risk-seeking agent.
Finally, let’s consider the utilities of and
, the middle outcomes of h. They will have the same value, so we need only consider the utility of
. Suppose her risk function is
; then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0024.png?pub-status=live)
Thus, again, the agent with risk function assigns higher utility to obtaining
as a result of h than to obtaining
as the result of
. But this is not generally true of risk-averse agents. Consider, for instance, a more risk-averse agent, who has a risk function
. Then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0025.png?pub-status=live)
Again, this seems right. As we said above, the risk-averse agent wants the highest utility, but she also cares about how sure she was to obtain it. The less risk-averse agent – whose risk function is – is sufficiently sure that h would obtain for her at least the utility of
and possibly more that she assigns higher value to getting
as a result of h than to getting it as a result of
. For the more risk-averse agent – whose risk function is
– she is not sufficiently sure. And reversed versions of these points can be made for risk-seeking agents with risk functions
and
, for instance. Thus, we can see why it makes sense to demand of an agent that her utility function
on
encodes attitudes to risk relative to a risk function in the sense that was made precise above – see Equation 1.
Since what we have just provided is a genuine redescription of Buchak’s REU Rule of Combination, we can see that Buchak’s representation theorem is agnostic between a version of REU in which utilities attach to elements of , and a version of EU in which utilities attach to elements of
.
Theorem 3 (Buchak) If satisfies the Buchak axioms, there is a unique probability function c, unique risk function, and unique-up-to-affine-transformation utility function u on
such that
is determined by r, c and u in line with the REU rule of combination.
And we have the following straightforward corollary:
Theorem 4 If satisfies the Buchak axioms, there is a unique probability function c and unique-up-to-affine
-transformation utility function
on
that encodes attitudes to risk relative to a risk function such that
is determined by c and
in line with the EU rule of combination (where
is unique-up-to-affine
-transformation if
is unique-up-to-affine-transformation).
Thus, by redescribing the set of outcomes to which our agent assigns utilities, we can see how her preferences in fact line up with her estimates of the utility of her acts, as required by the de Finetti-inspired argument for the EU Rule of Combination given in the previous section.
5. What’s wrong with redescription?
Although Buchak does not address precisely this particular version of the redescription strategy, she does consider others nearby. Against those, she raises what amount to two objections (Buchak Reference Buchak2013,Chapter 4). (Buchak raises a further objection against versions of the redescription strategy that attempt to identify certain outcome-act pairs to give a more coarse-grained outcome space; but these do not affect my proposal.)
5.1. The problem of proliferation
One potential problem that arises when one moves from assigning utilities to to assigning them to
is that an element in the new outcome space is never the outcome of more than one act: (x,f) is a possible outcome of act f but not of any act g other than f. Thus, this outcome never appears in the expected utility (or indeed risk-weighted expected utility) calculation of more than one act. The result is that very few constraints are placed on the utilities that must be assigned to these new outcomes and the probabilities that must be assigned to the propositions in number to recover a given preference numbering on the acts via the EU (or REU) rule of combination. Suppose
is a preference numbering on
. Then, for each act f in
, pick a real number
such that
iff
. Now there are many ways to do this, and they are not all affine transformations of one another – indeed, any strictly increasing
will take one such assignment to another. Now pick any probability function c on
. Now, given an act
, the only constraint on the values
,
,
is that
. And this of course permits many different values.Footnote 8 Buchak dubs this phenomenon belief and desire proliferation (Buchak Reference Buchak2013, 140).
Why is this a problem? There are a number of reasons to worry about belief and desire proliferation. There is the epistemological worry that, if utilities and probabilities are as loosely constrained as this, it is not possible to use an agent’s observed behaviour to predict her unobserved behaviour. Divining her preferences between two acts will teach us nothing about the utilities she assigns to the outcomes of any other acts since those outcomes are unique to those acts. Also, those who wish to use representation theorems for the purpose of radical interpretation will be concerned by the complete failure of the uniqueness of the rationalization of preferences that such a decision theory provides.
Neither of these objections seems fatal to me. But in any case, the version of the redescription strategy presented here avoids them altogether. The reason is that I placed constraints on the sort of utility function an agent can have over
: I demanded that
encode attitudes to risk; that is,
is defined in terms of
in a particular way (given by Equation 1). And we saw in Theorem4 above that, for any agent whose preferences satisfy the Buchak axioms, there is a unique probability function c and a unique utility function
on
that encodes attitudes to risk relative to a unique risk function such that together c and
generate the agent’s preferences in accordance with the EU Rule of Combination.
5.2. Ultimate ends and the locus of utility
Buchak’s second objection initially seems more worrying (Buchak Reference Buchak2013, 137–138). A theme running through Risk and Rationality is that decision theory is the formalization of instrumental or means-end reasoning. One consequence of this is that an account of decision theory that analyses an agent as engaged in something other than means-end reasoning is thereby excluded.
Buchak objects to the redescription strategy on these grounds. According to Buchak, to understand an agent as engaged in means-end reasoning, one must carefully distinguish the means and the ends: in Buchak’s framework, the means are the acts and the ends are the outcomes. One must then assign utilities to the ends only. Of course, in terms of these utilities and the agent’s probabilities and possibly other representations of internal attitude such as the risk function, one can then assign value or utility to the means. But the important point is that this value or utility that attaches to the means is assigned on the basis of the assignment of utility to the ultimate ends. Thus, while there is a sense in which we assign a value or utility to means – i.e. acts – in expected utility theory, this assignment must depend ultimately on the utility we attach to ends – i.e. outcomes.
Thus, a first pass at Buchak’s second complaint against the redescription strategy is this: the redescription strategy assigns utilities to something other than ends – it assigns utilities to outcome-act pairs, and these are fusions of means and ends. Thus, an agent analysed in accordance with the redescription strategy is not understood as engaged in means-end reasoning.
However, this seems problematic in two ways. Whether they constitute ultimate ends or not, there are at least two reasons why an agent must assign utilities to outcome-act pairs rather than outcomes on their own. That is, there are two reasons why at least this part of the redescription strategy – namely, the move from to
– is necessary irrespective of the need to accommodate risk in expected utility theory.
Firstly, utilities must attach to the true outcomes of an act. But these true outcomes aren’t the sort of thing we’ve been calling an outcome here. When I choose Safe over Risky and receive , the outcome of that act is not merely
; it is
as the result of Safe. Thus, the true outcomes of an act are in fact the elements of
– they are what we have been calling the outcome-act pairs.
Of course, at this point, Buchak might accept that utilities attach to outcome-act pairs, but insist that it is nonetheless a requirement of rationality that an agent assign the same utility to two outcome-act pairs with the same act component; that is, ; that is, while utilities attach to fusions of means and ends, they must be a function only of the ends. But the second reason for attaching utilities to outcome-act pairs tells against this claim in general. The reason is this: As Bernard Williams urges, it is neither irrational nor even immoral to assign higher utility to a person’s death as a result of something other than my agency than to that same person’s death as a result of my agency (Williams and Smart Reference Williams and Smart1973). This, one might hold, is what explains my hesitation in a Williams-style example in which I must choose whether or not to shoot a particular individual when I know that, if I don’t shoot him, someone else will. I assign higher utility to the death of that person at the hands of someone else than to the death of that person at my hands. Thus, it is permissible in at least some situations to care about the act that gives rise to the outcome and let one’s utility in an outcome-act pair be a function also of that act.
Nonetheless, this is not definitive. After all, Buchak could reply that this is peculiar to acts that have morally relevant consequences. Acts such as those in the Allais paradox do not have morally relevant consequences; but the redescription strategy still requires us to make utilities depend on acts as well as outcomes in those cases. Thus, for non-moral acts f and g, Buchak might say, it is a requirement of rationality that , even if it is not such a requirement for moral cases. And this would be enough to scupper the redescription strategy.
However, it is not clear why the moral and non-moral cases should differ in this way. Consider again the Williams-style example from above: I must choose whether to shoot an individual or not; I know that, if I do not shoot him, someone else will. I strictly prefer not shooting him to shooting him. My reasoning might be reconstructed as follows: I begin by assigning a certain utility to this person’s death as the result of something other than my agency – natural causes, for instance, or murder by a third party. Then, to give my utility for his death at my hand, I weight this original utility in a certain way, reducing it on the basis of the action that gave rise to the death. Thus, the badness of the outcome-act pair (X’s death, My agency) is calculated by starting with the utility of another outcome-act pair with the same outcome component – namely, (X’s death, Not my agency) – and then weighting that utility based on the act component. We might call (X’s death, Not my agency) the reference pair attached to the outcome X’s death. The idea is that the utility we assign to the reference pair attached to an outcome comes closest to what we might think of as the utility that attaches solely to the outcome; the reference pair attached to an outcome x is the outcome-act pair (x,f) for which the act f contributes least to the utility of the pair.
Now this is exactly analogous to what the redescription strategy proposes as an analysis of risk-sensitive behaviour. In that case, when you wish to calculate the utility of an outcome-act pair (x,f), you begin with the utility you attach to . Then you weight that utility in a certain way that depends on the riskiness of the act. This gives the utility of (x,f). Thus, if we take
to be the reference pair attached to the outcome x, then this is analogous to the moral case above. In both cases, we can recover something close to the notion of utility for ultimate ends or pure outcomes (i.e. elements of
): the utility of the pure outcome x – to the extent that such a utility can be meaningfully said to exist – is
, the utility of the reference pair attached to x. That seems right. Strictly speaking, there is little sense to asking an agent for the utility they assign to a particular person’s death; one must specify whether or not the death is the result of that agent’s agency. But we often do give a utility to that sort of outcome; and when we do, I submit, we give the utility of the reference pair. Similarly, we often speak as if we assign a utility to receiving
, even though the request makes little sense without specifying the act that gives rise to that pure outcome: again, when we do so, what we really do is give the utility of
for sure, that is, the utility of
.
Understood in this way, the analysis of a decision given by the redescription strategy still portrays the agent as engaged in means-end reasoning. Of course, there are no pure ultimate ends to which we assign utilities. But there is something that plays that role, namely, reference pairs. An agent’s utility for an outcome-act pair (x,f) is calculated in terms of her utility for the relevant reference pair, namely, ; and the agent’s value for an act f is calculated in terms of her utilities for each outcome-act pair (x,f) where x is a possible outcome of f. Thus, though the value of an act on this account is not ultimately grounded in the utilities of pure, ultimate outcomes of that act, it is grounded in the closest thing that makes sense, namely, the utilities of the reference pairs attached to the pure, ultimate outcomes of the act.
6. Conclusion
Buchak proposes a novel decision theory. It is formulated in terms of an agent’s probability function on , utility function on
and risk function. It permits a great many more preference numberings than orthodox expected utility theory. On Buchak’s theory, the utility that is assigned to an act is not the expectation of the utility of its outcome; rather it is the risk-weighted expectation. But the argument of Section 2 of this paper suggests that the value of an act for an agent should be her estimate of the utility of its outcome; and her estimate of a quantity should be her expectation of that quantity. And these, together, give the EU Rule of Combination. In this paper, we have tried to reconcile the preferences that Buchak endorses with the EU Rule of Combination. To do this, we redescribed the outcome space so that utilities were attached ultimately to outcome-act pairs rather than to outcomes themselves. This allowed us to capture precisely the preferences that Buchak permits, whilst letting the utility of an act be the expectation of the utility it will produce. The redescription strategy raises some questions: Does it prevent us from using decision theory for certain epistemological purposes? Does it fail to portray agents as engaged in means-end reasoning? In Section 5, we tried to answer these questions.
Proof of theorem 1
In this appendix, we prove Theorem1. We begin by giving a geometric characterization of the pairs (c,e), where c is a credence function and e is an estimate function, such that c is probabilistic and e is expectational relative to c.
Lemma 5 Suppose c is a credence function defined on and e is an estimate function defined on
. Then the following two propositions are equivalent:
(i) c is probabilistic and e is expectational with respect to c.
(ii) For each state s, there is
such that
and
(a)
, for each proposition A in
;
(b)
, for each quantity X in
.
Proof 1 First, we prove (ii) (i). Suppose (ii). First, we show that c is probabilistic. Recall that there are three conditions on being probabilistic: Range, Normalization, Additivity. We take them each in turn.
Range: Suppose A is in
. Then, note that: (1) each
lies between 0 and 1 inclusive; (2) all of the
s summed together give 1; (3)
or 1 for each s in
. Thus, it is certainly true that
lies between 0 and 1 inclusive.
Normalization: Since
is true at all states of the world,
for all s in
, so
.
Additivity: If there are no states s at which both A and B are true, then
.
Next, we show that e is expectational with respect to c. Suppose is a state of the world. Then note that
. But of course, since the states of the world form a partition,
if
and
if
. Thus,
. Thus,
, as required. This gives Expectation.
Second, we prove (i) (ii). Let
and the result follows easily from Additivity and Expectation.
The upshot of this result is the following: Suppose and
. And, if c is a credence function on
and e is an estimate function on
, represent the pair (c,e) by the following vector in
:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0026.png?pub-status=live)
And represent a state of the world s by the following vector in :
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0027.png?pub-status=live)
Then Lemma5 says that (c, e) is probabilistic and expectational iff lies in the convex hull of the vectors
for s in
– that is,
.Footnote 9
The second lemma that we require to prove Theorem1 is a geometric fact about the following measure of distance between two vectors in a real-valued vector space. If and
are vectors in
, then let
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0029.png?pub-status=live)
Thus, clearly,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201019011345104-0294:S0045509100024127:S0045509100024127_eqn0030.png?pub-status=live)
for any credence function c, estimate function e and state of the world s.
Lemma 6 Suppose . Then
(i) If
, then there is
such that
, for all
.
(ii) If
, then there is no
such that
, for all
.
I won’t provide a full proof of these geometric facts – proofs can be found in any geometry textbook. But here is a brief sketch. (i) is an easy consequence of the Hilbert Projection Theorem, since is the square of the Euclidean metric. (ii) is a consequence of the fact that, if we measure distance between vectors using the Euclidean metric or its square,
, then, for any two vectors, the set of vectors that are closer to the first than to the second is a convex set.
Putting these two results together and, in Lemma6, letting , Theorem1 follows.