1. Introduction
A public official is deliberating about whether to approve a marginal increase in the speed limit for autonomous vehicles. Surveys show that this increase would improve passenger satisfaction, but would not lead to any other substantial improvements. Against this move, however, is evidence that increasing the speed limit poses some risk of increasing the current incidence of pedestrian deaths: some people crossing the road may underestimate the speed of the self-driven vehicles as they quietly shuttle through the streets. There are no other relevant considerations. The decision is hers alone.
The official takes a moment to reflect on the moral considerations at hand. One such consideration is public safety: she ought to choose the option that minimises pedestrian deaths. The other consideration is people's pleasure: increasing passenger satisfaction will increase the amount of pleasure in the world. Although both considerations are morally significant, to her mind they are not equally important. In fact, she believes that no amount of passenger satisfaction could ever morally justify a pedestrian's death. As far as she is concerned, considerations of public safety are the primary moral consideration at hand.
The primacy of public safety in this case can be spelled out more precisely using the notion of lexical priority.Footnote 1 A moral consideration has lexical priority over another just in case, given the choice between upholding a higher-ranked consideration versus upholding any number of lower-ranked ones, we ought to uphold the higher-ranked consideration. Lower-ranked considerations are only relevant insofar as they help to break ties between options that uphold higher-ranked considerations equally well. Lexical priority theories thus prohibit trade-offs between different kinds of moral considerations. In the official's case, considerations of passenger satisfaction ought not to be traded-off against pedestrians’ lives.
To her disappointment, however, the official realises that her awareness of these moral considerations is not sufficient to guide her decision. The problem, of course, is that she does not know whether increasing the speed limit will in fact lead to more deaths. Since she is to some degree uncertain about whether her choices will lead her to violate or uphold particular moral considerations, she faces a moral decision under risk. What she requires is a moral decision rule: a rule that identifies which options are morally permissible, given her degrees of uncertainty. The question is: is there a moral decision rule that both maintains the lexical priority of public safety over additional passenger satisfaction, while also giving acceptable guidance in moral decisions under risk?
If the recent decision-theoretic critique of lexical priorities is correct, then it turns out that there is no acceptable moral decision rule available to the official (Huemer Reference Huemer2010; Jackson and Smith 2006, 2016). This is because lexical priority theories appear to commit a decision-maker to at least one of the following problems: the Permissiveness Problem, the Risk Problem, and the Agglomeration Problem. The Permissiveness Problem consists of cases where all of one's options are equally permissible, simply because they have some positive probability of violating a lexical priority. The Risk Problem arises when arbitrarily small probabilities make particular actions impermissible. The Agglomeration Problem involves cases where lexical priority theories give contradictory advice depending on whether the relevant moral considerations are viewed separately or together. Lexical priority theories must avoid each of these problems in order to have an acceptable moral decision rule for risky situations.
This paper uses lexicographic expected value theory to rebut the decision-theoretic critique. As I shall argue, many of the problems that have been raised for lexical priority theories are actually the result of inappropriate decision-theoretic modelling. Once a more appropriate model is applied, the problems do not arise. Nevertheless, although the decision-theoretic critique fails in this respect, it succeeds in revealing important but under-theorised aspects of lexical priority theories.
Part 2 briefly sets out the preliminaries for modelling a moral theory using decision theory. Part 3 responds to the Permissiveness Problem. Part 4 responds the Risk Problem. Part 5 responds to the Agglomeration Problem. Part 6 discusses the value and limits of decision-theoretic critiques of moral theories, and identifies further questions that lexical priority theories must answer to give a complete account of moral decision-making under risk. Conclusion follows.
2. Preliminaries
The decision-theoretic critique of lexical priority theories takes place on the following theoretical terrain: firstly, it proceeds on the assumption that expected value theory is an appropriate framework for representing moral theories; secondly, the objections to lexical priority views are premised on a particular way of modelling lexical priorities using expected value theory. I will explain these theoretical points in turn.
In its normative application, expected value theory determines what you ought to do, given your uncertainty. It identifies the best option relative to a description of a decision problem.Footnote 2 A decision problem consists of: the options available to you; the possible states of the world; the probabilities of those states (represented by some value in the [0, 1] interval), given that you perform some optionFootnote 3; and the value of the outcomes that result from choosing a particular option in a given state of the world. According to expected value theory, the best option is the one that maximises expected value, where this is the option whose possible outcomes together have the greatest sum of probability-weighted value.
To apply expected value theory to moral theory, we must assume that the relative importance of moral considerations can be numerically represented by a value function.Footnote 4 This ‘moral’ value function must assign equal value to equally important considerations, greater value to more important considerations, and lesser value to less important considerations.Footnote 5 However, to operate in a context of risk, a moral value function must also accurately represent the differences in the relative importance of moral considerations. That is, we must assume that the importance of moral considerations can be represented by a cardinal moral value function.
As we shall see, the precise specification of the cardinal value function is important to how a moral theory operates in an expected value framework. A common modelling assumption in discussions of lexical priority theories is that higher-ranked considerations are infinitely more important than lower-ranked considerations, and so should be represented by an infinite value difference.Footnote 6 As we shall see, this is a problematic and unnecessary assumption. Strictly speaking, lexical priority theories only subscribe to an ordinal ranking of moral considerations: given a choice between upholding either a higher-ranked consideration or any number of lower-ranked considerations, it is more important to uphold the higher-ranked consideration. Lexical priority theories need not make any commitment about how much more important it is to uphold a higher-ranked consideration than lower-ranked ones. Infinite values simply offer one way of representing the fact that lexical priority theories stratify some kinds of moral considerations over others.
To test whether lexical priority theories can give adequate guidance in cases of risk, we will model the official's case as follows. We will assume that she has only two available options: increase the speed limit (by some fixed amount) or maintain the current speed limit. There will be only two possible and mutually exclusive states of the world: either the pedestrians will be careful when crossing the road (Careful) or they will not (Careless). We will allow that the probabilities of these states may be affected by the option chosen. Following the critics of lexical priority theories, we will hold that outcomes in which a lexical priority is violated – for instance, cases where there is an increase in pedestrian deaths – have infinite disvalue (−∞), whereas outcomes in which a lexical priority is upheld are assigned some finite value (f).
The above framework sets out a particular way of determining what lexical priority theories require of us in cases of uncertainty. One might object to this framework on a number of fronts (Temkin Reference Temkin2012, chap. 8). For dialectical purposes, however, we will assume that it is appropriate. We will see that by making amendments within the expected value framework, lexical priority theories can satisfactorily deal with the decision-theoretic critique.
3. The Permissiveness Problem
Lexical priority theories hold that we are morally prohibited from trading-off higher considerations for the sake of lesser ones. Framing lexical priority theories in this way, an important question is: what are we to do when all of our options have some positive probability of being of this prohibited kind? Some worry that lexical priority theories may commit us to a life of implausibly strong moral dilemmas, whereby all of our options are prohibited, merely by virtue of having a positive probability of violating a lexical priority. They call this the Paralysis Problem.Footnote 7
However, when the problem is cast in an expected value framework, the Paralysis Problem is in fact better understood as the Permissiveness Problem. Expected value theory, after all, exhorts us to choose the option that maximises expected value, even if that option has low (or, indeed, negatively infinite!) expected value. As we shall see, when lexical priority theories are modelled using infinite values, the problem is that all of our options turn out to be equally permissible simply by virtue of having some positive probability of violating a lexical priority.
To illustrate, suppose that the official must decide on the basis of the following evidence: one report suggests that increasing the speed limit may greatly increase pedestrian deaths, whereas another report suggests that maintaining the current speed limit may also increase pedestrian deaths, albeit to a lesser extent. Given that both options are risky, what ought she to do? (Table 1).
Table 1. The Permissiveness Problem.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab1.png?pub-status=live)
As shown in Table 1, since both options have some positive probability of violating a lexical priority, both have negatively infinite expected value. According to the expected value model, both options are therefore permissible because they have the same expected value.
Clearly, however, this can lead to absurd results. For example, even if one option is far riskier than the other, it will have the same expected value and therefore will be deemed equally choice-worthy. The Permissiveness Problem seems to show that lexical priority theories, blinded by their own fanaticism, lack the ability to distinguish impermissible risky options from permissible ones.
As it turns out, however, the Permissiveness Problem does not pose a serious challenge to lexical priority theories. Rather, the problem is due to a flawed decision-theoretic model of lexical priority theories. Specifically, the Permissiveness Problem arises because infinite values swamp expected value calculations, rendering all options equally permissible regardless of their riskiness. A more appropriate model will allow lexical priority theories to use expected value calculations to distinguish between permissible and impermissible options. This gives us good reason to look for non-infinitistic, expected value representations of lexical priority theories under risk.
The rich literature on Pascal's wager offers many candidate approaches.Footnote 8 One example is lexicographic expected value theory. Unlike most other approaches, the lexicographic approach constitutes a minimal departure from orthodox expected value theory.Footnote 9 Indeed, it is simply a generalisation of standard expected value theory. Where standard expected value theory operates with a one-dimensional value function, the lexicographic theory operates with a multi-dimensional function. It also has the benefit of being well known and comprehensively theorised in contemporary economic and decision theory.Footnote 10 Famously, Rawls defended a version of lexicographic decision theory in his Theory of Justice, although he did not apply it to decision-making under risk (due to his denial that probabilities are available behind the veil of ignorance).Footnote 11 As we shall see, the lexicographic model allows lexical priority theories to systematically avoid the Permissiveness Problem.
Using the lexicographic model, rather than using infinite values, we will represent lexical priorities using a ranking of finite-valued value functions, <v1, v2, …, vn >. In the official's case, we have assumed for simplicity that there are only two kinds of moral considerations at hand: public safety and additional passenger satisfaction. As such, we will require only two value functions, <v1, v2>, where v1 represents the moral significance of the level of public safety in that outcome and v2 represents the moral significance of the level of passenger satisfaction in that outcome. We represent the priority of public safety over additional passenger satisfaction using the following moral decision rule: the official ought to choose the act that has the highest expected value for v1 and – just in case there is a tie among the options – then she ought to choose the act that has the highest expected value for v2. In the unlikely event that there is yet another tie, she is permitted to choose either option (since there are no other considerations in this example).Footnote 13
Without loss of generality, we will assume that the official has a confidence of 0.95 that the public will be careful and we will assign more-or-less arbitrary lexicographic values to the various outcomes. This will give us a clear demonstration that our chosen moral decision rule can guide decision-making in cases where all options are risky.
As shown in Table 2, the option that uniquely maximises expected moral value in this case is to maintain the speed limit. However, that is not the important point. The important point is that there is an option that uniquely maximises expected moral value, even though all options have some probability of violating a lexical priority. As the official is not permitted to pursue just any of her available options, the Permissiveness Problem has been avoided. This shows that infinite values were indeed at the root of the problem. By eschewing infinite values, the expectations of the official's options are now sensitive to, among other things, the probabilities of the states. The lexicographic model is therefore a more appropriate decision-theoretic representation of lexical priority theories.
Table 2. Lexicographic expected value.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab2.png?pub-status=live)
4. The Risk Problem
Although lexicographic decision theory solves the Permissiveness Problem, it seems to expose lexical priority theories to the Risk Problem. This involves cases where an intuitively acceptable option is ruled out simply because it has a positive – albeit negligible – probability of violating a lexical priority. To many, such cases give us grounds for rejecting lexical priority theories altogether.Footnote 14
To illustrate the Risk Problem, let us suppose that the official has reason to believe that increasing the speed limit will almost certainly cause the public to become more careful. By contrast, maintaining the current speed limit will encourage complacency, leading to an almost equal probability of carelessness or carefulness. According to the lexicographic model, what ought she to do?
As shown in Table 3, the official maximises expected moral value by maintaining the speed limit, despite the fact that it is very improbable that the public would be careless, were she to increase the speed limit. Critics of lexical priority theories hold that in these types of cases, there is a point at which the probabilities of violating a lexical priority are so low as to be negligible. It cannot be that any prospect of upholding a lexically prior consideration, no matter how slim, is more important than a certainty of upholding any number of lesser considerations.
Table 3. The Risk Problem.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab3.png?pub-status=live)
To appease proponents of the Risk Problem, lexical priority theories must give a principled explanation for why decision-makers should ignore very unlikely prospects of upholding a lexically prior consideration. One approach is to posit a probabilistic threshold, t, that governs which possibilities decision-makers should ignore and which they should attend to. Note that this option is available because expected value theory merely identifies the best option relative to a specification of the decision problem. It is silent with respect to what information is or is not included in the decision problem. It is therefore open to lexical priority theories to supplement the lexicographic decision model with rules for determining which possibilities are relevant to a decision and which should be ignored.
Critics of lexical priority theories doubt that there is any principled way of fixing a value for t (Jackson and Smith Reference Jackson and Smith2006, 276; Isaacs Reference Isaacs2014, 97). However, they seem to assume that the threshold value must be context invariant: that is, the same under all circumstances. Admittedly, it is difficult to see what could justify a universal choice of t. However, since there is no need to assume context invariance, I will offer a context-variant approach to assigning a value to t.
On this approach, the chosen threshold should be that which, when followed as a rule, maximises expected moral value.Footnote 15 For example, a policy of always carefully attending to all manner of highly dubious conspiracy theories is unlikely to prove the most effective way to make accurate and timely decisions about public safety. Rather, public policy decisions should only be based on ‘live’ possibilities, where these are possibilities that are sufficiently probable. The correct threshold of probability will be that which, relative to the stakes and other factors, has the greatest expected lexicographic value when employed in decision-making. This, in turn, will be a matter of balancing contingent factors about the stakes at play in such decision contexts, the quality of the information available, the abilities of the decision-maker to accurately weigh evidence and moral considerations, and the time and resource pressures surrounding the decision. Once this threshold is set, the official will adopt a probabilistic threshold that, when followed as a rule, leads her to maximise the expected moral value of public safety, with the expected moral value of passenger satisfaction acting as a tie-breaker between equally choice-worthy threshold values.Footnote 16
Note that lexical priority theories can adopt this approach because they are not, strictly speaking, absolutist moral theories. Absolutist theories categorically hold that some kinds of actions are always morally prohibited, irrespective of what other options are available to the decision-maker (Jackson and Smith Reference Jackson and Smith2006, 268). Since prohibition does not seem to come in degrees or to be amenable to aggregation, absolutists cannot optimise in the face of risk. By contrast, lexical priority theories allow moral considerations of the same ranking to be traded-off against each other, making room for decision-makers to adopt a probabilistic threshold to optimise their conformity to lexically prior considerations. Under all but the most rarefied circumstances (those involving theoretically ideal agents in ideal situations), a context variant, probabilistic threshold will best allow decision-makers to navigate moral decision-making under risk.
In what follows, we will equip lexical priority theories with a context variant, odds-based threshold for determining which states to include in deliberations and which to exclude (Lin and Kelly 2012a, 2012b). Where standard probability threshold approaches assess whether or not a given state is sufficiently probable, the odds-based threshold evaluates whether a state is sufficiently more probable than its most probable alternative. This latter approach ensures not only that a consistent set of states is included in the decision problem, but also that the probabilities of these states are updated in accordance with the Bayesian rules of belief revision. Using this threshold, the official will include in her deliberations the states that are not sufficiently less probable than any other states. She will exclude from her deliberations any state that is sufficiently less probable than some other state. She will then update her beliefs about the included states, normalising the probabilities according to the rules of belief revision, and then maximise expected moral value.
To illustrate, let the relevant odds-based threshold, t, equal 1:99. In the case of the Risk Problem (Table 3), the probability that the public will be careless if she raises the speed limit is less than 0.01. As such, it is insufficiently probable. Given this information, the official should exclude the possibility that the public will be careless if she increases the speed limit, update her credences, and frame her decision problem as follows (Table 4).
Table 4. The Risk Problem avoided.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab4.png?pub-status=live)
Even though increasing the speed limit has a positive probability of violating a lexical priority, this fact is excluded from the official's deliberations. Updating her beliefs on the fact that the public is sufficiently unlikely to be careless (given that she increases the speed limit), the official maximises expected moral value within the revised decision problem by increasing the speed limit.Footnote 17 In this way, the official upholds her commitment to lexical priorities while also avoiding the Risk Problem. However, as we shall see, positing any kind of probabilistic threshold seems to raise a further problem for lexical priority theories.
5. The Agglomeration Problem
The Agglomeration Problem consists of cases where a lexical priority theory offers inconsistent verdicts about what ought to be done, depending on whether moral considerations are responded to separately or together. There are two versions of the Agglomeration Problem: one conjunctive and the other disjunctive. As critics of lexical priority theories note, both versions of the problem result from positing a probabilistic threshold. They conclude that lexical priority theories therefore cannot appeal to probabilistic thresholds to avoid the Risk Problem. Below, I will explain how lexical priority theories can avoid both versions of the Agglomeration Problem, without abandoning a probabilistic threshold.Footnote 18
The Conjunctive Agglomeration Problem involves cases where each of an agent's options has some moral property (say, that of respecting a lexical priority), but the conjunction of those options has a different property (say, that of violating a lexical priority). To illustrate, suppose that the official has just received additional information about the carefulness of the public, indexed to weekday behaviour and weekend behaviour. To simplify the following discussion, we will assume that the official should raise the speed limit if and only if there is a live possibility that the public is careful all week; otherwise, maintaining the speed limit is the most appropriate option available.
Given this simplifying assumption, if the official were to apply a probabilistic threshold, t, she might face the following dilemma. Suppose that the probability that the public will be careful on weekdays is greater than or equal to t (hence, it is a live possibility) and the probability that the public will be careful on weekends is greater than or equal to t (hence, a live possibility). It therefore seems that the official should increase the speed limit. However, it also seems possible that these states are incompatible (or, at least, anti-correlated), such that the probability that the public will be careful on both weekdays and weekends could be less than t (and, hence, not a live possibility), in which case she should not be open to increasing the speed limit. What ought she to do? Lexical priority theories appear to offer no guidance about what should be done. As it stands, the official seems to face an especially implausible kind of moral dilemma: if she decides on the basis of the time periods taken separately, she should perform an action that would be ruled out if she were to decide on the basis of the time periods taken together (Table 5).Footnote 19
Table 5. The Conjunctive Agglomeration Problem.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab5.png?pub-status=live)
In response, it is first worth noting that nothing about this problem relies on the notion of lexical priority. The problem is more generic: when a threshold is posited, there can be cases where the relevant conjuncts fall on one side of the threshold, but the conjunction falls on the other.Footnote 20 This should be encouraging for lexical priority theorists: the Agglomeration Problem is, strictly speaking, orthogonal to the concept of lexical priority; the problem's source – as well as its solution – lies elsewhere.
Note, also, that unlike the previous problems, the Conjunctive Agglomeration Problem is based on an informal description of the decision situation. This is important because, as it turns out, the problem disappears when it is placed in a more formal decision-theoretic framework. The informality of the Conjunctive Agglomeration Problem exploits an ambiguity in how to describe the relevant states in a decision problem. The decision-theoretic model forces a resolution of this ambiguity, thereby preventing the dilemma from arising.
In this case, the ambiguity driving the Conjunctive Agglomeration Problem is that there are two ways for the public to be careful: by being careful during a particular time period only (say, weekdays) or by being careful during that period and the remaining time period (say, weekdays and weekends). Since different degrees and types of risk may be associated with these different possibilities, they should be distinguished. A more accurate specification of the states of the decision problem would be as follows (Table 6).
Table 6. A more accurate specification of the states.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab6.png?pub-status=live)
1 The value of s in this table is implied by the description of the decision situation.
Applying the relevant probabilistic threshold truncates the decision problem to rule out states whose probability is less than t. This rules out the possibility that the public is careful all week. For simplicity, we will also assume that the probability that the public is careless all week is also less than t. Having ruled out these possibilities from her deliberations, the official updates her beliefs and maximises expected moral value (Table 7).Footnote 21
Table 7. The Conjunctive Agglomeration Problem avoided.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20201017073747757-0490:S0045509100021202:S0045509100021202_tab7.png?pub-status=live)
Given the arbitrary values above, the official should, in this case, maintain the speed limit. The more important point, however, is that there is no dilemma. Moreover, there will be no dilemma, whatever values are plugged into the decision problem. By disambiguating the informal description of the case, we created a well-defined decision problem relative to which the official can maximise expected moral value. The Conjunctive Agglomeration Problem has been solved.
The Disjunctive Agglomeration Problem relies on a similar ambiguity about the relevant state-space of a decision problem. To illustrate, let us again assume that the official should raise the speed limit if and only if there is a live possibility that the public is careful all week; otherwise, maintaining the speed limit is the most appropriate option available.
Suppose now that the official's evidence indicates that it is sufficiently improbable that the public is careless on weekdays, such that she should exclude that possibility and be provisionally open to raising the speed limit. If it is also sufficiently improbable that the public is careless on weekdays, then it seems that she should therefore hold that there is a live possibility that the public is always careful and so should raise the speed limit.
Given this evidence, however, the following situation is also possible: it may also be sufficiently probable that the public is careless during weekdays or the weekend. Even though the probability of each disjunct falls below the threshold, the sum of their probabilities may be above it. In this case, the official cannot exclude the possibility that the public is careless during at least one of the time periods, whichever it happens to be. Given these assumptions, it appears that the official faces a strange kind of moral dilemma: if she decides on a ‘time-period-by-time-period’ basis, she ought to raise the speed limit; if she decides ‘all together’, then she ought to maintain the speed limit. It seems that whatever she chooses, she will be acting wrongfully.
As we shall see, the Disjunctive Agglomeration Problem exploits the fact that the concept of lexical priority underdetermines the appropriate decision-theoretic representation of a moral decision problem. The Disjunctive Agglomeration Problem consists of cases where there are conflicting but seemingly equally eligible ways of specifying the relevant states of a decision problem. For example, taken on a ‘time-period-by-time-period’ basis, the relevant states (S) of the official's decision problem are:
(S1) The public is always careful.
(S2) The public is careful on weekdays but careless on weekends.
(S3) The public is careless on weekdays but careful on weekends.
(S4) The public is always careless.
This specification of the decision problem suggests that considerations of public safety pertain to particular, identified, risks. For instance, it may be that different subgroups or individuals in the population are exposed to heightened risk at different times periods (such as children on weekends), and the official may be particularly concerned to ensure that disproportionate risk is not imposed on those subgroups or individuals. That the public is careless at some time or other is not relevant, given this specification of the decision problem.
However, having specified the decision problem as above, standard rehearsals of the Disjunctive Agglomeration Problem then make salient an alternative framing of the decision problem, whereby unidentified risk is morally relevant. This leads to a different specification of the decision problem's relevant possibilities, such as:
(S1*) The public is always careful.
(S2*) The public is at least sometimes careless.
On this approach, evidence about whether the public is careful is only relevant insofar as it indicates whether S1* or S2* is true. The official need not bother about which particular time period the public is careless during, so long as she is sufficiently certain that there is some time during which the public faces an increased risk of death.
The Disjunctive Agglomeration Problem thus raises a valuable question about what lexical priority theories, and moral theories in general, really care about. In the official's case, the question is: should she care about identified or unidentified risk imposition? Strictly speaking, the concept of lexical priority is silent about this. The Disjunctive Agglomeration Problem thus shows that lexical priority theories are under-theorised in this important respect. In order to provide determinate advice in risky situations, such theories must explain how we should frame decision problems. To do this, they must spell out their substantive commitments in more detail.
As it happens, when it comes to the ethics of distributing risk, it is usually not enough to know that some quantum of risk is being imposed, whoever it might befall; it is often important to know whether particular groups or individuals are bearing an unfair burden of the risk. Such fairness considerations would support a framing of the decision problem in terms of identified risks. On the other hand, in cases where there is a diffuse risk that affects the population equally, an official may be concerned solely with the probability that the risk will eventuate somehow or other (this might apply, for example, when evaluating the risk of a catastrophic nuclear disaster). The framing of the decision problem is determined, in the end, by the details of the background moral theory.
Once the framing is settled, however, the Disjunctive Agglomeration Problem does not arise: the odds-based threshold ensures that only a consistent set of possibilities is included in the decision problem.Footnote 22 Once the decision problem is specified, the official should maximise expected lexicographic moral value.
6. Further uncertainties for lexical priority theories
We have seen that lexical priority theories have the resources to avoid the various decision-theoretic objections that have been raised against them. The Permissiveness Problem relied on the mathematical oddities of infinite values, but lexical priorities can be modelled without infinite values. The Risk Problem was driven by the assumption that decision-makers should always attend to all possibilities, no matter how improbable. Lexical priority theories can explain why some possibilities should not be factored into a decision problem. The Agglomeration Problem exploited the fact that the concept of lexical priority underdetermines the appropriate framing of decision problems. Once lexical priority theories decide on how a given decision problem should be framed, they avoid the Agglomeration Problem. Each of these problems was rectified without having to retreat from the idea that some kinds of moral considerations cannot be defeated by any number of particular other considerations. In each case, the problem was actually a symptom of the chosen decision-theoretic representation. There are a few ways of interpreting this result.
One interpretation takes the failure of any such decision-theoretic critique as a foregone conclusion. After all, it seems that we know (indeed, a priori) that the truth of moral theories depends on the adequacy of their substantive justifications, not the adequacy of their decision-theoretic representations. Faced with a problematic decision-theoretic representation of a moral theory, it is open to the moral theorist to say: ‘So much the worse for your model!’ On this view, the decision-theoretic critique of lexical priority theories was bound to fail because it does not engage with the substantive justifications of the lexical priorities in question.
However, even if the above is true, this does not mean that decision-theoretic critiques of moral theories are without value. In the process of responding to these objections, lexical priority theories have been forced to clarify their substantive commitments: do they really consider some considerations to be infinitely more valuable than others? Do they require that we always attend to arbitrarily small probabilities in our decision-making? Do they care about identified risks (such as protecting the lives of particular people) or unidentified risks (protecting the lives of people in general)? Even if the decision-theoretic critique fails, as argued, it has nevertheless succeeded in revealing these under-theorised aspects of lexical priority theories.
Indeed, once we scratch below the surface of lexical priority theories, we discover further puzzling features, depending on the particular kind of substantive justification at play. For example, consider the view that lexical priorities outweigh other considerations. How is this to be modelled? We have seen that positing infinite values is highly problematic. Instead, lexical priority theories could perhaps adopt a different kind of value function. One idea is to hold that the marginal moral value of upholding a lesser consideration diminishes asymptotically towards a limit, such that upholding any number of lesser considerations never has as much moral value as upholding a higher consideration.Footnote 23 This kind of moral value function could ensure that, in the official's case, no matter how many millions of passengers may benefit, pleasing an additional passenger will never outweigh the moral importance of a person's life.Footnote 24 The question is: why should the moral importance of an additional passenger's pleasure diminish due to external factors, like the number of other passengers who happen to also be benefiting from slightly more exhilarating rides? If moral value is conditional in this way, what exactly are the conditions? These questions become salient only once we try to understand how lexical priorities operate in a decision-theoretic framework.
Or consider, instead, the cancellation approach to justifying lexical priorities. It holds that there is no moral value to acting upon some considerations unless we also conform to particular other considerations. For example, we might say that there is no moral value in the official deviating from her role as a protector of public safety to satisfy the passengers’ need for speed: optimising passenger satisfaction has no objective moral value when it involves disregarding the lives of those who may suffer the consequences.Footnote 25 This approach has its own mysteries: for instance, how do we weigh considerations when we are uncertain if they have been cancelled? Is there such thing as partial cancellation of moral value?Footnote 26
Finally, consider an exclusionary approach to justifying lexical priorities. This approach is silent with respect to whether lexical priorities outweigh or cancel other considerations; it instead argues that lexical priorities deliberatively exclude them.Footnote 27 That is, there are cases where we should act first and foremost on the balance of only some types of considerations (such as public safety), irrespective of how many other considerations of a particular kind (such as additional passenger satisfaction) are at stake. As with the cancellation approach, it is not at all obvious how exclusionary reasons operate in cases of uncertainty. Are they nothing more than useful guides for decision-making (similar to the threshold approach presented earlier)? Or, are they better understood as epistemic considerations that give us reason to change our credences that the world is one way or another? In any case, does an exclusionary reason's importance diminish with probability and, if so, can such ‘diminished’ reasons be weighed against the reasons they purport to exclude?
These are almost completely unexplored issues in moral theory. The reason is that they only become salient once we adopt a decision-theoretic perspective of lexical priority views. Although the recent decision-theoretic critique of lexical priorities fails, it has succeeded in revealing new, potentially important lines of inquiry in our moral theorising.
7. Conclusion
The decision-theoretic critique of lexical priority theories suggests that they have no acceptable moral decision rule for cases of uncertainty, and should therefore be abandoned. However, as shown, this critique relies on contentious modelling assumptions that lexical priority theories need not – and, indeed, should not – accept. By using additional resources in decision theory and spelling out their substantive moral commitments in more detail, lexical priority theories can guide us through risky situations.
Disclosure statement
No potential conflict of interest was reported by the author.
Acknowledgements
I am extremely grateful to the generous and vibrant philosophical community at the Australian National University for many helpful discussions and comments on this paper. Thanks in particular to Christian Barry, Paul Bartha, Justin Bruner, Edward Elliot, Frank Jackson, Avery Kolers, Carina Lee-Stronach, Leon Leontyev, Hanti Lin, Daniel Nolan, Alex Sandgren, Lachlan Umbers, David Wiens, Timothy L. Williamson, Shang Long Yeo, various anonymous referees, editors at the Canadian Journal of Philosophy, and an audience at the ANU Philsoc Seminar. Thanks also to Gabriel Watts and an audience at the University of Sydney Postgraduate Philosophy Seminar. I owe extra special thanks to Alan Hájek and Seth Lazar for their support and constructive advice on this paper. Any mistakes in this paper are my own.