Since there is so much we can do together — good and bad — we are subject to numerous normative requirements to perform certain actions and to abstain from others. In what follows I will argue that some intuitively feasible requirements, especially those that are collective, are not in fact feasible. I thereby aim to offer a revised account of what counts as a feasible action. In particular, I argue that we can best preserve the spirit of what is known as the conditional account of feasibility if we move to what I call the constrained account. Footnote 1
What is at stake in offering an account of feasibility is twofold. First, we gain conceptual intuitiveness — it’s better if the concepts we use seem intuitively right to us. Second, our account of feasibility has implications for the content of our normative requirements. This is so if, as many do, we accept that “ought” implies “is feasible.” But it is also the case even if we reject it, just as long as we agree, as I think we should, that feasibility bears on the content or the status of what is required. Footnote 2
I. What Is Feasibility?
What counts as a feasible action? Footnote 3 It might be tempting to see feasibility simply as a function of possibility and/or the likelihood of success of an agent performing the action. But any such account is subject to a powerful counterexample, most memorably illustrated by Estlund’s chicken dance case. Footnote 4 Suppose that there is nothing that anyone could do to induce me to dance like a chicken in front of my students. (As it happens, a high enough monetary reward would do the trick, but suppose it would not.) It is thus extremely unlikely, indeed, close to impossible, that I will so dance. Surely, however, we should think that it is, nonetheless, feasible for me to dance like a chicken. The possibility/likelihood account then delivers the counterintuive answer here.
The conditional account of feasibility, by contrast, deals with such cases by stipulating that feasibility — roughly put for now — is a function of the possibility and/or likelihood of success of an agent performing an action, conditional on trying. Depending on whose formulation we follow, an action is feasible if it is possible Footnote 5 or if it is reasonably probable Footnote 6 conditional on trying, and the more likely it is, conditional on trying, the more feasible it is. Thus Brennan and Southwood define feasibility as “reasonable probability of success conditional upon trying.” Footnote 7 Gilabert and Lawford-Smith adopt a “binary test” for what is feasible and a “scalar test” for what’s more or less feasible once we are in the domain of what’s feasible delineated by their binary test. They distinguish between hard constraints — that is, constraints that it would be impossible to overcome — and soft constraints — that is, constraints that are malleable. According to the binary test, “it is feasible for X to φ to bring about O in context Z only if X’s φ-ing to bring about O is not incompatible with any hard constraints.” According to the scalar test, “it is more feasible for X to bring about O1 than for Y to bring about O2 when it is more probable, given soft constraints, for X to bring about O1 given that he or she tries, than it is for Y to bring about O2 given that he or she tries.” Footnote 8
We have a number of elements here, which it will be useful to fix for later use (and which are all offered as a list at the end of the essay). Let me simplify “action Φ is feasible for agent X” as “action Φ is feasible.” I use “doing Φ,” “Φ-ing,” “performing Φ” and, if emphasis is needed, “succeeding in doing Φ” interchangeably. Let me also define “agent X does action Φ” as “in context C at time T, agent X succeeds in performing action Φ thereby bringing about state of affairs S.” Thus, an action Φ is always defined with reference to the state of affairs S it brings about. An agent may thus succeed in performing Φ1 that brings about S1 while also failing at the same time to perform Φ2 that brings about S2 (for example, “writing a long novel” versus “writing a gripping novel”; “giving CPR” versus “saving a life”). The conditional account implies that an agent may fail to do an action either because she does not try (and the account finds such failures irrelevant to the calculus of feasibility) or because something intervenes, so to speak, between the trying and the bringing about of the state of affairs S (that is, the action may be difficult the way building a long bridge is). I assume throughout that “trying” involves an appropriately sustained attempt so that any failure here is not due to the agent giving up too early. Footnote 9 Finally, the agent X in question can be a collective agent. By collective agent X, I mean all those individuals X1 . . . Xn whose each respective doing of Φ1 . . . Φn constitutes the collective action Φ.
The conditional account can thus be restated as:
Proposal 1 (the possibility version): Action Φ is feasible (and more feasible) if it is possible (and more likely Footnote 10 ) for agent X to Φ given that X tries to Φ.
Proposal 2 (the likelihood version): Action Φ is (more) feasible if agent X is (more) likely to Φ given that X tries to Φ.
The difference between the versions will not matter to me here but let me state for the record that I favor Proposal 2. We have good reason to resist branding as feasible actions whose performance is merely possible (if tried), however unlikely such performance. Footnote 11 In this I follow Brennan and Southwood and adapt their compelling example. Footnote 12 Fixing your heart in open heart surgery is possible for me — by accident, of course, if I and you are extraordinarily lucky — but it’s not a feasible course of action for me when you are having a heart attack in my house. I will come back to what might explain our intuitions that such an action is not feasible. For now, let me manage expectations by explaining that I do not plan to resolve here the issue of how likely the success needs to be for the action to count as feasible. For the simplicity of grammatical phrasing it affords, I refer to the success of such actions as “likely” rather than “of a certain degree of probability” and I will assume in what follows that the probability of success has to be at least 50 percent. That said, this is not where my battle lies and I am open to the possibility that it may have to be lower or higher than that.
The conditional clause of the conditional account neatly deals with the chicken dance case: the likelihood that you will dance like a chicken is very high conditional on trying — if you try, it is an easy thing to do. However, the use of the conditional clause gives rise to two further problems. First, in its zeal not to undercount feasible actions for unmotivated chicken dancers it seems to overcount them for those who are genuinely unable to motivate themselves to carry them out. This is the problem of how to deal with instances of motivational failure.
Second, it does not seem to give the right answer in some cases of collective action. To see this second problem, consider two broad types of collective action. The first type requires coordination, in the sense that what a given individual needs to do depends on what other individuals do. Playing football — at least successfully — is an example of a collective action that requires coordination: whether I kick the ball forward or backward depends on where else my teammates are on the pitch. I will refer to such actions as coordinated if the mechanisms of coordination are in place (for example, a football team exists, has a manager, and so on) and as uncoordinated if coordination is needed but the mechanisms are absent. The second type of collective action merely has a number of individuals joining in, but what each individual needs to do does not depend — is not responsive to — what the others do. Footnote 13 Erasing the right foot of the St. Peter statue in the St. Peter’s Basilica by repeated touching of it, is an example of what we might call joint collective action: many people simply need to touch the foot (no one, I assume, could, as it were, do it single-handedly).
The problem with the conditional account of feasibility — as it is stated above — is that it suggests that a joint collective action is more feasible than a coordinated collective action even in cases when the opposite seems intuitively true. For instance, take the joint collective action of everyone on the planet, who has the capacity, touching his or her own nose next Tuesday. Were everyone to try, they would likely succeed (since it is such an easy action for any individual who has the capacity). Thus, if feasibility tracks the likelihood of success conditional on trying, then this is a feasible action. Footnote 14 But it seems odd to think that an action requiring everyone to touch his or her own nose on the same day is feasible, let alone highly so. It is also odd to think that such an action is more feasible than the coordinated collective action of, say, executing a goal-scoring plan in a game of football but, again, this is suggested by the understanding outlined above since the likelihood of executing the plan in the game of football, conditional on trying, will still depend on skill and luck, making it less feasible. More generally, the conditional account suggests that for any action that, if tried, is bound to be successful (for example, the action of thinking of a bat) it is just as feasible for three billion people to perform it as it is for three people to perform it. It suggests, implausibly, that we should, say, select the policy of everyone alive, who can, treating each other with kindness, rather than the policy of reforming the UN peace keeping force as a way of securing world peace.
This leaves us with a dilemma. If feasibility is not conditional on trying, then it seems that it is not feasible for people to do all sorts of easy things simply because they do not want to, dancing like a chicken being one of those. But if feasibility is conditional on trying, then all sorts of collective actions implausibly become feasible and our concept of feasibility recommends the wrong type of policy selection. Below I suggest how to avoid the dilemma. In Section II, I focus on the problem of motivation and in Section III on the problem of collective doing.
II. The Feasibility of Unmotivated Action
The rationale for adding the conditional clause to any account of feasibility is, obviously, that merely lacking the motivation to act seems insufficient to signal that the agent lacks the ability to perform the action and thus that the action is not feasible. This is captured so well by the chicken dance objection. For the sake of exposition, let me assume that acting only happens if an agent is motivated to act and let me call all instances in which the agent is not motivated to act instances of motivational failure. Then the simplest and probably most natural way to understand cases when motivational failure does not seem to rule out feasibility is that the agent could make herself appropriately motivated — that is, she sufficiently controls her motivational states. We can thus describe such cases as ones in which the agent is merely unwilling to do Φ, but motivationally able to. Footnote 15
Many of us also worry, however, that some instances of motivational failure are not instances of mere unwillingness to do Φ, but of what we might call genuine motivational inability: the agent cannot will the action. This inability can manifest itself either in the agent’s trying being somehow faulty or in her not even being able to bring herself to try. Typical examples of such genuine motivational inability are thought to include “clinical” cases, such as phobias, addiction, compulsion — all psychological mechanisms that in the philosophical literature, rightly or wrongly (usually wrongly), are assumed to override the agent’s will, replacing it with something alien to her. The thought here is that some cases of motivational failure are on a par with situations in which an agent fails to perform action Φ because, say (to indulge in an illustrative biological fiction), a nail in her head makes it impossible for the right synapses to fire in a way that would allow the agent to be motivated to do Φ. Possibly the most common case about which it is hard to resist the conclusion that we are in the presence of a genuine motivational inability, rather than mere unwillingness, is willingness to stay awake despite utter exhaustion. Footnote 16
Contrary to Gilabert and Lawford-Smith, then, trying is not “something that people can always do.” Footnote 17 The difficulty, therefore, becomes to distinguish between cases of mere unwillingness (which do not block feasibility) and cases of genuine motivational inability (which do). It would not help to stipulate that we are interested only in nonclinical cases when discussing feasibility of action. Such a move merely pushes the question of how to distinguish between cases of mere unwillingness and genuine motivational inability one step back. This is because the move either begs the question — we exclude the clinical cases so that we can simply assume, without a good reason, that all remaining cases are cases of mere unwillingness — or, alternatively, the move simply removes some cases from deliberation without, however, settling the issue of how to distinguish between mere unwillingness and genuine motivational inability for the remaining cases. Footnote 18
There are numerous attempts to draw the distinction in question. Wiens, for example, has recently argued against Estlund that motivational failure in the presence of a “good faith” attempt to do Φ should be seen as an instance of what I called genuine motivational inability that blocks the agent’s ability to do Φ rather than as an instance of mere unwillingness that does not block her ability to do Φ (with further consequences for what it might be appropriate to require of agents who fail at such good faith attempts). Footnote 19 Without even attempting to do justice to Wiens’s perceptive analysis, let me offer a general reason why I think we can be skeptical that the strategy will work for my question here. What counts as a good faith attempt could be defined internally (it appears so to the agent who is making it) and/or externally with reference to conduct (she has made a certain number of sacrifices, and so on). Footnote 20 The problem with an internal definition is that it falls prey to the possibility of self-deception: the agent may be merely unwilling but self-deceiving herself that her attempt is genuine (think of someone who might be ‘struggling’ against indulging in pleasant temptations). The problem with an external definition, in turn, is that it always makes sense to ask of any externally defined criterion: “but what if despite the sacrifices, and so forth, she is really just unwilling to do Φ?” An external definition of motivational (as opposed to some other) failure simply cannot block the possibility — since it cannot point to anything beyond the observable conduct — that had the agent really tried, she would have done it: “If she really tried to be faithful, she would be.”
Ultimately, instead of reviewing the full range of possibilities for how to draw the distinction on offer, let me make two points. First, I think it is not unreasonable to hold that the issue could only be settled, if at all, through or in conjunction with future empirical work. That is, if it is the case that some motivational failure is an instance of genuine inability and some is not, then it is possible that only complex neurological data (or equivalent) could help settle the issue. Second, however, since we do not have the data, here is a working proposal of my own. Motivational failure is an instance of mere unwillingness when there exists a conceivable incentive that would bring the agent’s motivational state in line with what is needed to perform the action in question. The incentive exists, on this view, if it would have such a power whether in fact the incentive could be offered in our world. Thus the incentive of “eternal beauty” or “immortality” would count on this view. The working proposal, then, defines feasible action as follows:
Proposal 3: Action Φ is (more) feasible if there is an incentive I such that, given I, X will try to do Φ and, given I, is (more) likely to do Φ.
This proposal replaces the conditional clause with the existential clause about incentives: if the agent is responsive to incentives then she is not subject to genuine motivational inability. (Whether the action is feasible further depends, on this account, on other obstacles that might block the performance of Φ.) For most people, this makes chicken dancing feasible while competent open heart surgery unfeasible since the typical person would likely succeed in doing the former but not the latter. Numerous instances of motivational failure, including many “clinical” cases, will now count as instances of mere unwillingness, with corresponding conclusions about their feasibility. I will shortly address the objection that this overestimates the feasibility of actions and a further complication that it underestimates it. First, however, let me point out that the proposal still allows for the existence of genuine motivational inability. For example, usually there comes a point prior to death from exhaustion when even the promise of a world will not suffice for a person to motivate herself to stay awake.
Proposal 3, however, offers only a sufficient condition for feasibility. For, hopefully, there is no incentive, for example, that could make you murder anyone, but this does not mean, depending on other features of the case, that the action is not feasible for you. A fuller account of motivational ability, then, should ask not only whether an agent is responsive to any incentives but also about the reasons why she may not be. If her failure to respond is solely due to her seeing action Φ as (normatively) wrong, then we should not brand her as genuinely motivationally unable. This should be understood counterfactually such that were the agent not to see the action as wrong, she would be responsive to incentives. Footnote 21 The category of wrongness here is meant to include moral wrongs as well as ethical wrongs (for those who distinguish between them). Given this, actions such as abstaining from killing but also such as declining to switch religions or worshipping trees remain feasible (assuming no other obstacles) even if they are never attempted on account of one’s moral and ethical convictions. By contrast, an action that the agent has no normative objection to, but doing it terrifies her to such an extent that there is no incentive powerful enough to induce her to do it, qualifies as unfeasible. This strikes me as correct: phobias constrain what people are able to do while normative beliefs express who people are rather than constrain them. Revised, the proposal becomes:
Proposal 4: Action Φ is (more) feasible if there is an incentive I — or had the agent X not seen Φ as wrong there would be I — such that, given I, X will try to Φ and, given I, X is (more) likely to Φ. Footnote 22
Let me, then, return to the objection flagged up above that the proposal overidentifies actions as feasible due to it tracking people’s sensitivity to conceivable rather than actual incentives. For example, suppose, that under normal circumstances, I cannot make myself approach a spider, but if my child’s happiness were at stake, I would. We may worry, then, that under normal circumstances I am genuinely motivationally unable to approach the spider, even if I would be able to do it in other circumstances. But I think, instead, that the incentives test correctly reveals whether an agent would be able to will the action if its payoffs were different, in which case her unwillingness to will it under normal circumstances is not an instance of genuine motivational failure, it is an upshot of her assessment that the effort, so to speak, is not worth the payoff. This point can be restated in terms of a distinction used by G. A. Cohen. Footnote 23 An action can be assessed along two dimensions: its difficulty (with impossibility at one end of the continuum) and its costliness. Taking you on a bicycle to Heathrow from Oxford is difficult but (if I like you) not costly. Kissing a frog is not difficult but is costly (I would experience disgust). A view of feasibility that made feasibility depend on how an agent experiences the action would track its costliness. But, as my incentives idea suggests, we should avoid the conclusion that a very costly action is not feasible; if changing the costs means we can will ourselves to perform the action then we reveal the motivational failure as a case of mere unwillingness and we should have seen the action as feasible all along. Feasibility, in other words, is better thought of as tracking the dimension of difficulty.
By the same token, the proposal above suggests that it is no more feasible for me to do something I am hugely motivated to do (kiss my baby daughter) than something I would hate to attempt (kiss a frog). But to think otherwise is to confuse feasibility with enthusiasm. It may also mean getting on a slippery slope at the bottom of which is the judgment we want to avoid that if I am entirely unenthusiastic (and so it feels mega hard or even impossible) to dance like a chicken then it is not feasible for me to do so. The final slide would be natural if degrees of feasibility tracked degrees of willingness and could be avoided only by an ad hoc stipulation that complete unwillingness does not mean infeasibility.
Consider next the objection that Proposal 4 (like Proposal 3 before it) underestimates feasible action. I think the following objection has force and should lead us to revise the Proposal. Suppose the agent responds to the same incentive in the same context — for reasons outside her control — only some of the time. Suppose, for example, in nine out of ten cases the incentive of £1,000 reward makes her do an extra hour’s work, but in one out of ten cases, it does not. Assume away all other incentives and it seems that although, given the incentive, the agent is close to 90 percent likely to succeed to perform Φ (since once she tries to do the extra work, she invariably succeeds), Φ is not feasible for her. This seems counterintuitive. Although, by assumption, in the 10 percent of cases the agent is genuinely motivationally unable to Φ, the fact that in 90 percent of (identical) cases she is able, seems to suggest she in fact has a (bit faulty) more general ability to try to Φ. If so, we can simplify Proposal 4 as follows:
Proposal 5: Action Φ is feasible if there is an incentive I — or had the agent X not seen Φ as wrong there would be I — such that, given I, X will likely do Φ.
Proposal 5, however, only captures the binary case of whether an action is feasible or not. For the scalar case, though, the notation needs to remain more complicated to avoid the suggestion that the enthusiasm with which an agent responds to an incentive could affect the feasibility of the action. Let the sign * by the Proposal’s number denote that it is about scalar feasibility. Proposal 5* disambiguates whether the likelihood of doing Φ tracks greater likelihood of responding to an incentive or greater likelihood of successful Φ-ing following a response to the incentive.
Proposal 5*: Action Φ is (more) feasible if there is an incentive I — or had the agent X not seen Φ as wrong there would be I — such that, given I, X will try to Φ and, given I, X is (more) likely to Φ.
All in all, an account of feasibility should discount the lack of motivation in agents who merely opt not to follow a certain course of action: their lack of motivation does not render the action unfeasible. But it should not discount the lack of motivation in agents who suffer from genuine motivational inability. For this reason we should prefer the account of feasibility that constrains feasibility by responsiveness to incentives to the conditional account. The constrained account preserves the spirit of the conditional account while abandoning the implausible assumption that whether it is motivationally possible for someone to try to do something makes no difference to whether it is feasible for them to do it.
III. The Feasibility of Collective Action
Part of the reason why the joint collective action of everyone in the world (who has the capacity) to touch his or her nose seems so utterly unfeasible is that we are dealing with a collective action where there is no method of deciding, communicating and coordinating together: there is no collective agency of any sort. Footnote 24 So while, conditional on everyone trying, the action would be carried out, there is no realistic account on hand of what would be involved in having everyone try to touch his or her nose next Tuesday.
And, just to be clear, the difficulty here does not reside merely in a confusion about the question we are asking. We can, of course, be interested in two issues. First, is it feasible for everyone to touch his or her nose next Tuesday? Second, is it feasible for me (or someone else) to get everyone to touch his or her nose next Tuesday? Clearly, the second action is not feasible for me (or anyone else) and the conditional account does not imply that it is. Footnote 25 The first set of actions is not feasible either, however, even though the conditional account suggests it is. The difficulty is also not merely that next Tuesday is just around the corner but that the touching of the noses by all needs to happen on the same day but, by assumption, we are not in the presence of a collective agent here whose parts (the individuals involved) can come and work together.
By contrast, in typical single-agent cases that the “conditional on trying” clause is meant to solve — the chicken dance — we know what is involved in the agent trying to perform the action: the agent just has to decide to dance like a chicken. Footnote 26 There remains a mystery, of course, as the section above illustrates, of what exactly takes place when a person decides. There is also the mystery of what exactly takes place when a coordinated group, such as a football team or a University, decides: How does the coordination happen? But the scale of the mystery is of a different magnitude than in the mass nose touching case. For three billion people to simultaneously touch their noses we need either massive coordination — but, if so, then we are no longer in the realm of merely joint collective action — or we need magic. An account of feasibility should not classify as feasible actions whose performance depends on magic.
Given that the problem with (some) joint collective action and uncoordinated collective action is the mystery of what trying to perform the action would amount to for the individuals involved since they themselves are in the dark, we should obviously refine our understanding of feasibility to take care of that. In fact, if we go back to the case of individual action, we notice that knowledge of how to do Φ is essential for the feasibility of Φ. By knowing “how to do Φ” I mean two things. First, a person needs to know that the action, if successful, will attain a state of affairs S (operating will save a life rather than be part of a horrific art project). This is the case since, given my definition of action Φ above, actions are in part defined by the state of affairs they attain. An agent who does not know that pressing a stone (which she would never dream of touching otherwise) will open a secret escape passage does not know that she can escape and escaping is not feasible for her. Second, the person needs to know how to do whatever Φ involves (grafting a healthy artery to a coronary artery rather than just poking around in the human body to save a life; pressing the stone to open the passage). Here “knowing how to” is also meant to include cases where, say, a tennis player knows how to score an ace even if she does not know precisely how to arrange her body; her body arranges itself so to speak.
Of course, knowledge is only in part a function of cognitive ability and access to sources, it also depends on motivation. Thus, one has the ability to know how to spell “knowledge” in Polish even if one purposefully sticks fingers in ones’ ears to avoid hearing the letters in “wiedza.” That said, knowledge is, of course, also dependent on access to sources of knowledge. Thus, even the smartest and most motivated monolingual speaker of English on a desert island with no books and internet has no ability to know how to spell it. In line with the analysis of motivational ability in the previous section, we can say that one has the ability to know K if there is an incentive I (or would be such an incentive had the agent not seen seeking or obtaining such knowledge as wrong) such that, given I, she will know K. And, since doing Φ is not possible without knowing how to do Φ (or, in case of collective action, that Φn is needed), it would appear we can just cut out any reference to knowledge in our definition of feasibility and rely simply on the thought (roughly put here) that an action is feasible if there is an incentive I such that given I, the agent will likely do Φ.
Things are more complicated, however. The Proposal — as it is stated so far — is most naturally interpreted as presupposing that either an incentive has been offered to the agent to perform the action, thus rendering the action thinkable for the agent, or that the agent is in fact already attempting to perform the action and so, again, the action is thinkable for her. (An analogous problem appears for the conditional account since the account asks what happens if the agent is trying to perform the action, thus the agent is presumed to have knowledge of what doing it involves.) But a correct assessment of the feasibility of Φ for agent X should take into account whether X is in position in context C at time T to actually know what Φ involves (that is, to know that successful Φ-ing will result in S and how to Φ). Footnote 27 Given this, we should revise the Proposal as follows. I refer to it as the constrained account of feasibility since feasibility here is seen as constrained by agents’ responsiveness to incentives and their knowledge:
Proposal 6: Action Φ is feasible if there is an incentive I — or had the agent X not seen Φ as wrong there would be I — such that, given I, X will likely do Φ, and agents’ knowledge how to do Φ is not contained in the offer of this incentive.
Proposal 6*: Action Φ is (more) feasible if there is an incentive I — or had the agent X not seen Φ as wrong there would be I — such that, given I, X will try to Φ and, given I, X is (more) likely to Φ, and agents’ knowledge how to do Φ is not contained in the offer of this incentive. Footnote 28
Next, let me make the fact that the agent can be collective explicit in the Proposal. Once we do this, we can soften the final condition to reflect the fact that not each individual involved in the collective action needs to see the whole picture. That is, some instances of collective action are possible without all those involved being aware of how or even that they contribute to Φ; all the individual agent may need to know is that she needs to do Φn (that is, that her doing Φn is called for for whatever reason: because it will contribute to Φ, or because someone has ordered her to do it, and so on). Also note that in the case of collective action, by “an incentive I” I mean “a set of incentives such that there is an incentive for each agent whose actions constitute the collective action.”
Proposal 7 (collective action): Action Φ is feasible if there is an incentive I —- or had X1 . . . …Xn not seen doing Φ1 . . . Φn as wrong there would be I — such that, given I, all agents X1 . . . Xn whose individual actions Φ1 . . . Φn constitute action Φ are likely Footnote 29 to do Φ1 . . . Φn, and if each X1 . . . Xn’s knowledge of how to do Φ, or that Φn is needed is not contained in the offer of this incentive.
Proposal 7* (collective action): Action Φ is (more) feasible if there is an incentive I —- or had X1 . . . …Xn not seen doing Φ1 . . . Φn as wrong there would be I — such that, given I, all agents X1 . . . Xn whose individual actions Φ1 . . . Φn constitute action Φ are likely to try to do Φ1 . . . Φn and are (more) likely to do Φ1 . . . Φn, and if each X1 . . . Xn’s knowledge of how to do Φ or that Φn is needed is not contained in the offer of this incentive.
On this account, an action of each of three billion people touching his or her nose next Tuesday is correctly classified as unfeasible because there are not three billion people who could know next Tuesday how to do it or that their individual contributions are needed. (Of course it remains open that this may become possible in future, in which case the action will become feasible then.) But the account also allows us to make sense of the intuition that sometimes even joint or uncoordinated collective action is feasible. For example, it is possible that everyone who needs to be involved has the ability to know what it would take to eliminate child prostitution (for example, the customers would need to stop seeking it, children would need anonymous access to child protection services without triggering deportation, and so on). It is also possible that each person knows what he himself needs to do and that he needs to do it (for example, one law enforcement official may need to know how to talk to a child given how others have treated the child, but she need not know, say, what her colleague needs to say to the child’s client). If so, eliminating child prostitution seems feasible (if extremely unlikely, given the actual motivations of some of those involved).
But the constrained account may be subject to an objection that it rules out too many collective actions as unfeasible. This objection is best introduced with the help of the following hypothetical scenario constructed by Estlund:
Slice and Patch go golfing. Suppose that unless the patient is cut and stitched he will worsen and die (though not painfully). Surgery and stitching would save his life. If there is surgery without stitching, the death will be agonizing. . . . Slice and Patch are each going golfing whether the other attends to the patient or not. Does anyone act wrongly? Patch ought to stitch the patient if and only if Slice will be doing the surgery (stitching is possible, but pointless and harmful if there is no wound that needs stitching). But suppose that Slice will not be doing the surgery. Patch might as well go golfing. Ought Slice to cut? Well, no, because Patch will not be there to stitch, and so the surgery will only make the patient’s death more painful. Slice might as well go golfing. Neither has acted (or omitted) wrongly, despite the fact that the patient will needlessly die. Footnote 30
Estlund is interested here not in the feasibility of their actions, but in whether Slice and Patch fail by any moral requirement; but the scenario illuminates how we should think about feasibility (and, in fact, I think that we can gain purchase on the issue of moral requirements in this instance by considering the issue of feasibility). Estlund wonders whether there has been any moral failure, where for moral failure of a collective action to arise (he assumes and I agree) it must be the case that we can point to individuals who act wrongly. Estlund leaves open the question of whether there is actual moral failure, but he does assert that: “Many of us respond to this case with the intuition that there is some moral violation here. . . .” and he talks of “the common response of moral offense or outrage when the patient is left to die.” If we accept these intuitions, “the puzzle,” he points out, “is to find an agent who has committed it” so that we can close the “normativity gap” between the intuition of the presence of some moral violation and the difficulty to pin point how any individual involved violates any moral obligations. Footnote 31 Ultimately, Estlund identifies an inconsistent triad of beliefs, at least one of which must be given up: 1. “Moral failure: It is morally wrong if the patient is left to die.” 2. “No wrong without obligation: ‘If something is morally wrong, then there was an obligation on some agent to act or omit other than as they did.’” 3. “No violating agent: There is no agent in this case who is morally required to act (or omit) otherwise.” Footnote 32
As it happens, I think that we should discard the first of these beliefs, but my reason for reporting the example is not to fight that battle, but to consider the following question: Are Slice and Patch being asked to do something feasible? The answer may appear obvious: surely it is feasible for them to operate (and, consequently, to save the life). But I do not think the answer is so obvious. Here is why. I assume that each agent in question has a good reason to believe that the other will be absent from the operating table or, more weakly, no good reason to suppose that she should make herself available. Otherwise, if Slice had a reason to believe that Patch will be there, or a means of bringing Patch there, then she should not, relative to her evidence, go golfing: She would violate a moral obligation even if it turned out that Patch did go golfing. So we should assume that neither agent can communicate with the other about the other’s plans and each has a good reason to believe that the other will be absent. But if this is the case, then it is not clear that saving the patient is a feasible moral requirement. Of course, normally doctors can reasonably grasp what each needs to do to operate and save a life but in this specific instance this is not the case. So the requirement to operate and save a life is not feasible in light of my proposed account.
Given this, Slice and Patch are being asked to do what is not feasible for each to do. We should therefore also reject the description of the case in which “the patient is left to die” if being “left to die” is meant to refer to an action performed by a conscious agent as opposed to it being a state of the world. After all, since Slice has no reason to believe Patch will be there, he is not leaving the patient to die and nor is Patch. Hence no one is leaving the patient to die. They would have left the patient to die if they were able to communicate and coordinate, but they cannot. We could hang on to our sense that they together leave the patient to die if we rejected the (correct) assumption made by Estlund that for a group to fail it must be the case that individual members of the group fail. Footnote 33 Since we should not do that, we should give up on the belief that there is a moral failure here that goes beyond the possible failure by the protagonists to respond with the right attitudes to the situation. We can still condemn each of them, that is, if they delight in, or hope for, or even simply do not regret the situation that arises. But this is a condemnation of something other than whether they meet a normative requirement to operate.
The conclusion that saving the life of the patient is not feasible might be thought of as a reductio of the constrained account of feasibility. But to see the intuition driving my conclusions, consider a different case. Suppose there are two individuals, each located in a separate room equipped with a row of a thousand consecutively numbered buttons. To save the life of a third person, each agent must press the button of the same number as the other agent. And suppose they cannot communicate with each other. If so, the requirement that they save the life would, I take it, not strike us as feasible. It is certainly possible for them to save the life if, by accident, they press the right buttons, but it is no more feasible for each to press that button than it is for someone without any medical knowledge to perform open heart surgery. In neither case can the agents know what is required to pull off the action successfully.
Granted, as we reduce the number of buttons, the intuition that the requirement to save the life is not feasible appears weaker and weaker. But this is because we accept that some element of luck is compatible with an action being feasible; otherwise almost no action would qualify as such. So we think it is feasible to make a phonecall, even though one’s phone could give up, and we think it is feasible to drive to a shop even though one could be in an accident. I do not know how little or much control an agent needs to have for us to deem an action as feasible. If it is as little as 50 percent, that could explain why we may nonetheless think that the action of saving a life in the two button case remains feasible. But the fact that we allow judgments of feasibility to depend on the likelihood of success does not mean that we should discard the intuition that it must be the likelihood of success of an action — where there is an agent who is acting — rather than states of the world coming together to produce an outcome with no appropriate agency involvement. This is why we can say that it is feasible for me to walk on a tight rope if I have the skill to do it, but we should not say it is feasible for me if I can only do it when you hypnotize me and I cannot get you to hypnotize me (where I understand hypnosis in the psychologically inaccurate but philosophically familiar sense of something that overrides the agent’s will replacing it with that of the hypnotizer). In the latter case, it is perhaps feasible for you to make me walk on a tightrope, but walking on a tightrope is not feasible for me.
The feasibility of actions depends on there being an intentional agent, single or collective, who can perform the action in question. Thus going back, as advertised, to the earlier example of open heart surgery, the reason why it is not feasible for me to perform it (even though it is possible that I might succeed if I am extraordinarily lucky) is that I have no idea what such an action would involve. An account of feasibility should not collapse feasibility into possibility even at the extremes of the continuum of feasible action. The constrained account avoids doing so. Footnote 34 Of course, this still leaves unanswered the big question of what counts as an action performed by an intentional agent rather than a mere bodily movement, but I think we can begin to use the account of feasibility here proposed while relying on an imprecise understanding of agency. Or, to put the point more positively, the account of feasibility should work with a number of different accounts of agency that could be plugged into it.
All in all, we should adopt the following constrained account of feasible action, as outlined above: Footnote 35
Definitions:
-
[a.] “Action Φ is feasible” means “Action Φ is feasible for agent X.”
-
[b.] “Action Φ” means either “action Φ done by an individual” or “collective action Φ constituted by Φ1…Φn done by X1…Xn.”
-
[c.] “Agents X1 . . . Xn do Φ1 . . . Φn,” means “each agent X1 . . . Xn respectively does Φ1…Φn.”
-
[d.] “Agent X” means either “an individual agent” or “a collective agent composed of all those individuals X1 . . . Xn whose each respective doing of Φ1 . . . Φn constitutes the collective action Φ.”
-
[e.] “Agent X does Φ” means “in context C at time T, agent X succeeds in performing action Φ thereby bringing about state of affairs S.”
-
[f.] “Knowing” means “knowing in context C at time T when the action Φ is performed.”
-
[g.] “Knowing how to do Φ” means “knowing that succeeding in doing Φ results in S and how to Φ.”
-
[h.] “Trying to do Φ” means “engaging in an appropriately sustained attempt to do Φ so that any failure to do Φ is not due to the agent giving up too early.” Footnote 36
-
[i.] In case of collective action “an incentive I” means “a set of incentives such that there is an incentive for each agent whose actions constitute the collective action.”
The constrained account of feasibility: individual action. Action Φ is (more) feasible iff there is an incentive I (or had the agent X not seen doing Φ as wrong there would be I), such that, given I,
-
[1.] agent X is likely to try to do Φ
and
-
[2.] agent X is (more) likely to do Φ
and
-
[3.] agent X’s knowledge of how to do Φ is not contained in the offer of this I.
The constrained account of feasibility: collective action. Action Φ is (more) feasible iff there is a an incentive I (or had the agents X1 . . . Xn not seen doing Φ1 . . . Φn as wrong there would be I), such that, given I,
-
[1.] agents X1 . . . Xn are likely to try to do Φ1 . . . Φn
and
-
[2.] agents X1 . . . Xn are (more) likely to do Φ1 . . . Φn
and
-
[3.] agents X1 . . . Xn’s knowledge of how to do Φ, or that his or her Φn is needed is not contained in the offer of this I.
Our account of feasibility of individual and collective action should be constrained to take account of agents’ responsiveness to incentives and their ability to know how to pull off the action in question. The constrained account preserves the spirit of the conditional account in that it does not get people off the hook with regard to whether their acting is feasible merely if they are unwilling to so act. But it also avoids the pitfalls the conditional account faces when it incorrectly discounts the difficulties that arise due to genuine motivational inability and due to the fact that the collective agent who would need to do the action is unable to act together.