Hostname: page-component-745bb68f8f-cphqk Total loading time: 0 Render date: 2025-02-06T19:42:09.624Z Has data issue: false hasContentIssue false

The ecological benefits of being irrationally moral

Published online by Cambridge University Press:  25 October 2022

Elisabetta Sirgiovanni*
Affiliation:
Department of Molecular Medicine, Museum of the History of Medicine, Sapienza University of Rome, 00185 Rome, Italy elisabetta.sirgiovanni@uniroma1.it

Abstract

Trolley-like dilemmas are other cases of what Bermúdez refers to as (conscious) quasi-cyclical preferences. In these dilemmas, identical outcomes are obtained through morally non-identical actions. I will argue that morality is the context where descriptive invariance and ecological relevance may be crucially distinguished. Logically irrational moral choices in the short term may promote greater social benefits in the longer term.

Type
Open Peer Commentary
Copyright
Copyright © The Author(s), 2022. Published by Cambridge University Press

Framing effects offer a classic example of irrationality. Cases like Aeschylus's Agamemnon and Shakespeare's Macbeth, brilliantly discussed by José Luis Bermúdez, may subvert this consensus view. In my impression, the reason why this is evidently the case is that these are moral dilemmas as the agents face conflicts of moral nature. I will argue that morality is specifically the context where descriptive invariance and ecological relevance may be crucially distinguished.

The famous trolley-like dilemmas are other cases of what Bermúdez refers to as (conscious) quasi-cyclical preferences. Notoriously, switching the lever to deviate the trolley and sacrificing one person (F1o) is preferred to letting the five people on the tracks die (A), which is itself preferred to pushing a man out from the footbridge to stop the trolley (F2o). Even if frame-insensitive responses (F1o > A < F2o) can be primed through order presentation (Schwitzgebel & Cushman, Reference Schwitzgebel and Cushman2012), responses to “complex” frames are perceived as highly conflictual (Koenigs et al., Reference Koenigs, Young, Adolphs, Tranel, Cushman, Hauser and Damasio2007). Complex frames are the crying baby (i.e., would you smother your crying newborn to save yourself and four townspeople hiding from a Nazi incursion?) (Cushman & Greene, Reference Cushman and Greene2012), the transplant (i.e., would you harvest the organs from a pizza boy who entered the ER to save five patients in need of a transplant?) (Thomson, Reference Thomson1976), the cabin boy (i.e., would you kill and eat the cabin boy to save yourself and other starving shipwrecked sailors?) (Harris, Reference Harris2020), triage in a pandemic (Kneer & Hannikainen, Reference Kneer and Hannikainen2022), and many others (Edmonds, Reference Edmonds2014; Koenigs et al., Reference Koenigs, Young, Adolphs, Tranel, Cushman, Hauser and Damasio2007). These situations are logically indistinguishable from the bystander case (i.e., switching the lever). The irrationality of intuitive non-utilitarian responses in these complex frames (i.e., a “no” response) is the reason why these responses, associated with greater activation of emotional brain areas, are discarded by some as ethically unreliable (Greene, Reference Greene and Liao2016; Greene, Sommerville, Nystrom, Darley, & Cohen, Reference Greene, Sommerville, Nystrom, Darley and Cohen2001, Reference Greene, Nystrom, Engell, Darley and Cohen2004). Despite general appeals attributing a crucial role to emotions in morality (Greene & Haidt, Reference Greene and Haidt2002) and even if within a utilitarian (rather than a deontological) perspective (Greene, Reference Greene and Sinnott-Armstrong2008), this so-called dual-system approach ends up in a traditional merger between morality and rationality.

Emotionally different frames of the same outcome (one vs. five) in moral scenarios, I instead believe, may differ informationally by translating into dissociable moral attitudes. Framing it into Bermúdez's discussion, we may call this the non-extensionality of moral preferences, values, and actions. This discourse falls within the concept of “ecological rationality” proposed and defended over the years by Gigerenzer and colleagues (Gigerenzer, Reference Gigerenzer2015; Todd & Gigerenzer, Reference Todd and Gigerenzer2012). Accordingly, we may distinguish logical rationality or “axiomatic definitions of rationality that economic models draw upon” (Todd & Gigerenzer, Reference Todd and Gigerenzer2012, p. 425) from ecological rationality. The latter consists of contextual evaluations of decision processes that go beyond the identification of logical consistency, but rather refer “to how well they match the environments in which they are used” (ibid.). More precisely, ecological rationality expresses in encoding covert scenarios or surplus information between the lines (Sher & McKenzie, Reference Sher and McKenzie2006), a significant ability for moral cognition. In trolley-like dilemmas, identical outcomes are obtained through morally non-identical actions (Cushman, Reference Cushman2013). No less importantly, informational differentiation in these dilemmas is mostly evoked in a native language, where the emotionality of content sounds clearer (Costa et al., Reference Costa, Foucart, Hayakawa, Aparici, Apesteguia, Heafner and Keysar2014; Geipel, Hadjichristidis, & Surian, Reference Geipel, Hadjichristidis and Surian2016).

Sacrificial moral dilemmas are extraordinary cases where the logical accuracy of a frame does not necessarily imply its moral applicability. Interestingly enough, some utilitarian scholars (Kahane, Reference Kahane2012; Kahane, Everett, Earp, Farias, & Savulescu, Reference Kahane, Everett, Earp, Farias and Savulescu2015; Kahane et al., Reference Kahane, Everett, Earp, Caviola, Faber, Crockett and Savulescu2017) have defended this position. They reject the idea that cool harm-endorsing responses to complex frames are expressions of genuine utilitarianism, that is, “impartial concern for the greater good” (Kahane et al., Reference Kahane, Everett, Earp, Farias and Savulescu2015, p. 194), as they less arguably are in the bystander case. The scale they designed, the Oxford Utilitarian Scale (OUS), shows that responses to the two models through which utilitarianism is framed (impartial beneficence vs. instrumental harm) measure different individual traits and moral attitudes. Consistent with previous studies (Bartels & Pizarro, Reference Bartels and Pizarro2011; Glenn, Koleva, Iyer, Graham, & Ditto, Reference Glenn, Koleva, Iyer, Graham and Ditto2010; Koenigs, Kruepke, Zeier, & Newman, Reference Koenigs, Kruepke, Zeier and Newman2012), the scale associates non-clinical psychopathic tendencies with instrumentally harmful responses but, more importantly, not with an endorsement of impartial beneficence. Other morally questionable attitudes, such as rational egoism or lenient moral attitude, were also linked to “purely extensional approaches.” Besides, utility calculation – as intended by its originators – must consider also distal consequences of an action, referring to the class of actions to which the action belongs, and not only proximal consequences as if the action were single and insulated (Warnock, Reference Warnock and Warnock2003).

Sensitivity to emotionally salient frames in the moral domain, for example, to personal force (Greene et al., Reference Greene, Cushman, Stewart, Lowenberg, Nystrom and Cohen2009) and instrumental harm (Everett, Faber, Savulescu, & Crockett, Reference Everett, Faber, Savulescu and Crockett2018), is an efficient heuristic to prevent harm to others in real-life situations that also predicts expectations toward other people's moral behavior. Evolution may have selected for reactive harm-aversion as this favors cooperation, reciprocation, and trust (Cosmides, Guzmán, & Tooby, Reference Cosmides, Guzmán, Tooby, Zimmerman, Jones and Timmons2018). Harm-aversion is linked to a social cognition brain area (i.e., right temporoparietal junction [RTPJ]) (Young, Camprodon, Hauser, Pascual-Leone, & Saxe, Reference Young, Camprodon, Hauser, Pascual-Leone and Saxe2010), and it indeed increases the chances of being chosen as romantic or business partners (Everett, Pizarro, & Crockett, Reference Everett, Pizarro and Crockett2016).

We may also cast doubt on the association – mentioned by Bermúdez – between the recruitment of a brain area (i.e., dorsolateral prefrontal cortex [DLPFC]) and cognitive control processing, especially as this association justifies the appropriateness of utilitarian choices in complex frames (Greene, Nystrom, Engell, Darley, & Cohen, Reference Greene, Nystrom, Engell, Darley and Cohen2004, Reference Greene, Morelli, Lowenberg, Nystrom and Cohen2008). Brain stimulation studies suggest the opposite of what this model predicts, that is, disrupting the right-DLPFC actually increases utilitarian responses (Tassy, Oullier, Mancini, & Wicker, Reference Tassy, Oullier, Mancini and Wicker2013), while enhancing the left-DLPFC increases non-utilitarian responses (Kuehne, Heimrath, Heinze, & Zaehle, Reference Kuehne, Heimrath, Heinze and Zaehle2015). Agreeing with Bermúdez, the literature is now converging on the idea that self-control is closer to emotional regulation than cognitive effort (Fujita, Reference Fujita2011; Inzlicht & Friese, Reference Inzlicht and Friese2021; Magen, Kim, Dweck, Gross, & McClure, Reference Magen, Kim, Dweck, Gross and McClure2014). Accordingly, cognitive load delays reaction time, but does not determine any change in footbridge responses (Greene, Morelli, Lowenberg, Nystrom, & Cohen, Reference Greene, Morelli, Lowenberg, Nystrom and Cohen2008). We discern two aspects of self-control – inhibiting intuitive response and delaying gratification (Duckworth & Kern, Reference Duckworth and Kern2011). Failing to inhibit automatic non-utilitarian choices may be seen as gratification delay if looked at through ecological lenses. These choices may be logically irrational in the short term, as they imply a greater number of deaths, but promote greater social benefits in the longer term, as they limit pernicious attitudes like justifying the sacrifice of innocents.

Financial support

The author received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.

Conflict of interest

None.

References

Bartels, D. M., & Pizarro, D. A. (2011). The mismeasure of morals: Antisocial personality traits predict utilitarian responses to moral dilemmas. Cognition, 121, 154161. doi: 10.1016/j.cognition.2011.05.010CrossRefGoogle ScholarPubMed
Cosmides, L., Guzmán, R. A., & Tooby, J. (2018). The evolution of moral cognition. In Zimmerman, A., Jones, K., & Timmons, M. (Eds.), The Routledge handbook of moral epistemology (pp. 174228). Routledge Publishing.Google Scholar
Costa, A., Foucart, A., Hayakawa, S., Aparici, M., Apesteguia, J., Heafner, J., & Keysar, B. (2014). Your morals depend on language. PLoS ONE, 9(4), e94842. doi: 10.1371/journal.pone.0094842.Google ScholarPubMed
Cushman, F. (2013). Action, outcome, and value: A dual-system framework for morality. Personality and Social Psychology Review, 17(3), 273292. doi: 10.1177/1088868313495594CrossRefGoogle ScholarPubMed
Cushman, F., & Greene, J. D. (2012). Finding faults: How moral dilemmas illuminate cognitive structure. Social Neuroscience, 7(3), 269279. doi: 10.1080/17470919.2011.614000.CrossRefGoogle ScholarPubMed
Duckworth, A. L., & Kern, M. L. (2011). A meta-analysis of the convergent validity of self-control measures. Journal of Research in Personality, 45(3), 259268. doi: 10.1016/j.jrp.2011.02.004CrossRefGoogle ScholarPubMed
Edmonds, D. (2014). Would you kill the fat man? The trolley problem and what your answer tells us about right and wrong. Princeton University Press.Google Scholar
Everett, J. A. C., Faber, N. S., Savulescu, J., & Crockett, M. J. (2018). The costs of being consequentialist: Social inference from instrumental harm and impartial beneficence. Journal of Experimental Social Psychology, 79, 200216. doi: 10.1016/j.jesp.2018.07.004CrossRefGoogle ScholarPubMed
Everett, J. A. C., Pizarro, D. A., & Crockett, M. J. (2016). Inference of trustworthiness from intuitive moral judgments. Journal of Experimental Psychology, 145(6), 772787. doi: 10.1037%2Fxge0000165CrossRefGoogle ScholarPubMed
Fujita, K. (2011). On conceptualizing self-control as more than the effortful inhibition of impulses. Personality and Social Psychology Review, 15(4), 352366. doi: 10.1177/1088868311411165CrossRefGoogle Scholar
Geipel, J., Hadjichristidis, C., & Surian, L. (2016). Foreign language affects the contribution of intentions and outcomes to moral judgment. Cognition, 154, 3439. doi: 10.1016/j.cognition.2016.05.010CrossRefGoogle ScholarPubMed
Gigerenzer, G. (2015). On the supposed evidence for libertarian paternalism. Review of Philosophy and Psychology, 6, 361383. doi: 10.1007/s13164-015-0248-1CrossRefGoogle ScholarPubMed
Glenn, A. L., Koleva, S., Iyer, R., Graham, J., & Ditto, P. H. (2010). Moral identity in psychopathy. Judgment and Decision Making, 5, 497505.Google Scholar
Greene, J. D. (2008). The secret joke of Kant's soul. In Sinnott-Armstrong, W. (Ed.), Moral psychology, Vol. 3. The neuroscience of morality: Emotion, brain disorders, and development (pp. 3580). MIT Press.Google Scholar
Greene, J. D. (2016). Beyond point-and-shoot morality: Why cognitive (neuro)science matters for ethics. In Liao, S. M. (Ed.), Moral brains: The neuroscience of morality (pp. 119149). Oxford University Press.Google Scholar
Greene, J. D., Cushman, F. A., Stewart, L. E., Lowenberg, K., Nystrom, L. E., & Cohen, J. D. (2009). Pushing moral buttons: The interaction between personal force and intention in moral judgment. Cognition, 111(3), 364371.CrossRefGoogle ScholarPubMed
Greene, J. D., & Haidt, J. D. (2002). How does moral judgment work? Trends in Cognitive Sciences, 6(12), 517523. doi: 10.1016/S1364-6613(02)02011-9CrossRefGoogle ScholarPubMed
Greene, J. D., Morelli, S. A., Lowenberg, K., Nystrom, L. E., & Cohen, J. D. (2008). Cognitive load selectively interferes with utilitarian moral judgment. Cognition, 107(3), 11441154. doi: 10.1016/j.cognition.2007.11.004Google ScholarPubMed
Greene, J. D., Nystrom, L. E., Engell, A. D., Darley, J. M., & Cohen, J. D. (2004). The neural bases of cognitive conflict and control in moral judgment. Neuron, 44, 389400. doi: 10.1016/j.neuron.2004.09.027CrossRefGoogle ScholarPubMed
Greene, J. D., Sommerville, R. B., Nystrom, L. E., Darley, J. M., & Cohen, J. D. (2001). An fMRI investigation of emotional engagement in moral judgment. Science, 293(5537), 21052108. doi: 10.1126/science.1062872CrossRefGoogle ScholarPubMed
Harris, J. (2020). Why kill the cabin boy? Cambridge Quarterly of Healthcare Ethics, 30(1), 49. doi: 10.1017/S0963180120000420CrossRefGoogle ScholarPubMed
Inzlicht, M., & Friese, M. (2021). Willpower is overrated. Behavioral and Brain Sciences, 44, e42. doi: 10.1017/S0140525X20000795CrossRefGoogle ScholarPubMed
Kahane, G. (2012). On the wrong track: Process and content in moral psychology. Mind and Language, 25(5), 519545.Google Scholar
Kahane, G., Everett, J., Earp, B., Caviola, L., Faber, N., Crockett, M., & Savulescu, J. (2017). Beyond sacrificial harm: A two dimensional model of utilitarian psychology. Psychological Review, 125(2), 131164. doi: 10.1037/rev0000093CrossRefGoogle ScholarPubMed
Kahane, G., Everett, J., Earp, B., Farias, M., & Savulescu, J. (2015). Utilitarian judgment in sacrificial dilemmas does not reflect impartial concern for the greater good. Cognition, 134, 193209. doi: 10.1016/j.cognition.2014.10.005CrossRefGoogle Scholar
Kneer, M., & Hannikainen, I. R. (2022). Trolleys, triage and Covid-19: The role of psychological realism in sacrificial dilemmas. Cognition & Emotion, 36(1), 137153. doi: 10.1080/02699931.2021.1964940CrossRefGoogle ScholarPubMed
Koenigs, M., Kruepke, M., Zeier, J., & Newman, J. P. (2012). Utilitarian moral judgment in psychopathy. Social Cognitive and Affective Neuroscience, 7(6), 708714. doi: 10.1093/scan/nsr048CrossRefGoogle ScholarPubMed
Koenigs, M., Young, L., Adolphs, R., Tranel, D., Cushman, F., Hauser, M., & Damasio, A. (2007). Damage to the prefrontal cortex increases utilitarian moral judgements. Nature, 446(7138), 908911. doi: 10.1038/nature05631CrossRefGoogle Scholar
Kuehne, M., Heimrath, K., Heinze, H.-J., & Zaehle, T. (2015). Transcranial direct current stimulation of the left dorsolateral prefrontal cortex shifts preference of moral judgments. PLoS ONE, 10(5), e0127061. doi: 10.1371/journal.pone.0127061CrossRefGoogle ScholarPubMed
Magen, E., Kim, B., Dweck, C. S., Gross, J. J., & McClure, S. M. (2014). Behavioral and neural correlates of increased self-control in the absence of increased willpower. Proceedings of the National Academy of Sciences, 111(27), 97869791.CrossRefGoogle ScholarPubMed
Schwitzgebel, , E., & Cushman, , F. (2012). Expertise in moral reasoning? Order effects on moral judgment in professional philosophers and non-philosophers. Mind & Language, 27, 135–153. doi: 10.1111/j.1468-0017.2012.01438.xCrossRefGoogle Scholar
Sher, S., & McKenzie, C. R. (2006). Information leakage from logically equivalent frames. Cognition, 101(3), 467494.CrossRefGoogle ScholarPubMed
Tassy, S., Oullier, O., Mancini, J., & Wicker, B. (2013). Discrepancies between judgment and choice of action in moral dilemmas. Frontiers in Psychology, 4, 18. doi: 10.3389/fpsyg.2013.00250.CrossRefGoogle ScholarPubMed
Thomson, J. J. (1976). Killing, letting die, and the trolley problem. The Monist, 59(2), 204217. doi: 10.5840/monist197659224CrossRefGoogle ScholarPubMed
Todd, P. M., & Gigerenzer, G. (2012). Ecological rationality: Intelligence in the world. Oxford University Press.Google Scholar
Warnock, M. (2003). Introduction. In Warnock, M. (Ed.), Utilitarianism and on liberty (pp. 116). Blackwell.CrossRefGoogle Scholar
Young, L., Camprodon, J. A., Hauser, M., Pascual-Leone, A., & Saxe, R. (2010). Disruption of the right temporoparietal junction with transcranial magnetic stimulation reduces the role of beliefs in moral judgments. Proceedings of the National Academy of Sciences of the United States of America, 107, 67536758. doi: 10.1073/pnas.0914826107CrossRefGoogle ScholarPubMed