Hostname: page-component-745bb68f8f-hvd4g Total loading time: 0 Render date: 2025-02-11T20:38:26.576Z Has data issue: false hasContentIssue false

Post hoc rationalism in science

Published online by Cambridge University Press:  25 August 2011

Eric Luis Uhlmann
Affiliation:
HEC Paris–School of Management, Management and Human Resources Department, 78351 Jouy-en-Josas, France. eric.luis.uhlmann@gmail.comwww.socialjudgments.com

Abstract

In advocating Bayesian Enlightenment as a solution to Bayesian Fundamentalism, Jones & Love (J&L) rule out a broader critique of rationalist approaches to cognition. However, Bayesian Fundamentalism is merely one example of the more general phenomenon of Rationalist Fundamentalism: the tendency to characterize human judgments as rational and optimal in a post hoc manner, after the empirical data are already known.

Type
Open Peer Commentary
Copyright
Copyright © Cambridge University Press 2011

Jones & Love (J&L) are right to criticize what they term “Bayesian Fundamentalism” as not empirically grounded, uninformed by psychological data, open to multiple rational accounts of a task or decision, and conducive to post hoc explanations. However, in advocating Bayesian Enlightenment as a solution, they appear to rule out a broader critique of rationalist approaches to human cognition. Specifically, Bayesian Fundamentalism is one example of the more general phenomenon of Rationalist Fundamentalism: the tendency to characterize a given judgment as rational and optimal in a post hoc manner, after the empirical data are already known. Few researchers would argue that human behavior is perfectly optimal and rational. However, a desire to see the human mind as operating rationally, and the use of post hoc justifications to reach this conclusion, suggest we should be skeptical of after-the-fact “rational” explanations.

Decades of empirical studies show people are strongly motivated to see themselves as rational and objective (for reviews, see Armor Reference Armor1999; Pronin et al. Reference Pronin, Gilovich and Ross2004; Pyszczynski & Greenberg Reference Pyszczynski, Greenberg and Berkowitz1987; Ross & Ward Reference Ross, Ward, Reed and Turiel1996). Decision makers engage in motivated reasoning and psychological rationalizations designed to preserve this “illusion of objectivity” (Armor Reference Armor1999; Pronin et al. Reference Pronin, Lin and Ross2002) – for instance, changing their definition of what an optimal judgment is after the fact (Dunning & Cohen Reference Dunning and Cohen1992; Epstein et al. Reference Epstein, Lipson, Holstein and Huh1992; Kunda Reference Kunda1987; Norton et al. Reference Norton, Vandello and Darley2004; Uhlmann & Cohen Reference Uhlmann and Cohen2005). Evidence that general psychological processes are not rational or optimal represents a threat to this cherished illusion. Fundamentalist resistance to evidence of human irrationality further stems from economics and related disciplines, in which optimality and the maximization of utility are widely perceived as necessary assumptions about human behavior.

A rationalist defense can involve constructing a post hoc Bayesian account of an empirical finding predicted a priori from theories grounded in psychological limitations and motives. Consider the phenomenon of biased assimilation, in which participants rate a scientific study that supports their political beliefs (e.g., about the deterrent effects of capital punishment) as methodologically superior to a study that refutes their beliefs (Lord et al. Reference Lord, Ross and Lepper1979). The cognitive-rationalist interpretation is that decision makers are simply making Bayesian inferences, taking into account subjective probabilities (e.g., their prior political beliefs) when evaluating new evidence. However, further findings contradict the claim that biased assimilation is merely the product of Bayesian inferences. For instance, individuals whose positive self-image is affirmed are less likely to exhibit biased assimilation (Cohen et al. Reference Cohen, Aronson and Steele2000; see also Dunning et al. Reference Dunning, Leuenberger and Sherman1995; Sherman & Cohen Reference Sherman and Cohen2002). This is consistent with the idea that biased information processing stems from a motivated desire to dismiss evidence that threatens valued beliefs and, by extension, the self (Sherman & Cohen Reference Sherman, Cohen and Zanna2006; Steele Reference Steele and Berkowitz1988). When a decision maker is feeling good about herself, there is less need to be biased. In addition, would-be parents who believe day care is bad for children, but plan to use day care themselves (and therefore desire to conclude that day care is just as good as home care), show biased assimilation in favor of day care (Bastardi et al. Reference Bastardi, Uhlmann and Ross2011). What decision makers desire to be true seems to trump what they believe to be factually true – the ostensive basis for any Bayesian inferences.

As J&L point out, one of the most problematic aspects of rational models is how little attention can be paid to whether the assumptions of the statistical model correspond to what is actually going on in people's heads as they engage in a task or make a decision. I once debated an economist who argued that micro-level psychological data on what goals people pursue in the dictator game are irrelevant: The material self-interest account must be true if people's offers correspond to the predictions of the statistical model. However, it is dangerous to assume that because a rational statistical model can mimic or reproduce a pattern of data, the underlying psychological process is a rational one. That a computer can mimic some of the outputs of human thought does not necessarily mean the mind functions in the same way as a computer.

The last defense of post hoc rationalism is to swap normative models of rationality entirely. In other words, researchers can speculate post-hoc as to what alternative goals decision-makers may have been pursuing, in order to preserve the view that participants were acting rationally. Never mind the goals to optimize material outcomes or achieve accuracy: Judgmental biases can be defined as “rational” because they preserve the decision maker's personal self-image, psychological well-adjustment, public reputation, cherished religious beliefs, desire to punish norm violators, existential goals, likelihood of survival in ancestral environments, or even the happiness of their marriage (Cosmides & Tooby Reference Cosmides, Tooby, Hirschfeld and Gelman1994; Hamilton Reference Hamilton1980; Krueger & Funder Reference Krueger and Funder2004; Lerner & Tetlock Reference Lerner and Tetlock1999; Tetlock Reference Tetlock2002; Tetlock et al. Reference Tetlock, Kristel, Elson, Green and Lerner2000; Reference Tetlock, Visser, Singh, Polifroni, Elson, Mazzocco and Rescober2007).

It has been argued that the heuristics-and-biases approach to cognition is itself biased, in the direction of attributions to irrationality (Krueger & Funder Reference Krueger and Funder2004). Despite its shortcomings, however, the heuristics-and-biases research program is at least based on a priori theoretical hypotheses. There are few cases of “post hoc irrationalism” in which robust empirical effects predicted a priori by Bayesian or otherwise rationalist models are redefined post hoc as due to motives such as the need for self-esteem or control.

Although Bayesian Enlightenment, as advocated by J&L, is a major improvement on Bayesian Fundamentalism, it is still subject to post hoc rationalism. An interface between Bayesian or otherwise rationalist models and data on psychological processes leaves plenty of room for the former to distort interpretations of the latter. A wealth of evidence indicates that human beings are subject to a powerful illusion of rationality and objectivity they are strongly motivated to maintain and which influences their perceptions of scientific data. Researchers are also human beings. It would be remarkable indeed if scientists were immune to the empirical phenomena we study.

References

Armor, D. A. (1999) The illusion of objectivity: Bias in the belief in freedom from bias. Doctoral dissertation, University of California, Los Angeles.Google Scholar
Bastardi, A., Uhlmann, E. L. & Ross, L. (2011) Wishful thinking: Belief, desire, and the motivated evaluation of scientific evidence. Psychological Science 22:731–32.Google Scholar
Cohen, G. L., Aronson, J. & Steele, C. M. (2000) When beliefs yield to evidence: Reducing biased evaluation by affirming the self. Personality and Social Psychology Bulletin 26:1151–64.Google Scholar
Cosmides, L. & Tooby, J. (1994) Origins of domain specificity: The evolution of functional organization. In: Mapping the mind: Domain specificity in cognition and culture, ed. Hirschfeld, L. A. & Gelman, S., pp. 85116. Cambridge University Press.Google Scholar
Dunning, D. & Cohen, G. L. (1992) Egocentric definitions of traits and abilities in social judgment. Journal of Personality and Social Psychology 63:341–55.Google Scholar
Dunning, D., Leuenberger, A. & Sherman, D. A. (1995) A new look at motivated inference: Are self-serving theories of success a product of motivational forces? Journal of Personality and Social Psychology 69:5868.CrossRefGoogle Scholar
Epstein, S., Lipson, A., Holstein, C. & Huh, E. (1992) Irrational reactions to negative outcomes: Evidence for two conceptual systems. Journal of Personality and Social Psychology 62:328–39.Google Scholar
Hamilton, V. L. (1980) Intuitive psychologist or intuitive lawyer: Alternative models of the attribution process. Journal of Personality and Social Psychology 39:767–73.CrossRefGoogle Scholar
Krueger, J. I. & Funder, D. C. (2004) Towards a balanced social psychology: Causes, consequences, and cures for the problem-seeking approach to social behavior and cognition. Behavioral and Brain Sciences 27:313–27.Google Scholar
Kunda, Z. (1987) Motivated inference: Self-serving generation and evaluation of causal theories. Journal of Personality and Social Psychology 53:3754.CrossRefGoogle Scholar
Lerner, J. S. & Tetlock, P. E. (1999) Accounting for the effects of accountability. Psychological Bulletin 125:255–75.CrossRefGoogle ScholarPubMed
Lord, C. G., Ross, L. & Lepper, M. R. (1979) Biased assimilation and attitude polarization: The effects of prior theories on subsequently considered evidence. Journal of Personality and Social Psychology 37:2098–109.CrossRefGoogle Scholar
Norton, M. I., Vandello, J. A. & Darley, J. M. (2004) Casuistry and social category bias. Journal of Personality and Social Psychology 87:817–31.Google Scholar
Pronin, E., Gilovich, T. & Ross, L. (2004) Objectivity in the eye of the beholder: Perceptions of bias in self versus others. Psychological Review 111:781–99.Google Scholar
Pronin, E., Lin, D. Y. & Ross, L. (2002) The bias blind spot: Perceptions of bias in self versus others. Personality and Social Psychology Bulletin 28:369–81.Google Scholar
Pyszczynski, T. & Greenberg, J. (1987) Toward an integration of cognitive and motivational perspectives on social inference: A biased hypothesis-testing model. In: Advances in experimental social psychology, vol. 20, ed. Berkowitz, L., pp. 297340. Academic Press.Google Scholar
Ross, L. & Ward, A. (1996) Naive realism in everyday life: Implications for social conflict and misunderstanding. In: Values and knowledge, ed. Reed, E. S. & Turiel, E., pp. 103–35. Erlbaum.Google Scholar
Sherman, D. K. & Cohen, G. L. (2002) Accepting threatening information: Self-affirmation and the reduction of defensive biases. Current Directions in Psychological Science 11:119–23.Google Scholar
Sherman, D. K. & Cohen, G. L. (2006) The psychology of self-defense: Self-affirmation theory. In: Advances in Experimental Social Psychology, vol. 38, ed. Zanna, M. P., pp. 183242. Academic Press.Google Scholar
Steele, C. M. (1988) The psychology of self-affirmation: Sustaining the integrity of the self. In: Advances in Experimental Social Psychology, vol. 21, ed. Berkowitz, L., pp. 261302. Academic Press.Google Scholar
Tetlock, P. E. (2002) Social functionalist frameworks for judgment and choice: Intuitive politicians, theologians, and prosecutors. Psychological Review 109:451–71.Google Scholar
Tetlock, P. E., Kristel, O., Elson, B., Green, M. & Lerner, J (2000) The psychology of the unthinkable: Taboo trade-offs, forbidden base rates, and heretical counterfactuals. Journal of Personality and Social Psychology 78:853–70.Google Scholar
Tetlock, P. E., Visser, P., Singh, R., Polifroni, M., Elson, B., Mazzocco, P. & Rescober, P. (2007) People as intuitive prosecutors: The impact of social control motives on attributions of responsibility. Journal of Experimental Social Psychology 43:195209.Google Scholar
Uhlmann, E. L. & Cohen, G. L. (2005) Constructed criteria: Redefining merit to justify discrimination. Psychological Science 16:474–80.Google Scholar