Understanding more about how the brain functions should help us understand economic behaviour. But some would have us believe that it has done this already, and that insights from neuroscience have already provided insights in economics that we would not otherwise have. Much of this is just academic marketing hype, and to get down to substantive issues we need to identify that fluff for what it is. After we clear that away as a distraction, what is left? The answer is that a lot is left, but it is still all potential. That is not a bad thing, or a reason to stop the effort, but it does point to the need for a serious reconsideration of what neuroeconomics is and what passes for explanation in this literature. I argue that neuroeconomics can be a valuable field, but not the way it is being developed and “sold” now. The same is true more generally of behavioural economics, which shares many of the methodological flaws of neuroeconomics.Footnote 1
Epistemology is the study of knowledge claims in philosophy. My complaints are with the epistemological basis of neuroeconomics research as it has been practised. Most of the claims appear to build on poor experimental and statistical foundations, to the point where we end up with a mixture of ad hoc statistical evidence and story-telling, presented as truth or knowledge. In many cases pre-confused experimental designs have simply had neural data tacked on for academic novelty value. Again, this is fine as long as we do not present it as more than it is. But once we deal with what we have before us, we would like to ask the deeper question about whether neuroeconomics can ever add knowledge to economics. Some have already claimed that it cannot, and those claims have to be evaluated as well.
In section 1 I discuss some marketing bloopers of the neuroeconomics literature, since this promotional material is just a distraction from the main issues but is widely cited.Footnote 2 If this is taken for real science, then it is already “game over” in terms of the deeper contribution that neuroeconomics might, and should, make (Rubinstein Reference Rubinstein, Blundell, Newey and Persson2006). In section 2 I discuss some general concerns with applications of neuroeconomics, returning more precisely to some themes emerging in shadowy form in section 1. In section 3 I consider some of the debates over the validity of data from neuroeconomics collected in Caplin and Schotter (Reference Caplin, Schotter and Andrew2008). Section 4 offers a broader framework to think about the contributions that psychology and neuroscience can make to understanding economic behaviour, responding constructively to the concern that Smith (Reference Smith2007: ch.14) expressed about the role of neuroeconomics. By broadening the questions we ask, rather than just trying to answer the same questions with new toys, I believe we can see where neuroeconomics will eventually make a significant contribution to our scientific knowledge about economic behaviour. For now, that promise remains just that.
1. MARKETING BLOOPERS IN THE SELLING OF NEUROECONOMICS
Camerer, Loewenstein and Prelec (Reference Camerer, Loewenstein and Prelec2005) (CLP) provide a dramatic statement of “how neuroscience can inform economics”. They undertake a valuable counterfactual thought experiment: how would economics have evolved if it had been informed at the outset by what we now know from neuroscience? This is not just Monday Morning Quarterbacking, but exactly the kind of meta-analysis of the history of thought in economics that philosophy teaches us to be a valuable methodological tool. So, primed to the purpose of their task, what a disappointment is in store! They claim that neuroscience “. . . points to an entirely new set of constructs to understand economic decision-making” (p. 10). In virtually all cases, however, the claims are just recycled from earlier work in judgement and decision-making, and do not represent insights gained from neuroeconomics as such. But are they, indeed, insights? Let's see.
1.1 Fundamental new concepts?
CLP (p. 32) claim that understanding about how the brain works
. . . challenges some of our fundamental assumptions about how people differ from one-another when it comes to economic behavior. Economists currently classify individuals on such dimensions as “time preference,” “risk preference,” and “altruism.” These are seen as characteristics that are stable within an individual over time and consistent across activities; someone who is risk-seeking in one domain is expected to be risk-seeking in other domains as well. But empirical evidence shows that risk-taking, time discounting, and altruism are very weakly correlated or uncorrelated across situations. This inconsistency results in part from the fact that preferences are state-contingent (and that people may not recognize the state-contingency, which – if they did – would trigger overrides that impose more consistency than observed). But it also may point to fundamental problems with the constructs that we use to define how people differ from each other. [italics added]
So we learn from neuroeconomics that we should, in general, allow state-dependent preferences? This approach would allow a person to have different discount rates for saving, flossing of teeth, dieting, and the decision to get a tattoo, to take their examples.
Of course, economists have known this for decades, and book-length treatments by Hirshleifer and Riley (Reference Hirshleifer and Riley1992) and Chambers and Quiggin (Reference Chambers and Quiggin2000) review the massive literature. Andersen et al. (Reference Andersen, Harrison, Lau and Rutström2008b) illustrate how one can apply it to determine if preferences are stable over time, and the subtlety of putting operational structure on that question even in a controlled experimental setting.
Whether or not we use a state-dependent approach in analysis is a separate matter. The extreme alternative, and no straw man, is presented by Stigler and Becker (Reference Stigler and Becker1977). They are clearly proposing the view that assuming that preferences are stable, and common across individuals, is simply a more useful approach. Judgement day on that hypothesis can wait for our purposes, although there is a behaviourist undercurrent in much of the neuroeconomics literature that presumes that judgement has been made.
But consider the evidence provided to support the claim that we have a challenge to fundamental concepts. “For example, whether a person smokes is sometimes taken as a crude proxy for low rates of time discounting, in studies of educational investment or savings” (p. 32). The use of a crude proxy, no doubt openly acknowledged as same, in an empirical effort, is somehow evidence that we need a new concept of time preferences? I miss the point. The remaining examples are no less anecdotal, and simply irrelevant to the argument (e.g. flossing being correlated with feeding parking meters).
Camerer, Loewenstein and Prelec (Reference Camerer, Loewenstein and Prelec2004: 563) make the same point with a dripping of sarcasm that plays well in plenary lectures, but can be waved aside by any student of economics:
Another example suggests how concepts of economics can be even wider off the mark by neglecting the nature of biological state-dependence: Nobody chooses to fall asleep at the wheel while driving. Of course, an imaginative rational-choice economist – or a satirist – could posit a tradeoff between ‘sleep utility’ and ‘risk of plowing into a tree utility’ and infer that a dead sleeper must have had higher u(sleep) than u(plowing into a tree). But this ‘explanation’ is just tautology.
And lousy, sucker-punch, economics.
1.2 Utility for money?
CLP (p. 35) argue that
. . . neuroscience can point out commonalities between categories that had been viewed as distinct. An example of this with important implications for economics is the utility for money. The canonical economic model assumes that the utility for money is indirect – i.e., that money is a mere counter, only valued for the goods and services it can procure. Thus, standard economics would view, say, the pleasure from food or cocaine and the “pleasure” from obtaining money as two totally different phenomena. Neural evidence suggests, however, [. . .] that money provides direct reinforcement.
Huh?
First, there are several popular models in economics in which money is an argument of the utility function. Some economists grump about those models, but it is more a matter of modelling taste than a clear, defining line between “the one, true model of economics” and “reduced forms used in the heat of the expositional moment”. Important parts of economics written in terms of utility functions defined directly over payoffs include behavioural game theory (Camerer Reference Camerer2003), auction theory, contract theory, and so on. The opening sentence in Pratt's (Reference Pratt1964) classic defining measures of risk aversion is pretty clear on the matter: “Let u(x) be a utility function for money.” If that is not canonical enough for you, then you need to re-load your canon.
Second, it does not take much head scratching to envisage a two-stage, nested utility function in which money enters at some top level, and consumption of non-money goods are purchased with that money. It is simply self-serving formalism, common from the behavioural literature, to construct a straw-man model to attack. Thus, in a modern text on contract theory, Bolton and Dewatripont (Reference Bolton and Dewatripont2005: 4) get on with their work without formal semantics on the issue:
Let us denote the employer's utility function as U(l, t) where l is the quantity of employee time the employer has acquired and t denotes the quantity of “money” – or equivalently the “output” that this money can buy [footnote omitted] – that he has at his disposal. Similarly, employee utility is u(l, t), where l is the quantity of time the employee has kept for herself and t is the quantity of money that she has at her disposal.
The omitted footnote neatly clarifies the obvious to any practising economist: “Indeed, the utility of money here reflects the utility derived from the consumption of a composite good that can be purchased with money.” So if the same reward circuits of the brain fire when subjects get money or beer, that is fine with any number of representations of the utility function in economics.Footnote 3
1.3 Wants are not likes?
CLP (p.37) next take aim at an alleged assumption in economics:
Economists usually view behaviour as a search for pleasure (or, equivalently, escape from pain). The subfield of welfare economics, and the entire ability of economists to make normative statements, is premised on the idea that giving people what they want makes them better off. But, there is considerable evidence from neuroscience and other areas of psychology that the motivation to take an action is not always closely tied to hedonic consequences. Berridge (Reference Berridge1996) argues that decision making involves the interaction of two separate, though overlapping systems, one responsible for pleasure and pain (the “liking” system), and the other for motivation (the “wanting” system). This challenges the fundamental supposition in economics that one only strives to obtain what one likes.
No, this is not what economics assumes at all. We say that choices reveal preferences, on a good inferential day, which is not even close. Binmore (Reference Binmore2007a: 111) explains the methodological difference between these points of view well:
To speak of utility is to raise the ghost of a dead theory. Victorian economists thought of utility as measuring how much pleasure or pain a person feels. Nobody doubts that our feelings influence the decisions we make, but the time has long gone when anybody thought that a simple model of a mental utility generator is capable of capturing the complex mental process that swings into action when a human being makes a choice. The modern theory of utility has therefore abandoned the idea that a util can be interpreted as one unit more or less of pleasure or pain. One of these days, psychologists [neuroeconomists?] will doubtless come up with a workable theory of what goes on in our brains when we decide something. In the interim, economists get by with no theory at all of why people choose one thing rather than another. The modern theory of utility makes no attempt to explain choice behavior. It assumes that we already know what people choose in some situations and uses this data to deduce what they will chose in others – on the assumption that their behavior is consistent.
To non-economist readers, the “modern” theory in question began with Samuelson (Reference Samuelson1938).
1.4 Domain-specific expertise?
CLP (p. 33) argue that “Economics implicitly assumes that people have general cognitive capabilities that can be applied to any type of problem and, hence, that people will perform equivalently on problems that have similar structure.” Really? I thought we had a pretty good working theory of human capital and compensating wage differentials for between-subject comparisons.
Is this also true, however, interpreted as applying on a within-subjects basis? If somebody is presented with an abstract logic puzzle, such as the renowned Wason Selection Task they reference (Wason and Johnson-Laird Reference Wason and Johnson-Laird1972), and represents it differently than when it is presented as a concrete puzzle, it is not just idle semantics to say that they are solving different problems. One can argue that it is a different task when the subject perceives it differently, and then the notion of “similarity of structure” becomes a rich research question in search of a metric.Footnote 4
Of course, one does not want to take that argument too far, or else it guts theory of any generality. The real point of this example is that people tend to perform better at this task when it is presented with concrete, naturally occurring referents, as distinct from an abstract representation.Footnote 5 Moreover, there is important evidence that exposure to realistic instances fails to transfer to abstract instances (Johnson-Laird, Legrenzi and Legrenzi Reference Johnson-Laird, Legrenzi and Sonino Legrenzi1972). These remain important findings, and tell us that the ability to correctly apply principles of logical reasoning depends on syntax and semantics.
The contextual nature of performance is important, but it is a mistake to again think that it is intrinsic to economic reasoning, as distinct from something that is sometimes, or often, assumed away for convenience.
The final example is the Tower of Hanoi puzzle, extensively studied by cognitive psychologists (e.g. Hayes and Simon Reference Hayes, Simon and Gregg1974) and more recently by economists (McDaniel and Rutström Reference McDaniel and Rutström2001) in some fascinating experiments.Footnote 6
Casual observation of students in Montessori classrooms makes it clear how they (eventually) solve the puzzle, when confronted with the initial state. They shockingly violate the constraints and move all the disks to the goal state en masse, and then physically work backwards along the lines of the above thought experiment in backwards induction. The critical point here is that they temporarily violate the constraints of the problem in order to solve it “properly”.
Contrast this behaviour with the laboratory subjects in McDaniel and Rutström (Reference McDaniel and Rutström2001). They were given a computerized version of the game, and told to try to solve it. However, the computerized version did not allow them to violate the constraints. Hence the laboratory subjects were unable to use the classroom Montessori method, by which the student learns the idea of backwards induction by exploring it with physical referents. This is not a design flaw of these lab experiments, but simply one factor to keep in mind when evaluating the behaviour of their subjects. Without the physical analogue of the final goal state being allowed in the experiment, the subject was forced to visualize that state conceptually, and to likewise imagine the penultimate states. Although that might encourage more fundamental conceptual understanding of the idea of backwards induction, if attained, it is quite possible that it posed an insurmountable cognitive burden for some of the experimental subjects.Footnote 7
Harrison and List (Reference Harrison and List2004: 1024) use this example to illustrate one of several defining characteristics of field experiments, in contrast to the typical lab experiment:
It might be tempting to think of this as just two separate tasks, instead of a real commodity and its abstract analogue. But we believe that this example does identify an important characteristic of commodities in ideal field experiments: the fact that they allow subjects to adopt the representation of the commodity and task that best suits their objective. In other words, the representation of the commodity by the subject is an integral part of how the subject solves the task. One simply cannot untangle them, at least not easily and naturally. This example also illustrates that off-equilibrium states, in which one is not optimizing in terms of the original constrained optimization task, may indeed be critical to the attainment of the equilibrium state. Thus we should be mindful of possible field devices which allow subjects to explore off-equilibrium states, even if those states are ruled out in our null hypotheses.
So the point is that tasks might look similar, logically or nominally, from the metric of the equilibrium prediction of some theory given some assumption about the stock of knowledge the subject has, but not be at all similar when one considers off-equilibrium behaviour or heterogeneity in the stock of knowledge subjects have. It is simply inappropriate to claim that economics has no interest in the latter environments, even if it has had a lot more to say about the former environments. We return to this point later, since it alerts us to the importance of thinking about the process leading to choice, rather than the choice itself. This will be the key to seeing what role neuroeconomics can play.
2. BUT IS IT GOOD ECONOMICS?
So much for the marketing: what are we to make of some of the major claims in the neuroeconomics literature? Is it good, interesting economics? We first consider some general issues to do with samples, statistics and procedures. These are troubling, and need to be made explicit because they are blurred in the existing literature, but in the end not the stuff that we should use to pass judgment on the potential role of neuroeconomics. We then examine three illustrative areas of substantive research that should be of immediate interest to economists: discounting over time, social preferences and trust, and the strategic sense that we expect to see in subjects playing games. The theme to emerge is that many confounds that are known in the experimental economics literature are glossed, and the contribution is to simply tack on some neurological data to help tell the preferred story. This brief review provides a basis for asking, in section 3, if this is just a reflection of a nascent field and we should just be more patient, or if it is fundamental.
2.1 Questions about procedures
Sample sizes for many neuroeconomics studies relying on imaging are small if we count a brain as the unit of analysis.Footnote 8 It is common to see studies where the sample size is less than 10, and rarely does one see much more than a dozen or so.Footnote 9 To take a recent example, and something of an outlier at that, Bhatt and Camerer (Reference Bhatt and Camerer2005) used 16 subjects, and noted “To experimental social scientists, 16 seems like a small sample. But for most fMRI studies this is usually an adequate sample to establish a result because adding more subjects does not alter the conclusions much” (p. 432; fn.12). One has to wonder how one can possibly draw the final inference without actually doing it for this experiment and set of statistical inferences. There is presumably some good budgetary explanation for only using small numbers of subjects, particularly in an era in which many neuroeconomists now have their own imaging machinery or ready access to same.
But the unit of analysis is not actually the brain: it is a spatial location in the brain firing per unit of time. So we end up with a data set in which a few brains contribute many observations at each point in time, and in a time-series. Now, econometricians know a few things about handling data like this, but what role do those methods play in the analysis? No problem if they are called something else, but the statistical analysis of these neural data is currently a black box in expositions of neuroeconomics research. Bullmore et al. (Reference Bullmore, Brammer, Williams, Rabe-Hesketh, Janot, David, Mellers, Howard and Sham1995) and Rabe-Hesketh et al. (Reference Rabe-Hesketh, Bullmore and Brammer1997) provide an excellent exposition of the inner workings of this black box, which is a classic “sausage making factory”.
The statistical issues break down into inferences about a single brain, and then a host of new issues coming from inferences over a pooled sample of brains. To understand the significance of the former, here is a list of the estimation methods Rabe-Hesketh et al. (Reference Rabe-Hesketh, Bullmore and Brammer1997) walk through in a typical statistical analysis: filtered Fourier transformations of the raw images (p. 217), interpolation to ensure re-alignment to the first image for that subject (p. 218), simple polynomial regression to correct for intra-subject movement (p. 219 and Figure 2), possible corrections for magnetic field inhomgeneity (p. 219), linear regression of signal intensity value (which is an estimate obtained from prior steps) on a dummy to reflect treatments (p. 221), estimation of a response function to reflect haemodynamically mediated delay between stimulus and response (p. 222–5), and then we get to the point where we start really opening up the econometrics toolkit to worry about serial correlation and heteroscedasticity on a within-subject basis (p. 226ff.). This list varies from application to application, but is sufficiently representative.
Now open up the inferential can of worms involved in pooling across brains. The clinical need to do this arose from a desire to increase statistical power, and now has become the conditio sine qua non of neural comparisons of patients developing symptoms of Alzheimers (so the data become longitudinal). Andreasen et al. (Reference Andreasen, Arndt, Swayze, Cizadlo, Flaum, O'Leary, Ehrhardt and Yuh1994) famously identified the role of abnormalities of the thalamus in patients with schizophrenic disorders by conducting a meta-analysis of scans of 47 normal brains and 39 brains from known schizophrenics. Techniques for normalizing the brain scans, discussed below, were applied, and
an “average brain” was generated for the schizophrenic group and for the normal group. This average brain can be visualized and resampled three-dimensionally in the same manner as an individual MR data set. It has the advantage, however, of providing a concise numeric and visual summary of the group as a whole. (p. 295)
Note that the “summary” in question is not meant to be descriptive: the whole point is that it is to be used inferentially. But there are different ways to normalize brains, as one can imagine (and hope).Footnote 10 And it matters for inference. Tisserand et al. (Reference Tisserand, Pruessner, Sanz Arigita, Boxtel, Evans, Jolles and Uylings2002) examine the impact on inferences about the effect of ageing on regional frontal cortical volumes of using a labor-intensive, manual tracing method, a semi-automatic, partial-brain, volumetric method, and the now-popular whole-brain, voxel-based morphometric methods. They find significant differences across methods, and conclude (p. 667) that
. . . despite the clear advantages of automatic and voxel-based approaches (quick, perfectly reproducible, applicable to large datasets), the current findings suggest that, at present, the most accurate method is still an anatomically based manual tracing one.
The same point is developed in an important evaluation of “Zen and the art of medical image registration” by Crum et al. (Reference Crum, Griffin, Hill and Hawkes2003: 1435):
VBM (Voxel-Based Morphometry) is undoubtedly a powerful framework that successfully removes the need for expert labour-intensive segmentation but currently replaces it with a complicated problem of interpretation and validation which significantly reduces its efficacy. The validity of current published studies relying on NRR (Non-Rigid Registration) in this way is in most cases limited and in some cases suspect due to indiscriminate application of these poorly understood techniques. Until the technology can be made demonstrably more robust one possible solution is to accept that VBM should only be used as a prompting system to highlight regions of the brain worthy of further analysis by more manually intensive techniques. This approach combines the hypothesis-free advantages of VBM with the benefits of expert manual intervention but is by no means an ideal solution.
So it is not just that there are differences in the methods, with some inferential significance, but that the automatic methods have open issues of validation and interpretation. It is apparent from their earliest statements, such as Friston et al. (Reference Friston, Ashburner, Frith, Poline, Heather and Frackowiak1995), that these methods, apparently universally adopted in the neuroeconomics literature, “have deliberately sacrificed a degree of face validity (the ability to independently gauge the correctness of the approach) to ensure construct validity (the validity of the approach by comparison with other constructs)” (Crum et al. Reference Crum, Griffin, Hill and Hawkes2003: 1426). And this is only one of the statistical concerns about where the “left hand side” of the final analysis comes from in neuroeconomics.Footnote 11
A broader problem is that the statistical modelling of neural data is a sequential mixture of limited-information likelihood methods, cobbled together in a seemingly ad hoc manner: MacGyver econometrics. Brain scans do not light up like Xmas trees without lots of modelling assumptions being made. Recognizing this fact is not meant as a way of invalidating the clinical or research goals of the exercise, but an attempt to be sure that we understand the extremely limited extent to which modeling and estimation errors are correctly propagated throughout the chain of inferences. The end result is often a statistical test in which “left hand side” and “right hand side” variables are themselves estimates, often from a long chain of estimates, and are treated as if they are data that are known for certain. The overall implication is that one should be concerned that there is likely a significant understatement of standard errors on estimate of effects, implying a significant overstatement of statistical significant differential activation.
The statistical methods used in many cases are those that have received some acceptance in the neuroscience literature, but those norms and practices may have evolved from clinical needs (e.g. the detection of ischemia, to facilitate the early prevention of a stroke) rather than research needs. It is worth stressing that these issues are debated within the specialist field journals. Thus Crum et al. (Reference Crum, Griffin, Hill and Hawkes2003) note clearly that
. . . the most widely used methods are essentially dumb in that, for a particular registration task, they report only a measure of image similarity which does not allow a judgement of “success” or “failure” to be made. Worse, the magnitude and spatial distribution of errors in NRR are unknown and an understanding of exactly how image-similarity measures and arbitrary transformation models combine in the matching of complex sets of imaged features remains out of reach. (p. 1425) [. . .] these automated approaches are only of value if their performance can be evaluated, preferably in a way that allows correspondence error to be propagated through subsequent statistical analysis. This is currently not the case. . . (p. 1434)
Constructive statistical approaches that take more of a “full information” approach are also starting to be developed in the specialist literature, such as Bowman et al. (Reference Bowman, Caffo, Bassett and Kilts2008) and Wong et al. (Reference Wong, Grosenick, Uy, Guimaraes, Carvalhaes, Desain and Suppes2008), but none of these concerns seem to filter through to qualify the conclusions of the neuroeconomics literature.
A serious issue arises, however, from one of the dark secrets of this field: no neural data is apparently ever made public. For some years requests were made to prominent authors in neuroeconomics, particularly economists with significant and well-earned reputations in experimental economics. Not one has provided data. In some cases the economist could not get the neuroscience co-author to release data, perhaps reflecting traditions in that field.Footnote 12 In any event, one hopes this unfortunate professional habit ends quickly, particularly for government-funded research from tax dollars. No trips to Stockholm, presumably, until statistical methods can be replicated, extended or evaluated with alternatives.
The procedures of many neuroeconomics tasks seem to gloss some of the norms that have evolved, for good or purely historical reason, in experimental economics. For example, Knoch et al. (Reference Knoch, Pascual-Leone, Meyer, Treyer and Fehr2006) studied behaviour in an Ultimatum Bargaining game in which subjects' brains were zapped by “low-frequency repetitive transcranial magnetic stimulation” for 15 minutes prior to making their choices. Quite apart from the ethics of inflicting such an electrical tsunami on subjects, raised by Jones (Reference Jones2007) and responded to by Knoch et al. (Reference Knoch, Pascual-Leone and Fehr2007), the study used deception to lead subjects to think that human subjects were actually providing offers.Footnote 13 The actual offers made were originally made by humans, but in a prior experiment, and the aggregate distribution from that prior distribution was used to generate the distribution given to each subject in the main experiment. Thus the distribution actually received would have less variance than the expected predictive distribution of the subjects. The same deception was employed in studies of behaviour in simple games by Sanfey et al. (Reference Sanfey, Rilling, Aronson, Nystrom and Cohen2003) and Rilling et al. (Reference Rilling, Sanfey, Aronson, Nystrom and Cohen2004), and in their case the deception may have confounded inferences about whether the “theory of mind” regions of the brain are activated. These are the sorts of cheap tricks in design that experimental economists like to avoid.
A final procedural feature of many neuroeconomics studies is the exceedingly cryptic manner in which things are explained. The neuroscience jargon is not the issue: presumably that sub-field has an acronym-rich semantic structure because they find it efficient. Instead, basic behavioural procedures and statistical tests are minimally explained. Many of the journals in question, particularly the prominent general science journals, have severe page restrictions. But in an era of web appendices, that cannot be decisive. In many cases it is simply impossible to figure out, without an extraordinary amount of careful reading, exactly what was done in the economics parts of the experiment and statistical analysis. One wonders how incestuous and unquestioning the refereeing process has become in certain journals.
2.2 Discounting behaviour
The experimental study of time discounting has had a messy history, and procedures have evolved to mitigate concerns with previous studies. The first generation of studies used hypothetical tasks, varying principals, and cognitively difficult ways of eliciting valuations from subjects. The second generation used real rewards and simplified the task in various ways to separate out subjective discount rates from the cognitive burden of evaluating alternatives that differed in several dimensions (principal, front end delay, time horizon). The third generation elicited risk attitudes and discount rates jointly, to allow the latter to be inferred as the rate at which time-dated utility streams were being valued rather than as the rate at which time-dated monetary streams were being valued; see Andersen et al. (Reference Andersen, Harrison, Lau and Rutström2008a) for a review of the literature.
A key procedural issue has been the use of a front end delay on the earlier option. That is, if the subject is asked to choose between an amount of money m in d days and an amount of money M in D days, where M > m and D > d, does d refer to the current experimental session (d = 0) or some time in the future (d > 0)? The reason that this procedural matter assumes such importance is because of the competing explanations for apparently huge discount rates when there is no front end delay (d = 0). By “huge” we mean off the charts: this is a literature that earnestly catalogues annual discount rates in the hundreds or thousands when there is no front end delay.
One explanation for this outcome is that behaviour is consistent with “hyerbolicky” discounting, an expression which conjoins at the hip various families of hyperbolic, generalized hyperbolic and quasi-hyperbolic discounting. The most popular exemplar is quasi-hyperbolic discounting, which posits that the individual has one extremely high rate of time preference for the present versus the future (i.e. maintaining d = 0) and a much lower rate of time preference for time-dated choices in the future (i.e. maintaining d > 0). This hypothesis suggests to McClure et al. (Reference McClure, Laibson, Loewenstein and Cohen2004, Reference McClure, Ericson, Laibson, Loewenstein and Cohen2007) that there could be different parts of the brain activated when considering d = 0 choices and d > 0 choices. They do not claim that it is necessary for there to be different parts activated for the quasi-hyperbolic specification to be supported, just that it would be consistent with that specification if there was such differential brain activity.
An alternative explanation, recognized variously in comments in the earlier literature, and first stated clearly by Coller and Williams (Reference Coller and Williams1999), is that the evidence is an experimental artefact of the credibility and transactions costs of the subject receiving money at the session rather than some other time in the future. This artefact may be what proponents of hyperbolicky specifications mean by a “passion for the present”, but it is distinct from any notions of time preference deriving solely from delay in consumption.Footnote 14 Again, however, it could be that the parts of the brain that light up when doing discounting calculations are different from the parts of the brain that evaluate immediate rewards. The concept of the counterfactual of “the bigger catch we might get tomorrow if we rest today” presumably evolved in our brains separately from the visceral senses that motivate us to tuck in today to tasty food when it is available.
McClure et al. (Reference McClure, Laibson, Loewenstein and Cohen2004) provide evidence that when subjects face decisions in which “money today” is involved, compared with decisions in which “money in the future” is also involved, different parts of the brain light up. For money today, the same regions identified for “rewards” were differentially activated (ventral striatum, medial orbitofrontal cortex, and median prefontal cortex).Footnote 15 However, for decisions over money today and money in the future together, they identified differential activity, inter alia, in the lateral prefontal cortex, which is one of the regions of the brain classically associated with “executive functions” involving trade-offs between competing goals.
So far so good, but how far have we come? It is premature to then conclude (p. 506) that “In economics, intertemporal choice has long been recognize as a domain in which ‘the passions’ can have large sway in affecting our choices (cites omitted). Our findings lend support to this intuition.” Quite apart from the premise not being ascribed to by all economists, as distinct from those that push one version of what is happening with intertemporal choice behaviour, the findings do not speak to “passions” at all, or at least not obviously. The reward regions of the brain reflect the fact that money today will let me buy a sandwich today if I want, or something else as visceral as you like. And money today is surely more credible than money in the future, quite apart from any possible role of the executive function of the brain in translating it into a present value of money today. So the results are equally consistent with the view of hyerbolicky discounting as an artefact of asking subjects to compare “m good apples today” with “M poorer apples tomorrow”, where apples are good or poor in terms of the credibility of me getting to eat them.
McClure et al. (Reference McClure, Laibson, Loewenstein and Cohen2004: 507, fn. 29) make a show of addressing this issue, and ruling out the possibility that the credibility of immediate reward dominates choice:
One possible explanation for increased activity associated with choice sets that contain immediate rewards is that the discounted value for these choice sets is higher than the discounted value of choice sets that contain only delayed rewards.
This could arise if the former choice sets were simply more salient, ceteris paribus the nominal amount of money, because they were more credible to the subject (or, equivalently, not subject to the subjective transactions costs of receiving them). But the manner in which this possibility is addressed is opaque at best:
To rule out this possibility, we estimated discounted value for each choice as the maximum discounted value among the two options. We made the simplifying assumption that subjects maintain a constant weekly discount rate and estimated this value based on expressed preferences (best-fitting value was 7.5% discount rate per week). We then regressed out effects of value from our data with two separate mechanisms. [. . .] Both of these procedures indicate that value has minimal effects on our results, with all areas of activation remaining significant . . .
The procedures in question, and omitted here, are alternative ways of including this estimate of “value” in the statistical model explaining differential voxel activation. There are many unclear aspects of this test. It appears to be making the maintained assumption that the subjects have exponential discounting functions, but attach a higher discount rate to the sooner options simply because they have high discount rates. So the implicit idea is that unless subjective value gets above some threshold, the reward regions of the brain will not differentially fire, and that the only reason they did fire with immediate payments is because of the higher subjective value. But then why assume that every subject had the same discount rate of 7.5% per week?
2.3 Other applications
Many applications of neuroeconomics follow the same pattern as the discounting application just considered. The use of neural data does not provide any insight in relation to the hypotheses being proposed, but is used to promote one favoured explanation even if it does not provide evidence that favours it in relation to known alternative explanations. There are some important exceptions to this pattern, where neural data is examined when subjects are “in equilibrium” as defined by some theory and “out of equilibrium”, to detect possible differences in the processes at work and thereby generate testable hypotheses (e.g. Bhatt and Camerer Reference Bhatt and Camerer2005; Grether et al. Reference Grether, Plott, Rowe, Sereno and Allman2007).
The Trust Game
The Trust Game offers a witches’ brew of confounds. One player receives money from the experimenter (e.g. $10). He can then transfer some or all of that to another player. The experimenter then adds in some money to the transferred pie, typically scaling up the transferred amount by a factor of 3. Then the recipient decides what fraction of the enhanced pie, if any, to send back to the first player. The initial transfer is labelled as “trust”, and the transfer back is labelled “trustworthiness”, with the same skill that phrenologists of old used to label different lumps in the skull.Footnote 16 Of course, several things could motivate someone to send money, or send money back, and these are well known. Either player might be altruistic towards the other player. Or they might be spiteful towards the experimenter, wanting to extract more from him. Or they might be risk loving in the case of “trusting behaviour”. Or they might view the one-shot game through the lens of a repeated game. The implication is that unless one wants to use the words “trust” and “trustworthiness” in the compounded sense of an amalgam of all of these motivations, one needs to design experiments to control for some or all of these other possible confounds, as in the designs of Cox (Reference Cox2004). Any trust experiment that does not do so is just adding to the pile of confusion. Does neuroeconomics improve on things?
Kosfeld et al. (Reference Kosfeld, Heinrichs, Zak, Fischbacher and Fehr2005) and Zak et al. (Reference Zak, Kurzban and Matzner2005) administer oxytocin to some subjects to see if it enhances trust and trustworthiness, respectively. Oxytocin is a neurally active hormone associated with bonding and social recognition. They find evidence that dosed subjects do send more in the Trust game, and that is not associated with them sending more in a nice control in which there is only risk involved. So the risk confound is controlled for, in aggregate “representative agent” terms.Footnote 17 They also find evidence that dosed subjects send more back in the Trust game, and here there is no need for a control for risk. But what about the other confounds? To take the obvious one, altruism, Zak et al. (Reference Zak, Stanton and Ahmadi2007) subsequently demonstrated that oxytocin had no effect on altruism, as measured in a simple Dictator game in which there is no strategic response involved, even if there is a social relationship.Footnote 18 So we later learn that it is probably not altruism, at least in terms of the representative agent. But there are no controls for other known confounds.
DeQuervain et al. (Reference DeQuervain, Fischbacher, Treyer, Schellhammer, Schnyder, Buck and Fehr2004) also examine the Trust game, and trumpet the discovery of a “taste for punishment of unfair behaviour”. They focus on the striatum, which we can stipulate for present purposes as being closely correlated with rewards in the brain. They see the striatum light up when subject A punishes subject B in a way that hurts B's payoffs but costs A nothing, and compare it to how the striatum lights up when the punishment is symbolic, and does not actually hurt B or cost A anything. They find that it lights up more in the initial condition than the latter. So what? If subjects viewed the game as repeated, then it could be a rational strategy to punish those that defect from a profitable strategy (Samuelson Reference Samuelson2005). The underlying game was against different opponents in each round, so the subjects should have taken this into account and realized that it was not a repeated game. But many experimental economists, perhaps most notably Binmore (Reference Binmore2007b: 1–22), would argue that we cannot easily displace field-hardened heuristics for playing repeated games when we drop subjects into an artefactual lab game. These are important methodological issues to resolve before we start adding (costly) brain scans to the analysis, since they perfectly confound inference. In effect, let's get an answer to this exercise in Binmore's (Reference Binmore2007a) text on game theory before we start adding neural correlates:
In laboratory studies, real people don't play the subgame-perfect equilibrium [in a game strategically similar to the Trust game]. The Humean explanation is that people are habituated to playing the fair equilibrium in repeated versions of the game. [. . .] comment on how people would use the words fairness, reputation, and reciprocity if the Humean explanation were correct. Why would this explanation be difficult to distinguish from the claim that people have a taste for a good reputation, fairness, or reciprocity built into their utility functions? (p. 349/350).
This is not to say that one or other of these competing explanations is better, or that indeed we should insist on just one explanation for these data. But it is clear that we have conceptual work to do before we fire up the scanner.
Game theory and the theory of mind
Rilling et al. (Reference Rilling, Sanfey, Aronson, Nystrom and Cohen2004) scanned subjects playing the Ultimatum Bargaining game as well as the Prisoner's Dilemma. In each case the subject played one-shot games against a range of opponents. Some were supposedly human opponents, and a photo shown of the opponent, but the subjects were in fact deceived and the responses generated by computer. The responses were actually drawn from a distribution that reflected true human responses “in a typical uncontrolled version of the game, in which actual human partners” (p. 1696) made decisions. The scanned subjects also played against the computer, and were honestly told that.
Rilling et al. (Reference Rilling, Sanfey, Aronson, Nystrom and Cohen2004) find that areas of the brain associated with a “theory of mind” differentially light up as the scanned subject receives feedback about the choices of the other player.Footnote 19 These areas received positive activation even when computers were generating the responses, which might seem odd. But consider the actual design, where actual human offers were fed to the scanned subjects by the computer: why would subjects not process them in the same way as they would if they had been generated there and then by humans, rather than by some distant human? Rilling et al. (Reference Rilling, Sanfey, Aronson, Nystrom and Cohen2004) report greater activation in the deceptive treatment in which the scanned subjects were led to believe the responses came from actual humans, but they note (p. 1701) several plausible confounds: greater cognitive engagement when told they have a human opponent, which would also explain the activation; changing photos from round to round in the “human” treatments compared to the computer treatments, which instead had the same photograph of a boring computer instead.
3. THE TROUBLE WITH THOUGHT EXPERIMENTS
Gul and Pesendorfer (Reference Gul, Pesendorfer, Caplin and Schotter2008) stirred a debate, collected in Caplin and Schotter (Reference Caplin, Schotter and Andrew2008), on the methodological status of neuroeconomics as positive economics.Footnote 20 They argued that positive economics is defined over choices, and makes no assumptions or predictions about the physiology or processes of the brain. Hence research generating any data on the brain, other than the choices that thinking agents make, is observationally irrelevant to the domain of economics. It is irrelevant, they claim, in the strong sense that it cannot be called on to support or refute economic theories about choice behaviour. Their argument deserves attention, because it quickly brings us to a better understanding of the potential role of neural data, as part of a broader drive to end the unproductive separation between theory and empirics in economics. The separation is not just a matter of specialization by comparative advantage, but has become embedded in the rhetoric used by economists.
3.1 Revealed preference and naked emperors
The beginning of the argument is familiar, from our earlier discussion of what utility functions are and what they are not (§ 1.2), and there is nothing to quarrel with here. But then they move beyond theory and implicitly consider the role of revealed preference as an empirical strategy for recovering preferences from observed choices in order to explain observed behaviour and, presumably, test theory:
In the standard approach, the term utility maximization and choice are synonymous. A utility function is always an ordinal index that describes how the individual ranks various outcomes and how he behaves (chooses) given his constraints (available options). The relevant data are revealed preference data; that is, consumption choices given the individual's constraints. These data are used to calibrate the model (i.e., to identify the particular parameters) and the resulting calibrated models are used to predict future choices and perhaps equilibrium variables such as prices. Hence, standard (positive) theory identifies choice parameters from past behavior and relates these parameters to future behavior and equilibrium variables.
One has the distinct feeling, however, that this is advice coming from non-practitioners.Footnote 21
The problem comes when we try to make the theory operationally meaningful, in the sense of writing out explicit instructions on testing it, and the conditions under which it might ever be refuted. The separation between theory and empiricism cannot just be replaced by casual references to “revealed preference” or, worse, treated with an appended error term. This division of intellectual labor in economics, between theory and empirics, has become a serious hindrance to doing good economics.
Take revealed preference, a beautiful and powerful idea to help sort out misconceptions of behaviour, as illustrated in § 1.2, but practically useless as an empirical tool. First, it imposes very few constraints on behaviour without auxiliary assumptions about the domain of choice. Varian (Reference Varian1988) shows that if one only observes choices of a subset of goods, when the individual is actually making choices over a larger set, then revealed preference places essentially no restrictions on behaviour over the observed subset. Second, we have virtually no systematic theory of how to relate errors in the implications of revealed preference to a degree of belief in the validity of the underlying theory.Footnote 22 Varian (Reference Varian1985) is the only attempt to address this matter, noting (p. 445) that in revealed preference tests the “. . . data are assumed to be observed without error, so that the tests are ‘all or nothing’: either the data satisfy the optimization hypothesis or they don't”. The methods proposed, finding the smallest variations in the observed data to account for violations, is explicitly ad hoc: do you vary income levels, relative prices, quantities chosen, or some (weighted?) combinations of these?
Estimation of parameters is also something that involves a lot more than “calibration” using more and more refinements of revealed preference bounds. Many theorists are seduced by the attractive bounding of indifference curves that we present to our undergraduates, and think this ends up as an effective substitute for explicit econometric methods. Take the example that Gul and Pesendorfer (Reference Gul, Pesendorfer, Caplin and Schotter2008) offer:
The standard approach provides no methods for utilizing nonchoice data to calibrate preference parameters. The individual's coefficient of risk aversion, for example, cannot be identified through a physiological examination; it can only be revealed through choice behavior. If an economist proposes a new theory based on nonchoice evidence then either the new theory leads to novel behavioral predictions, in which case it can be tested with revealed preference evidence, or it does not, in which case the modification is vacuous. In standard economics, the testable implications of a theory are its content; once they are identified, the nonchoice evidence that motivated a novel theory becomes irrelevant.
Consider, then, the estimation of risk aversion under standard Expected Utility Theory, essentially following the standard, pioneering approach of Camerer and Ho (Reference Camerer and Ho1994) and Hey and Orme (Reference Hey and Orme1994).Footnote 23
Assume some parametric utility function in which risk attitudes are determined by one parameter, such as a power function. Conditional on values for this parameter, Expected Utility Theory predicts that one lottery will be selected over the other, or that the subject will be indifferent (so nothing is predicted). Define a latent index of the difference in expected utility, ∇EU, favouring the lottery on the right (R) hand side of a choice. This latent index, based on latent preferences defined by the risk aversion parameter, is then commonly linked to the observed choices using a standard cumulative normal distribution function Φ(∇EU). This “probit” function takes any argument between ±∞ and transforms it into a number between 0 and 1 using the function shown in Figure 1. Thus we have the probit link function,
The logistic function is very similar, as illustrated in Figure 1, and leads instead to the “logit” specification.
Even though Figure 1 is common in econometrics texts, it is worth noting explicitly and understanding. It forms the critical statistical link between observed binary choices, the latent structure generating the index y*, and the probability of that index y* being observed. Thus the likelihood of the observed responses, conditional on the validity of the Expected Utility Theory model and the specific utility parameterizations being true, depends on the estimates of the risk aversion parameter given this statistical specification and the observed choices. The “statistical specification” here includes assuming some functional form for the cumulative density function, such as one of the two shown in Figure 1. Formal maximum likelihood methods for estimating the preference parameter are reviewed in Harrison and Rutström (Reference Harrison, Rutström, Cox and Harrison2008), and are not important here.
An important and popular extension of the core model is to allow for subjects to make some errors. The general notion of error is one that has already been encountered in the form of the statistical assumption that the probability of choosing a lottery is not 1 when the EU of that lottery exceeds the EU of the other lottery. This assumption is clear in the use of a link function between the latent index ∇EU and the probability of picking one or other lottery. If the subject exhibited no errors from the perspective of Expected Utility Theory, this function would be the step function shown in Figure 2: zero for all values of y*<0, anywhere between 0 and 1 for y* = 0, and 1 for all values of y* > 0. In contrast to (1), we then have the connection between preferences and data that a Hardnose Theorist would endorse:
By varying the shape of the link function in Figure 1, one can informally imagine subjects that are more sensitive to a given difference in the index ∇EU and subjects that are not so sensitive. This is what a structural error parameter allows.Footnote 24
3.2 Hardnose theorists or mere theorists?
The problem with the cumulative density function of the Hardnose Theorist is immediate: it predicts with probability one or zero. The likelihood approach asks the model to state the probability of observing the actual choice, conditional on some trial values of the parameters of the theory. Maximum likelihood then locates those parameters that generate the highest probability of observing the data. For binary choice tasks, and independent observations, we know that the likelihood of the sample is just the product of the likelihood of each choice conditional on the model and the parameters assumed, and that the likelihood of each choice is just the probability of that choice. So if we have any choice that has zero probability, and it might be literally 1-in-a-million choices, the likelihood for that observation is not defined. Even if we set the probability of the choice to some arbitrarily small, positive value, the log-likelihood zooms off to minus infinity. We can reject the theory without even firing up any statistical package.
Of course, this implication is true for any theory that predicts deterministically, including Expected Utility Theory. This is why one needs some formal statement about how the deterministic prediction of the theory translates into a probability of observing one choice or the other, and then perhaps also some formal statement about the role that structural errors might play. In short, one cannot divorce the job of the theorist from the job of the econometrician, and some assumption about the process linking latent preferences and observed choices is needed. That assumption might be about the mathematical form of the link, as in (1), and does not need to be built from the neuron up, but it cannot be avoided. Even the very definition of risk aversion needs to be specified using stochastic terms unless we are to impose absurd economic properties on estimates (Wilcox Reference Wilcox, Cox and Harrison2008a, Reference Wilcox2008b).
However, since one has to make some such assumption, why not see if insights can be gained from neuroscience? To some extent the emerging neuroscientific literature on the effect of reward systems on differential brain activition is relevant, since it promises to provide data that can allow us to decide between alternative propensities to make choices. Thus one subject, in one domain, might behave in the manner of Figure 2, one subject more in the manner of one of the curves in Figure 1, and another subject more in the manner of the other curve in Figure 1. With structural noise parameters, and more flexible specifications of the cumulative density function, one could allow a myriad of linking functions to be data driven. Nobody is saying that neural data is sufficient to do this now, and I can think of more efficient ways to go about estimating it before I would put a subject in a scanner, but the methodological point is to identify what role neural data could play. The general point is to see thought experiments for what they are, and to see why we do experiments with real subjects.
3.3 The limits of thought experiments
Sorenson (Reference Sorenson1992) presents an elaborate defense of the notion that a thought experiment is really just an experiment “that purports to achieve its aim without the benefit of execution” (p. 205), and that such experiments can be viewed as “slimmed-down experiments – ones that are all talk and no action”. This lack of execution leads to some practical differences, such as the absence of any need to worry about luck affecting outcomes.Footnote 25 A related trepidation with treating a thought experiment as just a slimmed-down experiment is that it is untethered by the reality of “proof by data” at the end. But this has more to do with the aims and rhetorical goals of doing experiments. As Sorenson (Reference Sorenson1992: 205) notes:
The aim of any experiment is to answer or raise its question rationally. As stressed [earlier. . .], the motives of an experiment are multifarious. One can experiment in order to teach a new technique, to test new laboratory equipment, or to work out a grudge against white rats. (The principal architect of modern quantum electrodynamics, Richard Feynman, once demonstrated that the bladder does not require gravity by standing on his head and urinating.) The distinction between aim and motive applies to thought experiments as well. When I say that an experiment ‘purports’ to achieve its aim without execution, I mean that the experimental design is presented in a certain way to the audience. The audience is being invited to believe that contemplation of the design justifies an answer to the question or (more rarely) justifiably raises its question.
In effect, then, it is caveat emptor with thought experiments – but the same homily surely applies to any experiment, even if executed.
This might all seem like an exceedingly fine point until we consider the link between theory and evidence. We rejoice in an intellectual division of labour between theorists and applied economists, but the absence of explicit econometric instructions on how to test theory has led to some embarrassing debates in economics.Footnote 26 To avoid product liability litigation, it is standard practice to sell commodities with clear warnings about dangerous use, and operating instructions designed to help one get the most out of the product. Unfortunately, the same is not true of economic theories. When theorists undertake thought experiments about individual or market behaviour they are positing “what if” scenarios which need not be tethered to reality. Sometimes theorists constrain their propositions by the requirement that they be “operationally meaningful”, which only requires that they be capable of being refuted, and not that anyone has the technology or budget to actually do so.
4. ECONOMIC BEHAVIOUR AS ALGORITHMIC PROCESS
If the current promotional material and substantive examplars of neuroeconomics leave us disappointed, should we conclude that neuroeconomics is unlikely to ever contribute anything valuable? Smith (Reference Smith2007: 313) draws the conclusion that
Neuroeconomics will not achieve distinction in a focus confined to correcting the ‘errors’ believed to pervade professional economics of the past, an exercise of interest to a narrow few. Nor will notoriety likely stem from better answers to the traditional questions. Rather, neuroeconomic achievement more likely will be determined by its ability to bring a new perspective and understanding to the examination of important economic questions that have been intractable for, or beyond the reach of, traditional economics. Initially new tools tend to be applied to the old questions, but [. . .] their ultimate importance emerges when the tools change how people think about their subject matter, ask new questions, and pursue answers that would not have been feasible before the innovation. Neuroeconomics has this enormous nonstandard potential, but it is far too soon to judge how effective it will be in creating new pathways of comprehension.
The time is nigh to start being explicit about these “new questions,” and to think about what form these “previously infeasible answers” will take. The approach proposed here is to to formally view economic behavior as the outcome of algorithmic processes, and is consistent with many of the intellectual paths leading to neuroeconomics. The idea is familiar to many behavioural economists already, but only as metaphor.Footnote 27
4.1 Behaviour as the outcome of one or more processes
How do individuals make good decisions, and why do they also make bad decisions? Standard economic theory has a relatively simple hypothesis that generates answers to the first question up to a point, but that is silent with respect to the second question. Since experimental data appears to provide a wealth of observations of each type of decision, a good deal of frustration has swirled around the effort to interpret experimental data in terms of standard economic theory. Indeed, it is fair to say that experimental data has provided the lightning rod for debates over the behavioural relevance of standard economic theory.Footnote 28
The general answer offered by economic theory to the first question is that individuals act as if they have optimized a well-behaved objective function subject to some well-defined constraints. This general hypothesis, along with some further structure on the nature of the objective functions and/or the constraints, provides testable restrictions on observable behaviour.Footnote 29 Moreover, there is no presumption that there is only one formulation of the objective function and constraints that can account for the observed behaviour.
Whenever there appears to be evidence that individuals make bad decisions, the data tend to be interpreted in one of two ways. Either the data are questioned as being a valid test of the theory, or the theory is discarded. These are extreme positions, and most serious researchers appropriately qualify their views with the usual double negatives, but in general terms this seems to be the way the empirical tension is resolved.
One possible alternative is to relax the perspective that economic theory has on the behavioural process being observed, and to interpret behaviour as if the subjects are optimizing a well-behaved objective function subject to some well-defined constraints. Keep the “as if” preface, but just shift the tense slightly so that we view the observed subject behavior as potentially being iterations of some algorithmic process rather than as the end-point of that process. An alternative way of viewing this slight shift is to think of the researcher as exploiting the way that the researcher might himself solve the problems that the subject is modelled as solving so as to gain insights into how the subject might be solving the problem. The obvious methodological parallel is to the “rational expectations” insight into modelling the beliefs of subjects in a system.
This emphasis on behaviour as reflecting an algorithmic process is one that is consistent with the rhetoric of many neuroeconomists. One difference is to take the idea of algorithms, and their epistemologically poor cousins, heuristics, seriously as a guiding framework for analysis. Thus one thinks of algorithms as more than metaphors, and exploits the fact that we know a lot about when algorithms work well and when they do not.
4.2 Algorithms versus heuristics?
If one adopts an algorithmic perspective on the decision-making process, it does not follow that one must rule out attention to heuristics. The key distinction between an algorithm and a heuristic has to do with the knowledge claim that they each allow one to make. If an algorithm has been applied correctly, then the result will be a solution that we know something about. For example, we may know that it is a local optimum, even if we do not know that it is a global optimum. Heuristics are lesser epistemological beasts: the solution provided by a heuristic has no claim to be a valid solution in the sense of meeting some criteria. In the computational literature, if not some parts of the psychological literature, heuristics are akin to “rules of thumb” that simply have good or bad track records for certain classes of problems.Footnote 30 The track record may be defined in terms of the speed of arriving at a candidate solution, or the ease of application.
The line between the two is not always clear. Many algorithms can be heuristically applied. For example, one of the most popular ways to start solving an integer programming problem is to define a “relaxed” version of the problem that does not constrain the solution variables to take on discrete values, solve it using some appropriate algorithm for the relaxed problem, and then do a systematic local search for discrete solutions in the neighbourhood of the solution to the relaxed problem. If the result of this procedure is a proposed solution that is not in fact a global optimum, then one should not blame that on the application of the algorithm to the relaxed problem (assuming it was implemented correctly).
Similarly, most of the algorithms in widespread use to solve real problems utilize one or more heuristics to guide their behaviour “under the hood.” Most of the solvers for the GAMS software package, for example, are robust precisely because they have built in much of the experience gleaned by their authors from solving a wide array of instantiations of their problem class. The efficient choice of “stepping size”, for example, can be critical to the speedy application of algorithms for non-linear programming problems. In some difficult problems these algorithms spend a great deal of time explicitly “back-tracking” when iteration steps do not produce the desired improvement in the objective function; but most of the time simple adjustment rules suffice to detect multidimensional escapes from (very) local plateaus. Above all, differences in the knowledge claims that results from applying heuristics or algorithms should not divert from the essential complementarity of the two.
4.3 Homotopies and path-following
The homotopy approach to solving non-linear systems of equations provides a powerful generalization of many existing solution methods, and a general framework to see the complementary roles of economics and several other sciences.Footnote 31 To fix ideas, consider a system that has n variables and n equations, and let x = (x1, x2, . . ., xn) define the values of the variables. We seek the solution values of x, denoted x*, that solve the n × n system of non-linear equations F(x) = 0.
Let H(x(t), t): IRn+1 IRn define a homotopy function over x, where t is a scalar parameter that determines the path to be followed from some initial solution at t = 0 to the final solution at t = 1. Specifically, we define the homotopy function so that at x = x0 the original system F(x) has a “simple solution”. Define this system by E(x) = 0, where we know that x0 solves it, so that H(x(t), 0) = H(x0, 0) = E(x). We examine below what we mean by a simple solution. We also want the homotopy function to solve the original system F(x) = 0 when t = 1, so that H(x(t), 1) = H(x*, 1) = F(x). And along the path defined by tracing out values of t between 0 and 1, we want the homotopy function to define the deformed system obtained by “taking some weighted average” of E(x) and F(x), so that H(x(t), t) = 0 for 0<t<1 as well. The idea of taking a weighted average of the initial solution and the final solution is only a metaphor, but a useful one. When we specify particular homotopy functions in examples below it will be clear in what respects it is a metaphor and in what respects it is not.
The requirements for the homotopy function are (i) that H(x, 0) be easily solved for x0; (ii) that the solution to H(x, 1) = 0, x*, is the solution to F(x) = 0; and (iii) that the path x(t) leads from x0 to x* as t increases. If we can find some representation of the original system F(x) = 0 that is easy to solve, and if we can define a path from that initial representation to the original system that is well-behaved in the sense that we can easily follow it, then we will have a way of finding the solution to the original system.
What do we mean by a simple solution to the initial, deformed system E(x)? Literally, any solution that the human machine, aided with field referents, firing executive-region neurons, and even a dose of cognitive serendipity, can ascertain. To take some concrete and familiar examples (Garcia and Zangwill Reference Garcia and Zangwill1981: ch.1), the Newton homotopy lets one pick an arbitrary x0, providing it has the same dimensionality as the final solution, and then start from it. Given x0, you calculate F(x0), adopt E(x) ≡ F(x) – F(x0), and the homotopy function then becomes H(x, t) = F(x) – (1–t) F(x0). So this method would be an appropriate formalization if one had some reason to think that subjects had some focal point solution for their calculations: in a first-price sealed-bid auction over private values, for example, the bidder's private value itself is an obvious suggestion. An alternative homotopy, also attractive because it allows one to start at an arbitrary x0, is the Fixed-Point homotopy. In this case let E(x) ≡ x – x0, and use the function H(x, t) = (1–t)(x – x0) + tF(x). Or the final system F(x) might be seen, psychologically or mathematically, to be similar to some special-case with known solution E(x), and the Linear homotopy can be used in which H(x, t) = tF(x) + (1–t)E(x) = E(x) + t[F(x) – E(x)]. This might be appropriate if there was some special case of the final system that was relatively easy to solve, or familiar from field contexts, such as infinitely repeated versions of a one-shot game. Note well that in each example E(x) is a deformation and representation of F(x).
It is easy to see how this formalization neatly admits many of the interests of psychologists, behavioural economists, neuroeconomists and mainstream economists.
Psychologists have spent a lot of time studying task representation and recognition, which can be taken as the psychological counterpart to selecting the starting point of the homotopy. There are relatively few formal restrictions on what constitutes a good starting point: for instance, Newton-type homotopies and Fixed-Point homotopies allow virtually arbitrary values for the solution variables to be used. So this is fertile ground for similarity relations to be applied to explain the process by which individuals adopt “deformed representations” of the original problem that can be solved quickly by the neural hardware and software that the brain provides. Sometimes this may be a conscious choice of a deformation, or it might be a representation that is dredged up by the brain from the evolutionary bog without any executive function operating at all. Or it might be conscious and deliberate for some individuals, or some task domains, and automatic for others. In any event, we have the beginnings of a process that can lead to considerable heterogeneity in the path to be followed.
Without naming it as such, experimental economists and cognitive psychologists have spent a lot of time studying when subjects apply more or less effort to solving problems. These measures of effort can be viewed as indicia of the fraction of the path followed and the effort assigned to coming up with the initial representation. In some cases the measures are as direct as response time (Luce Reference Luce1986: Wilcox Reference Wilcox1993; Rubinstein Reference Rubinstein2007) or mouse clicking (Johnson et al. Reference Johnson, Camerer, Sankar and Tymon2002).
Insights about the type of path being followed also come from applied mathematics. Allgower and Georg (Reference Allgower and Georg2003) draw a distinction between two types of ways in which one can follow a homotopy path. One is the class of “predictor-corrector” methods, which attempt to closely trace out some smooth path defined by the homotopy function. These methods generally rely on the ability of the system to undertake lots of very small steps if the path is non-linear, and presume smoothness of the path. On the other hand, one might use “piecewise linear” methods that derive from convenient triangulations of the space over which the path is defined. These triangulations have the advantage that they might reflect efficient ways to store information: all that is needed is the information on the current simplex, and the rules needed to go to the next one (the “pivoting step”). So they might reflect pre-existing neuronal hardware developed for some other purpose, reflecting the evolutionary nature of the human brain (Dehaene et al. Reference Dehaene and Tsivkin1999; Linden Reference Linden2007).
Similarly, the issue of “stopping rules” has been extensively studied by decision theorists (von Winterfeldt and Edwards Reference von Winterfeldt and Edwards1982, Reference von Winterfeldt and Edwards1986) and experimental economists (Harrison Reference Harrison1989, Reference Harrison1992: Merlo and Schotter Reference Merlo and Schotter1992). The major issue has been to identify the metric of evaluation that subjects employ when evaluating alternative off-equilibrium choices, but there also remain mundane algorithmic issues such as identifying solution tolerances.
Homotopy methods can be used to solve an extraordinarily wide range of problems of interest to economists. Constrained optimization problems are standard fare, but equilibrium problems encountered in general equilibrium analysis and applied game theory can also be solved using path-following methods. Indeed, many existing solution methods are effectively path-following methods even if they have not traditionally been presented that way. For example, the “tracing procedure” of Harsanyi (Reference Harsanyi1975) is readily seen to be a homotopy method. Garcia and Zangwill (Reference Garcia and Zangwill1981: Part II) provide an extensive series of examples. Therefore, one advantage of the general path-following approach is that it offers a relatively unified approach to wide classes of problems in economics.
An explicitly algorithmic framework admits of the latent process assumed by economists as a special case, but allows space for insights from psychology, cognitive science, and, yes, even neuroscience.Footnote 32
4.4 Hardware, software, and explanation
One feature of the performance of algorithms, of some significance for the manner in which neuroeconomics seeks to explain behaviour, is the relationship between hardware and software. Understanding the physics of computers does not help us much to understand what makes one algorithm better than another. But once we know what we want an algorithm to do, and the various ways it can achieve its goal, we do care about the hardware. We care about the speed of the processor, we care about the available of concurrent processors, we care about input-output speed and hard disk capacity, and we might even care about rendering speed for graphics. This is exactly the concern of Marr (Reference Marr1982: 27ff.):
Trying to understand perception by studying only neurons is like trying to understand bird flight by understanding only feathers: It just cannot be done. In order to understand bird flight, we have to understand aerodynamics; only then do the structure of feathers and the different shapes of birds’ wings make sense. More to the point, we cannot understand why retinal ganglion cells and lateral geniculate neurons have the receptive fields that they do just by studying their anatomy and physiology. We can understand how these cells and neurons behave as they do by studying their wiring and interactions, but in order to understand why the receptive fields are as they are – why they are circularly symmetrical and why their excitatory and inhibitory regions have characteristic shapes and distributions – we have to know a little of the theory of differential operators, band-pass channels, and the mathematics of the uncertainty principle.
So look at brains, but understand why we do so. This perspective is a “top down” one from a functional perspective, not a reductionist “bottoms up” approach.Footnote 33 But it does not ignore or belittle the role of the hardware; instead, it just puts it in its place when it comes to trying to explain why behaviour occurs in the manner in which it does.
In fact, neuroeconomics should have been ideally positioned to pursue this approach, since economists have such a strong sense of what problems economic behaviour can be usefully viewed as solving. This is not to take a position on whether Expected Utility Theory is the best model of choice under uncertainty for every individual in every task domain, or whether the straw-man of self-regarding preferences is the best one to assume, to take two examples, but to point out that we already had a lot of that work done. This caveat is critical, since it is one of the later criticisms of the approach developed in Marr's name,Footnote 34 that it presumed the definition of “the” computational goal of the “agent”. Economists have learned, from internal debates, how to extend their standard paradigm in astonishing ways (Stigler and Becker Reference Stigler and Becker1977), how to apply their own optimization paradigm to more elaborate objective functions and constraints (e.g. the alternatives to Expected Utility Theory), and even how to contemplate the possibility that there might be multiple computational goals at work (mixture specifications, as in Harrison and Rutström Reference Harrison and Rutström2005).
4.5 Agency
One way to couch the debate over the role of neuroeconomics is to see it as a debate over the meaning of “agency” in economics: who is the economic agent? To many economists this seems like a non-question, but it is front and centre in philosophical debates over economics and cognitive science. Ross (Reference Ross2005) provides a detailed review of the issues, which run deep in philosophy.Footnote 35
One perspective is to think of the brain as made up of multiple selves or agents, so that agency is defined in terms of the part of the brain that makes specific decisions.Footnote 36 Some use this as metaphor, and others see it as more literal. The concept of “dual selves” has a long lineage in behavioural economics, and findings from neuroscience certainly suggest that multiple brain systems interact when subjects make economic decisions (Cohen Reference Cohen2005). An alternative interpretation of the concept “dual selves” would be a single decision maker that has dual cognitive processes that are activated under different conditions. This interpretation is consistent with the literature on dual process theories of mind in psychology and economics (e.g. see Lopes Reference Lopes, Busemeyer, Hastie and Medin1995; Barrett et al. Reference Barrett, Tugade and Engle2004; Benhabib and Bisin Reference Benhabib and Bisin2005; and their references to the older literature). It also lends itself to formalization in certain structured settings (e.g. Andersen et al. Reference Andersen, Harrison, Lau and Rutström2008a) and more generally using mixture specifications defined over multiple latent decision-making processes (e.g. Harrison and Rutström Reference Harrison and Rutström2005; Andersen et al. Reference Andersen, Harrison, Lau and Rutström2007).
4.6 Multiple levels of selection
A fundamental tenet of neuroeconomics appears to be the idea that it is neuronal activity that ultimately determines if we behave in one way or another in the economic domain. The algorithmic approach offers another model of selection of behaviour: in some cases the process is causal in the direction suggested by neuroeconomists, but in other cases the process might have the reverse causality. These processes might operate concurrently, sometimes over very short periods of time, and sometimes over extended evolutionary time (e.g., Keller Reference Keller1999). Sunder (Reference Sunder2006: 322) identified this as one of the insights from an algorithmic perspective on economic behaviour:
The marriage of economics and computers led to a serendipitous discovery: there is no internal contradiction in suboptimal behaviour of individuals yielding aggregate-level outcomes derivable from assuming individual optimization. Individual behaviour and aggregate outcomes are related but distinct phenomena. Science does not require integration of adjacent disciplines into a single logical structure. As the early-twentieth-century unity of science movement discovered, if we insist on reducing all sciences to a single integrated structure, we may have no science at all. In Herbert Simon's (Reference Simon1996: 16) words: “This skyhook-skyscraper construction of science from the roof down to the yet unconstructed foundations was possible because the behavior of the system at each level depended on only a very approximate, simplified, abstracted characterization of the system at the level next beneath. This is lucky; else the safety of bridges and airplanes might depend on the correctness of the ‘Eightfold Way’ of looking at elementary particles.” This is the story of how we found that economists can have their cake while psychologists eat it too.
And there might even be some cake left for neuroscientists: if not, blame (cognitive) psychologists!
5. CONCLUSIONS
From the perspective of economists, the neuroeconomics literature seems to have used a production function with a sub-optimal mix of human capital and physical capital, in a blushing fascination with the toys of neuroscience. The result has been dazzling images of light bulbs popping on in different parts of the brain, but unimpressive economics. Straw-men are erected as null hypotheses, multiple alternative hypotheses are ignored and left behind as the literature cites itself in a spiral, and known confounds are glossed. As the behavioural economics literature demonstrated, however, we already knew how to do poor economics (and get it published). The fear is that the impressive and important machinery of neuroscience will make it even harder for anyone to know what passes for scientific knowledge in economics and what is just great story-telling.
We can put the academic marketing of neuroeconomics aside. It almost seems unfair to put some of those claims on display, but, like the assertions of the behaviourists, they have taken hold in many quarters as knowledge claims when they are just “chloroform in print”. Since economists have important and serious questions to get on with, the opportunity cost of these diversions has just become too great to ignore.
The more important business is to decide what to make of the substantive claims of neuroeconomics. In this respect the evaluation is mixed.
As an economist I do not learn much on the broad substantive issues reviewed, and these cover what should be the low-hanging fruit for neuroeconomics. The lack of insight does not primarily come from unfamiliarity with the details of the methods: some of those details bother me, and need exposition by the economists on these teams, but that is not in the end decisive. My main concern is whether neuroeconomists have added insight to already-confused experimental designs, or just covered up those confusions and promoted one plausible story over another. I conclude the latter, unfortunately. Obviously there is no need to add neural correlates to pre-confused designs.
But the potential remains. I reject the view that neural data must be irrelevant to economics as needlessly isolationist. I do not take the free-disposability view that any data is useful data until proven otherwise, implying that we should just collect it anyway and decide later if it was useful; that is a poor model for advancement of study in any field.Footnote 37 Instead, I encourage a restatement of the formal processes by which agents make economic decisions, so that we can better see what questions neural data can provide insight into. This restatement does not mean rejecting what we have already in mainstream economics, but viewing it as a special case which may or may not be applicable in certain domains. The framework I have in mind leaves a clear role for neural data, but a more urgent role for proper, sustained communication between economics and cognitive psychology.