Hostname: page-component-745bb68f8f-f46jp Total loading time: 0 Render date: 2025-02-06T10:05:39.246Z Has data issue: false hasContentIssue false

Neural reuse: A fundamental organizational principle of the brain

Published online by Cambridge University Press:  22 October 2010

Michael L. Anderson
Affiliation:
Department of Psychology, Franklin & Marshall College, Lancaster, PA 17604, and Institute for Advanced Computer Studies, Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD 20742. michael.anderson@fandm.eduhttp://www.agcognition.org
Rights & Permissions [Opens in a new window]

Abstract

An emerging class of theories concerning the functional structure of the brain takes the reuse of neural circuitry for various cognitive purposes to be a central organizational principle. According to these theories, it is quite common for neural circuits established for one purpose to be exapted (exploited, recycled, redeployed) during evolution or normal development, and be put to different uses, often without losing their original functions. Neural reuse theories thus differ from the usual understanding of the role of neural plasticity (which is, after all, a kind of reuse) in brain organization along the following lines: According to neural reuse, circuits can continue to acquire new uses after an initial or original function is established; the acquisition of new uses need not involve unusual circumstances such as injury or loss of established function; and the acquisition of a new use need not involve (much) local change to circuit structure (e.g., it might involve only the establishment of functional connections to new neural partners). Thus, neural reuse theories offer a distinct perspective on several topics of general interest, such as: the evolution and development of the brain, including (for instance) the evolutionary-developmental pathway supporting primate tool use and human language; the degree of modularity in brain organization; the degree of localization of cognitive function; and the cortical parcellation problem and the prospects (and proper methods to employ) for function to structure mapping. The idea also has some practical implications in the areas of rehabilitative medicine and machine interface design.

Type
Target Article
Copyright
Copyright © Cambridge University Press 2010

Although an organ may not have been originally formed for some special purpose, if it now serves for this end we are justified in saying that it is specially contrived for it. On the same principle, if a man were to make a machine for some special purpose, but were to use old wheels, springs, and pulleys, only slightly altered, the whole machine, with all its parts, might be said to be specially contrived for that purpose. Thus throughout nature almost every part of each living being has probably served, in a slightly modified condition, for diverse purposes, and has acted in the living machinery of many ancient and distinct specific forms.

— Charles Darwin (Reference Darwin1862), p. 348.

1. Introduction and background

Research in the cognitive neurosciences has long been guided by the idealization that brain regions are highly selective and specialized, and that function can be mapped to local structure in a relatively straightforward way. But the degree of actual selectivity in neural structures is increasingly a focus of debate in cognitive science (Poldrack Reference Poldrack2006). It appears that many structures are activated by different tasks across different task categories and cognitive domains. For instance, although Broca's area has been strongly associated with language processing, it turns out to also be involved in many different action- and imagery-related tasks, including movement preparation (Thoenissen et al. Reference Thoenissen, Zilles and Toni2002), action sequencing (Nishitani et al. Reference Nishitani, Schürmann, Amunts and Hari2005), action recognition (Decety et al. Reference Decety, Grezes, Costes, Perani, Jeannerod, Procyk, Grassi and Fazio1997; Hamzei et al. Reference Hamzei, Rijntjes, Dettmers, Glauche and Weiller2003; Nishitani et al. Reference Nishitani, Schürmann, Amunts and Hari2005), imagery of human motion (Binkofski et al. Reference Binkofski, Amunts, Stephan, Posse, Schormann, Freund, Zilles and Seitz2000), and action imitation (Nishitani et al. Reference Nishitani, Schürmann, Amunts and Hari2005; for reviews, see Hagoort Reference Hagoort2005; Tettamanti & Weniger Reference Tettamanti and & Weniger2006). Similarly, visual and motor areas – long presumed to be among the most highly specialized in the brain – have been shown to be active in various sorts of language processing and other higher cognitive tasks (Damasio & Tranel Reference Damasio and Tranel1993; Damasio et al. Reference Damasio, Tranel, Hichwa and Damasio1996; Glenberg & Kaschak Reference Glenberg and Kaschak2002; Hanakawa et al. Reference Hanakawa, Honda, Sawamoto, Okada, Yonekura, Fukuyama and Shibasaki2002; Martin et al. Reference Martin, Haxby, Lalonde, Wiggs and Ungerleider1995; Reference Martin, Wiggs, Ungerleider and Haxby1996; Reference Martin, Ungerleider, Haxby and Gazzaniga2000; Pulvermüller Reference Pulvermüller2005; see sect. 4 for a discussion). Excitement over the discovery of the Fusiform Face Area (Kanwisher et al. Reference Kanwisher, McDermott and Chun1997) was quickly tempered when it was discovered that the area also responded to cars, birds, and other stimuli (Gauthier et al. Reference Gauthier, Skudlarski, Gore and Anderson2000; Grill-Spector et al. Reference Grill-Spector, Sayres and Ress2006; Rhodes et al. Reference Rhodes, Byatt, Michie and Puce2004). The ensuing debates over the “real” function of these areas have still not been resolved.

This is just a short list of some highly-studied regions for which the prospect of a clear-cut mapping of function to structure appears dim. In this target article, I will review a great deal more evidence that points in a similar direction. But if selectivity and localization are not in fact central features of the functional organization of the brain, how shall we think about the function-structure relationship? This target article reviews an emerging class of theories that suggest neural circuits established for one purpose are commonly exapted (exploited, recycled, redeployed) during evolution or normal development, and put to different uses, often without losing their original functions. That is, rather than posit a functional architecture for the brain whereby individual regions are dedicated to large-scale cognitive domains like vision, audition, language, and the like, neural reuse theories suggest instead that low-level neural circuits are used and reused for various purposes in different cognitive and task domains.

In just the past five years, at least four different, specific, and empirically supported general theories of neural reuse have appeared. Two of these theories build on the core notion of the sensorimotor grounding of conceptual content to show how it could implicate many more aspects of human cognitive life: Vittorio Gallese's “neural exploitation” hypothesis (Gallese Reference Gallese2008; Gallese & Lakoff Reference Gallese and Lakoff2005) and Susan Hurley's “shared circuits model” (Hurley Reference Hurley, Hurley and Chater2005; Reference Hurley2008). Two other theories suggest that reuse could be based on even more universal foundations: Dehaene's “neuronal recycling” theory (Dehaene Reference Dehaene, Dehaene, Duhamel, Hauser and Rizolatti2005; Reference Dehaene2009; Dehaene & Cohen Reference Dehaene and Cohen2007) and my own “massive redeployment” hypothesis (M. L. Anderson Reference Anderson2007a; Reference Anderson2007c).Footnote 1 . These latter two suggest reuse might in fact constitute a fundamental developmental (Dehaene's recycling theory) or evolutionary (my redeployment hypothesis) strategy for realizing cognitive functions. Others are clearly thinking along similar lines, for example, Luiz Pessoa (Reference Pessoa2008), Gary Marcus (Reference Marcus2004; Reference Marcus2008), Steven Scher (Reference Scher2004), William Bechtel (Reference Bechtel, Scher and Rauscher2003), and Dan Lloyd (Reference Lloyd2000). These models have some interesting similarities and equally interesting differences, but taken together they offer a new research-guiding idealization of brain organization, and the potential to significantly impact the ongoing search for the brain basis of cognition.

I discuss each model, and what these models might collectively mean for cognitive science, in sections 6 and 7, after reviewing some of the broad-based evidence for neural reuse in the brain (sects. 4 and 5). In order to better appreciate that evidence and its implications, however, it will be useful to have before us a more concrete example of a theory of neural reuse, and some sense of where such theories fit in the landscape of cognitive science. To this end, the next subsection briefly details one of the theories of reuse – the massive redeployment hypothesis – and sections 2 through 5 serve to situate reuse with respect to some other well-known accounts of the functional structure of the brain.

1.1. The massive redeployment hypothesis

The core of the massive redeployment hypothesis is the simple observation that evolutionary considerations might often favor reusing existing components for new tasks over developing new circuits de novo. At least three predictions follow from this premise. Most generally, we should expect a typical brain region to support numerous cognitive functions in diverse task categories. Evidence to the contrary would tend to support the localist story that the brain evolved by developing dedicated circuits for each new functional capacity. More interestingly, there should be a correlation between the phylogenetic age of a brain area and the frequency with which it is redeployed in various cognitive functions; older areas, having been available for reuse for longer, are ceteris paribus more likely to have been integrated into later-developing functions. Finally, there should be a correlation between the phylogenetic age of a cognitive function and the degree of localization of its neural components. That is, more recent functions should generally use a greater number of and more widely scattered brain areas than evolutionarily older functions, because the later a function is developed, the more likely it is that there will already be useful neural circuits that can be incorporated into the developing functional complex; and there is little reason to suppose that the useful elements will happen to reside in neighboring brain regions. A more localist account of the evolution of the brain would instead expect the continual development of new, largely dedicated neural circuits, and would predict that the resulting functional complexes would remain tightly grouped, as this would minimize the metabolic cost of wiring the components together and communicating among them.

In a number of recent publications (M. L. Anderson Reference Anderson2007a; Reference Anderson2007c; Reference Anderson2008a) I report evidence for all of these predictions. Consider, for instance, some data demonstrating the first prediction, that a typical brain region serves tasks across multiple task categories. An empirical review of 1,469 subtraction-based fMRI experiments in eleven task domains reveals that a typical cortical regionFootnote 2 is activated by tasks in nine different domains. The domains investigated were various – action execution, action inhibition, action observation, vision, audition, attention, emotion, language, mathematics, memory, and reasoning – so this observation cannot be explained by the similarity of the task domains. And because the activations were post-subtraction activations, the finding is not explained by the fact that most experimental tasks have multiple cognitive aspects (e.g., viewing stimuli, recalling information, making responses). Control tasks would (mostly) ensure that the reported brain activity was supporting the particular cognitive function under investigation. Finally, the observation is not explained by the size of the regions studied. As recounted in more detail in section 5, below, one gets the same pattern of results even when dividing the cortex into nearly 1,000 small regions.Footnote 3

In evaluating the second prediction, one is immediately faced with the trouble that there is little consensus on which areas of the brain are older. I therefore employed the following oversimplification: All things being equal, areas in the back of the brain are older than areas in the front of the brain (M. L. Anderson Reference Anderson2007a). Thus, the prediction is for a relationship between the position of a brain region along the Y-axis in Talairach space (Talairach & Tournaux Reference Talairach and Tournaux1988) and the frequency with which it is used in cognitive functions. The study reports the expected negative correlationFootnote 4 between the Y-position and the number of tasks in which it is active (r=−0.412, p=.003, t=−3.198, df=50). A similar analysis using the data set mentioned above reveals a negative correlation between the number of domains in which an anatomical region is activated and the Y-position of the region (r=−0.312, p=0.011, t=−2.632, df=65). Although the amount of variance explained in these cases is not especially high, the findings are nevertheless striking, at least in part because a more traditional theory of functional topography would predict the opposite relation, if there were any relation at all. According to traditional theories, older areas – especially those visual areas at the back of the brain – are expected to be the most domain dedicated. But that is not what the results show.

As for the last prediction, that more recently evolved functions will be supported by more broadly scattered regions of activation, in (M. L. Anderson Reference Anderson2007a), I reported that language tasks activate more and more broadly scattered regions than do visual perception and attention. This finding was corroborated by a larger study (M. L. Anderson Reference Anderson2008a), which found that language was the most widely scattered domain of those tested, followed (in descending order) by reasoning, memory, emotion, mental imagery, visual perception, action, and attention. The significant differences in the degree of scatter were observed between attention and each of the following domains: language, reasoning, memory, emotion, and mental imagery; and between language and each of the following domains: visual perception, action, and attention. No other pair-wise comparisons showed significant differences.

Note that, in addition to supporting the main contentions of the massive redeployment hypothesis, this last finding also corroborates one of the main assumptions behind most theories of neural reuse: that cortical regions have specific biases that limit the uses to which they can be put without extensive rewiring. If neural circuits could be easily put to almost any use (that is, if small neural regions were locally poly-functional, as some advocates of connectionist models suggest), then given the increased metabolic costs of maintaining long-distance connections, we would expect the circuits implementing functions to remain relatively localized. That this is not the observed pattern suggests that some functionally relevant aspect of local circuits is relatively fixed. The massive redeployment hypothesis explains this with the suggestion that local circuits may have low-level computational “workings” that can be put to many different higher-level cognitive uses.Footnote 5

If this is the right sort of story, it follows that the functional differences between task domains cannot be accounted for primarily by differences in which brain regions get utilized – as they are reused across domains. And naturally, if one puts together the same parts in the same way, one will get the same functional outcomes. So, the functional differences between cognitive domains should reveal themselves in the (different) ways in which the (shared) parts are assembled. I explored this possibility using a co-activation analysis – seeing which brain regions were statistically likely to be co-active under what task conditions. The results indicated that although different domains do indeed tend to be supported by overlapping neural regions, each task domain was characterized by a distinctive pattern of co-activation among the regions (M. L. Anderson Reference Anderson2008a). This suggests an overall functional architecture for the brain that is quite different from that proposed by anatomical modularity and functional localization (see Fig. 1).

Figure 1. Expected patterns of co-activation in a simple six-region brain for two cognitive functions (solid vs. dashed lines). Anatomical modularity and localization (top) predicts largely non-overlapping sets of regions will contribute to each function, whereas reuse (bottom) suggests that many of the same cortical regions will be activated in support of both functions, but that they will co-activate (cooperate) in different patterns.

Keeping this substantive introduction to the concept of neural reuse in view, I will devote the next three sections to situating neural reuse with respect to three relevant classes of theory in cognitive science, and return to both neural reuse theory and supporting data in sections 5 and 6. For the purposes of this review, it is important to note that neural reuse theories are not full-fledged theories of how the brain (or mind) works. Rather, they are theories of how neural resources are (typically) deployed in support of cognitive functions and processes. Given this, there are at least three relevant comparison classes for neural reuse, each of which I discuss in turn in the sections that follow.

First, in section 2, I briefly discuss some other theories – anatomical modularity and global wiring optimization theory – for how neural resources are typically deployed in support of the brain's function. Then, in section 3, I turn to some theories of overall cognitive architecture – ACT-R, massive modularity, and both classic and contemporary parallel distributed processing models – and what they may imply for neural reuse and vice versa. And finally, in section 4, I examine at some length some other theories that predict neural reuse, notably concept empiricism and conceptual metaphor theory, as part of an argument that these established theories are not adequate to account for the full range of neural reuse that can be observed in the brain.

2. How are neural resources deployed in the brain?

There are two prominent theories for how neural resources are deployed in the function and structure of the brain: anatomical modularity and global wiring optimization theory. We will see that neural reuse is deeply incompatible with anatomical modularity, but compatible with wiring optimization theory. In fact, in combination neural reuse and wiring optimization theory make some novel predictions for cortical layout.

2.1. Anatomical modularity

Anatomical modularity is functional modularity plus a strong thesis about how the functional modules are implemented in the brain. Functional modularity is (minimally) the thesis that our cognitive systems are composed of separately modifiable (or “nearly decomposable”; Simon Reference Simon1962/1969) subsystems, each typically dedicated to specific, specialized functions (see sect. 3.1 for a discussion). Anatomical modularity is the additional thesis that each functional module is implemented in a dedicated, relatively small, and fairly circumscribed piece of neural hardware (Bergeron Reference Bergeron2007).

Simply put, neural reuse theories suggest anatomical modularity is false. According to the picture painted by reuse, even if there is functional modularity (see sect. 3.1), individual regions of the brain will turn out to be part of multiple functional modules. That is, brain regions will not be dedicated to single high-level tasks (“uses”), and different modules will not be implemented in separate, small, circumscribed regions. Instead, different cognitive functions are supported by putting many of the same neural circuits together in different arrangements (M. L. Anderson Reference Anderson2008a). In each of these arrangements, an individual brain region may perform a similar information-processing operation (a single “working”), but will not be dedicated to that one high-level use.

Although there are few defenders of a strong anatomical modularity hypothesis, Max Coltheart (Reference Coltheart and Rapp2001) goes so far as to include it as one of the fundamental assumptions guiding cognitive neuropsychology. The idea is that the success of neuropsychological research – relying as it does on patients with specific neurological deficits, and the discovery of double-dissociations between tasks – both requires and, in turn, supports the assumption that the brain is organized into anatomical modules. For if it were not, we wouldn't observe the focal deficits characteristic of some brain injuries, and nor would we be able to gather evidentiary support for double-dissociations between tasks.

If this argument were sound, then the success of neuropsychology as a discipline would itself be prima facie evidence against neural reuse. In fact, the inference is fairly weak. First, it is possible for focal lesions to cause specific functional deficits in non-modular systems (Plaut Reference Plaut1995), and double-dissociations do not by themselves support any inference about the underlying functional architecture of the brain (Van Orden et al. Reference Van Orden, Pennington and Stone2001). In any event, such deficits are the exception rather than the rule in human brain injuries. Even some of the patients most celebrated for having specific behavioral deficits often have multiple problems, even when one problem is the most obvious or debilitating (see Bergeron Reference Bergeron2007; Prinz Reference Prinz and Stainton2006 for discussions). The evidence coming from neuropsychology, then, is quite compatible with the truth of neural reuse. But is neural reuse compatible with the methodological assumptions of cognitive neuropsychology? Section 7 will discuss some of the specific methodological changes that will be needed in the cognitive neurosciences in light of widespread neural reuse.

2.2. Optimal wiring hypotheses

The layout of neurons in the brain is determined by multiple constraints, including biomorphic and metabolic limitations on how big the brain can be and how much energy it can consume. A series of studies by Christopher Cherniak and others has reported that the layout of the nervous system of C. elegans, the shape of typical mammalian neuron arbors, and the placement of large-scale components in mammalian cortex are all nearly optimal for minimizing the total length of neurons required to achieve the structure (Cherniak et al. Reference Cherniak, Mokhtarzada, Rodrigues-Esteban and Changizi2004; see also Wen & Chklovskii Reference Wen and Chklovskii2008). The last finding is of the highest relevance here. Cherniak et al. examined the 57 Brodmann areas of cat cortex. Given the known connections between these regions, it turns out that the Brodmann areas are spatially arranged so as to (nearly) minimize the total wiring length of those connections.

This is a striking finding; and even though this study examined physical and not functional connectivity, the two are undoubtedly related – at least insofar as the rule that “neurons that fire together wire together” holds for higher-level brain organization. In fact, Cherniak et al. (Reference Cherniak, Mokhtarzada, Rodrigues-Esteban and Changizi2004) predict that brain areas that are causally related – that co-activate, for instance – will tend to be physically adjacent. The data reviewed above did not exactly conform to this pattern. In particular, it seems that the neural regions supporting more recent cognitive functions tended to be less adjacent – farther apart in the brain – than those supporting older cognitive functions. Neverthless, neural reuse and the global optimization of component layout appear broadly compatible, for four reasons. First, wiring length can hardly be considered (and Cherniak et al. do not claim that it is) the only constraint on cortical structure. The total neural mass required to achieve the brain's function should also be kept minimal, and reuse would tend to serve that purpose. Second, it should be kept in mind that Cherniak et al. (Reference Cherniak, Mokhtarzada, Rodrigues-Esteban and Changizi2004) predict global optimization in component layout, and this is not just compatible with, but also positively predicts that subsets of components will be less optimal than the whole. Third, there is no reason to expect that all subsets will be equally suboptimal; global optimality is compatible with differences in the optimality of specific subsets of components. Fourth, when there is a difference in the optimality of component subsets, neural reuse would predict that these differences would track the evolutionary age of the supported function. That is, functionally connected components supporting recently evolved functions should tend to be less optimally laid out than those supporting older functions. More specifically, one would expect layout optimality to correlate with the ratio of the age of the cognitive function to the total evolutionary age of the organism. When functional cooperation emerged early in the evolution of the cortex, there is a greater chance that the components involved will have arrived at their optimal locations, and less chance of that for lower ratios, as overall brain morphology will not have had the same evolutionary opportunity to adjust.

This notion is not at all incompatible with the thesis of global (near-) optimality and indeed might be considered a refinement of its predictions. Certainly, this is a research direction worth pursuing, perhaps by merging the anatomical connectivity data-sets from Hagmann et al. (Reference Hagmann, Cammoun, Gigandet, Meuli, Honey, Wedeen and Sporns2008) with functional databases like BrainMap (Laird et al. Reference Laird, Lancaster and Fox2005) and the NICAM database (M. L. Anderson et al. Reference Anderson, Brumbaugh, Şuben, Chaovalitwongse, Pardalos and Xanthopoulos2010). In fact, I am currently pursuing a related project, to see whether co-activation strength between regions predicts the existence of anatomical connections.

3. Cognitive architectures

In this section, I review four of the most commonly adopted approaches to understanding how the mind is functionally structured, and the implications of these approaches for the functional structure of the brain: massive modularity; ACT-R; and classic and contemporary parallel distributed processing models. Neural reuse appears to undermine the main motivation for positing massive modularity, and although reuse is broadly compatible with the other three theories, it seems likely to somewhat modify the direction of each research program.

3.1. Massive modularity

As noted above, functional modularity is minimally the thesis that the mind can be functionally decomposed into specialized, separately modifiable subsystems – individual components charged with handling one or another aspect of our mental lives. Carruthers (Reference Carruthers2006) follows this formulation:

In the weakest sense, a module can just be something like: a dissociable functional component. This is pretty much the everyday sense in which one can speak of buying a hi-fi system on a modular basis, for example. The hi-fi is modular if one can purchase the speakers independently of the tape-deck, say, or substitute one set of speakers for another with the same tape deck. (Carruthers Reference Carruthers2006, p. 2)

Massive modularity, which grows largely out of the modularity movement in evolutionary psychology (Pinker Reference Pinker1997; Sperber Reference Sperber1996; Tooby & Cosmides Reference Tooby, Cosmides, Barkow, Cosmides and Tooby1992) is the additional thesis that the mind is mostly, if not entirely, composed of modules like this – largely dissociable components that vary independently from one another. Is such a vision for the mind's architecture compatible with widespread neural reuse? Carruthers (Reference Carruthers2006) certainly thinks so:

If minimizing energetic costs were the major design criterion, then one would expect that the fewer brain systems that there are, the better. But on the other hand the evolution of multiple functionality requires that those functions should be underlain by separately modifiable systems, as we have seen. As a result, what we should predict is that while there will be many modules, those modules should share parts wherever this can be achieved without losing too much processing efficiency (and subject to other constraints: see below). And, indeed, there is now a great deal of evidence supporting what Anderson [2007c] calls “the massive redeployment hypothesis”. This is the view that the components of brain systems are frequently deployed in the service of multiple functions. (Carruthers Reference Carruthers2006, pp. 23–24; emphasis his)

As much as I appreciate Carruthers' swift adoption of the redeployment hypothesis, I am troubled by some aspects of this argument. First, it appears to contain a false premise: Energetic constraints predict more compact or localized, not necessarily fewer brain systems. Second, it may be logically invalid, because if functions must be underlain by separately modifiable systems, then they cannot be built from shared parts. That is, it appears that this apparently small concession to neural reuse in fact undermines the case for massive modularity.

Consider Carruthers' hi-fi system analogy. There it is true that the various components might share the amplifier and the speakers, the way many different biological functions – eating, breathing, communicating – “share” the mouth. But if neural reuse is the norm, then circuit sharing in the brain goes far beyond such intercommunication and integration of parts. The evidence instead points to the equivalent of sharing knobs and transistors and processing chips. A stereo system designed like this would be more like a boom-box, and its functional components would therefore not be separately modifiable. Changing a chip to improve the radio might well also change the performance of the tape player.Footnote 6

To preview some of the evidence that will be reviewed in more detail in section 4, the brain may well be more boom-box than hi-fi. For instance, Glenberg et al. (Reference Glenberg, Sato and Cattaneo2008a) report that use-induced motor plasticity also affects language processing, and Glenberg et al. (Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008b) report that language processing modulates activity in the motor system. This connection is confirmed by the highly practical finding that one can improve reading comprehension by having children manipulate objects (Glenberg et al. Reference Glenberg, Brown and Levin2007). And of course there are many other such examples of cognitive interference between different systems that are routinely exploited by cognitive scientists in the lab.

This does not mean that all forms of functional modularity are necessarily false – if only because of the myriad different uses of that term (see Barrett & Kurzban Reference Barrett and Kurzban2006 for a discussion). But it does suggest that modularity advocates are guided by an idealization of functional structure that is significantly at odds with the actual nature of the system. Instead of the decompose-and-localize approach to cognitive science that is advocated and exemplified by most modular accounts of the brain, neural reuse encourages “network thinking” (Mitchell Reference Mitchell2006). Rather than approach a complex system by breaking functions into subfunctions and assigning functions to proper parts – a heuristic that has been quite successful across a broad range of sciences (Bechtel & Richardson Reference Bechtel and Richardson1993; Reference Bechtel and Richardson2010) – network thinking suggests one should look for higher-order features or patterns in the behavior of complex systems, and advert to these in explaining the functioning of the system. The paradigm exemplars for this sort of approach come from the discovery of common, functionally relevant topological structures in various kinds of networks, ranging from human and insect social networks to the phone grid and the Internet, and from foraging behaviors to the functioning of the immune system (Barabási & Albert Reference Barabási and Albert1999; Barabási et al. Reference Barabási, Albert and Jeong2000; Boyer et al. Reference Boyer, Miramontes, Ramos-Fernández, Mateos and Cocho2004; Brown et al. Reference Brown, Liebovitch and Glendon2007; Jeong et al. Reference Jeong, Tombor, Albert, Oltvai and Barabási2000; Newman et al. Reference Newman, Barabasi and Watts2006). Although it is hardly the case that functional decomposition is an ineffective strategy in cognitive science, the evidence outlined above that patterns of neural co-activation distinguish between cognitive outcomes more than the cortical regions involved do by themselves suggests the need for a supplement to business as usual.

Even so, there are (at least) two objections that any advocate of modularity will raise against the picture of brain organization that is being painted here: Such a brain could not have evolved, because (1) the structure would be too complex, and (2) it would be subject to too much processing interference and inefficiency.

Carruthers (Reference Carruthers2006) follows Simon (Reference Simon1962/1969) in making the first argument:

Simon [Reference Simon1962/1969] uses the famous analogy of the two watchmakers to illustrate the point. One watchmaker assembles one watch at a time, attempting to construct the whole finished product at once from a given set of micro components. This makes it easy for him to forget the proper ordering of parts, and if he is interrupted he may have to start again from the beginning. The second watchmaker first builds a set of sub-components out of given micro component parts and then combines those into larger sub-component assemblies, until eventually the watches are complete … . Simon's argument is really an argument from design, then, whether the designer is natural selection (in the case of biological systems) or human engineers (in the case of computer programs). It predicts that, in general, each element added incrementally to the design should be realized in a functionally distinct sub-system, whose properties can be varied independently of the others (to a significant degree, modulated by the extent to which component parts are shared between them). It should be possible for these elements to be added to the design without necessitating changes within the other systems, and their functionality might be lost altogether without destroying the functioning of the whole arrangement. (Carruthers Reference Carruthers2006, pp. 13, 25; emphasis in original)

The argument from design set forth here is more convincing when it is applied to the original emergence of a complex system than when it is applied to its subsequent evolutionary development. What the argument says is that it must be possible for development to be gradual, with functional milestones, rather than all-or-nothing; but neural reuse hardly weakens the prospect of a gradual emergence of new functions. And the possibility that new functionality can be achieved by combining existing parts in new ways – which undermines independent variation and separate modifiability, as Carruthers (Reference Carruthers2006) admits, here – suggests that a modular architecture is only one possible outcome from such gradualism.

Moreover, the strong analogy between natural selection and a designer may not be the most helpful conceptual tool in this case. When one thinks about the brain the way a human designer would, the problem that neural reuse presents is one of taking a given concrete circuit with a known function and imagining novel uses for it. That this process can be very difficult appears to place a heavy burden on reuse theories: How could such new uses ever be successfully designed? But suppose instead that, in building a given capacity, one is offered a plethora of components with unknown functions. Now the task is quite different: Find a few components that do something useful and can be arranged so as to support the current task – whatever their original purpose. Thus is a problem of design imagination turned into a problem of search. Evolution is known to be quite good at solving problems of the latter sort (Newell & Simon Reference Newell and Simon1976), and it is useful to keep this alternate analogy for the evolutionary process in mind here.

This brings us to the second objection, that non-modular systems would suffer from disabling degrees of interference and processing inefficiency. Here, it may be useful to recall some of the main findings of the situated/embodied cognition movement (M. L. Anderson Reference Anderson2003; Chemero Reference Chemero2009; Clark Reference Clark1997; Reference Clark, Bechtel and Graham1998). Central to the picture of cognition offered there is the simple point that organisms evolve in a particular environment to meet the particular survival challenges that their environment poses. Situated/embodied cognition emphasizes that the solutions to these problems often rely in part on features of the environments themselves; for example, by adopting heuristics and learning biases that reflect some of the environments' structural invariants (Gigerenzer et al. Reference Gigerenzer and Todd1999; Gilovitch et al. 2002). One such useful feature of most environments is that they don't pose all their problems all at once – inclement weather rarely comes along with predator abundance, pressing mating opportunities, and food shortages, for instance. And often when there are competing opportunities or challenges, there will be a clear priority. Thus, an organism with massively redeployed circuitry can generally rely on the temporal structure of events in its environment to minimize interference. Were this environment-organism relationship different – or if it were to change – then neural reuse does predict that increased interference will be one likely result.

Interestingly, contemporary humans encounter just such a changed organism-environment relationship in at least two arenas, and the effect of reused circuitry can often be seen as a result: First, in the labs of some cognitive scientists, who carefully engineer their experiments to exploit cognitive interference of various sorts; and, second, at the controls of sophisticated machinery, where the overwhelming attentional demands have been observed to cause massive processing bottlenecks, often with dangerous or even deadly results (Fries Reference Fries2006; Hopkin Reference Hopkin1995). It is no coincidence that, in addition to designing better human-machine interfaces, one important way of minimizing the problems caused by processing bottlenecks is to engineer the environment, including, especially, changing its task configuration and social structure, for instance by designing more efficient teams (Hutchins Reference Hutchins1995).

3.2. ACT-R

At the core of ACT-R is the notion of a cognitive architecture, “a specification of the structure of the brain at a level of abstraction that explains how it achieves the function of the mind” (J. R. Anderson Reference Anderson2007, p. 7). ACT-R is explicitly modular. As of ACT-R 6.0, it consisted of eight functionally specialized, domain-specific, relatively encapsulated, independently operating, and separately modifiable components. Given J. R. Anderson's definition of a cognitive architecture, it might seem to directly follow that ACT-R is committed to the notion that the brain, too, consists of functionally specialized, domain-specific, relatively encapsulated, independently operating, and separately modifiable regions that implement the functional modules of the ACT-R model. Certainly, recent experiments meant to associate ACT-R components with specific brain regions encourage this impression (J. R. Anderson Reference Anderson2007; J. R. Anderson et al. Reference Anderson, Qin, Junk and Carter2007). As he argues:

As discussed above, modular organization is the solution to a set of structural and functional constraints. The mind needs to achieve certain functions, and the brain must devote local regions to achieving these functions. This implies that if these modules reflect the correct division of the functions of the mind, it should be possible to find brain regions that reflect their activity. Our lab has developed a mapping of the eight modules … onto specific brain regions … (J. R. Anderson Reference Anderson2007, p. 74)

Given that neural reuse implies that anatomical modularity is false (see sect. 2.1), success in assigning ACT-R modules to specific brain regions would seem to be a problem for neural reuse, and evidence for neural reuse would appear to create problems for ACT-R. But the conclusion does not follow quite so easily as it seems. First, ACT-R does not strictly imply anatomical modularity. ACT-R is committed to the existence of functional modules, and to the existence of elements of the brain that implement them. If it turned out that activity in the ACT-R goal module was a better fit to the coordinated activity of some non-contiguous set of small brain regions than it was to the anterior cingulate (to which they currently have the goal module mapped), then this would count as progress for ACT-R, and not a theoretical setback. Similarly, if it turned out that some of the brain regions that help implement the goal module also help implement the imaginal module, this would pose no direct challenge to ACT-R theory.Footnote 7

Therefore, although J. R. Anderson is at pains to deny he is a functionalist – not just any possible mapping of function to structure will count as a success for ACT-R – there is a good deal of room here for alternatives to the simple 1:1 mapping that he and other ACT-R theorists are currently exploring. For its part, neural reuse predicts that the best fit for ACT-R modules, or any other high-level functional components, is much more likely to be some cooperating complex of multiple brain regions than it is a single area, and that brain regions involved in implementing one ACT-R function are likely to be involved in implementing others as well. Interestingly, this is more or less what J. R. Anderson et al. (Reference Anderson, Qin, Junk and Carter2007) found. For every task manipulation in their study, they found several brain regions that appeared to be implicated. And every one of their regions of interest was affected by more than one factor manipulated in their experiment. Thus, despite their methodological commitment to a 1:1 mapping between modules and brain regions, J. R. Anderson et al. (Reference Anderson, Qin, Junk and Carter2007) are quite aware of the limitations of that approach:

Some qualifications need to be made to make it clear that we are not proposing a one-to-one mapping between the eight regions at the eight functions. First, other regions also serve these functions. Many areas are involved in vision and the fusiform gyrus has just proven to be the most useful to monitor. Similarly, many regions have been shown to be involved in retrieval, particularly the hippocampus. The prefrontal region is just the easiest to identify and seems to afford the best signal-to-noise ratio. Equally, we are not claiming these regions only serve one function. This paper has found some evidence for multiple functions. For instance, the motor regions are involved in rehearsal as well as external action. (J. R. Anderson et al. Reference Anderson, Qin, Junk and Carter2007, pp. 213–14)

Here, the regulative idealization promoted by decomposition and localization may have unduly limited the sorts of methodological and inferential tools that they initially brought to bear on the project. As noted already in section 1, one of the contributions neural reuse may be able to make to cognitive science is an alternate idealization that can help guide both experimental design and the interpretation of results (M. L. Anderson et al. Reference Anderson, Brumbaugh, Şuben, Chaovalitwongse, Pardalos and Xanthopoulos2010).

Going forward there is at least one other area where we can expect theories of neural reuse and modular theories like ACT-R to have significant, bidirectional critical contact. Right now, ACT-R is not just theoretically, but also literally modular: it is implemented as a set of independent and separately modifiable software components. It does not appear, however, that separate modifiability is theoretically essential to ACT-R (although it is no doubt a programming convenience). Therefore, implementing overlaps in ACT-R components in light of the evidence from neuroimaging and other studies of the sort recounted here is likely to offer scientific opportunities to both research communities (see Stewart & West Reference Stewart and West2007 for one such effort). For example, overlaps in implementation might offer a natural explanation and a convenient model for certain observed instances of cognitive interference, such as that between language and motor control (Glenberg & Kaschak Reference Glenberg and Kaschak2002) or between memory and audition (Baddeley & Hitch Reference Baddeley, Hitch and Bower1974), helping to refine current hypotheses regarding the causes of the interference.

The ACT-R community is already investigating similar cases, where different concurrent tasks (dialing the phone while driving) require the use of the same ACT-R module, and thus induce performance losses (Salvucci Reference Salvucci2005). Altering ACT-R so that different modules share component parts might enable it to model some cognitive phenomena that would otherwise prove more difficult or perhaps impossible in the current system, such as the observation that object manipulation can improve reading comprehension (Glenberg et al. Reference Glenberg, Brown and Levin2007). Finally, observations of interference in a modified ACT-R but not in human data, might suggest that the ACT-R modules did not yet reflect the correct division of the mind's functions. Such conflicts between model and data could be leveraged to help ACT-R better approximate the high-level functional structure of the mind.

3.3. Classic parallel distributed processing

It is of course true that from a sufficiently abstract perspective, the idea of neural reuse in cognitive functioning is nothing new. It has been a staple of debates on brain architecture at least since the advent of parallel distributed processing (PDP) models of computation (Rummelhart & McClelland Reference Rumelhart and McClelland1986). For one widely cited example, consider the following from Mesulam (Reference Mesulam1990). He writes:

A central feature of networks is the absence of a one-to-one correspondence among anatomical site, neural computation and complex behavior … Figure [2] implies that each behavior is represented in multiple sites and that each site subserves multiple behaviors, leading to a distributed and interactive but also coarse and degenerate (one-to-many and many-to-one) mapping of anatomical substrate onto neural computation and computation onto behavior. This distributed and degenerate mapping may provide an advantage for computing complex and rapid cognitive operations and sets the network approach sharply apart from theories that postulate a nondegenerate one-to-one relationship between behavior and anatomical site. (Mesulam Reference Mesulam1990, pp. 601–602)

Broadly speaking, neural reuse theories are one of a family of network approaches to understanding the operation of the brain. They share with these an emphasis on cooperative interactions, and an insistence on a non-modular, many-to-many relationship between neural-anatomical sites and complex cognitive functions/behaviors. But there are also some important differences that set neural reuse apart.

First is a better appreciation of the computational work that can be done by very small groups of, or even individual, neurons (Koch & Segev Reference Koch and Segev2000). Neural reuse theories all agree that most of the interesting cognitive work is done at higher levels of organization, but they also emphasize that local circuits have specific and identifiable functional biases. In general, these models make a strong distinction between a “working” – whatever specific computational contribution local anatomical circuits make to overall function – and a “use,” the cognitive purpose to which the working is put in any individual case. For neural reuse theories, anatomical sites have a fixed working, but many different uses.

In contrast, note that in Figure 2 “neural computations” are located at Plane 2, parallel distributed processing. This reflects the belief that computational work can only be done by fairly large numbers of neurons, and that responsibility for this work can only be assigned to the network as a whole. Put differently, on PDP models there are no local workings. Classic PDP models are indeed a powerful way to understand the flexibility of the brain, given its reliance on relatively simple, relatively similar, individual elements. But the trouble for PDP models in this particular case is that there is no natural explanation for the data on increasing scatter of recently evolved functions, nor for the data on the cross-cultural invariance in the anatomical locations of acquired practices (see sect. 6.3). Indeed, on PDP models, investigating such matters is not even a natural empirical avenue to take. This represents a significant distinction between PDP and neural reuse.

Figure 2. Detail of Figure 3 from Mesulam (Reference Mesulam1990). Reprinted with permission of the author.

Other important differences between neural reuse and classic PDP models flow from the above considerations, including the way neural reuse integrates the story about the cognitive architecture of the brain into a natural story about the evolution and development of the brain. In a sense, neural reuse theories make some more specific claims than generalized PDP – not just that the brain is a kind of network, but that it is a kind of network with functional organization at more levels than previously thought. As can be seen already in the evidence outlined above, and will be seen in greater detail in sections 5 and 6, this specificity has led to some interesting and empirically testable implications for the brain's overall functional organization.

3.4. Contemporary parallel distributed processing models

More contemporary versions of network models, such as Leabra (O'Reilly Reference O'Reilly1998; O'Reilly & Munakata Reference O'Reilly and Munakata2000) tend to be composed of densely connected, locally specialized networks that are sparsely connected to one another (see Fig. 3).

Figure 3. Overview of the Leabra architectural organization. Reprinted from Jilk et al. (Reference Jilk, Lebiere, O'Reilly and Anderson2008) with permission of the authors.

In one sense, Leabra appears to be more compatible with neural reuse than classic PDP models are, as Leabra explicitly allows for regional functional biases. But insofar as this new architecture reflects the influence of the selectivity assumption, and represents a more modularist approach to understanding the brain, then there are potentially the same points of conflict with Leabra as there are with those theories. Consider the following, from a recent paper describing Leabra:

The brain is not a homogenous organ: different brain areas clearly have some degree of specialized function. There have been many attempts to specify what these functions are, based on a variety of theoretical approaches and data. In this paper, we summarize our approach to this problem, which is based on the logic of computational tradeoffs in neural network models of brain areas. The core idea behind this approach is that different brain areas are specialized to satisfy fundamental tradeoffs in the way that neural systems perform different kinds of learning and memory tasks. (Atallah et al. Reference Atallah, Frank and O'Reilly2004, p. 253)

There is nothing here that explicitly commits the authors to the idea that large brain regions are dedicated to specific tasks or cognitive domains – something the data presented here throw into question – although that is certainly one possible reading of the passage. Moreover, O'Reilly (Reference O'Reilly1998) tends to focus more on modeling processes over modeling parts, an approach that need not commit one to a specific story about how and where such processes are implemented in the brain – it needn't be the case that individual brain regions implement the processes being modeled, for instance.

And yet, O'Reilly and his collaborators have assigned these processes to specific regions:

The large-scale architectural organization of Leabra includes three major brain systems: the posterior cortex, specialized for perceptual and semantic processing using slow, integrative learning; the hippocampus, specialized for rapid encoding of novel information using fast arbitrary learning; and the frontal cortex/basal ganglia complex, specialized for active and flexible maintenance of goals and other context information, which serves to control or bias processing throughout the system. (Jilk et al. Reference Jilk, Lebiere, O'Reilly and Anderson2008, p. 204)

And, in fact, the Leabra team has gone further than this by recently integrating Leabra with ACT-R to form the SAL architecture:

When the ACT-R and Leabra research teams started to work together in 2006, they came to a startling realization: the two theories, despite their origins in virtually opposite paradigms (the symbolic and connectionist traditions, respectively) and widely different levels of abstraction, were remarkably similar in their view of the overall architecture of the brain. (Jilk et al. Reference Jilk, Lebiere, O'Reilly and Anderson2008, p. 205)

So it is not clear just what commitments Leabra has to modularity and localization. As with ACT-R, there doesn't seem to be anything essential to Leabra that would prevent it from explicitly incorporating neural reuse as one of its organizing principles. In particular, the functional specializations being ascribed to the brain regions mentioned are general enough to plausibly have many different cognitive uses, as predicted by neural reuse theories. But, as with ACT-R, more research will be needed before it becomes clear just how compatible these visions for the functional organization of the brain in fact are. The notion of neural reuse cuts across some old divisions – localization versus holism; modular versus connectionist – and whether theories falling on one or another side of each dichotomy are compatible with the notion of neural reuse will ultimately depend on how their advocates interpret the theories, and how flexible their implementations turn out to be.

4. Other theories predicting forms of neural reuse

One of the most successful theoretical paradigms in cognitive science has been the conceptual metaphor theories originating with Lakoff and Johnson (Reference Lakoff and Johnson1980; Reference Lakoff and Johnson1999) and extended by many others, perhaps most notably Fauconnier and Turner (Reference Fauconnier and Turner2002).Footnote 8 As is well known, conceptual metaphor theories suggest that cognition is dominated by metaphor-based thinking, whereby the structure and logical protocols of one or more domains, combined in various ways, guide or structure thinking in another. For a simple case, consider the Love Is War mapping taken from Lakoff and Johnson (Reference Lakoff and Johnson1980; Reference Lakoff and Johnson1999). When employing this metaphorical mapping, people use their understanding of war – of how to interpret events and how to respond to them – to guide their thinking about love: One fights for a partner, makes advances, fends off suitors, or embarks on a series of conquests. Similarly, the Life Is a Journey mapping allows people to leverage their extensive experience and competence in navigating the physical world in order to facilitate planning for life more generally: We plan a route, overcome obstacles, set goals, and reach milestones. The theory has been widely discussed and tested, and enjoys a raft of supporting evidence in linguistics and cognitive psychology.

A natural question that arises for such theories, however, is how the structured inheritance from one domain to another is actually achieved by the brain. Is it done abstractly, such that mental models (Gentner & Stevens Reference Gentner and Stevens1983; Johnson-Laird Reference Johnson-Laird1983) of war or navigation are used as prototypes for building other models of love or life? Or is there a more basic biological grounding, such that the very neural substrates used in supporting cognition in one domain are reused to support cognition in the other? Although some researchers favor the first possibility – notably Lera Boroditsky (e.g., Boroditsky & Ramscar Reference Boroditsky and Ramscar2002) – it seems fair to say that the greater effort has been focused on investigating the second.

This is at least in part because the debate over the biological basis of conceptual metaphors dovetails with another over the nature and content of cognitive representations – symbols, concepts, and (other) vehicles of thought – that has also played out over the last twenty years or so. At issue here is the degree to which the vehicles of thought – our mental carriers of meaning – are tied to sensory experience (Barsalou Reference Barsalou2008; Reference Barsalou1999). Concept empiricists (as they are called in philosophy) or supporters of modal theories of content (as they are called in psychology) are generally committed to some version of the thesis that “the vehicles of thought are re-activated perceptual representations” (Weiskopf Reference Weiskopf2007, p. 156). As one of the core statements of the modal position puts it, perceptual symbols, which “constitute the representations that underlie cognition,” are “record[s] of the neural activation that arises during perception” (Barsalou Reference Barsalou1999, pp. 578, 583; see Prinz Reference Prinz2002 for a general discussion). This position is meant to contrast with a rationalist or amodal one in which the vehicles of thought are inherently nonperceptual, abstract, logical, linguistic, or computational structures for which (as the classic semiotics line goes) the relation between signifier and signified is established arbitrarily (see, e.g., Fodor Reference Fodor1975; Fodor & Pylyshyn Reference Fodor and Pylyshyn1988).

In the case of both debates, it looked as if information about what neural resources were actually deployed to support cognitive tasks could provide evidence favoring one side or another. If planning a task used brain regions different from those used in planning (or imagining) a journey, then this would be prima facie evidence against the notion that the two were related via direct neural grounding. Similarly, if perceptual tasks and cognitive tasks appeared to be handled by distinct brain regions, this would appear to favor the amodal view.

In the event, a series of early findings bolstered the case for modal concepts, on the one hand, and for the idea that direct neural substrates supported metaphorical mappings, on the other. For example, a series of papers from the labs of Antonio Damasio and Alex Martin offered evidence that verb retrieval tasks activated brain areas involved in motor control functions, and naming colors and animals (that is, processing nouns) activated brain regions associated with visual processing (Damasio & Tranel Reference Damasio and Tranel1993; Damasio et al. Reference Damasio, Tranel, Hichwa and Damasio1996; Martin et al. Reference Martin, Haxby, Lalonde, Wiggs and Ungerleider1995; Reference Martin, Wiggs, Ungerleider and Haxby1996; Reference Martin, Ungerleider, Haxby and Gazzaniga2000). Similarly, it was discovered that perceiving manipulable artifacts, or even just seeing their names, activates brain regions associated with grasping (Chao & Martin Reference Chao and Martin2000). All this suggested that class concepts like HAMMER, RED, and DOG might be stored using a sensory and/or motor code, and, more generally, that high-level, conceptual-linguistic understanding might involve the reactivation of perceptuomotor experiences. This dovetailed nicely with the general idea behind direct neural support for metaphorical mappings, whereby understanding in one domain would involve the reactivation of neural structures used for another. Thus, the finding that mental planning can activate motor areas even when the task to be planned itself involves no motor activity (Dagher et al. Reference Dagher, Owen, Boecker and Brooks1999) has long been taken to support the case that mappings like Life Is a Journey are mediated by the direct sharing of neural resources by both domains.Footnote 9

It seems fair to say that these early discoveries prompted a much larger effort to uncover the neural underpinnings of high-level cognitive functions, one specifically focused on revealing the ways in which these underpinnings were shared with those of the sensorimotor system. The result is literally hundreds of studies detailing the various ways in which neural substrates are shared between various cognitive functions. A representative sample of these studies will be reviewed further on in sections 4.1 through 4.6, but to presage the argument to follow: The effort to uncover instances of neural reuse has been so successful that even a cursory examination of the breadth and frequency of reuse suggests that there is much more reuse than can be accounted for by modal concepts or conceptual metaphor theory. Any explanation of the phenomenon must therefore articulate a broader framework within which the prevalence of reuse naturally fits, and which in turn can explain such individual cognitive phenomena.Footnote 10 We will review some of the evidence for this claim in the next subsections.

4.1. Reuse of motor control circuits for language

A great deal of the effort to discover the specific neural underpinnings of higher cognitive functions has focused on the involvement of circuits long associated with motor control functions. In a typical example of this sort of investigation, Pulvermüller (Reference Pulvermüller2005) reports that listening to the words “lick,” “pick,” and “kick” activates successively more dorsal regions of primary motor cortex (M1). The finding is consistent both with the idea that comprehending these verbs relies on this motor activation, insofar as the concepts are stored in a motoric code, and also with the idea that understanding these verbs might involve (partial) simulations of the related actions. Either interpretation could easily be used as part of the case for concept empiricism.

Similarly, Glenberg and Kaschak (Reference Glenberg and Kaschak2002) uncover an interesting instance of the entanglement of language and action that they call the “action-sentence compatibility effect” (ACE). Participants are asked to judge whether a sentence makes sense or not and to respond by pressing a button, which requires a move either toward or away from their body. In one condition “yes” is away and “no” is toward; another condition reverses this. The sentences of interest describe various actions that would also require movement toward or away, as in “put a grape in your mouth,” “close the drawer,” or “you gave the paper to him.” The main finding is of an interaction between the two conditions, such that it takes longer to respond that the sentence makes sense when the action described runs counter to the required response motion. More striking, this was true even when the sentences described abstract transfers, such as “he sold his house to you,” which imply a direction without describing a directional motor action. Following the reasoning originally laid out by Sternberg (Reference Sternberg1969), an interaction between two manipulated factors implies at least one shared component between these two different processes – movement and comprehension. A likely candidate for this component would be a neural circuit involved in motor control, a supposition confirmed by Glenberg (2008b).Footnote 11 Thus, this seems another clear case in which motor control circuits are involved in, and perhaps even required for, language comprehension, whether via simulation (e.g., in the concrete transfer cases), metaphorical mapping (e.g., in the abstract transfer cases), or by some other mechanism. Glenberg has suggested both that the effect could be explained by the activation of relevant action schemas (Glenberg et al. Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008b) and by the activation and combination of appropriate affordances (Glenberg & Kaschak Reference Glenberg and Kaschak2002; Glenberg et al. Reference Glenberg, Becker, Klötzer, Kolanko, Müller and Rinck2009). Whatever the precise mechanism involved, the finding has been widely interpreted as support for both concept empiricism and for conceptual metaphor theory (although see M. L. Anderson Reference Anderson, Calvo and Gomila2008c for a dissent).

4.2. Reuse of motor control circuits for memory

Another interesting description of the motor system's involvement in a different cognitive domain comes from Casasanto and Dijkstra (Reference Casasanto and Dijkstra2010), who describe bidirectional influence between motor control and autobiographical memory. In their experiment, participants were asked to retell memories with either positive or negative valence, while moving marbles either upward or downward from one container to another. Casasanto and Dijkstra found that participants retrieved more memories and moved marbles more quickly when the direction of movement was congruent with the valence of the memory (upward for positive memories, downward for negative memories). Similarly, when participants were asked simply to relate some memories, without prompting for valence, they retrieved more positive memories when instructed to move marbles up, and more negative memories when instructed to move them down. Because the effect is mediated by a mapping of emotional valence on a spatial schema, the finding seems to most naturally support conceptual metaphor theory. The fact that the effect was bidirectional – recounting memories affected movement and movement affected memory retrieval – is a striking detail that seems to suggest direct neural support for the mapping.Footnote 12

4.3. Reuse of circuits mediated by spatial cognition

Many of the apparent overlaps between higher-order cognition and sensorimotor systems appear to be mediated by spatial schemas in this way. For example, Richardson et al. (Reference Richardson, Spivey, Barsalou and McRae2003) report that verbs are associated with meaning-specific spatial schemas. Verbs like “hope” and “respect” activate vertical schemas, whereas verbs like “push” and “argue” activate horizontal ones. As the authors put it, “language recruits spatial representations during real-time comprehension.” In a similar vein, Casasanto and Boroditsky (Reference Casasanto and Boroditsky2008) suggest that our mental representations of time are built upon the foundations of our experience with space. These findings appear to provide strong and relatively unproblematic support for conceptual metaphor theory, and perhaps also for a generic theory of concept empiricism, according to which the content of our concepts is grounded in (but does not necessarily constitute a simulation or reactivation of) sensorimotor experiences.

On the other hand, even when simulation is an important aspect of the reuse of resources between different domains, it does not always play the functional role assigned it by concept empiricism or conceptual metaphor theory. For some time, there has been growing evidence that doing actions, imagining actions, and watching actions done by others all activated similar networks of brain regions (Decety et al. Reference Decety, Sjoholm, Ryding, Stenberg and Ingvar1990; Decety et al. Reference Decety, Grezes, Costes, Perani, Jeannerod, Procyk, Grassi and Fazio1997; Jeannerod Reference Jeannerod1994). This has suggested to many that social cognition – understanding the actions and intentions of others – could involve simulating our own behaviors, a notion that attracted even more widespread interest after the discovery of mirror neurons (Decety & Grèzes Reference Decety and Grèzes1999; Gallese et al. Reference Gallese, Fadiga, Fogassi and Rizzolatti1996; Gallese & Goldman Reference Gallese and Goldman1998; Rizzolati et al. 1996). The trouble for concept empiricism and conceptual metaphor theory is that the logic governing the reuse of resources for multiple purposes is quite different in this case. Here, the idea is that circuits associated with behavioral control can be used to build predictive models of others, by inputting information about another agent into the system that would normally be used to guide one's own actions (and reactions). Although it could be argued that using simulation in support of such “mindreading” (Gallese & Goldman Reference Gallese and Goldman1998) requires a kind of metaphorical mapping (he is like me in relevant ways), in fact this is simply a necessary assumption to make the strategy sensible, and does not play the role of a domain-structuring inheritance.

Even some of the evidence for the reuse of spatial operations in other cognitive domains – which has been a mainstay of research into concept empiricism and conceptual metaphor theory – suggests the existence of more kinds of reuse than can be accounted for by these theoretical frameworks. Consider just a few of the various manifestations of the spatial-numerical association of response codes (SNARC) effect (Dehaene et al. Reference Dehaene, Bossini and Giraux1993): (1) When participants are asked to judge whether numbers are even or odd, responses are quicker for large numbers when made on the right side of space (canonically with the right hand, although the effect remains if responses are made while hands are crossed) and quicker for smaller numbers when responses are made on the left side of space. (2) Participants can accurately indicate the midpoint of a line segment when it is composed of neutral stimuli (e.g., XXXXX), but are biased to the left when the line is composed of small numbers (e.g., 22222 or twotwotwo) and to the right when the line is composed of large numbers (e.g., 99999 or nineninenine). (3) The presentation of a number at the fixation point prior to a target detection task will speed detection on the right for large numbers and to the left for small numbers. Hubbard et al. (Reference Hubbard, Piazza, Pinel and Dehaene2005) hypothesize that the SNARC effect can be accounted for by the observed reuse in numeric cognition of a particular circuit in left inferior parietal sulcus that plays a role in shifting spatial attention. Briefly, the idea is that among the representational formats we make use of in numerical cognition there is a mental “number line,” on which magnitudes are arrayed from left to right in order of increasing size. Once numerals are arrayed in this format, it is natural to reuse the circuit responsible for shifting spatial attention for the purpose of shifting attention between positions on this line. The resulting magnitude-influenced attentional bias can explain the SNARC effect.

This redeployment of visuo-spatial resources in support of alternate cognitive uses is somewhat difficult to explain from the standpoint of either concept empiricism or conceptual metaphor theory. In these examples, the effects would not be accounted for by the fact that numbers might be grounded in or involve simulations of basic sensorimotor experience, nor is it immediately obvious what metaphorical mapping might be implicated here. In fact, if the reuse of spatial schemas were in support of some semantically grounding structural inheritance from one domain to the other, we would expect the numbers to be arrayed vertically, with magnitude increasing with height. Instead, the reuse in this case appears driven by more abstract functional considerations. When doing certain numerical tasks, a number line is a useful representational format, and something like the visuo-spatial sketchpad (Baddeley Reference Baddeley1986) offers a convenient and functionally adequate storage medium. Similarly, reusing the spatial shifting mechanism is a sensible choice for meeting the functional requirements of the task, and need not ground any semantic or structural inheritance between the domains.

4.4. Reuse of circuits for numerical cognition

In fact, several such examples can be found in the domain of numerical cognition. Zago et al. (Reference Zago, Pesenti, Mellet, Crivello, Mazoyer and Tzourio-Mazoyer2001) found increased activation in the premotor strip in a region implicated in finger representation during multiplication performance compared to a digit reading condition. Similar findings were reported by Andres et al. (Reference Andres, Seron and Oliver2007), who found that hand motor circuits were activated during adults' number processing in a dot counting task. That these activations play a functional role in both domains was confirmed by Roux et al. (Reference Roux, Boetto, Sacko, Chollet and Tremoulet2003), who found that direct cortical stimulation of a site in the left angular gyrus produced both acalculia and finger agnosia (a disruption of finger awareness), and by Rusconi et al. (Reference Rusconi, Walsh and Butterworth2005), who found that repetitive Transcranial Magnetic Stimulation (rTMS) over the left angular gyrus disrupted both magnitude comparison and finger gnosis in adults.

Here again, this reuse of a basic sensorimotor function in an alternate cognitive domain does not seem to follow the logic of conceptual metaphor theory or concept empiricism. These theories are not making the claim that magnitudes inherit their meanings from finger representations, nor is any mathematical metaphor built in any straightforward way on our finger sense. Rather, the idea is that this neural circuit, originally developed to support finger awareness, is offering some functionally relevant resource in the domain of numerical cognition. For instance, Butterworth (Reference Butterworth1999c) suggests that the fingers provide children a useful physical resource for counting, with the neural result that the supporting circuits now overlap, while Penner-Wilger and Anderson (Reference Penner-Wilger, Anderson, Love, McRae and Sloutsky2008; submitted) suggest instead that the circuit in question might itself offer useful representational resources (such as a storage array).Footnote 13 This is not to question the notion that mathematical concepts and procedures are in some way grounded in sensorimotor experience (Lakoff & Núñez Reference Lakoff and Núñez2000), but this specific overlap in neural circuitry isn't straightforward to explain in the context of such grounding, nor is it anything that would have been predicted on the basis of either conceptual metaphor theory or concept empiricism. In fact, proponents of conceptual metaphor theory in mathematics tend to focus on relatively higher-level concepts like sets and investigate how our understanding of them is informed by such things as our experience with physical containers.

A similar argument can be made when considering the interrelations of speech and gesture, and the cognitive importance of the latter (see, e.g., Goldin-Meadow Reference Goldin-Meadow2003). According to Goldin-Meadow (Reference Goldin-Meadow2003), gesture is typically used not just to signal different moments in the learning process (e.g., to index moments of decision or reconsideration in a problem-solving routine), but also appears to have utility in advancing the learning process by providing another representational format that might facilitate the expression of ideas currently unsuited (for whatever reason) to verbal expression. The motor control system is here being used for a specific cognitive purpose not because it is performing semantic grounding or providing metaphorically guided domain structuring, but because it offers an appropriate physical (and spatiotemporal) resource for the task.

4.5. Reuse of perceptual circuits to support higher-order cognition

There are examples of the reuse of circuits typically associated with perception that also make the same point. Although there have certainly been studies that appear to unproblematically support concept empiricism – for example, Simmons et al. (Reference Simmons, Ramjee, Beauchamp, McRae, Martin and Barsalou2007) report the discovery of a common neural substrate for seeing colors, and for knowing about (having concepts for) color – other studies suggest that such cases represent only a small subset of a much broader phenomenon. Consider one of the earliest and most discussed cases of the reuse of neural circuits for a new purpose, the Baddeley and Hitch model of working memory (Baddeley & Hitch Reference Baddeley, Hitch and Bower1974; Reference Baddeley and Hitch1994; Baddeley Reference Baddeley1986; Reference Baddeley and Gazzaniga1995). One strategy for remembering the items on a grocery list or the individual numbers in a phone number involves (silently) saying them to one's self (producing a “phonological loop”), which engages brain areas typically used both in speech production and in audition.

A pattern of findings supports the existence of a phonological loop, and the engagement of both inner “speaking” and inner “hearing” to support working memory (see Wilson Reference Wilson2001 for a review). First, there is poor recall of similar sounding terms; second, there is poor recall of longer words; third, there is poor recall if the subject is made to speak during the maintenance period; and fourth, there is poor recall when the subject is exposed to irrelevant speech during the maintenance period. Moreover, imaging studies have found that such memory tasks cause activation in areas typically involved in speech production (Broca's area, left premotor cortex, left supplementary motor cortex, and right cerebellum) and in phonological storage (left posterior parietal cortex) (Awh et al. Reference Awh, Jonides, Smith, Schumacher, Koeppe and Katz1996).

In this interesting and complicated case, we have something of a triple borrowing of resources. First is the use of a culturally specific, acquired representational system – language – as a coding resource, and second is the application of a particular learned skill – silent inner speech – as a storage medium. These two together imply the third borrowing – of the neural resources used to support the first two functions. And note that all of this borrowing is done in support of what is likely an enhancement of a basic evolved function for storing small amounts of information over short periods. This raises the obvious question of whether and to what degree evolutionary pressures might have shaped the language system so that it was capable of just this sort of more general cognitive enhancement (Carruthers Reference Carruthers2002). In any case, it seems clear that this sort of borrowing is very hard to explain in terms of concept empiricism or conceptual metaphor theory. In the case of sensorimotor coding in working memory, the phonological loop is not metaphorically like speech; rather, it is a form of speech. In this, it is another instance of a straightforward functional redeployment – the reuse of a system for something other than its (apparent) primary purpose because it happens to have an appropriate functional structure.

4.6. Reuse is not always explained by conceptual metaphor theory or concept empiricism

These various examples suggest something along the following lines: One of the fundamental principles guiding reuse is the presence of a sufficient degree of functional relatedness between existing and newly developing purposes. When these functional matches result in the reuse of resources for both purposes, this history sometimes – but not always – reveals itself in the form of a metaphorical mapping between the two task domains, and sometimes, but not always, results in the inheritance or grounding of some semantic content. This way of thinking makes conceptual metaphors and “grounded” symbols into two possible side-effects of the larger process of reuse in cognition. It also muddies the causal story a bit: Planning is like locomotion because it inherits the structure of the existing domain via neural overlap; but planning also overlaps with the neural implementation base of locomotion to the degree that it is like locomotion.

The suggestion here is not that planning or communication or any other cognitive function has some predetermined Platonic structure that entirely reverses the causal direction typically supposed by conceptual metaphor theory. Rather, the idea is to point out the need to be open to a more iterative story, whereby a cognitive function finds its “neural niche” (Iriki & Sakura Reference Iriki and Sakura2008) in a process codetermined by the functional characteristics of existing resources, and the unfolding functional requirements of the emerging capacity (Deacon Reference Deacon1997).

Consider, in this regard, the particular phonemic character of human speech. A phoneme is defined by a certain posture of the vocal apparatus, and is produced by moving the apparatus toward that posture while making some noise (Fowler et al. Reference Fowler, Rubin, Remez, Turvey and Butterworth1980). Why should speech production be this way? In an article outlining their discoveries regarding the postural organization of the motor-control system, Graziano et al. (Reference Graziano, Taylor, Moore and Cooke2002b) write:

One possibility is that the mechanisms for speech were built on a preexisting mechanism for motor control, one that emphasized the specification of complex, behaviorally useful postures. When we stimulated the ventral part of the precentral gyrus, in the mouth and face representation, we often caused the lips and tongue to move toward specific postures (Graziano et al. Reference Graziano, Taylor and Moore2002a). For example, at one site, stimulation caused the mouth to open about 2cm and the tongue to move to a particular location in the mouth. Regardless of the starting posture of the tongue or jaw, stimulation evoked a movement toward this final configuration. This type of posture may be useful to a monkey for eating, but could also be an evolutionary precursor to the phoneme. (Graziano et al. Reference Graziano, Taylor, Moore and Cooke2002b, p. 355)

There are certainly functional characteristics that a unit of acoustic communication must have in order to adequately perform its communicative purpose, and not just any neural substrate would have had the required characteristics. But there remain degrees of freedom in how those characteristics are implemented. Speech production, then, developed its specific phonemic character as the result of the circuits on which it was built. Had the motor control system been oriented instead around (for example) simple, repeatable contractions of individual muscles – or had there been some other system with these functional characteristics available for reuse as acoustic communication was evolving – the result of the inheritance might have been a communication code built of more purely temporal elements, something closer to Morse code.Footnote 14

Finally, consider what may be a case not of the reuse of a basic sensorimotor area for higher cognitive functions, but rather the reverse. Broca's area has long been associated with language processing, responsible for phonological processing and language production, but what has recently begun to emerge is its functional complexity (Hagoort Reference Hagoort2005; Tettamanti & Weniger Reference Tettamanti and & Weniger2006). For instance, it has been shown that Broca's area is involved in many different action- and imagery-related tasks, including movement preparation (Thoenissen et al. Reference Thoenissen, Zilles and Toni2002), action sequencing (Nishitani et al. Reference Nishitani, Schürmann, Amunts and Hari2005), action recognition (Decety et al. Reference Decety, Grezes, Costes, Perani, Jeannerod, Procyk, Grassi and Fazio1997; Hamzei et al. Reference Hamzei, Rijntjes, Dettmers, Glauche and Weiller2003; Nishitani et al. Reference Nishitani, Schürmann, Amunts and Hari2005), imagery of human motion (Binkofski et al. Reference Binkofski, Amunts, Stephan, Posse, Schormann, Freund, Zilles and Seitz2000), and action imitation (Nishitani et al. Reference Nishitani, Schürmann, Amunts and Hari2005). Note that Müller and Basho (Reference Müller and Basho2004) suggest that these functional overlaps should not be understood as the later reuse of a linguistic area for other purposes, but are rather evidence that Broca's area already performed some sensorimotor functions that were prerequisites for language acquisition, and which made it a candidate for one of the neural building blocks of language when it emerged. That seems reasonable; but on the other hand, Broca's area is also activated in domains such as music perception (Tettamanti & Weniger Reference Tettamanti and & Weniger2006). While it is possible that this is because processing music requires some of the same basic sensorimotor capacities as processing language, it seems also possible that this reuse was driven by functional features that Broca's acquired as the result of its reuse in the language system, and thus by some more specific structural similarity between language and music (Fedorenko et al. Reference Fedorenko, Patel, Casasanto, Winawer and Gibson2009). Whatever the right history, this clearly represents another set of cases of functional reuse not explained by conceptual metaphor theory or concept empiricism.

Assuming the foregoing is sufficient to establish the existence of at least some cases of neural reuse that cannot be accounted for by these theoretical frameworks alone, the question naturally arises as to whether these anomalous cases should be dealt with by post-hoc elaborations of these theories (and/or by generating one or a few similarly specific theories), or whether this is a situation that calls for a global theory of reuse that supersedes and at least partially subsumes these existing frameworks. Far be it from me to argue a priori that one tack must be the correct one to take – science works best when we pursue multiple competing research paths – but one thing it might be useful to know when deciding how to spend one's research time is exactly how widespread neural reuse is. That is, the more widespread reuse appears, and the more instances of reuse that can be identified that do not involve the sensorimotor system, the stronger the justification would seem for trying to formulate a more global theory of neural reuse.

5. Further evidence that neural reuse is a pervasive feature of brain organization

Given the success of the theoretical frameworks just mentioned, as well as the growing interest in embodied cognition (M. L. Anderson Reference Anderson2003; Chemero Reference Chemero2009; Clark Reference Clark1997; Reference Clark, Bechtel and Graham1998), it is quite easy to find studies reporting that the neural implementations of higher cognitive functions overlap with those of the sensorimotor system. Indeed, this was the theme of a recent Attention and Performance Symposium, culminating in the 27-essay volume Sensorimotor Foundations of Higher Cognition (Haggard et al. Reference Haggard, Rossetti and Kawato2008). In contrast, there are only a few examples of reuse not involving the sensorimotor system that are reported as such in the literature. This fact would seem to favor the post-hoc elaboration approach to explaining the sorts of cases outlined above. On the other hand, the lack of such reports could simply be because people are not looking in the right place, or looking in the right way; after all, nobody is trying to establish a theory of attention-grounded, mathematics-grounded, or music-grounded cognition (as interesting as that sounds!). Absence of evidence of these cases, this is to say, is not evidence of absence. A typical literature search, then, will not help answer our question.

The literature can, however, be used in a somewhat different way. There are many, many thousands of studies in the neuroimaging literature that purport to uncover the neural underpinnings of various cognitive functions. If one were to compile a number of these studies in various task domains, one could ask, for each region of the brain, whether it supported functions in multiple domains, and whether such reuse was typically limited to regions of the brain implicated in supporting sensorimotor tasks.

The NICAM database (M. L. Anderson et al. Reference Anderson, Brumbaugh, Şuben, Chaovalitwongse, Pardalos and Xanthopoulos2010) currently contains information from 2,603 fMRI studies reported in 824 journal articles. All the studies involve healthy adults and use a within-subjects, subtraction-based, whole-brain design. That is, for all the studies in the database, brain activity during an experimental task was observed over the whole brain (not just a region of interest), and then compared to and subtracted from activity observed in the same participant during a control task. The logic of subtraction method is such that it should uncover only the regions of activation that support the specific mental function that best captures the difference between the experimental and control task. The neural activations supporting the mental operation that the two tasks have in common – the visual process allowing one to see the stimuli in a language task, for example – should be subtracted out. The database lists, among other things, the locations of the 21,553 post-subtraction fMRI activations observed during those 2,603 studies – that is, the regions of activation that are purported to specifically support those 2,603 mental operations. These features make the database ideal for investigating whether and to what degree specific brain regions support multiple functions across various task domains.

The general methodology for this sort of study is simple and straightforward. First, choose a spatial subdivision of the brain, then see which experiments, in which (and how many) domains, showed activity in each of the regions. To get the results reported in the next paragraph, below, I used the same 998 anatomical regions of interest (ROIs) used by Hagmann et al (Reference Hagmann, Cammoun, Gigandet, Meuli, Honey, Wedeen and Sporns2008).Footnote 15 The study was restricted to the following eleven task domains: three action domains – execution, inhibition, and observation – two perceptual domains – vision and audition – and six “cognitive” domains – attention, emotion, language, mathematics, memory, and reasoning.Footnote 16 Any study that was assigned to more than one domain was excluded. Activations were assigned to the ROI with the closest center; any activation that was more than 13mm from the center of one of the ROIs was excluded. This left 1,469 experiments collectively reporting 10,701 eligible activations.Footnote 17

There were 968 regions that were active in at least one experiment (and thus in one domain). Of these, 889 (91.8%) were active in at least two domains – that is, were reused at least once. On average, these 968 regions were active in 4.32 different domains (SD 1.99), and 555 of the regions were active in action tasks, with 535 of these “action” areas also active in an average of 3.97 (SD 1.58) non-action domains, and 530 active in an average of 3.16 (SD 1.23) cognitive domains. There were 565 regions active in perception tasks; 555 of these “perception” regions were also active in an average of 4.00 (SD 1.61) non-perception domains, and 550 were active in an average of 3.20 (SD 1.24) cognitive domains. There were 348 regions active in both action and perception tasks. On average, these were reused in 3.33 (SD 1.22) cognitive domains. There were also 196 regions not active in either perception or action tasks; 143 of these (72.96%) were active in two or more domains and averaged 2.97 (SD 0.95) domains. With all 196 regions included, the average is 2.43 (SD 1.19) of the six cognitive domains.Footnote 18

Naturally, if one uses larger regions – for instance, the 66 cortical ROIsFootnote 19 used by Hagmann et al (Reference Hagmann, Cammoun, Gigandet, Meuli, Honey, Wedeen and Sporns2008) – the average amount of reuse increases accordingly. All 66 regions were active in at least one domain; 65 (98.5%) were active in two or more domains.Footnote 20 As noted already above, the 66 regions were active in an average of 9.09 (SD 2.27) different domains. The 60 regions active in action tasks were also active in an average 7.38 (SD 0.98) non-action domains and 5.5 (SD 0.81) cognitive domains. The 64 regions active in perception tasks were also active in 7.39 (SD 1.87) non-perceptual domains and 5.34 cognitive domains. The 59 regions active in both perception and action tasks were also active in an average of 5.53 (SD 0.80) other domains, and the 7 regions not active in both perception and action tasks were active in an average of 3.00 (SD 1.41) of the cognitive domains. Only one region was active in only cognitive tasks, and that region was active only in memory.

These data appear to support the following claims: (1) Regions of the brain – even fairly small regions – are typically reused in multiple domains. (2) If a region is involved in perception tasks, action tasks, or both, it is more likely to be reused than if it is not involved in such tasks.Footnote 21 (3) Regions not involved in such tasks are nevertheless more likely than not to be reused in multiple domains. Note that the way of counting adopted above makes the best possible case for the “action and perception are special” position, by classifying as an “action” or “perception” region every region that is active in any such task. But it seems unlikely that there are 60 large cortical “action areas” and 64 “perception areas” in the way this term is usually understood. If instead some of these regions in fact contain instances of the reuse of “cognitive” circuitsFootnote 22 for action or perception tasks, then this way of counting likely overestimates the relatively higher reuse frequency of action and perception circuits. That is, neural reuse appears to be a pervasive feature of the functional organization of the brain, and although circuits that support action and perception may be favored targets for reuse, reuse is by no means restricted to sensorimotor circuits. Therefore, the situation appears to call for an assimilative, global theory, rather than the elaboration of existing theoretical frameworks.

6. Global theories of neural reuse

As mentioned at the outset, there are currently four candidates for a broad, general theory of neural reuse (or for the core around which such a theory could be built): Gallese's neural exploitation hypothesis, Hurley's shared circuit model, Dehaene's neuronal recycling hypothesis, and my massive redeployment hypothesis (already outlined in sect. 1.1 of this article). In this section, I will discuss each theory in turn and explore some of their similarities and differences.

6.1. Neural exploitation hypothesis

The neural exploitation hypothesis is a direct outgrowth of conceptual metaphor theory and embodied cognition, and largely sits at the intersection of these two frameworks. The main claim of the framework is that “a key aspect of human cognition is … the adaptation of sensory-motor brain mechanisms to serve new roles in reason and language, while retaining their original function as well.” (Gallese & Lakoff Reference Gallese and Lakoff2005, p. 456) This claim is the conclusion of an argument about the requirements of understanding that runs roughly as follows:

  1. 1 Understanding requires imagination. In the example most extensively developed by Gallese and Lakoff (Reference Gallese and Lakoff2005), understanding a sentence like “He grasped the cup” requires the capacity to imagine its constituent parameters, which include the agent, the object, the action, its manner, and so on.

  2. 2 Imagination is simulation. Here, the neural exploitation hypothesis dovetails with concept empiricism in arguing that calling to mind individuals, objects, actions, and the like involves reactivating the traces left by perceiving, doing, or otherwise experiencing instances of the thing in question.

  3. 3 Simulation is therefore neural reuse. Simulation involves reuse of the same functional clusters of cooperating neural circuits used in the original experience(s).

As much of the evidence for these claims has been laid out already in earlier sections, it won't be recounted here. The reader will of course notice that the theory as stated is limited to the adaptation of sensorimotor circuits, and we have already seen that reuse in the brain is much more broad-based than this. This is indeed a drawback of the theory, but it is nevertheless included here for two reasons: first, because it has been expanded to include not just the case of concept understanding, but also of human social understanding (Gallese Reference Gallese2008); and, second, because it incorporates a detailed computational model for how the reuse of circuitry might actually occur, based on work by Feldman and Narayanan (Reference Feldman and Narayanan2004). This model has broader applicability than is evidenced in the two main statements of the neural exploitation hypothesis (Gallese Reference Gallese2008; Gallese & Lakoff Reference Gallese and Lakoff2005).

The core of the computational model is a set of schemas, which are essentially collections of features in two layers: descriptions of objects and events and instructions regarding them. These two layers are systematically related to one another and to the sensorimotor system, such that event schemas can be used both to recognize events and to guide their execution, and object schemas can be used both to recognize objects and also to guide actions with respect to them.Footnote 23 . The schemas are also connected to the conceptual system, such that the contents of our concepts are built from the same features that form the schemas. The general idea is that the features' connections to the sensorimotor system give semantic substance to the concepts, as well as a natural model for understanding as the activation of neurally (or, in the current case, neural-network-ly) instantiated features and schemas. Like Gallese and Lakoff (Reference Gallese and Lakoff2005), Feldman and Narayanan (Reference Feldman and Narayanan2004) focus primarily on cases of understanding that can be directly (“He grabbed the cup”) or metaphorically (“He grabbed the opportunity”) mapped to basic perception-action domains.

But there is no reason in principle that the model need be limited in that way. As the authors note, by adding layers of abstraction, one can move from concrete action execution plans to abstract recipes like mathematical algorithms. Given this flexibility, it seems that action schemas need not be limited to providing guidance for the manipulation of independent objects (whether concrete or abstract) but could presumably also become control systems for the manipulation of neural circuits. That is, the same action schema that might normally be used to control rhythmic speech production, could be reused to guide silent memory rehearsal, and more abstract schemas might form the basis of control systems for predictive modeling or other applications.Footnote 24

Of course, this emendation would constitute a significant departure from the model as originally formulated.Footnote 25 . In particular, it would turn a system in which neural reuse was driven by grounding – the inheritance of semantic content from one level to another – into one in which reuse was driven by the need to create control systems for functionally relevant outcomes. Although it is far from clear that this switch precludes the possibility that grounding plays a role in driving neural reuse, it certainly moves it from center stage, which may have undesirable theoretical consequences for the theory as a whole, and for the way it interfaces with related ideas in linguistics, philosophy, and psychology. On the other hand, without some emendation that significantly broadens the kinds of cases that it can cover, the neural exploitation hypothesis risks being inadequate to the full range of available empirical evidence. We will return to these issues when we come to our general discussion of the four candidate theories.

6.2. The shared circuits model

The shared circuits model (Hurley Reference Hurley, Hurley and Chater2005; Reference Hurley2008) is organized around five control layers of similar structure, which are differentiated by the increasing abstraction of inputs and outputs. Each layer consists of an adaptive feedback loop that takes state information as input and generates control information as output. The first, lowest layer is a simple perception-action feedback loop that monitors progress toward action goals (reaching a target) and adjusts motor output in light of perceptually generated state information. It is, in this sense, a model of the simplest sort of thermostat; and the idea is that behavioral control systems might consist, at the most basic level, of multiple such control systems – or circuits. Layer 2 takes input from the external world, but also from layer 1, and becomes in essence an adaptive feedback loop monitoring the original layer. That is, layer 2 is in essence a forward model of layer 1. As is well known, incorporating such models into adaptive control systems tightens overall control by allowing for the prediction of state information, so appropriate action can be taken without waiting for the (typically slower) external feedback signal.Footnote 26 . The more hysteresis in the system – the longer it takes control interventions to produce expected results – the more improvement forward models can offer.

Circuit sharing really begins with layer 3, in which the same control circuits described by layers 1 and 2 take as input observations of the actions (or situations) of other agents. Hurley's suggestion is that the mirror system (Decety & Grèzes Reference Decety and Grèzes1999; Gallese et al. Reference Gallese, Fadiga, Fogassi and Rizzolatti1996; Rizzolati et al. 1996) should be modeled this way, as the activation of basic control circuits by state information relevant to the situations of other agents. Layer 3 also implements output inhibition, so agents don't automatically act as if they were in another agent's situation whenever they observe another agent doing something. Layer 4 incorporates monitoring of the output inhibition, supporting a self-other distinction; and layer 5 allows the whole system to be decoupled from actual inputs and outputs, to allow for counter-factual reasoning about possible goals and states and about what actions might follow from those assumptions. The idea is that the same circuits normally used to guide action in light of actual observations can also be fed hypothetical observations to see what actions might result; this can be the basis of predictive models. By the time we achieve the top layer, then, we have the outline for a model both of deliberation about possible actions, and also of multi-agent planning, which could serve as the basis for high-level social awareness and intelligence.

Like the neural exploitation hypothesis, one of the main explanatory targets of the shared circuits model is the possibility of mindreading and intelligent social interaction. And like the neural exploitation hypothesis, it is built entirely on the foundation of sensorimotor circuits. However, unlike the neural exploitation hypothesis, the shared circuits model does not revolve around the inheritance of semantic content from one level to another, but rather around the inheritance of function. The core capacities of the higher layers are based on exploiting the functional properties of the lower layers; all the layers are essentially control loops containing predictive models because they are reusing the basic implementation of the lowest levels. This is an advantage in that it is easier to see how the shared circuits model could be used to explain some of the specific instances of function-driven inheritance canvassed above; for, although Hurley models layer 1 on low-level sensorimotor circuits, there seems no reason in principle that the general approach couldn't allow for other kinds of basic circuits, on which other functional layers could be built.Footnote 27 It is also a potential weakness, in that it is less easy to see how it could be used to account for the central findings of concept empiricism or conceptual metaphor theory; can the sort of functional inheritance allowed by this model also allow for semantic inheritance? The inheritance of a basic feedback structure does not seem to lend itself to any obvious examples of this sort. This is not a criticism of the model as it stands – it was meant only to account for our understanding of instrumental actions; but it suggests that there is no very simple way to generalize the model to a wider set of cases. On the other hand, there seems no fundamental conflict between inheriting a function and thereby inheriting semantic content or domain structure.

I mentioned at the outset that the hierarchy of levels was characterized by an increasing abstraction of input and output. Consider layer 3 in this regard – at this level, input will be both impoverished and abstract as compared with lower layers. It will be impoverished because it will be missing a great deal of the richness of embodied experience – tactile experience, proprioceptive feedback, and efference copy are all absent when observing as opposed to acting. One is left with the visual experience of an action. And note that an action viewed from the first-person perspective looks different from the same action viewed from the third-person perspective. This points to one reason that the information must be abstract: since the visual experience of another agent's action will differ in most, if not all, of its low-level particulars, the system must be sensitive not to these, but to high-level features of the action that are common to the two situations.Footnote 28 Moreover, by hypothesis, layer 3 responds not just to actions, but to situations in which actions are possible – not just to another agent reaching for a banana, but to the banana being within the reach of another agent. This requires imputing possible goals to the observed agent, as well as encoding the high-level features of situations (relations between other agents, their capacities, and the objects in a scene). Here, the shared circuits model may need to be supplemented with something like the feature schemas from the neural exploitation model, itself expanded to allow for situation schemas, and not just object-action ones.

Similarly, if layer 4 is to appropriately and selectively inhibit the control outputs, it must take as input information about the relationships among the actions, agents, goals, and situations – who is in which situation doing what – which requires at least a rudimentary self/other distinction. And if layer 5 is going to be useful at all, the predictions it provides as output must be abstract, high-level action descriptions, not low-level motor commands.

These facts might seem to be nothing more than interesting and functionally useful features of the model, but in fact the requirement for abstraction at higher levels raises a puzzle: If low-level circuits respond to high-level features as inputs, and can produce high-level commands as outputs, might this not imply that layers 1 and 2 are more abstract than the model assumes? The trouble this raises is not with the coherence of the model, but with the evidence for it: All the evidence for layer 1 and 2 type controllers comes from on-line control systems dealing with real-time effector-specific, low-level feedback and control information, and not with abstract, feature-based information.

One obvious way to address this puzzle is to say that each layer is in fact a separate control structure that takes input from and delivers output to the layer below it, but this would undercut the entire premise of the model, since it would no longer be clear in what sense circuits were being “shared.” That high-level control systems are structurally like low-level ones is a perfectly reasonable hypothesis, but this is not the hypothesis put forward by this model, nor is it one for which there is a great deal of biological evidence.

A different approach would be to retain the central hypothesis that control circuits are shared among layers – that layer 3 reuses the control circuit defined by layers 1 and 2, and layer 5 reuses the control circuit defined by layers 1–4 – but suggest that the inputs between layers must be mediated by translators of various kinds. That is, layer 3 takes high-level feature information and translates this into the low-level information favored by layers 1 and 2 before passing it on. Indeed, one might hypothesize it does this by reusing other circuits, such as those that translate abstract plans into successive low-level motor actions. Similarly, layer 5 accepts the low-level motor commands natively output by layer 1, but translates them into high-level action descriptions. This picture is pretty plausible in the case of layer 3 observations of abstract action features, but it is much less clear how situations might get translated appropriately; and it is especially unclear how the reverse inference from low-level motor commands to high-level action descriptions might work. Just as a high-level action might be implemented any number of ways, a specific motor movement might be put in the service of innumerable high-level actions. The fact that part of the sensory information used to retroduct the action/intention from motor movement is the observed effect of the motor movement will help somewhat, but the basic problem still remains: There is a many-to-many relationship between movement and actions, so the valid deduction of a movement from an intention, and the valid retroduction of an intention from a movement need not follow the same paths in opposite directions.

These are hard problems to address; and because they originate from the fact that the shared circuits model requires that different kinds of inputs be fed to the same neural circuits, they may be problems that will surface for any theory of neural reuse (see discussion in section 6.4). Hence, it seems that the best approach to this puzzle may be to bite the bullet and say that, in at least some cases, circuit reuse is arranged such that different data – both information pertaining to different targets, as well as information about the same targets but at different levels of abstraction – can be fed without translation to the same circuits and still produce useful outputs.Footnote 29 Many sorting algorithms can just as easily sort letters as numbers; and if you feed a given algorithm pictures instead, it will do something with them. Naturally, this raises some pressing questions that seem ready-made for an enterprising theorist of neural computation: Under what conditions might useful things be done by circuits working with non-standard data? What kinds of implementations increase the chances of functionally beneficial outcomes given the fact of reuse? We will return to these issues in sections 6.4 and 7.

At its core, the shared circuits model offers an approach to understanding how high-level function could possibly be enabled by low-level circuits – and specifically by the reuse of low-level circuits for various purposes. Unfortunately, it is left fairly unclear exactly how they might actually be so enabled, given the different input-output requirements for each level; I have tried to sketch a solution that does the least damage to the intentions of the model, but I have to admit that some deep puzzles potentially remain.

Nevertheless, the model is interesting as an example of what might come from adopting a fairly classical “boxological” approach to cognitive modeling – understanding information processes via decomposition and interrelation – but without the underlying assumption of anatomical modularity.Footnote 30 If neural reuse is indeed a pervasive feature of the functional organization of the brain – as the current article is arguing – we will need to see more such work in the future.

6.3. The neuronal recycling hypothesis

The neuronal recycling hypothesis (Dehaene Reference Dehaene, Dehaene, Duhamel, Hauser and Rizolatti2005; Dehaene & Cohen Reference Dehaene and Cohen2007) originates from a set of considerations rather different from those motivating the two theories just discussed (i.e., the neural exploitation hypothesis and the shared circuits model). While those are neutral on the question of how and over what timescales the brain organization they propose came about, Dehaene is interested specifically in those cognitive capacities – such as reading and mathematics – that have emerged too recently for evolution to have generated cortical circuits specialized for these purposes. Such cultural practices must be learned, and the brain structures that support them must therefore be assigned and/or shaped during development.

There are two major ways to explain how recent cultural acquisitions, which emerge and are maintained in a population only by learning and not via genetic unfolding, can be supported by neural structures, as of course they must partly be. One way is to take our capacity to acquire such practices as reading and arithmetic as evidence for domain-general learning mechanisms (Barkow et al. Reference Barkow, Cosmides and Tooby1992) and fairly unconstrained neural plasticity (Quartz & Sejnowski Reference Quartz and Sejnowski1997). The other way is to suggest that cultural acquisitions must find a “neuronal niche”– a network of neural structures that already have (most of) the structure necessary to support the novel set of cognitive and physical procedures that characterize the practice. The neuronal recycling hypothesis is of the latter sort.

Note the interesting implication that the space of possible cultural acquisitions is partly constrained by cortical biases. The phrase “neuronal niche” is clearly meant to echo the idea of an ecological niche, and suggests both that acquired cognitive abilities “belong” in specific neural locations (i.e., can only survive where the neural climate is appropriate) and that the neural ecology may partly determine the characteristics that these cultural acquisitions possess, by limiting what is even possible to learn (and therefore which cognitive animals survive). Assuming the set of evolutionarily determined cortical biases is consistent across the species, we should expect to find evidence of at least three things: First, the neural manifestations of acquired abilities should be relatively consistent across individuals and even cultures; second, these practices should have some common cross-cultural characteristics; and third, the same sorts of cortical biases, as well as some of the same possibilities for learning, should be present in nonhuman primates.

As evidence for the first expectation, Dehaene and Cohen (Reference Dehaene and Cohen2007) note that the visual word form area, functionally defined as a region specifically involved in the recognition and processing of written words, appears in the same location in the brain across participants, whether the participants in question are using the same language and writing system or using different ones. Similarly, the intraparietal sulcus has been implicated in numeric tasks, regardless of the culture or number representation system used by the participants. As evidence for the second expectation, they point to work by Changizi and colleagues (Changizi & Shimojo Reference Changizi and Shimojo2005; Changizi et al. Reference Changizi, Zhang, Ye and Shimojo2006) that writing systems are characterized by two cross-cultural invariants: an average of three strokes per written letter; and a consistent frequency distribution for the types of contour intersections among the parts of those letters (T, Y, Z, etc.). Finally, the third expectation has been supported by some interesting and groundbreaking work by Atsushi Iriki and colleagues (Iriki Reference Iriki, Dehaene, Duhamel, Hauser and Rizzolati2005; Iriki & Sakura Reference Iriki and Sakura2008) who uncover evidence for real-time neural niche construction in primate brains (specifically Macaca fuscata) as the result of learning to use simple tools. The location of the observed neuro-morphological changes following tool training is roughly homologous to the regions associated with tool-use in the human brain (Culham & Valyear Reference Culham and Valyear2006). Thus, the theory suggests a novel pathway by which Homo sapiens may have achieved its current high-level cognitive capacities.

The neuronal recycling hypothesis outlines a universal developmental process that, although illustrated with specific examples, is meant to describe the way any acquired ability would come to have a neural instantiation. In this sense, it is broader in conception than the neural exploitation and shared circuits theories described in sections 6.1 and 6.2, respectively (although as noted their scope might well be increased with a few modifications). How much neural plasticity would be required in any given case will vary with the specifics of the acquisition, but one strength of the neuronal recycling theory is it makes clear some of the limits and costs that would be involved. The greater the distance between the function(s) required by a given practice, and the existing cortical biases, the harder the learning process will be, and the more likely that the learning process will disrupt whatever other functions the affected brain regions support.

On the other hand, the more the requirements of the acquisition match what is already possible, the less novel and potentially less valuable the cultural practice is likely to be – unless, that is, it is possible to combine existing capacities in new ways, to use old wheels, springs, and pulleys to form new machines. It is interesting to note in this regard that while the neural exploitation and shared circuits theories outlined earlier tend to envision neural circuits being put to fairly similar new uses – for example, forward models in motor control being used to support forward models in social interaction – the neuronal recycling hypothesis suggests that neural circuits might be put to uses quite other than the ones for which they were originally developed. As already noted above, this notion is central to the massive redeployment hypothesis, which we will briefly review next.

6.4. The massive redeployment hypothesis

Since the massive redeployment hypothesis has already been discussed in section 1.1, I will only review the main idea here. The primary distinction between massive redeployment and neuronal recycling is the time course over which each is supposed to operate. Massive redeployment is a theory about the evolutionary emergence of the functional organization of the brain, whereas neuronal recycling focuses on cognitive abilities for which there has been insufficient time for specialized neural circuits to have evolved. Both, however, suggest that the functional topography of the brain is such that individual circuits are put to various cognitive uses, across different task domains, in a process that is constrained in part by the intrinsic functional capacities (the “workings” or “cortical biases”) of local circuitry.

It is worth noting that the concepts of a “working” and of a “cortical bias” are not identical. The workings envisioned by the massive redeployment hypothesis commit that theory to the existence of cortical biases – that is, limitations on the set of functions it is possible for the circuit to perform in its present configuration. However, Dehaene is not committed to the notion of a local working in virtue of belief in cortical biases. Although it would be natural to understand cortical biases as the result of fixed local workings, a region could in fact perform more than one working and still have a cortical bias. However, the more flexible regions are, the less their individual biases will differ, and the harder it will be to explain the findings that recently evolved cognitive functions use more and more widely scattered neural components. On the other hand, as noted already above, the data are consistent with a number of functionally relevant constraints on local operation. For example, it could be that the dynamic response properties of local circuits are fixed, and that cognitive function is a matter of tying together circuits with the right (relative) dynamic response properties (for a discussion, see Anderson & Silberstein, submitted). In this sense, “cortical bias” is perhaps useful as a more generic term for denoting the functional limitations of neural regions.

In any event, both theories are committed to the notion that putting together the same neural bits in different ways can lead to different – in some cases very different – functional outcomes. In the discussion of the shared circuits model (sect. 6.2), I raised the issue of whether and how a single circuit could be expected to deal with various different kinds of data, as reuse theories seem to require. The question arises here as well: Exactly how is such reuse possible? It must be considered a weakness of both the massive redeployment and the neuronal recycling hypotheses that they lack any semblance of a functional model. In describing my theory (M. L. Anderson Reference Anderson2007a; Reference Anderson2007b; Reference Anderson2007c), I have used the metaphor of component reuse in software engineering, which may be useful as a conceptual heuristic for understanding the proposed functional architecture but cannot be taken as a model for the actual implementation of the envisioned reuse. In software systems, objects are reused by making virtual copies of them at run-time, so that there can be multiple, separately manipulable tokens of each object type. With wetware systems, no such process is possible. What is reused is the actual circuit.

In general, how such reuse is actually effected must be considered an open question for the field. Going forward, supporters of recycling and redeployment need to provide at least three things: specific models of how information could flow between redeployed circuits; particular examples of how different configurations of the same parts can result in different computations; and a more complete discussion of how (and when and whether) multiple uses of the same circuit can be coordinated. Penner-Wilger and Anderson (Reference Penner-Wilger, Anderson, Love, McRae and Sloutsky2008; submitted) have taken some tentative steps in this direction, but much more such work is needed. It is to the credit of both Hurley and Gallese that they each offer a (more or less concrete) proposal in this regard (see Gallese 1996; Reference Gallese2008; Hurley Reference Hurley, Hurley and Chater2005; Reference Hurley2008). That neither seems wholly adequate to the task should not be surprising nor overemphasized; the neurosciences are replete with what must be considered, at best, partial models of the implementation of function by neural structure. More important by far is that neural reuse offers a unique guide to discovery – a sense of what to look for in understanding brain function, and how to put the pieces together into a coherent whole. If neural circuits are put to many different uses, then the focus on explaining cognitive outcomes should shift from determining local circuit activity and single-voxel effects to uncovering the complex and context-dependent web of relations between the circuits that store, manipulate, or otherwise process and produce information and the functional complexes that consume that information, putting it to diverse purposes.Footnote 31

One way this effort might be abetted is via the formulation of an even more universal theory of neural reuse than is offered by any of the four theories canvassed above. As should be clear from the discussion, none of the four proposals can explain all the kinds of reuse in evidence: reuse supporting functional inheritance, reuse supporting semantic inheritance, reuse that occurs during development, and reuse that occurs during evolution. In fact, each is strongest in one of these areas, and weaker in the others. This opens the obvious possibility that the four theories could be simply combined into one.Footnote 32 While it is true that there seems no obvious barrier to doing so, in that none of the theories clearly contradicts any of the others, this lack of conflict is in part an artifact of the very under-specification of the theories that leaves them covering distinct parts of the phenomenon. As mentioned already, it may turn out that the kind of functional inheritance required by the shared circuits model precludes the kinds of semantic inheritance required by the neural exploitation hypothesis, or that the schemas envisioned by neural exploitation cannot be modified and expanded along the necessary lines. Likewise, it could turn out that the processes driving massive redeployment are in tension with those driving neuronal recycling; or that one, or the other, but not both can explain semantic and/or functional inheritance.

Should such problems and conflicts arise, no doubt solutions can be found. The point here is simply: We don't yet even know if there will be problems, because no one has yet even tried to find a solution. I would encourage all those interested in the general topic of brain organization to ponder these issues – how does the fact of reuse change our perspective on the organization, evolution, development, and function of the brain? Within what framework should findings in neuroscience ultimately be placed? There is enough work here for many hands over many years.

7. Implications

Although the question of how neural reuse is actually effected must be considered open, the question of whether there is significant, widespread, and functionally relevant reuse must be considered closed. In light of all the evidence discussed above, it is clear that there is neural reuse, and there is a lot of it. Neural reuse is a real feature of brain organization, but it is also a novel concept – something about the brain that we are just now beginning to notice. What might it mean? What is the way forward? I close the article with a few thoughts on these topics.

First, and most obviously, the fact of widespread neural reuse seems to favor modal and “embodied” accounts of cognition – and of representational content, in particular – over amodal or more abstract accounts. On the other hand, the neuroscientific evidence for these theories has generally been over-read (M. L. Anderson Reference Anderson, Calvo and Gomila2008c). Especially in light of the many different kinds of reuse, and the many potential mechanisms by which it may have come about, the claims made on behalf of concept empiricism and embodied cognition need close examination. Although a lack of neural reuse would have been evidence against embodied cognition, concept empiricism, and conceptual metaphor theory, the fact that it is even more widespread than these theories predicted means that neural overlaps are not by themselves evidence for these theories, and do not fully explain the relationships between cognitive domains that are at the heart of these ideas. In particular, it needs to be asked what kinds of reuse will, and will not, support the kinds of inheritance of structure and content these theories require; and whether the evidence actually points specifically to that sort of reuse. In fact, this is one of the main open areas of research for neural reuse: How is functional inheritance possible, and what kinds of implementations of reuse can lead to semantic inheritance of the sort described in concept empiricism, conceptual metaphor theory, and other theories of cognitive grounding? Providing this sort of story would offer the possibility of unifying these different theories of grounding with one another, under the umbrella of general neural reuse. In the absence of such a story, general neural reuse instead threatens to undermine some of the justification for these accounts.

If regions of the cortex are indeed put to many different cognitive uses, this suggests that cortical parcellation and function-to-structure mapping should be approached via multiple- or cross-domain investigations (Penner-Wilger & Anderson Reference Penner-Wilger, Anderson, Love, McRae and Sloutsky2008; submitted). One way to move forward on this task is via the increased use of effect location meta-analysis, in which multiple imaging studies, each reporting significant effects, are analyzed together to get more accurate information about the brain locations of mental operations (Fox et al. Reference Fox, Parsons and Lancaster1998). Although such studies are increasingly common, they are also typically limited to one task domain. There is nothing intrinsic to effect-location meta-analysis or cognitive modeling in general that militates against cross-domain modeling, but in practice it is very rarely done. This is, I presume, because there remains a very strong, and perhaps typically unconscious, assumption that brain regions are both unifunctional and domain dedicated.Footnote 33 Widespread neural reuse suggests that this assumption must be given up.

Neural reuse offers an alternative to these assumptions, as well as to the more general selectivity and localization assumptions that have long been the guiding idealization for research in the cognitive neurosciences. In their place, neural reuse offers the strong distinction between working (or local cortical bias) and cognitive use, which can help guide the (re-)interpretation of experimental results, especially those based on single brain-imaging experiments. It also offers the suggestion that attention paid to the interactions of multiple regions over the activity of single ones will be well rewarded. Methodological tools that take us beyond single-voxel effects – such as functional connectivity analysis and multi-voxel pattern analysis – may have an important role to play in supporting these efforts (Anderson & Oates Reference Anderson, Oates, Ohlsson and Catrambone2010; M. L. Anderson et al. Reference Anderson, Brumbaugh, Şuben, Chaovalitwongse, Pardalos and Xanthopoulos2010; Honey et al. Reference Honey, Kötter, Breakspear and Sporns2007; Pereira et al. Reference Pereira, Mitchell and Botvinick2009; Sporns et al. Reference Sporns, Tononi and Edelman2000; Reference Sporns, Chialvo, Kaiser and Hilgetag2004).

Once we give up these assumptions, our vocabulary of cognitive function might need specific revision to include fewer domain-specific concepts. In current practice, cortical regions are assigned visual functions by vision researchers, memory functions by memory researchers, attention functions by attention researchers, and so on (Cabeza & Nyberg Reference Cabeza and Nyberg2000). But if cortical circuits contribute to multiple task domains, then this practice will not lead to the accurate attribution of workings to these circuits. In light of neural reuse, it appears that this practice can at best reveal one of the uses to which a region is put, but is unlikely to hit upon the actual local working (see M. L. Anderson Reference Anderson2007b; Bergeron Reference Bergeron2008 for discussions). This best-case scenario requires that the process models are themselves accurate, but it seems implausible to suppose that these models – also typically generated on the basis of domain-focused experimentation – will themselves survive widespread acceptance of neural reuse without significant revision. In this sense neural reuse is a potentially disruptive finding, although hopefully in the service of increased theoretical fertility.

Widespread neural reuse makes it quite clear that there is not and cannot be anatomical modularity in the brain. Whether this means there is no functional modularity is an open question. Can cognitive functions be independent when they have overlapping neural implementations? Questions about what functional modularity requires are vexed, and different researchers have come to many different conclusions on the matter (Barrett & Kurzban Reference Barrett and Kurzban2006; Carruthers Reference Carruthers2006). Whether and precisely how neural reuse constrains this debate is a matter that deserves careful attention.

There are some practical upshots as well. Maps of the overlaps among the circuits supporting cognitive function will support robust predictions regarding cognitive processes and tasks that are likely to interfere with one another. Not only does this offer leverage to the experimentalist in designing inquiries into brain function, it also offers advice to the system designer in designing work flows and machine interfaces. As consumer devices, medical instruments, and heavy machinery become more sophisticated and powerful, increasing attention will need to be paid to the cognitive demands of operating them, and information about neural overlaps will be one important tool in the designers' toolbox (Rasmussen & Vicente Reference Rasmussen and Vicente1989; Ritter & Young Reference Ritter and Young2001), especially as leading cognitive models start incorporating information about reuse into their systems (Stewart & West Reference Stewart and West2007).

Similarly, knowledge of neural overlaps might suggest novel therapies for brain injury. Many therapies for traumatic brain injury are based on the “use it or lose it” principle – the more tasks that stimulate a brain region, the more likely patients are to recover function. Knowledge about the range of different tasks that potentially stimulate each region may serve as the basis for unexpected therapeutic interventions, ways of indirectly recovering function in one domain by exercising capacities in another. Indeed, there is evidence from healthy subjects that such indirect approaches to strengthening neural function can in fact work – for example, the finding that object manipulation can increase reading comprehension in school-age children (Glenberg et al. Reference Glenberg, Brown and Levin2007).

Finally, given that brain functions are apparently supported by multiuse components, there are possible implications for how cognition might be engineered and reproduced in robotic artificial intelligence (AI) (M. L. Anderson Reference Anderson2008a). That is, neural reuse might recommend a shift from building intelligent systems out of separate, specialized modules dedicated to language, motor-control, vision, and such, to engineering low-level multi-use components that offer services to many different high-level functions. There has been some theoretical and practical work in this direction (Hall Reference Hall, Goertzel, Hitzler and Hutter2009; Stewart & West Reference Stewart and West2007), but much more is needed. Such work is probably the necessary precursor to any satisfactory theory of how it is that component reuse can engender both functional and semantic inheritance. I hope the present article gives some sense that such efforts will be rewarded.

Reuse or re-function? Aisenberg, Daniela and Henik, Avishai Department of Psychology, Ben-Gurion University of the Negev, Beer-Sheva, 84105, Israel. http://www.bgu.ac.il/~henik From the physical to the psychological: Mundane experiences influence social judgment and interpersonal behavior Bargh, John A., Williams, Lawrence E., Huang, Julie Y., Song, Hyunjin and Ackerman, Joshua M. Department of Psychology, Yale University, New Haven, CT 06520. www.yale.edu/acmelab; Leeds School of Business, University of Colorado at Boulder, Boulder, CO 80309-0419. leeds-faculty.colorado.edu/lw/; Sloan School of Management, Massachusetts Institute of Technology, Cambridge, MA 02142. web.mit.edu/joshack/www/ Neural reuse and cognitive homology Bergeron, Vincent Department of Philosophy, University of Ottawa, Ottawa, ON K1N 6N5, Canada. Neural reuse implies distributed coding Bridgeman, Bruce Department of Psychology, University of California at Santa Cruz, Santa Cruz, CA 95064. http://people.ucsc.edu/~bruceb/ Sensorimotor grounding and reused cognitive domains Brincker, Maria Department of Philosophy, Graduate Center, City University of New York, New York, NY 10016. sites.google.com/site/mariabrincker/ The importance of ontogenetic change in typical and atypical development Dekker, Tessa M. and Karmiloff-Smith, Annette Centre for Brain and Cognitive Development, Birkbeck College, University of London, London WC1 7HX, United Kingdom. http://www.psyc.bbk.ac.uk/research/DNL/personalpages/tessa.html http://www.bbk.ac.uk/psyc/staff/academic/annettekarmilofsmith How and over what timescales does neural reuse actually occur? Donnarumma, Francesco, Prevete, Roberto and Trautteur, Giuseppe Dipartimento di Scienze Fisiche, Università di Napoli Federico II, Complesso Universitario Monte Sant'Angelo, I-80126 Napoli, Italy. http://vinelab.na.infn.it Sleep, neural reuse, and memory consolidation processes Fishbein, William, Lau, Hiuyan, DeJesús, Rafael and Alger, Sara Elizabeth Laboratory of Cognitive Neuroscience and Sleep, The City College and Graduate Center, The City University of New York, New York, NY 10031. Reuse (neural, bodily, and environmental) as a fundamental organizational principle of human cognition Foglia, Lucia and Grush, Rick Dipartimento di Studi Storico, Sociali e Filosofici, Università degli Studi di Siena, 52100 Arezzo, Italy. ; Philosophy Department, University of California – San Diego, La Jolla, CA 92093-0119. http://mind.ucsd.edu Understanding brain circuits and their dynamics Gomila, Antoni and Calvo, Paco Department of Psychology, University of the Balearic Islands, 070XX Palma, Spain. ; Department of Philosophy, University of Murcia, 30003 Murcia, Spain. Neural reuse in the social and emotional brain Immordino-Yang, Mary Helen, Chiao, Joan Y. and Fiske, Alan P. Brain and Creativity Institute and Rossier School of Education, University of Southern California, Los Angeles, CA 90089. http://rossier.usc.edu/faculty/mary_helen_immordinoyang.html; Psychology Department, Northwestern University, Evanston, IL 60208. http://culturalneuro.psych.northwestern.edu; Anthropology Department, University of California Los Angeles, Los Angeles, CA 90095. http://www.sscnet.ucla.edu/anthro/faculty/fiske/ Neural reuse: A polysemous and redundant biological system subserving niche-construction Iriki, Atsushi Laboratory for Symbolic Cognitive Development, RIKEN Brain Science Institute, Wako 351-0198, Japan. http://www.brain.riken.jp/en/a_iriki.html Multi-use and constraints from original use Jungé, Justin A. and Dennett, Daniel C. Center for Cognitive Studies, Tufts University, Medford, MA 02155. http://www.tufts.edu Comparative studies provide evidence for neural reuse Katz, Paul S. Neuroscience Institute, Georgia State University, Atlanta, GA 30302-5030. http://neuroscience.gsu.edu/pkatz.html No bootstrapping without semantic inheritance Kiverstein, Julian School of Philosophy, Psychology, and Language Sciences, University of Edinburgh, Edinburgh EH8 7PU, Scotland, United Kingdom. http://www.artandphilosophy.com/philosophy.html Redeployed functions versus spreading activation: A potential confound Klein, Colin Department of Philosophy, University of Illinois at Chicago, Chicago, IL 60607. http://tigger.uic.edu/~cvklein/links.html Implications of neural reuse for brain injury therapy: Historical note on the work of Kurt Goldstein Lia, Barry Dizziness and Balance Center, Otolaryngology/Head and Neck Surgery, University of Washington Medical Center, Seattle, WA 98195-6161. Reuse in the brain and elsewhere Lindblom, Björn Department of Linguistics, Stockholm University, 10691 Stockholm, Sweden. http://www.ling.su.se Let us redeploy attention to sensorimotor experience Michaux, Nicolas, Pesenti, Mauro, Badets, Arnaud, Di Luca, Samuel and Andres, Michael Institut de Recherche en Sciences Psychologiques, Université catholique de Louvain, 1348 Louvain-la-Neuve, Belgium http://www.uclouvain.be/315041.html; Centre de Recherches sur la Cognition et l'Apprentissage, CNRS UMR 6234, France. http://cerca.labo.univ-poitiers.fr Neural reuse as a source of developmental homology Moore, David S. and Moore, Chris Department of Psychology, Pitzer College and Claremont Graduate University, Claremont, CA 91711. http://pzacad.pitzer.edu/~dmoore/; Department of Psychology, Dalhousie University, Halifax, NS B3H4J1, Canada. http://myweb.dal.ca/moorec/index.html Reuse of identified neurons in multiple neural circuits Niven, Jeremy E. and Chittka, Lars Department of Zoology, University of Cambridge, Cambridge, CB2 3EJ, United Kingdom. http://www.neuroscience.cam.ac.uk/directory/profile.php?jen22; Research Centre for Psychology, School of Biological and Chemical Sciences, Queen Mary University of London, London E1 4NS, United Kingdom. http://chittkalab.sbcs.qmul.ac.uk/ The Leabra architecture: Specialization without modularity Petrov, Alexander A., Jilk, David J. and O'Reilly, Randall C. Department of Psychology, Ohio State University, Columbus, OH 43210. http://alexpetrov.com; eCortex, Inc., Boulder, CO 80301. http://www.e-cortex.com; Department of Psychology and Neuroscience, University of Colorado, Boulder, CO 80309. http://psych.colorado.edu/~oreilly Neural reuse and human individual differences Rabaglia, Cristina D. and Marcus, Gary F. Department of Psychology, New York University, New York, NY 10003. Reuse of molecules and of neural circuits Reimers, Mark Department of Biostatistics, Virginia Commonwealth University, Richmond, VA 23284. http://www.people.vcu.edu/~mreimers Massive modularity is consistent with most forms of neural reuse Ritchie, J. Brendan and Carruthers, Peter Department of Philosophy, University of Maryland, College Park, MD 20742. https://sites.google.com/site/jbrendanritchie/Home http://www.philosophy.umd.edu/Faculty/pcarruthers/ More than modularity and metaphor: The power of preadaptation and access Rozin, Paul Department of Psychology, University of Pennsylvania, Philadelphia, PA 19104-6241. Optical holography as an analogue for a neural reuse mechanismFootnote 1 Speed, Ann, Verzi, Stephen J., Wagner, John S. and Warrender, Christina Sandia National Laboratories2 Albuquerque, NM 87185-1188. www.sandia.gov Massive redeployment or distributed modularity? Toskos Dils, Alexia and Flusberg, Stephen J. Department of Psychology, Stanford University, Stanford, CA 94305. Belling the cat: Why reuse theory is not enough Vilarroya, Oscar Unitat de Recerca en Neurociència Cognitiva, Departament de Psiquiatria i Medicina Legal, Universitat Autònoma de Barcelona, and Fundació IMIM, Barcelona 08193, Spain. Cortex in context: Response to commentaries on neural reuse Anderson, Michael L. Department of Psychology, Franklin & Marshall College, Lancaster, PA 17603, and Institute for Advanced Computer Studies, Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD 20742. http://www.agcognition.org

ACKNOWLEDGMENT

Several students have played important roles in building the NICAM database used for some of the analyses reported here: Joan Brumbaugh, Kristen Calapa, Thomas Ferro, Justin Snyder, and Aysu Şuben. This project would not have been possible without their efforts. Many colleagues made helpful remarks on earlier drafts of the essay, including Paco Calvo, Cristóbal Pagán Cánovas, Tony Chemero, Andy Clark, Antoni Gomila, Julian Kiverstein, Marcie Penner-Wilger, Michael Silberstein, and Terry Stewart. The anonymous reviewers for BBS also made detailed, extensive, and helpful comments. The preparation of this article was made possible by several kinds of support from Franklin & Marshall College, including generous lab start-up funds, support for student research assistants, and a junior faculty research leave. All of this support is gratefully acknowledged.

Footnotes

1. It is perhaps worth mentioning that, although the first publications on the massive redeployment hypothesis did not appear in print until 2007, the original article detailing the theory was received by Philosophical Psychology in 2005. It hence appears likely that all the neural reuse theories of cognition discussed here were independently developed in the very same year.

2. The cortical regions studied were the same as those used in Hagmann et al (Reference Hagmann, Cammoun, Gigandet, Meuli, Honey, Wedeen and Sporns2008): “The 66 cortical regions are labeled as follows: each label consists of two parts, a prefix for the cortical hemisphere (r=right hemisphere, l=left hemisphere) and one of 33 designators: BSTS=bank of the superior temporal sulcus, CAC=caudal anterior cingulate cortex, CMF=caudal middle frontal cortex, CUN=cuneus, ENT=entorhinal cortex, FP=frontal pole, FUS=fusiform gyrus, IP=inferior parietal cortex, IT=inferior temporal cortex, ISTC=isthmus of the cingulate cortex, LOCC=lateral occipital cortex, LOF=lateral orbitofrontal cortex, LING=lingual gyrus, MOF=medial orbitofrontal cortex, MT=middle temporal cortex, PARC=paracentral lobule, PARH=parahippocampal cortex, POPE=pars opercularis, PORB=pars orbitalis, PTRI=pars triangularis, PCAL=pericalcarine cortex, PSTS=postcentral gyrus, PC=posterior cingulate cortex, PREC=precentral gyrus, PCUN=precuneus, RAC=rostral anterior cingulate cortex, RMF=rostral middle frontal cortex, SF=superior frontal cortex, SP=superior parietal cortex, ST=superior temporal cortex, SMAR=supramarginal gyrus, TP=temporal pole, and TT=transverse temporal cortex.”

3. If cognitive scientists are very bad at categorizing their experiments – at knowing what cognitive domains or tasks their experiments in fact explore – that could explain the simple finding that regions are activated by multiple tasks, because some experiments that belonged in one category would have instead been placed in another. I don't doubt we are pretty bad at this. But this fact alone would not explain the specific patterns of findings reported in support of the other predictions of redeployment. Moreover, Tony Chemero and I have performed a clustering analysis on the data to see if there is a way of dividing experiments into groups so that the neural activations do not overlap. There does not seem to be any clustering that avoids overlaps (unpublished data). We have not yet determined whether and to what degree it is possible to minimize overlap with alternate clusterings of the experiments.

4. In Talairach space, the origin is located deep in the center of the brain, and regions anterior of that are increasingly positive, and posterior to that are increasingly negative.

5. The terms “working” and “use” are adopted from Bergeron (Reference Bergeron2008). That brain regions have fixed low-level functions (“workings”) that are put to many high-level “uses” is the assumption followed by most work on the massive redeployment hypothesis (M. L. Anderson Reference Anderson2007a; Reference Anderson2007b; Reference Anderson2007c; Reference Anderson2008a; Penner-Wilger & Anderson Reference Penner-Wilger, Anderson, Love, McRae and Sloutsky2008), but it should be noted that there are other possibilities consistent with the data. For example, it could be that the dynamic response properties of local circuits are fixed, and that cognitive function is a matter of tying together circuits with the right (relative) dynamic response properties. See Anderson and Silberstein (submitted) for a discussion.

6. Terry Stewart (personal communication) suggests that an even better analogy might be modern Graphics Processing Units (GPUs). GPUs were initially intended as specialized devices to offload computationally intensive graphics rendering from the main CPU, but it has turned out they are useful for many other tasks. He writes: “it's turning out that they're extremely useful for general parallel processing, and lots of people (including us) are using them to run neural simulations. And, it's an interesting balancing task for the GPU developers to support this new use of the same working while maintaining the graphics use as well.” (See, e.g., Ho et al. Reference Ho, Lama and Leung2008; Nvidia 2007, sect. 1.1.)

7. ACT-R modules are separately modifiable, and, if neural reuse is true, the functional components of the brain will often not be. But separate modifiability does not appear to be an essential aspect of ACT-R theory, the way it is at the core of massive modularity (see sect. 3.1).

8. Some proponents of blending have argued to me that Conceptual Blending Theory (CBT) and Conceptual Metaphor Theory (CMT) are much more different than this brief description allows. For instance, Cristóbal Pagán Cánovas (personal communication) writes that:

Fauconnier and Turner argue that double-scope blending is a defining capacity of our species, of which metaphor is just a surface product, emergent from complex integration network that cannot be described by binary unidirectional mappings. I think that: a) this makes CBT and CMT hardly compatible; b) CBT (unlike CMT) accounts for frame shifting, bidirectional or multidirectional conceptual mappings, emergence of new meanings not present in their inputs, opportunistic re-use of conceptual materials, etc. and thus constitutes a change of paradigm; c) CBT is much more compatible with the massive redeployment hypothesis; d) a deeper debate about CMT and CBT is necessary. (For more on this, see Pagán Cánovas Reference Pagán Cánovas2009.)

This is certainly a very interesting issue, and I would be especially pleased if Conceptual Blending turned out to be more compatible with the observed extent of neural reuse than CMT appears to be (although whether it could account for all of it is a different matter), but space constraints dictate that we leave the matter for future discussion.

9. Note the reuse of the same neural circuits to support abstract planning wouldn't necessarily mean that one simulates motor experience as part of the planning process. Rather, for conceptual metaphor theory, the neural overlap would support the inheritance of elements of one domain (e.g., its inferential structure) by the other. The discovery of such circuit reuse therefore does offer support for both theories – although, as I have complained elsewhere (M. L. Anderson Reference Anderson, Calvo and Gomila2008c), little attention has been paid to the fact that concept empiricists and conceptual metaphor theorists in fact need to interpret this evidence in quite different ways for it to support their specific claims.

10. Apropos of which it should be noted that this approach is broadly compatible with the developmental theories of Piaget, according to which abstract thought depends on the acquisition of sensorimotor skills and concrete operations (e.g., Piaget Reference Piaget1952).

11. Glenberg et al. (Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008b) confirmed that motor regions were involved by applying TMS over the motor areas and measuring a motor-evoked potential (MEP) at the hand while having a subject judge both action sentences, describing concrete and abstract transfers, and neutral sentences. A larger MEP response was seen during transfer sentences as compared with non-transfer sentences, consistent with the notion that the motor areas are specifically activated by action sentences.

12. If it were the case that emotional valence was metaphorically mapped to movement in space without direct neural sharing, we would be more likely to see that emotions affected movement, but not the reverse, for presumably movement is not metaphorically mapped to anything. The fact that the effect is bidirectional suggests that it is mediated by the activation of something shared by and necessary to both systems, and a shared neural circuit seems a likely (although naturally not the only) possibility.

13. Note that on both views the neural overlaps could remain even if numbers were entirely amodally represented. A complete review of the evidence for and the various theories regarding the nature of this resource would take us too far afield to include here. For a discussion, see (Penner-Wilger Reference Penner-Wilger2009; Penner-Wilger & Anderson, submitted).

14. Interestingly, this inheritance by the language system of the postural organization of motor control circuits also has the potential to help explain why even American Sign Language (ASL) seems to have a phonemic structure, despite differences in modality that might otherwise have predicted a rather different organization (Sandler & Lillo-Martin Reference Sandler and Lillo-Martin2006).

15. The advantages of using this subdivision are that it ensures a neutral choice of ROIs, and lays the groundwork for future studies in which the domain-related topology of the cortex can be directly compared to the cortical connection matrix reported in that study. Thanks to the authors for sharing their ROI data.

16. The domains follow the standards defined by the BrainMap database (Fox & Lancaster Reference Fox and Lancaster2002; Laird et al. Reference Laird, Lancaster and Fox2005), and are generally determined by the authors of the study. Where available, we adopted the classification entered into the BrainMap database itself.

17. The disadvantage of using this set of ROIs is that it is based on 1.5cm2 regions of the cortical surface; hence, many activations deeper in the brain are not captured by this subdivision. One can mitigate this problem by defining a set of cubes of roughly the same size as those from Hagmann et al. (Reference Hagmann, Cammoun, Gigandet, Meuli, Honey, Wedeen and Sporns2008) – 12mm on a side – but distributed equally through the entire brain. This brings the eligible total of 12,279 activations in 1,486 experiments. For the sort of counting we are presenting here, this addition of only 17 new experiments does not materially change the results.

18. These are averages of the raw counts. If the averages are normalized to 11 (the number of possible domains in the overall average), the numbers are as follows: Action areas are active in the equivalent of 5.46 (SD 2.17) nonaction domains and 5.79 (SD 2.26) cognitive domains; perception areas are active in 4.90 (SD 1.97) non-perception domains and 5.87 (SD 2.28) cognitive domains; perception-action areas are active in the equivalent of 6.11 (SD 2.23) cognitive domains; and cognitive areas are active in 4.46 (SD 2.18) cognitive domains.

19. See Note 2.

20. The one region active in only one domain was left Frontal Pole, which was active only in memory.

21. The differences are indeed significant, 2-tailed student's t-test, p << 0.01, whether one uses the raw or normalized counts. Note that the massive redeployment hypothesis would explain this finding in terms of the relative age of the brain regions involved. Perceptual and motor circuits are more frequently reused because they are older, and not necessarily because they are functionally special.

22. Note that for the purposes of this article, the term “circuit” is more-or-less interchangeable with “small neural region.” I take the evidence of this section to indicate that small neural regions are activated in multiple tasks across multiple domains, which for current purposes is interpreted to indicate that local neural structures – that is, neural circuits – are reused in these tasks and domains. One certainly could reserve the term “circuit” for larger neural structures, such as might be revealed by combining fMRI results with Diffusion Tensor Imaging data that can reveal the physical connectivity underlying function (see, e.g., Behrens & Johansen-Berg Reference Behrens and Johansen-Berg2005; Honey et al. Reference Honey, Sporns, Cammoun, Gigandet, Thiran, Meuli and Hagmann2009; Sporns et al. Reference Sporns, Tononi and Edelman2000), but this lexical preference would not materially alter the claims of this section. And although I do believe that one of the upshots of this article as a whole is that much more attention should be paid to functional connectivity and other measures of the cooperation between cortical regions, rather than making functional attributions primarily on the basis of differential activation, following out this implication in detail will have to wait for some future paper (but see, e.g., M. L. Anderson Reference Anderson2008a).

23. The authors explicitly relate this latter aspect to the concept of affordances (Gibson Reference Gibson1979), the perceived availability of objects for certain kinds of uses or other interactions.

24. Hurford (Reference Hurford2003) suggested something like this when he hypothesized that the division between ventral stream and dorsal stream vision provide the biological basis for predicate-argument structure.

25. There is the further drawback that, in contrast to the model actually built by Feldman and Narayanan (Reference Feldman and Narayanan2004), there is no proof it is actually possible to build such a control system.

26. In fact, most household electronic thermostats contain such forward models, one reason they are more efficient than the older mercury-switch models.

27. This might push the architecture in the direction of something like the “servo stacks” concept (Hall Reference Hall, Goertzel, Hitzler and Hutter2009), which imagines building diverse high-level cognitive components from the iterative combination of simple, relatively homogenous, low-level building blocks.

28. The problem remains even given (1) observed actions will be associated with motor commands and those commands may be simulated by the observer, and (2) part of the observation is not just the movements of an agent, but also the effects of the agent's actions. Even if motor simulations kick in, enriching our observational experience, one must begin with the visual experience of the action – it is that which drives the initial categorization. And the sensory effects of an action will still differ for actor and observer, so that abstraction – attention to high-level features – will still be required.

29. A somewhat different approach to this problem is offered by Wolpert et al. (Reference Wolpert, Doya and Kawato2003). In this model, there are multiple predictors and multiple controllers arranged in an abstraction hierarchy. Actions and observations activate different controllers and predictors to different degrees, and the ones that generate the fewest errors (of prediction or of movement) over time are the ones that come to dominate. That is, they are the ones that come to drive action, or action understanding. Miali (Reference Miali2003) describes how such a model might be instantiated in a large brain circuit involving F5 mirror neurons cooperating with cerebellum and cortical motor areas. In this model, there is no need for translation between levels because there are multiple specialist modules, each corresponding to some (class of) actions or situations, already arranged in an appropriate hierarchy; but there is also little need for reuse.

30. Hurley appears to accept functional modularity, but explicitly denies anatomical modularity.

31. The producer/consumer distinction is based on Millikan's (Reference Millikan1984) as it pertains to the content of representations. It will surely need to be part of the model for how circuit reuse is possible. The “same” representation can have different content depending on the characteristics of the representation consumer. Similarly, the same neural activity or output can have different functional significance depending on the nature of the neural partners.

32. My colleague Tony Chemero and I are developing one such model, by adapting and combining insights from the literature on niche construction (Odling-Smee et al. Reference Odling-Smee, laland and Geldman2005) and the evolution of food-webs (Quince et al. Reference Quince, Higgs, McKane, Laessig and Valleriani2002), but the space of illuminating models of this process is surely quite large.

33. Consider the titles of some recent meta-analyses of imaging data: “Functional neuroanatomy of emotion: A meta-analysis of emotion activation studies in PET and fMRI” (Phan et al. Reference Phan, Wager, Taylor and Liberzon2002); “Meta-analysis of the functional neuroanatomy of single-word reading: Method and validation” (Turkeltaub et al. Reference Turkeltaub, Eden, Jones and Zeffiro2002); “Functional neuroanatomy of emotions: A meta-analysis” (Murphy et al. Reference Murphy, Nimmo-Smith and Lawrence2003); “The functional neuroanatomy of autobiographical memory: A meta-analysis” (Svoboda et al. Reference Svoboda, McKinnon and Levine2006); “A systematic review and quantitative appraisal of fMRI studies of verbal fluency: Role of the left inferior frontal gyrus” (Costafreda et al. Reference Costafreda, Fu, Lee, Everitt, Brammer and David2006). In fact, of the 51 papers that cite Fox et al. (Reference Fox, Parsons and Lancaster1998), the only one to consider studies in more than one task domain was a paper analyzing the functional connectivity of the basal ganglia (Postuma & Dagher Reference Postuma and Dagher2006).

References

Anderson, J. R. (2007) How can the human mind occur in the physical universe? Oxford University Press.CrossRefGoogle Scholar
Anderson, J. R., Qin, Y., Junk, K. -J. & Carter, C. S. (2007) Information processing modules and their relative modality specificity. Cognitive Psychology 57:185217.CrossRefGoogle Scholar
Anderson, M. L. (2003) Embodied cognition: A field guide. Artificial Intelligence 149(1):91103.CrossRefGoogle Scholar
Anderson, M. L. (2007a) Evolution of cognitive function via redeployment of brain areas. The Neuroscientist 13:1321.CrossRefGoogle ScholarPubMed
Anderson, M. L. (2007b) Massive redeployment, exaptation, and the functional integration of cognitive operations. Synthese 159(3):329–45.CrossRefGoogle Scholar
Anderson, M. L. (2007c) The massive redeployment hypothesis and the functional topography of the brain. Philosophical Psychology 21(2):143–74.CrossRefGoogle Scholar
Anderson, M. L. (2008a) Circuit sharing and the implementation of intelligent systems. Connection Science 20(4):239–51.CrossRefGoogle Scholar
Anderson, M. L. (2008c) On the grounds of x-grounded cognition. In: The Elsevier handbook of cognitive science: An embodied approach, ed. Calvo, P. & Gomila, T., pp. 423–35. Elsevier.CrossRefGoogle Scholar
Anderson, M. L., Brumbaugh, J. & Şuben, A. (2010) Investigating functional cooperation in the human brain using simple graph-theoretic methods. In: Computational neuroscience, ed. Chaovalitwongse, A., Pardalos, P. M. & Xanthopoulos, P., pp. 3142. Springer.CrossRefGoogle Scholar
Anderson, M. L. & Oates, T. (2010) A critique of multi-voxel pattern analysis. Proceedings of the 32nd Annual Meeting of the Cognitive Science Society, ed. Ohlsson, S. and Catrambone, R., pp. 1511–16. Cognitive Science Society.Google Scholar
Anderson, M. L. & Silberstein, M. D. (submitted) Constraints on localization as an explanatory strategy in the biological sciences.Google Scholar
Andres, M., Seron, X. & Oliver, E. (2007) Contribution of hand motor circuits to counting. Journal of Cognitive Neuroscience 19:563–76.CrossRefGoogle ScholarPubMed
Atallah, H. E., Frank, M. J. & O'Reilly, R. C. (2004) Hippocampus, cortex, and basal ganglia: Insights from computational models of complementary learning systems. Neurobiology of Learning and Memory 82(3):253–67.CrossRefGoogle ScholarPubMed
Awh, E., Jonides, J., Smith, E. E., Schumacher, E. H., Koeppe, R. A. & Katz, S. (1996) Dissociation of storage and rehearsal in verbal working memory: Evidence from positron emission tomography. Psychological Science 7:2531.CrossRefGoogle Scholar
Baddeley, A. D. (1986) Working memory. Oxford University Press.Google ScholarPubMed
Baddeley, A. D. (1995) Working memory. In: The cognitive neurosciences, ed. Gazzaniga, M. S., pp. 755–64. MIT Press.Google Scholar
Baddeley, A. D. & Hitch, G. (1974) Working memory. In: The psychology of learning and motivation, ed. Bower, G. H., pp. 647–67. Erlbaum.Google Scholar
Baddeley, A. D. & Hitch, G. (1994) Developments in the concept of working memory. Neuropsychology 8:485–93.CrossRefGoogle Scholar
Barabási, A.-L. & Albert, R. (1999) Emergence of scaling in random networks. Science 286:509–12.CrossRefGoogle ScholarPubMed
Barabási, A.-L., Albert, R. & Jeong, H. (2000) Scale-free characteristics of random networks: The topology of the World Wide Web. Physica A 281:6977.CrossRefGoogle Scholar
Barkow, J. H., Cosmides, L. & Tooby, J., eds. (1992) The adapted mind: Evolutionary psychology and the generation of culture. Oxford University Press.CrossRefGoogle Scholar
Barrett, H. C. & Kurzban, R. (2006) Modularity in cognition: Framing the debate. Psychological Review 113(3):628–47.CrossRefGoogle ScholarPubMed
Barsalou, L. W. (1999) Perceptual symbol systems. Behavioral and Brain Sciences 22:577660.Google Scholar
Barsalou, L. W. (2008) Grounded cognition. Annual Review of Psychology 59:617–45.CrossRefGoogle ScholarPubMed
Bechtel, W. (2003) Modules, brain parts, and evolutionary psychology. In: Evolutionary psychology: Alternative approaches, ed. Scher, S. J. & Rauscher, F., pp. 211–27. Kluwer.CrossRefGoogle Scholar
Bechtel, W. & Richardson, R. C. (1993) Discovering complexity: Decomposition and localization as strategies in scientific research. Princeton University Press.Google Scholar
Bechtel, W. & Richardson, R. C. (2010) Discovering complexity: Decomposition and localization as strategies in scientific research, 2nd edition. MIT Press/Bradford Books.CrossRefGoogle Scholar
Behrens, T. E. & Johansen-Berg, H. (2005) Relating connectional architecture to grey matter function using diffusion imaging. Philosophical Transactions of the Royal Society of London, B: Biological Sciences 360:903–11.CrossRefGoogle ScholarPubMed
Bergeron, V. (2007) Anatomical and functional modularity in cognitive science: Shifting the focus. Philosophical Psychology 20(2):175–95.CrossRefGoogle Scholar
Bergeron, V. (2008) Cognitive architecture and the brain: Beyond domain-specific functional specification. Unpublished doctoral dissertation, Department of Philosophy, University of British Columbia. Available at: http://circle.ubc.ca/handle/2429/2711.Google Scholar
Binkofski, F., Amunts, K., Stephan, K. M., Posse, S., Schormann, T., Freund, H.-J., Zilles, K. & Seitz, R. J. (2000) Broca's region subserves imagery of motion: A combined cytoarchitectonic and fMRI study. Human Brain Mapping 11:273–85.3.0.CO;2-0>CrossRefGoogle Scholar
Boroditsky, L. & Ramscar, M. (2002) The roles of body and mind in abstract thought. Psychological Science 13(2):185–88.CrossRefGoogle ScholarPubMed
Boyer, D., Miramontes, O., Ramos-Fernández, G., Mateos, J. L. & Cocho, G. (2004) Modeling the searching behavior of social monkeys. Physica A 342:329–35.CrossRefGoogle Scholar
Brown, C. T., Larry S. Liebovitch, L. S. & Glendon, R. (2007) Lévy flights in dobe ju/'hoansi foraging patterns. Human Ecology 35:129–38.CrossRefGoogle Scholar
Butterworth, B. (1999c) What counts – How every brain is hardwired for math. The Free Press.Google Scholar
Cabeza, R. & Nyberg, L. (2000) Imaging cognition II: An empirical review of 275 PET and fMRI studies. Journal of Cognitive Neuroscience 12:147.CrossRefGoogle Scholar
Carruthers, P. (2002) The cognitive functions of language. Behavioral and Brain Sciences 25(6):657–74.CrossRefGoogle ScholarPubMed
Carruthers, P. (2006) The architecture of the mind: Massive modularity and the flexibility of thought. Clarendon Press/Oxford University Press.CrossRefGoogle Scholar
Casasanto, D. & Boroditsky, L. (2008) Time in the mind: Using space to think about time. Cognition 106:579–93.CrossRefGoogle ScholarPubMed
Casasanto, D. & Dijkstra, K. (2010) Motor action and emotional memory. Cognition 115(1):179–85.Google Scholar
Changizi, M. A. & Shimojo, S. (2005) Character complexity and redundancy in writing systems over human history. Proceedings of the Royal Society of London B: Biological Sciences 272:267–75.Google ScholarPubMed
Changizi, M. A., Zhang, Q., Ye, H. & Shimojo, S. (2006) The structures of letters and symbols throughout human history are selected to match those found in objects in natural scenes. American Naturalist 167:E117–39.CrossRefGoogle ScholarPubMed
Chao, L. L. & Martin, A. (2000) Representation of manipulable man-made objects in the dorsal stream. NeuroImage 12:478–84.CrossRefGoogle ScholarPubMed
Chemero, A. (2009) Radical embodied cognitive science. MIT Press.CrossRefGoogle Scholar
Cherniak, C., Mokhtarzada, Z., Rodrigues-Esteban, R. & Changizi, K. (2004) Global optimization of cerebral cortex layout. Proceedings of the National Academy of Sciences USA 101:1081–86.Google Scholar
Clark, A. (1997) Being there: Putting brain, body, and world together again. MIT Press.Google Scholar
Clark, A. (1998) Embodied, situated, and distributed cognition. In: A companion to cognitive science, ed. Bechtel, W. & Graham, G., pp. 506–17. Blackwell.Google Scholar
Coltheart, M. (2001) Assumptions and methods in cognitive neuropsychology. In: The handbook of cognitive neuropsychology, ed. Rapp, B., pp. 321. Psychology Press.Google Scholar
Costafreda, S. G., Fu, C. H. Y., Lee, L., Everitt, B., Brammer, M. J. & David, A. S. (2006) A systematic review and quantitative appraisal of fMRI studies of verbal fluency: Role of the left inferior frontal gyrus. Human Brain Mapping 27(10):799810.CrossRefGoogle ScholarPubMed
Culham, J. C. & Valyear, K. F. (2006) Human parietal cortex in action. Current Opinion in Neurobiology 16:205–12.CrossRefGoogle ScholarPubMed
Dagher, A., Owen, A., Boecker, H. & Brooks, D. (1999) Mapping the network for planning. Brain 122:1973–87.CrossRefGoogle Scholar
Damasio, A. & Tranel, D. (1993) Nouns and verbs are retrieved with differently distributed neural systems. Proceedings of the National Academy of Sciences USA 90:4957–60.Google Scholar
Damasio, H., Grabowski,T. J., Tranel, D., Hichwa, R. D. & Damasio, A. R. (1996) A neural basis for lexical retrieval. Nature 380:499505.CrossRefGoogle ScholarPubMed
Darwin, C. (1862) On the various contrivances by which British and foreign orchids are fertilised by insects, and on the good effects of intercrossing. John Murray.Google Scholar
Deacon, T. (1997). The symbolic species. Norton.Google Scholar
Decety, J. & Grèzes, J. (1999) Neural mechanisms subserving the perception of human actions. Trends in Cognitive Sciences 3:172–78.CrossRefGoogle ScholarPubMed
Decety, J., Grezes, J., Costes, N., Perani, D., Jeannerod, M., Procyk, E., Grassi, F. & Fazio, F. (1997) Brain activity during observation of actions. Influence of action content and subject's strategy. Brain 120:1763–77.Google Scholar
Decety, J., Sjoholm, H., Ryding, E., Stenberg, G. & Ingvar, D. (1990) The cerebellum participates in cognitive activity: Tomographic measurements of regional cerebral blood flow. Brain Research 535:313–17.Google Scholar
Dehaene, S. (2005) Evolution of human cortical circuits for reading and arithmetic: The “neuronal recycling” hypothesis. In: From monkey brain to human brain, ed. Dehaene, S., Duhamel, J.-R., Hauser, M. D. & Rizolatti, G., pp. 133–57. MIT Press.Google Scholar
Dehaene, S. (2009) Reading in the brain. Viking.Google Scholar
Dehaene, S., Bossini, S. & Giraux, P. (1993) The mental representation of parity and numerical magnitude. Journal of Experimental Psychology: General 122:371–96.Google Scholar
Dehaene, S. & Cohen, L. (2007) Cultural recycling of cortical maps. Neuron 56:384–98.CrossRefGoogle ScholarPubMed
Fauconnier, G. & Turner, M. (2002) The way we think: Conceptual blending and the mind's hidden complexities. Basic Books.
Fedorenko, E., Patel, A., Casasanto, D., Winawer, J. & Gibson, T. (2009) Structural integration in language and music: Evidence for a shared system. Memory and Cognition 37(1):19.CrossRefGoogle ScholarPubMed
Feldman, J. & Narayanan, S. (2004) Embodied meaning in a neural theory of language. Brain and Language 89:385–92.Google Scholar
Fodor, J. (1975) The language of thought. Harvard University Press.Google Scholar
Fodor, J. & Pylyshyn, Z. W. (1988) Connectionaism and cognitive architecture: A critical analysis. Cognition 28:371.CrossRefGoogle ScholarPubMed
Fowler, C. A., Rubin, P., Remez, R. E. & Turvey, M. T. (1980) Implications for speech production of a general theory of action. In: Language production, vol. 1: Speech and talk, ed. Butterworth, B., pp. 373420. Academic Press.Google Scholar
Fox, P. T. & Lancaster, J. L. (2002) Mapping context and content: The BrainMap model. Nature Reviews Neuroscience 3:319–21.CrossRefGoogle ScholarPubMed
Fox, P. T., Parsons, L. M. & Lancaster, J. L. (1998) Beyond the single study: Function-location meta-analysis in cognitive neuroimaging. Current Opinions in Neurobiology 8:178–87.CrossRefGoogle ScholarPubMed
Fries, R. C. (2006) Reliable design of medical devices. CRC Press.Google Scholar
Gallese, V. (2003) A neuroscientific grasp of concepts: From control to representation. Philosophical Transactions of the Royal Society London, B: Biological Sciences 358(1435):1231–40.CrossRefGoogle ScholarPubMed
Gallese, V. (2008) Mirror neurons and the social nature of language: The neural exploitation hypothesis. Social Neuroscience 3 (3–4):317–33.Google Scholar
Gallese, V., Fadiga, L., Fogassi, L. & Rizzolatti, G. (1996) Action recognition in the premotor cortex. Brain 119:593609.CrossRefGoogle ScholarPubMed
Gallese, V. & Goldman, A. (1998) Mirror neurons and the simulation theory of mind-reading. Trends in Cognitive Sciences 2(12):493501.CrossRefGoogle ScholarPubMed
Gallese, V. & Lakoff, G. (2005) The brain's concepts: The role of the sensory-motor system in conceptual knowledge. Cognitive Neuropsychology 22 (3–4):455–79.Google Scholar
Gauthier, I., Skudlarski, P., Gore, J. C. & Anderson, A. W. (2000) Expertise for cars and birds recruits brain areas involved in face recognition. Nature Neuroscience 3(2):191–97.CrossRefGoogle ScholarPubMed
Gentner, D. & Stevens, A. L., eds. (1983) Mental models. Erlbaum.Google Scholar
Gibson, J. J. (1979) The ecological approach to visual perception. Erlbaum.Google Scholar
Gigerenzer, G., Todd, P. M. & The ABC Research Group (1999) Simple heuristics that make us smart. Oxford University Press.Google Scholar
Gilovich, T., Griffin, D. & Kahneman, D., eds. (2002) Heuristics and biases: The psychology of intuitive judgment. Cambridge University Press.Google Scholar
Glenberg, A. M., Becker, R., Klötzer, S., Kolanko, L., Müller, S. & Rinck, M. (2009) Episodic affordances contribute to language comprehension. Language and Cognition 1:113–35.Google Scholar
Glenberg, A. M., Brown, M. & Levin, J. R. (2007) Enhancing comprehension in small reading groups using a manipulation strategy. Contemporary Educational Psychology 32:389–99.Google Scholar
Glenberg, A. M. & Kaschak, M. P. (2002) Grounding language in action. Psychonomic Bulletin and Review 9:558–65.CrossRefGoogle ScholarPubMed
Glenberg, A. M., Sato, M. & Cattaneo, L. (2008a) Use-induced motor plasticity affects the processing of abstract and concrete language. Current Biology 18:R290–91.Google Scholar
Glenberg, A. M., Sato, M., Cattaneo, L., Riggio, L., Palumbo, D. & Buccino, G. (2008b) Processing abstract language modulates motor system activity. Quarterly Journal of Experimental Psychology 61:905–19.CrossRefGoogle ScholarPubMed
Goldin-Meadow, S. (2003) Hearing gesture: How our hands help us think. Belknap Press.Google Scholar
Graziano, M. S. A., Taylor, C. S. R. & Moore, T. (2002a) Complex movements evoked by microstimulation of precentral cortex. Neuron 34:841–51.Google Scholar
Graziano, M. S. A., Taylor, C. S. R., Moore, T. & Cooke, D. F. (2002b) The cortical control of movement revisited. Neuron 36:349–62.Google Scholar
Grill-Spector, K., Sayres, R. & Ress, D. (2006) High-resolution imaging reveals highly selective nonface clusters in the fusiform face area. Nature Neuroscience 9(9):1177–85.CrossRefGoogle ScholarPubMed
Haggard, P., Rossetti, Y. & Kawato, M., eds. (2008) Sensorimotor foundations of higher cognition. Oxford University Press.Google Scholar
Hagmann, P., Cammoun, L., Gigandet, X., Meuli, R., Honey, C. J., Wedeen, V. J. & Sporns, O. (2008) Mapping the structural core of human cerebral cortex. PLoS Biology 6(7):e159. Available at: http://biology.plosjournals.org/perlserv/?request=get-document. doi:10.1371/journal.pbio.0060159.Google Scholar
Hagoort, P. (2005) On Broca, brain and binding. Trends in Cognitive Sciences 9(9): 416–23.CrossRefGoogle ScholarPubMed
Hall, J. S. (2009) The robotics path to AGI using servo stacks. In: Proceedings of the Second Conference on Artificial General Intelligence, ed. Goertzel, B., Hitzler, P. & Hutter, M., pp. 4954. Atlantis Press. doi:10.2991/agi.2009.5.Google Scholar
Hamzei, F., Rijntjes, M., Dettmers, C., Glauche, V., Weiller, C. & Büchel (2003) The human action recognition system and its relationship to Broca's area: An fMRI study. Neuroimage 19:637–44.Google Scholar
Hanakawa, T., Honda, M., Sawamoto, N., Okada, T., Yonekura, Y., Fukuyama, H. & Shibasaki, H. (2002) The role of rostral Brodmann area 6 in mental-operation tasks: An integrative neuroimaging approach. Cerebral Cortex 12:1157–70.Google Scholar
Ho, T. -Y., Lama, P. -M. & Leung, C. -S. (2008) Parallelization of cellular neural networks on GPU. Pattern Recognition 41(8):2684–92.CrossRefGoogle Scholar
Honey, C. J., Kötter, R., Breakspear, M. & Sporns, O. (2007) Network structure of cerebral cortex shapes functional connectivity on multiple time scales. Proceedings of the National Academy of Sciences USA 104:10240–45.Google Scholar
Honey, C. J., Sporns, O., Cammoun, L., Gigandet, X., Thiran, J. P., Meuli, R. & Hagmann, P. (2009) Predicting human resting-state functional connectivity from structural connectivity. Proceedings of the National Academy of Sciences USA 106(6):2035–40.Google Scholar
Hopkin, V. D. (1995) Human factors in air traffic control. CRC Press.Google Scholar
Hubbard, E. M., Piazza, M., Pinel, P. & Dehaene, S. (2005) Interactions between number and space in parietal cortex. Nature Reviews Neuroscience 6(6):435–48.Google Scholar
Hurford, J. (2003) The neural basis of predicate-argument structure. Behavioral and Brain Sciences 26(3):261–83.Google Scholar
Hurley, S. L. (2005) The shared circuits hypothesis: A unified functional architecture for control, imitation and simulation. In: Perspectives on imitation: From neuroscience to social science, ed. Hurley, S. & Chater, N., pp. 7695. MIT Press.Google Scholar
Hurley, S. L. (2008) The shared circuits model (SCM): How control, mirroring, and simulation can enable imitation, deliberation, and mindreading. Behavioral and Brain Sciences 31(1):158.CrossRefGoogle ScholarPubMed
Hutchins, E. (1995) Cognition in the wild. MIT Press.Google Scholar
Iriki, A. (2005). A prototype of homo-faber: A silent precursor of human intelligence in the tool-using monkey brain. In: From monkey brain to human brain, ed. Dehaene, S., Duhamel, J. R., Hauser, M. & Rizzolati, G., pp. 133–57. MIT Press.Google Scholar
Iriki, A. & Sakura, O. (2008) Neuroscience of primate intellectual evolution: Natural selection and passive and intentional niche construction. Philosophical Transactions of the Royal Society of London, B: Biological Science 363:2229–41.CrossRefGoogle ScholarPubMed
Jeannerod, M. (1994) The representing brain: Neural correlates of motor intention and imagery. Behavioral and Brain Sciences 17:187245.Google Scholar
Jeong, H., Tombor, B., Albert, R., Oltvai, Z. N. & Barabási, A. -L. (2000) The large-scale organization of metabolic networks. Nature 407:651–54.Google Scholar
Jilk, D. J., Lebiere, C., O'Reilly, R. C. & Anderson, J. R. (2008) SAL: An explicitly pluralistic cognitive architecture. Journal of Experimental and Theoretical Artificial Intelligence 20:197218.Google Scholar
Johnson-Laird, P. N. (1983) Mental models: Towards a cognitive science of language, inference, and consciousness. Harvard University Press.Google Scholar
Kanwisher, N., McDermott, J. & Chun, M. (1997) The fusiform face area: A module in human extrastriate cortex specialized for face perception. Journal of Neuroscience 17(11):4302–11.CrossRefGoogle ScholarPubMed
Koch, C. & Segev, I. (2000) The role of single neurons in information processing. Nature Neuroscience 3:1171–77.Google Scholar
Laird, A. R., Lancaster, J. L. & Fox, P. T. (2005) BrainMap: The social evolution of a functional neuroimaging database. Neuroinformatics 3:6578.Google Scholar
Lakoff, G. & Johnson, M. (1980) Metaphors we live by. University of Chicago Press.Google Scholar
Lakoff, G. & Johnson, M. (1999) Philosophy in the flesh: The embodied mind and its challenge to western thought. Basic Books.Google Scholar
Lakoff, G. & Núñez, R. (2000) Where mathematics comes from: How the embodied mind brings mathematics into being. Basic Books.Google Scholar
Lloyd, D. (2000) Terra cognita: From functional neuroimaging to the map of the mind. Brain and Mind 1(1):93116.CrossRefGoogle Scholar
Marcus, G. F. (2004) The birth of the mind: How a tiny number of genes creates the complexities of human thought. Basic Books.Google Scholar
Marcus, G. F. (2008) Kluge: The haphazard construction of the human mind. Houghton Mifflin.Google Scholar
Martin, A., Haxby, J. V., Lalonde, F. M., Wiggs, C. L. & Ungerleider, L. G. (1995) Discrete cortical regions associated with knowledge of color and knowledge of action. Science 270:102105.Google Scholar
Martin, A., Ungerleider, L. G. & Haxby, J. V. (2000) Category-specificity and the brain: the sensorymotor model of semantic representations of objects. In: The new cognitive neurosciences, 2nd edition, ed. Gazzaniga, M. S., pp. 1023–36. MIT Press.Google Scholar
Martin, A., Wiggs, C. L., Ungerleider, L. G. & Haxby, J. V. (1996) Neural correlates of category-specific knowledge. Nature 379:649–52.CrossRefGoogle ScholarPubMed
Mesulam, M.-M. (1990) Large-scale neurocognitive networks and distributed processing for attention, language and memory. Annals of Neurology 28:597613.Google Scholar
Miali, R. C. (2003) Connecting mirror neurons and forward models. NeuroReport 14(17):2135–37.Google Scholar
Millikan, R. G. (1984) Language, thought and other biological categories. MIT Press.CrossRefGoogle Scholar
Mitchell, M. (2006). Complex systems: Network thinking. Artificial Intelligence 170:1194–212.Google Scholar
Müller, R.-A. & Basho, S. (2004) Are nonlinguistic functions in “Broca's area” prerequisites for language acquisition? fMRI findings from an ontogenetic viewpoint. Brain and Language 89(2):329–36.Google Scholar
Murphy, F. C., Nimmo-Smith, I. & Lawrence, A. D. (2003) Functional neuroanatomy of emotions: A meta-analysis. Cognitive, Affective and Behavioral Neuroscience 3(3):207–33.CrossRefGoogle ScholarPubMed
Newell, A. & Simon, H. A. (1976) Computer science as empirical enquiry. Communications of the ACM 19(3):113–26.CrossRefGoogle Scholar
Newman, M., Barabasi, A.-L. & Watts, D. J. (2006) The structure and dynamics of networks. Princeton University Press.Google Scholar
Nishitani, N., Schürmann, M., Amunts, K. & Hari, R. (2005) Broca's region: From action to language. Physiology 20:6069.Google Scholar
Nvidia Corporation. (2007) CUDA Programming Guide, version 1.1. Santa Clara, CA. Available at: http://developer.download.nvidia.com/compute/cuda/1_1/NVIDIA_CUDA_Programming_Guide_1.1.pdf.Google Scholar
Odling-Smee, F. J., laland, K. N. & Geldman, M. W. (2005) Niche construction: The neglected process in evolution. Princeton University Press.Google Scholar
O'Reilly, R. C. (1998) Six principles for biologically based computational models of cortical cognition. Trends in Cognitive Sciences 2:455–62.Google Scholar
O'Reilly, R. C. & Munakata, Y. (2000) Computational explorations in cognitive neuroscience: Understanding the mind by simulating the brain. MIT Press.Google Scholar
Pagán Cánovas, C. (2009) La emisión erótica en la poesía griega: una familia de redes de integración conceptual desde la Antigüedad hasta el siglo XX. Departmento Filología Clásica, Universidad de Murcia, Spain. http://www.tesisenred.net/TDR-0519110-103532/index.html.Google Scholar
Penner-Wilger, M. (2009) Subitizing, finger gnosis, and finger agility as precursors to the representation of number. Unpublished doctoral dissertation, Department of Cognitive Science, Carleton University, Ottawa, Canada. http://gradworks.umi.com/NR/52/NR52070.Google Scholar
Penner-Wilger, M. & Anderson, M. L. (2008) An alternative view of the relation between finger gnosis and math ability: Redeployment of finger representations for the representation of number. In: Proceedings of the 30th Annual Meeting of the Cognitive Science Society, Austin, TX, July 23–26, 2008, ed. Love, B. C., McRae, K. & Sloutsky, V. M., pp. 1647–52. Cognitive Science Society.Google Scholar
Penner-Wilger, M. & Anderson, M. L. (submitted) The relation between finger recognition and mathematical ability: Why redeployment of neural circuits best explains the finding.Google Scholar
Pereira, F., Mitchell, T. & Botvinick, M. M. (2009) Machine learning classifiers and fMRI: A tutorial overview. NeuroImage 45:S199209.CrossRefGoogle ScholarPubMed
Pessoa, L. (2008) On the relationship between emotion and cognition. Nature Reviews Neuroscience 9:148–58.CrossRefGoogle ScholarPubMed
Phan, K. L, Wager, T., Taylor, S. F. & Liberzon, I. (2002) Functional neuroanatomy of mmotion: A meta-analysis of emotion activation studies in PET and fMRI. NeuroImage 16(2):331–48.Google Scholar
Piaget, J. (1952) The child's conception of number. Routledge and Kegan Paul.Google Scholar
Pinker, S. (1997) How the mind works. Norton.Google Scholar
Plaut, D. C. (1995) Double dissociation without modularity: Evidence from connectionist neuropsychology. Journal of Clinical and Experimental Neuropsychology 17:291321.CrossRefGoogle ScholarPubMed
Poldrack, R. A. (2006) Can cognitive processes be inferred from neuroimaging data? Trends in Cognitive Sciences 10:5963.Google Scholar
Postuma, R. B. & Dagher, A. (2006) Basal ganglia functional connectivity based on a meta-analysis of 126 PET and fMRI publications. Cerebral Cortex 16(10):1508–21.Google Scholar
Prinz, J. (2002) Furnishing the mind: Concepts and their perceptual basis. MIT Press.Google Scholar
Prinz, J. (2006) Is the mind really modular? In: Contemporary debates in cognitive science, ed. Stainton, R. J., pp. 2236. Blackwell.Google Scholar
Pulvermüller, F. (2005) Brain mechanisms linking language and action. Nature Reviews Neuroscience 6:576–82.Google Scholar
Quartz, S. R. & Sejnowski, T. J. (1997) The neural basis of cognitive development: A constructivist manifesto. Behavioral and Brain Sciences 20:537–56.Google Scholar
Quince, C., Higgs, P. G. & McKane, A. J. (2002) Food web structure and the evolution of ecological communities. In: Biological evolution and statistical physics: Lecture notes in Physics 585, ed. Laessig, M. & Valleriani, A., pp. 281–98. Springer-Verlag.Google Scholar
Rasmussen, J. & Vicente, K. J. (1989) Coping with human errors through system design: implications for ecological interface design. International Journal of Man-Machine Studies 31:517–34.Google Scholar
Rhodes, G., Byatt, G., Michie, P. T. & Puce, A. (2004) Is the Fusiform Face Area specialized for faces, individuation, or expert individuation?. Journal of Cognitive Neuroscience 16(2):189203.CrossRefGoogle ScholarPubMed
Richardson, D., Spivey, M., Barsalou, L. & McRae, K. (2003) Spatial representations activated during real-time comprehension of verbs. Cognitive Science 27:767–80.Google Scholar
Ritter, F. E. & Young, R. M., eds. (2001) Using cognitive models to improve interface design. International Journal of Human-Computer Studies 55(1):1107. (Special Issue.)Google Scholar
Rizzolatti, G., Fadiga, L., Gallese, V. & Fogassi, L. (1996) Premotor cortex and the recognition of motor actions. Cognitive Brain Research 3:131–41.Google Scholar
Roux, F. -E., Boetto, S., Sacko, O., Chollet, F. & Tremoulet, M. (2003) Writing, calculating, and finger recognition in the region of the angular gyrus: A cortical stimulation study of Gerstmann syndrome. Journal of Neurosurgery 99:716–27.Google Scholar
Rumelhart, D. E. & McClelland, J. L (1986) Parallel distributed processing: Explorations in the microstructure of cognition. MIT Press.Google Scholar
Rusconi, E., Walsh, V. & Butterworth, B. (2005) Dexterity with numbers: rTMS over left angular gyrus disrupts finger gnosis and number processing. Neuropsychologia 43:1609–24.Google Scholar
Salvucci, D. D. (2005) A multitasking general executive for compound continuous tasks. Cognitive Science 29:457–92.Google Scholar
Sandler, W. & Lillo-Martin, D. (2006) Sign languages and linguistic universals. Cambridge University Press.Google Scholar
Scher, S. J. (2004) A lego model of the modularity of the mind. Journal of Cultural and Evolutionary Psychology 2(21):248–59.Google Scholar
Simmons, W. K., Ramjee, V., Beauchamp, M. S., McRae, K., Martin, A. & Barsalou, L. W. (2007) A common neural substrate for perceiving and knowing about color. Neuropsychologia 45(12): 2802–10.CrossRefGoogle ScholarPubMed
Simon, H. A. (1962/1969) The architecture of complexity. Proceedings of the American Philosophical Association 106:467–82. Reprinted in: H. Simon, The sciences of the artificial, 1st edition, pp. 192–229. MIT Press, 1969.Google Scholar
Sperber, D. (1996) Explaining culture. Blackwell.Google Scholar
Sporns, O., Chialvo, D. R., Kaiser, M. & Hilgetag, C. C. (2004) Organization, development and function of complex brain networks. Trends in Cognitive Sciences 8:418–25.Google Scholar
Sporns, O., Tononi, G. & Edelman, G. M. (2000) Theoretical neuroanatomy: Relating anatomical and functional connectivity in graphs and cortical connection matrices. Cerebral Cortex 10:127–41.Google Scholar
Sternberg, S. (1969) The discovery of processing stages: Extensions of Donders' method. Acta Psychologica 30: 276315.Google Scholar
Stewart, T. C. & West, R. L. (2007) Cognitive redeployment in ACT-R: Salience, vision, and memory. Paper presented at the 8th International Conference on Cognitive Modelling, Ann Arbor, MI, July 26–29, 2007.Google Scholar
Svoboda, E., McKinnon, M. C. & Levine, B. (2006) The functional neuroanatomy of autobiographical memory: A meta-analysis. Neuropsychologia 44(12):2189–208.Google Scholar
Talairach, J. & Tournaux, P. (1988) Co-planar stereotaxic atlas of the human brain. Thieme.Google Scholar
Tettamanti, M. & & Weniger, D. (2006) Broca's area: A supramodal hierarchical processor? Cortex 42:491–94.Google Scholar
Thoenissen, D., Zilles, K. & Toni, I. (2002) Differential involvement of parietal and precentral regions in movement preparation and motor intention. Journal of Neuroscience 22:9024–34.Google Scholar
Tooby, J. & Cosmides, L. (1992) The psychological foundations of culture. In: The adapted mind: Evolutionary psychology and the generation of culture, ed. Barkow, J., Cosmides, L. & Tooby, J., pp. 19136. Oxford University Press.Google Scholar
Turkeltaub, P. E., Eden, G. F., Jones, K. M. & Zeffiro, T. A. (2002) Meta-analysis of the functional neuroanatomy of single-word reading: Method and validation. NeuroImage 16:765–80.CrossRefGoogle ScholarPubMed
Van Orden, G. C., Pennington, B. F. & Stone, G. O. (2001) What do double dissociations really prove? Cognitive Science 25:111–72.CrossRefGoogle Scholar
Weiskopf, D. (2007) Concept empiricism and the vehicles of thought. The Journal of Consciousness Studies 14:156–83.Google Scholar
Wen, Q. & Chklovskii, D. B. (2008) A cost–benefit analysis of neuronal morphology. Journal of Neurophysiology 99:2320–28.Google Scholar
Wilson, M. (2001) The case for sensorimotor coding in working memory. Psychonomic Bulletin and Review 8:4457.Google Scholar
Wilson, M. (2002) Six views of embodied cognition. Psychonomic Bulletin and Review 9(4):625–36.Google Scholar
Wolpert, D. M., Doya, K. & Kawato, M. (2003) A unifying computational framework for motor control and social interaction. Philosophical Transactions of the Royal Society B: Biological Sciences 358:593602.Google Scholar
Zago, L., Pesenti, M., Mellet, E., Crivello, F., Mazoyer, B. & Tzourio-Mazoyer, N. (2001) Neural correlates of simple and complex mental calculation. NeuroImage 13:314–27.Google Scholar
Figure 0

Figure 1. Expected patterns of co-activation in a simple six-region brain for two cognitive functions (solid vs. dashed lines). Anatomical modularity and localization (top) predicts largely non-overlapping sets of regions will contribute to each function, whereas reuse (bottom) suggests that many of the same cortical regions will be activated in support of both functions, but that they will co-activate (cooperate) in different patterns.

Figure 1

Figure 2. Detail of Figure 3 from Mesulam (1990). Reprinted with permission of the author.

Figure 2

Figure 3. Overview of the Leabra architectural organization. Reprinted from Jilk et al. (2008) with permission of the authors.