Hostname: page-component-745bb68f8f-v2bm5 Total loading time: 0 Render date: 2025-02-06T19:50:24.992Z Has data issue: false hasContentIssue false

Toward mechanistic models of action-oriented and detached cognition

Published online by Cambridge University Press:  30 June 2016

Giovanni Pezzulo*
Affiliation:
Institute of Cognitive Sciences and Technologies, National Research Council, 00185 Rome, Italy. giovanni.pezzulo@istc.cnr.ithttp://www.istc.cnr.it/people/giovanni-pezzulo

Abstract

To be successful, the research agenda for a novel control view of cognition should foresee more detailed, computationally specified process models of cognitive operations including higher cognition. These models should cover all domains of cognition, including those cognitive abilities that can be characterized as online interactive loops and detached forms of cognition that depend on internally generated neuronal processing.

Type
Open Peer Commentary
Copyright
Copyright © Cambridge University Press 2016 

There is increasing consensus in cognitive science and neuroscience that we need a novel view of the brain that is more action-oriented; in this perspective, it has been argued that we might be facing a “pragmatic turn” in cognitive science (Engel et al. Reference Engel, Maye, Kurthen and König2013). One central proposal of action-oriented views is that the brain evolved to be a control system for the interaction with the external environment. We should look at cognitive (and brain) functions, including the most advanced (or “higher-cognitive”) functions, within an interacted and control-theoretic framework – as activities that an organism performs in interaction with its environment, rather than in terms of modular computational operations over discrete symbols independent from perception and action systems. This idea can be traced back to early theories in cybernetics, pragmatism, and ecological psychology (Ashby Reference Ashby1952; Craik Reference Craik1943; Gibson Reference Gibson, Shaw and Bransford1977; Wiener Reference Wiener1948) and has been often reproposed, in slightly different forms, in disciplines such as cognitive science, neuroscience, robotics, and philosophy (Cisek Reference Cisek1999; Cisek & Kalaska Reference Cisek and Kalaska2010; Clark Reference Clark1998; Engel et al. Reference Engel, Maye, Kurthen and König2013; Pezzulo Reference Pezzulo2011; Pezzulo & Castelfranchi Reference Pezzulo and Castelfranchi2009; Pezzulo et al. Reference Pezzulo, Rigoli and Friston2015; Pfeifer & Scheier Reference Pfeifer and Scheier1999; Scott Reference Scott2012; Stoianov et al. Reference Stoianov, Genovesio and Pezzulo2016; Varela et al. Reference Varela, Thompson and Rosch1992). Anderson (Reference Anderson2014) contributes to this debate both theoretically and empirically. He proposes a view of brain organization in which regions form functional coalitions dynamically and implement control loops for agent–environment interactions. He also discusses how neuronal coalitions could be reused to implement higher cognitive skills, forming possible control loops to support mathematical and language processing.

Despite its appeal, the impact of the action-oriented view is mitigated by various factors. One reason is the level of ambition of the novel proposal, which requires a unitary perspective on brain function and behavior. It is easier to study the brain if one assumes that each area or network is the locus of a single (or a small set of) cognitive operation(s) such as perception, memory, language, and so forth, and that each of these functions can be largely studied in isolation. But this perspective is not fully compatible with an action-oriented view, which assumes that the function of the brain as a whole is engaging in interactive control loops. Because these loops extend beyond the brain, we should also consider the contribution of the body and the external environment, including other agents, to cognition.

Given this level of ambition, advocating a generic “control metaphor” for cognition that replaces the older metaphors of the brain – the brain as a computer or as a serial transducer from stimuli to behavior – might be insufficient. What might be more effective is the realization of a coherent set of computationally specified process models of action-oriented cognitive processing – from sensorimotor transformations and situated choice to higher cognitive skills such as mathematics, language processing, and problem solving – that can guide empirical research and provide a more compelling and unifying explanation of empirical findings.

Indeed, despite noteworthy empirical demonstrations of the relevance of action-based concepts to understand cognition, including higher cognition (Barsalou Reference Barsalou2008; Glenberg Reference Glenberg1997; Jeannerod Reference Jeannerod2006), we still largely lack detailed process models. For example, several studies illustrate the importance of bodily processes and visuomotor strategies for decision making or problem solving (Spivey Reference Spivey2007), but the (relative) lack of mechanistic models hinders a complete understanding of these phenomena. Progress in the field will benefit from a “new alliance” between proponents of embodied and action-oriented views cognition and computational modelers including roboticists (Pezzulo et al. Reference Pezzulo, Barsalou, Cangelosi, Fischer, McRae and Spivey2011; Reference Pezzulo, Barsalou, Cangelosi, Fischer, McRae and Spivey2013). The idea that cognitive robotics can shed light on brain and cognition is relatively new, but it should not be too surprising if one considers that robots are well suited to implement the kind of brain-body-environment (and social) interactions that are considered essential for action-based cognition (Verschure et al. Reference Verschure, Pennartz and Pezzulo2014). Furthermore, developing complete models (robotic or nonrobotic) of cognitive operations can counteract an excess of specialization in the field – say, the focus on the functioning of just one brain area.

If we consider the theoretical debate within action-based theories, it often revolves around (the need for) internal representation. Various competing proposals include: abandoning the notion of internal representation to see cognition as online control and interaction with the environment (Chemero Reference Chemero2009; Gallagher Reference Gallagher2005); externalizing representation (Clark & Chalmers Reference Clark and Chalmers1998; Kirsh Reference Kirsh2010); replacing it with other constructs such as sensorimotor contingencies (O'Regan & Noe Reference O'Regan and Noe2001) or dispositions (Engel et al. Reference Engel, Maye, Kurthen and König2013); amending it in a more embodied and action-oriented view, in terms of, for example, perceptual symbols (Barsalou Reference Barsalou1999), action-oriented representations, or emulators that support control structures (Clark & Grush Reference Clark and Grush1999; Grush Reference Grush2004), to mention just a few.

Although important, this debate might be limited if not accompanied by the realization of mechanistic models, which answer specific questions such as: What kind of control system is the brain? What control loops are instantiated between the brain, the body, the environment, and with other agents, during the implementation of tasks of various complexity, such as walking, solving a puzzle, or planning a trip together? How can these and other tasks be mechanistically described in terms of agent–environment interactions, and which variables, if any, are technically “controlled” in these tasks? What is the contribution of different brain areas to each of the control loops required by different tasks, how are their contributions dynamically assembled, and which are their neuronal signatures? Which loops require a continuous engagement with external variables and which ones their internal, endogenous generation? Which aspects of the structure of the agent–environment interactions can be exploited online, which can be offloaded externally, and which, if any, need to be internalized? How can we better simulate, measure, and empirically study these interactive loops? And so on.

Several researchers are already addressing these and other relevant questions. However, the space of possibilities has not been systematically explored, with many interesting constructs from control theory, cybernetics, and related fields that remain untested. These include early proposals such as hierarchical perceptual control (Powers Reference Powers1973), the test-operate-test-exit architecture (Rosenblueth et al. Reference Rosenblueth, Wiener and Bigelow1943), and several other constructs from early cybernetics (Ashby Reference Ashby1952; Wiener Reference Wiener1948), as well as more recent developments such as, for example, optimal feedback control, model-based control, and risk-sensitive control, which are often mentioned in theoretical debates but not systematically tested (Shadmehr & Mussa-Ivaldi Reference Shadmehr and Mussa-Ivaldi2012).

Incorporating these ideas into specific process models could guide progress in psychology and neuroscience and illustrate where exactly an action-based theory differs from traditional ones. One interesting success case is the affordance competition hypothesis (Cisek Reference Cisek2006), in which action specification and selection proceed in parallel and compete – on the basis of the various sensory, contextual, affective, and motivational biases they continuously receive – until one action pattern is selected, hence dispensing from a central decision module that processes abstract values prior to action. Another recent example is the active inference framework (Friston et al. Reference Friston, Daunizeau, Kilner and Kiebel2010; Reference Friston, Rigoli, Ognibene, Mathys, FitzGerald and Pezzulo2015). It can be considered a modern development of cybernetic ideas (Seth 2014), although it casts control as a (probabilistic) inference problem (see also Attias Reference Attias, Bishop and Frey2003; Toussaint Reference Toussaint2009). It challenges traditional models of perception and action control in several ways – for example, by natively incorporating top-down processes and by highlighting the importance of predictive processing and error correction over and above stimulus-response associations – and it paves the way to the realization of models of more elaborated cognitive operations, suggesting that they might be based on the same principles (Clark Reference Clark2013b; Pezzulo Reference Pezzulo2012).

Developing mechanistic models can help advance our theoretical understanding of action-based cognition, too, because some of the aforementioned models cut across dichotomies (e.g., interactivity vs. internal modeling) that are often debated, including in Anderson (Reference Anderson2014). For example, model-based control and active inference offer a dynamical view of online action control and interaction but include internal generative and predictive modeling loops to make those interactions more effective. These systems learn the statistical structure of the environment and of agent–environment interactions, and they perform state estimation. However, structure learning and estimation are functional to effective control rather than having as a final goal a veridical representation of the external environment. These models remain to be fully tested. However, these examples suggest that the development and empirical testing of specific process models might contribute to the theoretical debate on action-based cognition – for example, by clarifying the possible contribution of predictive processing and structure learning to control and cognition.

These arguments are also important to meet another challenge of action-oriented cognition: developing process models of the brain's ability to temporarily detach (or disengage) from the here-and-now of the overt sensorimotor loop, as in the case of imagining or planning the future, or mental time travel. These and other detached forms of cognition have been traditionally considered difficult to explain from an action-based perspective and especially in terms of online control loops. In many parts of After Phrenology, Anderson (Reference Anderson2014) suggests that cognition consists in online interaction with the external environment, with no need for neuronal tissue to internalize the structure of such interaction. He also describes aspects of cognition as the manipulation of external symbols (Clark & Chalmers Reference Clark and Chalmers1998) and briefly alludes to the possibility that, if external symbols are unavailable, internal resources such as memory and imagination might be deployed in their stead. This latter possibility, and in general the contribution of internally generated neuronal processing to cognition, would require much more attention to fully understand some aspects of higher cognition as detached cognition.

One domain where internally generated neuronal processing has been studied in great detail is rodent navigation. It has been consistently reported that hippocampal place cells, whose firing is normally associated with the animal's spatial position, can also fire when the animal is outside its standard “place field,” especially during periods of rest or sleep, and at decision points. This “out-of-field” neuronal activity cannot be driven by external stimuli (consider the case of the sleeping animal) but needs to be internally generated based on intrinsic network dynamics. Still, it unfolds in highly organized forms and is neurophysiologically similar to the activity observed during overt action. One example is internally generated sequences of place cell activations observed when the animal rests or sleeps. These sequences can, for example, “replay” spatial trajectories that the animal has experienced (or recombined trajectories), forward or backward, in a time-compressed manner (e.g., during sharp wave ripple complexes, at about 140–200 Hz). These and other forms of internally generated sequences (e.g., preplays) have been associated with various functions such as memory consolidation and planning (Diba & Buzsáki Reference Diba and Buzsáki2007; Dragoi & Tonegawa Reference Dragoi and Tonegawa2011; Gupta et al. Reference Gupta, van der Meer, Touretzky and Redish2010; Pezzulo et al. Reference Pezzulo, van der Meer, Lansink and Pennartz2014; Pfeiffer & Foster Reference Pfeiffer and Foster2013). Another kind of internally generated neuronal sequence is expressed in the theta rhythm (8–12 Hz), when the animals are engaged in decision tasks, and has been associated with “what-if” loops and the anticipation of the consequences of possible choices in order to select, say, one of two arms of a maze (Johnson & Redish Reference Johnson and Redish2007; Wikenheiser & Redish Reference Wikenheiser and Redish2015).

These covert phenomena are not unique to the hippocampus but have been reported in several other brain areas and can at least potentially support a variety of detached or covert cognitive operations (Buzsáki et al. Reference Buzsáki, Peyrache and Kubie2015; Lisman Reference Lisman2015). Importantly, the covert activity recruits the same (or closely related) neurophysiological mechanisms and neuronal resources (e.g., neural assembly sequences) as those implied in overt action, but through internally generated processing. This raises the possibility that detached forms of cognition such as planning, but also potentially others such as imagination and mental time travel, can be explained within an action-based framework rather than require a distinct ontology of neural constructs – if we allow control loops to extend beyond online interactions to also cover internally generated neuronal dynamics.

In this vein, the possibility that the same brain networks can operate and realize control loops in two distinct modes – one stimulus-driven and another (detached) based on internally generated dynamics – has been raised by several researchers and is often associated with the functioning of predictive and generative models (Grush Reference Grush2004; Maye & Engel Reference Maye and Engel2011; Pezzulo & Castelfranchi Reference Pezzulo and Castelfranchi2009) and to theories of reuse (Anderson Reference Anderson2014; Dehaene Reference Dehaene, Dehaene, Duhamel, Hauser and Rizzolatti2005). However, these hypotheses remain to be systematically tested with the help of mechanistic models. If these theories are on the right track, (some) detached forms of cognition might be based on internalized neuronal loops that recapitulate the same dynamics as those learned during overt action (e.g., replays of spatial trajectories in hippocampal ripples) and hence remain fundamentally action-oriented, albeit in an indirect or covert form. The progressive internalization of overt loops, possibly linked to the functioning of internal predictive models, might be one way the neuronal architecture for action control of our ancient evolutionary ancestors progressively developed more sophisticated cognitive abilities in continuity with its more basic sensorimotor skills (e.g., planning and mental time travel abilities on top of the systems that supported spatial navigation, Buzsáki and Moser Reference Buzsáki and Moser2013). This hypothesis would also explain why these apparently disconnected abilities have partially shared neuronal circuits (Schacter et al. Reference Schacter, Addis and Buckner2007; Reference Schacter, Addis, Hassabis, Martin, Spreng and Szpunar2012). Clearly, more specific predictions on the similarities and differences between these abilities are required that can be derived from mechanistic models, which could possibly be tested within the neurophysiological framework for neural coalitions and neural reuse discussed in Anderson (Reference Anderson2014).

References

Anderson, M. L. (2014) After phrenology: Neural reuse and the interactive brain. MIT Press.CrossRefGoogle Scholar
Ashby, W. R. (1952) Design for a brain. Wiley.Google Scholar
Attias, H. (2003) Planning by Probabilistic Inference. Proceedings of the Ninth International Conference on Artificial Intelligence and Statistics, Key West, FL. January 2003, ed. Bishop, C. M. & Frey, B. J.. Society for Artificial Intelligence and Statistics.Google Scholar
Barsalou, L. W. (1999) Perceptual symbol systems. Behavioral and Brain Sciences 22(4):577660.CrossRefGoogle ScholarPubMed
Barsalou, L. W. (2008) Grounded cognition. Annual Review of Psychology 59:617–45.Google Scholar
Buzsáki, G. & Moser, E. I. (2013) Memory, navigation and theta rhythm in the hippocampal-entorhinal system. Natural Neuroscience 16:130–38. doi: 10.1038/nn.3304.Google Scholar
Buzsáki, G., Peyrache, A. & Kubie, J. (2015) Emergence of cognition from action. Cold Spring Harbor Symposia on Quantitative Biology 79:4150. doi: 10.1101/sqb.2014.79.024679.CrossRefGoogle Scholar
Chemero, A. (2009) Radical embodied cognitive science. MIT Press.Google Scholar
Cisek, P. (1999) Beyond the computer metaphor: Behaviour as interaction. Journal of Consciousness Studies 6(11–12):125–42.Google Scholar
Cisek, P. (2006) Integrated neural processes for defining potential actions and deciding between them: A computational model. The Journal of Neuroscience 26:9761–70.CrossRefGoogle ScholarPubMed
Cisek, P. & Kalaska, J. F. (2010) Neural mechanisms for interacting with a world full of action choices. Annual Review of Neuroscience 33:269–98.Google Scholar
Clark, A. (1998) Being there. Putting brain, body, and world together. MIT Press.Google Scholar
Clark, A. (2013b) Whatever next? Predictive brains, situated agents, and the future of cognitive science. Behavioral and Brain Sciences 36(3):181204. doi: 10.1017/S0140525X12000477.Google Scholar
Clark, A. & Chalmers, D. J. (1998) The extended mind. Analysis 58:1023.CrossRefGoogle Scholar
Clark, A. & Grush, R. (1999) Towards a cognitive robotics. Adaptive Behavior 7:516.CrossRefGoogle Scholar
Craik, K. (1943) The nature of explanation. Cambridge University Press.Google Scholar
Dehaene, S. (2005) Evolution of human cortical circuits for reading and arithmetic: The “neuronal recycling” hypothesis. In: From monkey brain to human brain: A Fyssen Foundation symposium, ed. Dehaene, S., Duhamel, J.-R., Hauser, M. D. & Rizzolatti, G., pp. 133–58. MIT Press.CrossRefGoogle Scholar
Diba, K. & Buzsáki, G. (2007) Forward and reverse hippocampal place-cell sequences during ripples. Nature Neuroscience 10:1241–42. doi: 10.1038/nn1961.CrossRefGoogle ScholarPubMed
Dragoi, G. & Tonegawa, S. (2011) Preplay of future place cell sequences by hippocampal cellular assemblies. Nature 469:397401.Google Scholar
Engel, A. K., Maye, A., Kurthen, M. & König, P. (2013) Where's the action? The pragmatic turn in cognitive science. Trends in Cognitive Sciences 17:202209.Google Scholar
Friston, K., Daunizeau, J., Kilner, J. & Kiebel, S. J. (2010) Action and behavior: A free-energy formulation. Biological Cybernetics 102:227–60.Google Scholar
Friston, K., Rigoli, F., Ognibene, D., Mathys, C., FitzGerald, T. & Pezzulo, G. (2015) Active inference and epistemic value. Cognitive Neuroscience 6(4):187214.Google Scholar
Friston, K. J., Daunizeau, J., Kilner, J. & Kiebel, S. J. (2010) Action and behavior: A free-energy formulation. Biological Cybernetics 102:227–60. doi: 10.1007/s00422-010-0364-z.Google Scholar
Gallagher, S. (2005) How the body shapes the mind. Oxford University Press.Google Scholar
Gibson, J. J. (1977) The theory of affordances. In: Perceiving, acting, and knowing: Toward an ecological psychology, ed. Shaw, R. & Bransford, J., pp. 6282. Erlbaum.Google Scholar
Glenberg, A. (1997) What memory is for. Behavioral and Brain Sciences 20:155.Google Scholar
Grush, R. (2004) The emulation theory of representation: Motor control, imagery, and perception. Behavioral and Brain Sciences 27:377–96.Google Scholar
Gupta, A. S., van der Meer, M. A., Touretzky, D. S. & Redish, A. D. (2010) Hippocampal replay is not a simple function of experience. Neuron 65:695705.Google Scholar
Jeannerod, M. (2006) Motor cognition. Oxford University Press.Google Scholar
Johnson, A. & Redish, A. D. (2007) Neural ensembles in CA3 transiently encode paths forward of the animal at a decision point. Journal of Neuroscience 27:12176–89.Google Scholar
Kirsh, D. (2010) Thinking with external representations. AI Soc.Google Scholar
Lisman, J. (2015) The challenge of understanding the brain: Where we stand in 2015. Neuron 86:864–82. doi: 10.1016/j.neuron.2015.03.032.Google Scholar
Maye, A. & Engel, A. K. (2011) A discrete computational model of sensorimotor contingencies for object perception and control of behavior. IEEE International Conference on Robotics and Automation (ICRA 2011), Shanghai, China, May 9–13, 2011, pp. 3810–15. IEEE. doi: 10.1109/ICRA.2011.5979919.Google Scholar
O'Regan, K. & Noe, A. (2001) A sensorimotor account of vision and visual consciousness. Behavioral and Brain Sciences 24(5):939–73.Google Scholar
Pezzulo, G. (2011) Grounding procedural and declarative knowledge in sensorimotor anticipation. Mind and Language 26:78114.Google Scholar
Pezzulo, G. (2012) An active inference view of cognitive control. Frontiers in Psychology 3:478. doi: 10.3389/fpsyg.2012.00478.CrossRefGoogle ScholarPubMed
Pezzulo, G., Barsalou, L. W., Cangelosi, A., Fischer, M. H., McRae, K. & Spivey, M. (2011) The mechanics of embodiment: A dialogue on embodiment and computational modeling. Frontiers in Cognition 2:121.Google Scholar
Pezzulo, G., Barsalou, L. W., Cangelosi, A., Fischer, M. H., McRae, K. & Spivey, M. J. (2013) Computational grounded cognition: A new alliance between grounded cognition and computational modeling. Frontiers in Psychology 3:612. doi: 10.3389/fpsyg.2012.00612.Google Scholar
Pezzulo, G. & Castelfranchi, C. (2009) Thinking as the control of imagination: A conceptual framework for goal-directed systems. Psychological Research 73:559–77.Google Scholar
Pezzulo, G., Rigoli, F. & Friston, K. (2015) Active inference, homeostatic regulation and adaptive behavioural control. Progress in Neurobiology 134:1735. doi: 10.1016/j.pneurobio.2015.09.001.Google Scholar
Pezzulo, G., van der Meer, M. A. A., Lansink, C. S. & Pennartz, C. M. A. (2014) Internally generated sequences in learning and executing goal-directed behavior. Trends in Cognitive Sciences 18:647–57. doi: 10.1016/j.tics.2014.06.011.CrossRefGoogle ScholarPubMed
Pfeifer, R. & Scheier, C. (1999) Understanding intelligence. MIT Press.Google Scholar
Pfeiffer, B. E. & Foster, D. J. (2013) Hippocampal place-cell sequences depict future paths to remembered goals. Nature 497:7479. doi: 10.1038/nature12112.Google Scholar
Powers, W. (1973) Behavior, the control of perception. Aldine de Gruyter.Google Scholar
Rosenblueth, A., Wiener, N. & Bigelow, J. (1943) Behavior, purpose and teleology. Philosophy of Science 10(1):1824.Google Scholar
Schacter, D. L., Addis, D. R. & Buckner, R. L. (2007) Remembering the past to imagine the future: The prospective brain. Nature Reviews Neuroscience 8:657–61.Google Scholar
Schacter, D. L., Addis, D. R., Hassabis, D., Martin, V. C., Spreng, R. N. & Szpunar, K. K. (2012) The future of memory: Remembering, imagining, and the brain. Neuron 76:677–94. doi: 10.1016/j.neuron.2012.11.001.Google Scholar
Scott, S. H. (2012) The computational and neural basis of voluntary motor control and planning. Trends in Cognitive Sciences 16:541–49. doi: 10.1016/j.tics.2012.09.008.Google Scholar
Seth, A. K. (2015) The cybernetic Bayesian brain – From interoceptive inference to sensorimotor contingencies. In: Open MIND: 35(T), ed. Metzinger, T. & Windt, J. M.. MIND Group. doi: 10.15502/9783958570108.Google Scholar
Shadmehr, R. & Mussa-Ivaldi, F. A. (2012) Biological learning and control. How the brain builds representations, predicts events and makes decisions. MIT Press.Google Scholar
Spivey, M. (2007) The continuity of mind. Oxford University Press.Google Scholar
Stoianov, I., Genovesio, A. & Pezzulo, G. (2016) Prefrontal goal codes emerge as latent states in probabilistic value learning. Journal of Cognitive Neuroscience 28(1):140–57. doi: 10.1162/jocn_a_00886.CrossRefGoogle ScholarPubMed
Toussaint, M. (2009) Probabilistic inference as a model of planned behavior. Künstliche Intelligenz. 3(9):2329.Google Scholar
Varela, F. J., Thompson, E. T. & Rosch, E. (1992) The embodied mind: Cognitive science and human experience. MIT Press.Google Scholar
Verschure, P., Pennartz, C. M. A. & Pezzulo, G. (2014) The why, what, where, when and how of goal-directed choice: Neuronal and computational principles. Philosophical Transactions of the Royal Society B: Biological Sciences 369:20130483.Google Scholar
Wiener, N. (1948) Cybernetics: Or control and communication in the animal and the machine. MIT Press Google Scholar
Wikenheiser, A. M. & Redish, A. D. (2015) Hippocampal theta sequences reflect current goals. Nature Neuroscience 18:289–94.Google Scholar