Hostname: page-component-745bb68f8f-l4dxg Total loading time: 0 Render date: 2025-02-12T02:40:30.107Z Has data issue: false hasContentIssue false

Generative models as parsimonious descriptions of sensorimotor loops

Published online by Cambridge University Press:  28 November 2019

Manuel Baltieri
Affiliation:
EASY Group – Sussex Neuroscience, Department of Informatics, University of Sussex, BrightonBN1 9RH, United Kingdom. m.baltieri@sussex.ac.ukc.l.buckley@sussex.ac.ukhttps://manuelbaltieri.comhttps://christopherlbuckley.com/
Christopher L. Buckley
Affiliation:
EASY Group – Sussex Neuroscience, Department of Informatics, University of Sussex, BrightonBN1 9RH, United Kingdom. m.baltieri@sussex.ac.ukc.l.buckley@sussex.ac.ukhttps://manuelbaltieri.comhttps://christopherlbuckley.com/

Abstract

The Bayesian brain hypothesis, predictive processing, and variational free energy minimisation are typically used to describe perceptual processes based on accurate generative models of the world. However, generative models need not be veridical representations of the environment. We suggest that they can (and should) be used to describe sensorimotor relationships relevant for behaviour rather than precise accounts of the world.

Type
Open Peer Commentary
Copyright
Copyright © Cambridge University Press 2019

In the target article, Brette questions the use of the neural coding metaphor in the neurosciences. One of the main arguments is related to the criticism of approaches that overemphasise the role of perception as opposed to motor control for accounts of cognition. As suggested by Brette, while the sense-model-plan-act paradigm has long been criticised (Brooks Reference Brooks1991b), it still survives in modern approaches to neuroscience. He then extends this criticism to the notion of efficient coding and its most recent heir, predictive coding. Specifically, he argues that models that describe perception as a process of minimising redundancy (efficient coding) or prediction error (predictive coding) between incoming sensations and a generative model of sensory data are often used to ascribe the “goal” of minimising redundancy/error to an agent, to account for the rich repertoire of behaviours of living organisms. The shortcomings of this approach have previously been discussed, and resulted in the emergence of problems such as the dark-room paradox (Friston et al. Reference Friston, Thornton and Clark2012): why should agents show complex behaviour when, to minimise redundancy/prediction error, they could easily find the most predictable state in the world, for example, a dark room?

Recently, the ideas of predictive coding/processing have been extended to include accounts of action, for example, active inference. On this account, while perception is a process of changing predictions to better account for sensory data, action is a process of changing the world to better meet predictions. Normative behaviour then arises in this framework if an agent predicts rewarding states (Friston et al. Reference Friston, Thornton and Clark2012). Active inference moves the goal of cognitive agents from inferring properties of sensory data to acting in order to meet their goals. This extension of predictive coding is thus not in conflict with the ideas in the target article but, rather, directly supports them.

Furthermore, adaptive behaviour can emerge even if generative models are far divorced from a veridical representation of the environment. These action-oriented generative models, described in the context of radical predictive processing (Clark Reference Clark2015), can operate on the basis of linear approximations of world dynamics (Baltieri & Buckley Reference Baltieri and Buckley2019b) or simple sensorimotor couplings rather than objective properties of the world (Baltieri & Buckley Reference Baltieri and Buckley2017). These ideas are derived from 4E (embodied, enactive, embedded, and extended) theories of cognition which have long sought to address different issues of computationalism (Newen et al. Reference Newen, De Bruin and Gallagher2018), including the misuse of properties metaphors such as neural coding.

Thus, the idea of predictive processing/coding that the author describes, based on models that generate accurate representations of observed data and inherited from statistics and machine learning, is not the only game in town. Instead, we advocate for ideas of predictive coding that include generative models that are parsimonious descriptions of sensorimotor contingencies (Baltieri & Buckley Reference Baltieri and Buckley2017). This may sound like an unnecessary stretch of the definition of a generative model (Bruineberg et al. Reference Bruineberg, Kiverstein and Rietveld2018) but, we argue, is far from being just a semantic argument. The mathematical definition of these models is still entirely consistent with the more familiar notion of “generative” models and constitutes a valuable framework for the modelling of sensorimotor loops because of its strong and established relationships with (optimal) control theory (Todorov Reference Todorov and Gazzaniga2009). In this context, “inference” can be best understood as a process of estimating actions necessary to attain future goals, generating expectations of desired states of affairs (rather than objective truths about the world) that are brought into existence by means of active behaviour.

In active inference, generative models further diverge in some fundamental ways from the more traditional ideal-observer-based forward models used in the context of motor neuroscience (Friston Reference Friston2011) and discussed in the target article. Forward models rely largely on a Kalman-like approach where all the variables affecting a system (parameters and inputs or causes, in a state-space formulation sense) and observations of said system (outputs) are available for an ideal observer to infer is latent states. On the other hand, generative models in active inference explicitly take the perspective of an agent into account. This includes discarding the idea that all causes affecting observations are known to an agent, suggesting instead the presence of approximate mechanisms to implement actions purely based on incoming sensations (not their estimates or “predictions”) (Baltieri & Buckley Reference Baltieri and Buckley2019a).

The target article further argues that theories based on Shannon communication theory (including efficient and predictive coding, but more in general all notions of neural coding) cannot in principle explain meaning in biological systems, as this definition explicitly precludes a study of the semantics of information. We agree with Brette that notions of semantic information (with meaning for an agent, not necessarily for an experimenter) are largely neglected in neuroscience but his claim seems to overlook some of the efforts made to extend Shannon's work (Kolchinsky & Wolpert Reference Kolchinsky and Wolpert2018). To understand the implications of the value of information in biological systems, we suggest not to ignore frameworks based on Shannon information, but rather to further look into their connections to causal frameworks. In this light, extensions of predictive coding models such as active inference based on Bayesian networks could, in principle, describe biological agents as intervening in the world using motor actions to learn causal relationships between their actions and their sensations (Hohwy Reference Hohwy2013), that is, the “subjective physics of the world” described by Brette.

The neural coding metaphor has long exhausted its appeal to explain cognition. This is because metaphors that attempt to describe living organisms as passively gathering and representing information from the environment cannot account for the complexity of the behaviour of biological systems. Similarly, we believe that a narrow interpretation of generative models as veridical world models is also a misdirection for cognitive science. Instead, we argue that the generative models at the heart of active inference, which are not to be seen as accurate maps of the world but as descriptions of useful information and desires for an agent in the form of parsimonious actions/percepts relationships, will be a valuable tool for the study of cognitive systems.

Acknowledgments

This work was supported in part by BBSRC Grant BB/P022197/1. MB thanks Martin Biehl and Simon McGregor for insightful discussions.

References

Baltieri, M. & Buckley, C. L. (2017) An active inference implementation of phototaxis. In: CAL 2017: The Fourteenth European Conference on Artificial Life. Artificial Life Conference Proceedings 14:36–43.Google Scholar
Baltieri, M. & Buckley, C. L. (2019a) Nonmodular architectures of cognitive systems based on active inference. arXiv preprint arXiv:1903.09542. DOI: 10.1109/IJCNN.2019.8852048CrossRefGoogle Scholar
Baltieri, M. & Buckley, C. L. (2019b) PID Control as a process of active inference with linear generative models. Entropy 21(3):257.CrossRefGoogle Scholar
Brooks, R. A. (1991b). New approaches to robotics. Science 253(5025):1227–32.CrossRefGoogle Scholar
Bruineberg, J., Kiverstein, J. & Rietveld, E. (2018) The anticipating brain is not a scientist: The free-energy principle from an ecological-enactive perspective. Synthese 195(6):2417–44.CrossRefGoogle Scholar
Clark, A. (2015) Surfing uncertainty: Prediction, action, and the embodied mind. Oxford University Press.Google Scholar
Friston, K. (2011) What is optimal about motor control? Neuron 72(3):488–98.CrossRefGoogle ScholarPubMed
Friston, K., Thornton, C. & Clark, A. (2012) Free-energy minimization and the dark-room problem. Frontiers in Psychology 3:130.CrossRefGoogle Scholar
Hohwy, J. (2013) The predictive mind, Oxford University Press.CrossRefGoogle Scholar
Kolchinsky, A. & Wolpert, D. H. (2018) Semantic information, autonomous agency and non-equilibrium statistical physics. Interface Focus 8(6):20180041.CrossRefGoogle Scholar
Newen, A., De Bruin, L. & Gallagher, S., eds. (2018) The Oxford handbook of 4E cognition. Oxford University Press.CrossRefGoogle Scholar
Todorov, E. (2009) Parallels between sensory and motor information processing. In: The Cognitive Neurosciences (4th edition), ed. Gazzaniga, M. S., pp. 613–24. MIT Press.Google Scholar