Hostname: page-component-745bb68f8f-b95js Total loading time: 0 Render date: 2025-02-06T14:51:01.539Z Has data issue: false hasContentIssue false

The neurobiology of receptive-expressive language interdependence

Published online by Cambridge University Press:  24 June 2013

Anthony Steven Dick
Affiliation:
Department of Psychology, Florida International University, Miami, FL 33199. adick@fiu.eduwww.fiu.edu/~adick
Michael Andric
Affiliation:
Center for Mind/Brain Sciences (CIMeC), The University of Trento, Trento 38122, Italy. michael.andric@unitn.itmichaelandric.tumblr.com

Abstract

With a focus on receptive language, we examine the neurobiological evidence for the interdependence of receptive and expressive language processes. While we agree that there is compelling evidence for such interdependence, we suggest that Pickering & Garrod's (P&G's) account would be enhanced by considering more-specific situations in which their model does, and does not, apply.

Type
Open Peer Commentary
Copyright
Copyright © Cambridge University Press 2013 

The classical Lichtheim–Broca–Wernicke neurobiological model of language proposed distinct neuroanatomical pathways for language comprehension and production. Recent evidence suggests abandoning this model's classical form, and although there is not yet an established replacement (Dick & Tremblay Reference Dick and Tremblay2012; Price Reference Price2010; Reference Price2012 for review), we think much of the data support P&G's proposal. However, we also think P&G could be clearer about whether there are situations in which their model does not apply. For example, they state that “comprehenders make whatever linguistic predictions they can” (target article, sect. 3.2, para. 1), but this is so broad as to be unfalsifiable.

Neurobiological evidence suggests production and perception system interdependence occurs in specific situations. By highlighting emerging models and findings in the neurobiology of receptive language, we suggest that P&G's proposal could be fine-tuned to make more-specific, testable predictions.

Neurobiological evidence for the interdependence of receptive-expressive language in speech perception

The most widely adopted model of language neurobiology is a dual-stream model analogous to the visual system (Ungerleider & Haxby Reference Ungerleider and Haxby1994). Within this model, during receptive language, auditory speech sounds map to articulatory (motor) representations in a dorsal stream and to meaning in a ventral stream (Hickok Reference Hickok2009b; Hickok & Poeppel Reference Hickok and Poeppel2000; Reference Hickok and Poeppel2004; Reference Hickok and Poeppel2007; Rauschecker Reference Rauschecker2011; Rauschecker & Scott Reference Rauschecker and Scott2009; Rauschecker & Tian Reference Rauschecker and Tian2000; Rogalsky & Hickok Reference Rogalsky and Hickok2011). If this is correct, models like P&G's must account for the way these processing streams interact with the motor system involved in language production.

This problem is easier to solve within the dorsal stream, as many of the same brain regions are active during speech planning and execution, and during speech perception (Callan et al. Reference Callan, Jones, Callan and Akahane-Yamada2004; Eickhoff et al. Reference Eickhoff, Heim, Zilles and Amunts2009; Hickok & Poeppel Reference Hickok and Poeppel2007; Pulvermüller et al. Reference Pulvermüller, Huss, Kherif, Moscoso del Prado Martin, Hauk and Shtyrov2006; Vigneau et al. Reference Vigneau, Beaucousin, Hervé, Duffau, Crivello, Houdé, Mazoyer and Tzourio-Mazoyer2006; Wilson et al. Reference Wilson, Saygin, Sereno and Iacoboni2004). In fact, a primary contention is not whether the motor system is recruited during speech perception but in what situations it occurs. Some argue the motor system is essential (D'Ausilio et al. Reference D'Ausilio, Pulvermüller, Salmas, Bufalari, Begliomini and Fadiga2009; Iacoboni Reference Iacoboni2008; Meister et al. Reference Meister, Wilson, Deblieck, Wu and Iacoboni2007), whereas others argue that it is only involved when auditory-only speech is difficult to parse (e.g., during noisy situations, or when discriminating between similar phonemic units; Hickok Reference Hickok2009a; Hickok et al. Reference Hickok, Houde and Rong2011; Sato et al. Reference Sato, Tremblay and Gracco2009; Tremblay & Small Reference Tremblay and Small2011).

The latter situation appears to be the case for audiovisual speech perception, when visual information from the lips and mouth is present. Moreover, a forward-modeling architecture consistent with P&G's proposal has been suggested to explain the neurobiology of audiovisual speech perception (Callan et al. Reference Callan, Jones, Callan and Akahane-Yamada2004; Skipper et al. Reference Skipper, Nusbaum and Small2005; Skipper et al. Reference Skipper, van Wassenhove, Nusbaum and Small2007b; van Wassenhove et al. Reference van Wassenhove, Grant and Poeppel2005; Wilson & Iacoboni Reference Wilson and Iacoboni2006). Here, visual information, temporally preceding the auditory signal by several hundred milliseconds (Chandrasekaran et al. Reference Chandrasekaran, Trubanova, Stillittano, Caplier and Ghazanfar2009), provides a “forward model” of the speech sound. These models draw on the listener's articulatory representations to provide possible phonetic targets of the talker's speech (Callan et al. Reference Callan, Jones, Callan and Akahane-Yamada2004; Skipper et al. Reference Skipper, van Wassenhove, Nusbaum and Small2007b; van Wassenhove et al. Reference van Wassenhove, Grant and Poeppel2005). Findings that visual speech influences the auditory neural response's latency and amplitude (van Wassenhove et al. Reference van Wassenhove, Grant and Poeppel2005), and recruits motor-speech regions (Callan et al. Reference Callan, Jones, Callan and Akahane-Yamada2004; Dick et al. Reference Dick, Solodkin and Small2010; Hasson et al. Reference Hasson, Skipper, Nusbaum and Small2007; Sato et al. Reference Sato, Buccino, Gentilucci and Cattaneo2010; Skipper et al. Reference Skipper, Nusbaum and Small2005; Reference Skipper, van Wassenhove, Nusbaum and Small2007b; Watkins et al. Reference Watkins, Strafella and Paus2003), support predictive coding via forward models of the kind P&G propose.

Neurobiological evidence for the interdependence of receptive-expressive language in language and gesture comprehension

Although the neurobiological evidence for receptive-expressive language interdependence is compelling in speech perception, it is mixed for higher-level language comprehension, which involves brain regions along a ventral language pathway (Binder et al. Reference Binder, Desai, Graves and Conant2009; Hickok & Poeppel Reference Hickok and Poeppel2007; Vigneau et al. Reference Vigneau, Beaucousin, Hervé, Duffau, Crivello, Houdé, Mazoyer and Tzourio-Mazoyer2006). There is evidence – for example, in processing verbs – that the motor system contributes to understanding, and this is cited to support “motor simulation” theories (Cappa & Pulvermüller Reference Cappa and Pulvermüller2012; Fischer & Zwaan Reference Fischer and Zwaan2008; Glenberg Reference Glenberg2011; Glenberg & Gallese Reference Glenberg and Gallese2012). Notably, some authors interpret these findings without adhering to motor simulation theories (Bedny & Caramazza Reference Bedny and Caramazza2011; Mahon & Caramazza, Reference Mahon and Caramazza2009). Indeed, motor (production) system contribution to language comprehension is a contentious issue (e.g., this was a topic of an organized debate at the 2011 Neurobiology of Language Conference).

Additional evidence suggests that involvement of the motor system is specific to the task. For example, Tremblay et al. (Reference Tremblay, Sato and Small2012) applied repetitive transcranial magnetic stimulation (rTMS) to the ventral premotor cortex during a sentence comprehension task. The rTMS interfered with sentences describing manual actions, but not with other types of sentences, suggesting that predictive motor encoding is not always called upon. Another example is gesture comprehension. Some studies have shown that the act of viewing gestures recruits areas associated with a putative “mirror neuron” system thought to covertly simulate others' actions (Green et al. Reference Green, Straube, Weis, Jansen, Willmes, Konrad and Kircher2009; Holle et al. Reference Holle, Gunter, Rüschemeyer, Hennenlotter and Iacoboni2008; Skipper et al. Reference Skipper, Goldin-Meadow, Nusbaum and Small2007a; Reference Skipper, Goldin-Meadow, Nusbaum and Small2009; Willems et al. Reference Willems, Özyürek and Hagoort2007; Xu et al. Reference Xu, Gannon, Emmorey, Smith and Braun2009), but others show no evidence that this correlates with comprehension (Andric & Small Reference Andric and Small2012; Dick et al. Reference Dick, Goldin-Meadow, Hasson, Skipper and Small2009; Reference Dick, Goldin-Meadow, Solodkin and Small2012; Straube et al. Reference Straube, Green, Bromberger and Kircher2011; Willems et al. Reference Willems, Özyürek and Hagoort2009).

In closing, we note that within P&G's model it may not be necessary to elicit motor activation. For example, P&G state that “embodied accounts assume that producers and comprehenders use perceptual and motor representations associated with the meaning of what they are communicating. Our account does not require such embodiment but is compatible with it” (sect. 4, para. 9). Hence, the model seems able to account for motor activity, or lack of it, during receptive language. If this is the case, P&G should clarify what neurobiological findings could help decide between competing accounts that call upon interdependent receptive and expressive language systems.

References

Andric, M. & Small, S. L. (2012) Gesture's neural language. Frontiers in Psychology 3:99. DOI: 10.3389/fpsyg.2012.00099.CrossRefGoogle ScholarPubMed
Bedny, M. & Caramazza, A. (2011) Perception, action, and word meanings in the human brain: The case from action verbs. Annals of the New York Academy of Sciences 1224:8195. DOI:10.1111/j.1749-6632.2011.06013.x.Google Scholar
Binder, J. R., Desai, R. H., Graves, W. W. & Conant, L. L. (2009) Where is the semantic system? A critical review and meta-analysis of 120 functional neuroimaging studies. Cerebral Cortex 19:2767–96.Google Scholar
Callan, D. E., Jones, J. A., Callan, A. M. & Akahane-Yamada, R. (2004) Phonetic perceptual identification by native- and second-language speakers differentially activates brain regions involved with acoustic phonetic processing and those involved with articulatory-auditory/orosensory internal models. NeuroImage 22:1182–94.CrossRefGoogle ScholarPubMed
Cappa, S. F. & Pulvermüller, F. (2012) Cortex special issue: Language and the motor system. Cortex 48(7):785.Google Scholar
Chandrasekaran, C., Trubanova, A., Stillittano, S., Caplier, A. & Ghazanfar, A. A. (2009) The natural statistics of audiovisual speech. PLoS Computational Biology 5(7):e1000436. DOI:10.1371/journal.pcbi.1000436.Google Scholar
D'Ausilio, A., Pulvermüller, F., Salmas, P., Bufalari, I., Begliomini, C. & Fadiga, L. (2009) The motor somatotopy of speech perception. Current Biology 19:381–85.Google Scholar
Dick, A. S., Goldin-Meadow, S., Hasson, U., Skipper, J. I. & Small, S. L. (2009) Co-speech gestures influence neural activity in brain regions associated with processing semantic information. Human Brain Mapping 30:3509–26.Google Scholar
Dick, A. S., Goldin-Meadow, S., Solodkin, A. & Small, S. L. (2012) Gesture in the developing brain. Developmental Science 15:165–80.Google Scholar
Dick, A. S., Solodkin, A. & Small, S. L. (2010) Neural development of networks for audiovisual speech comprehension. Brain and Language 114:101–14.CrossRefGoogle ScholarPubMed
Dick, A. S. & Tremblay, P. (2012) Beyond the arcuate fasciculus: Consensus and controversy in the connectional anatomy of language. Brain: A Journal of Neurology 135:3529–50. doi: 10.1093/brain/aws222.Google Scholar
Eickhoff, S. B., Heim, S., Zilles, K. & Amunts, K. (2009) A systems perspective on the effective connectivity of overt speech production. Philosophical Transactions of the Royal Society A: Mathematical, Physical, and Engineering Sciences 367(1896):2399–421. DOI:10.1098/rsta.2008.0287.CrossRefGoogle ScholarPubMed
Fischer, M. H. & Zwaan, R. A. (2008) Embodied language: A review of the role of the motor system in language comprehension. Quarterly Journal of Experimental Psychology (2006) 61(6):825–50. DOI:10.1080/17470210701623605.CrossRefGoogle Scholar
Glenberg, A. M. (2011) How reading comprehension is embodied and why that matters. International Electronic Journal of Elementary Education 4:518.Google Scholar
Glenberg, A. M. & Gallese, V. (2012) Action-based language: A theory of language acquisition, comprehension, and production. Cortex 48(7):905–22. DOI:10.1016/j.cortex.2011.04.010.CrossRefGoogle Scholar
Green, A., Straube, B., Weis, S., Jansen, A., Willmes, K., Konrad, K. & Kircher, T. (2009) Neural integration of iconic and unrelated coverbal gestures: A functional MRI study. Human Brain Mapping 30:3309–24.Google Scholar
Hasson, U., Skipper, J. I., Nusbaum, H. C. & Small, S. L. (2007) Abstract coding of audiovisual speech: Beyond sensory representation. Neuron 56:1116–26.Google Scholar
Hickok, G. (2009a) Eight problems for the mirror neuron theory of action understanding in monkeys and humans. Journal of Cognitive Neuroscience 21(7):1229–43. DOI: 10.1162/jocn.2009.21189.Google Scholar
Hickok, G. (2009b) The functional neuroanatomy of language. Physics of Life Reviews 6(3):121–43.CrossRefGoogle ScholarPubMed
Hickok, G., Houde, J. & Rong, F. (2011) Sensorimotor integration in speech processing: Computational basis and neural organization. Neuron 69(3):407–22. DOI:10.1016/j.neuron.2011.01.019.Google Scholar
Hickok, G. & Poeppel, D. (2000) Towards a functional neuroanatomy of speech perception. Trends in Cognitive Sciences 4(4):131–38.CrossRefGoogle ScholarPubMed
Hickok, G. & Poeppel, D. (2004) Dorsal and ventral streams: A framework for understanding aspects of the functional anatomy of language. Cognition 92(1–2):6799.CrossRefGoogle ScholarPubMed
Hickok, G. & Poeppel, D. (2007) The cortical organization of speech processing. Nature Reviews Neuroscience 8(5):393402.Google Scholar
Holle, H., Gunter, T. C., Rüschemeyer, S. A., Hennenlotter, A. & Iacoboni, M. (2008) Neural correlates of the processing of co-speech gestures. NeuroImage 39(4):2010–24.Google Scholar
Iacoboni, M. (2008) The role of premotor cortex in speech perception: Evidence from fmri and rtms. Journal of Physiology (Paris), 102:3134.Google Scholar
Mahon, B. Z. & Caramazza, A. (2009) Concepts and categories: A cognitive neuropsychological perspective. Annual Review of Psychology 60:2751. DOI:10.1146/annurev.psych.60.110707.163532.Google Scholar
Meister, I. G., Wilson, S. M., Deblieck, C., Wu, A. D. & Iacoboni, M. (2007) The essential role of premotor cortex in speech perception. Current Biology 17:1692–96.Google Scholar
Price, C. J. (2010) The anatomy of language: A review of 100 fMRI studies published in 2009. Annals of the New York Academy of Sciences 1191(1):6288.Google Scholar
Price, C. J. (2012) A review and synthesis of the first 20 years of PET and fMRI studies of heard speech, spoken language and reading. NeuroImage 62(2):816–47. DOI: 10.1016/j.neuroimage.2012.04.062.Google Scholar
Pulvermüller, F., Huss, M., Kherif, F., Moscoso del Prado Martin, F., Hauk, O., & Shtyrov, Y. (2006) Motor cortex maps articulatory features of speech sounds. Proceedings of the National Academy of Sciences 103(20):7865–70.Google Scholar
Rauschecker, J. P. (2011) An expanded role for the dorsal auditory pathway in sensorimotor control and integration. Hearing Research 271(1–2):1625. DOI:10.1016/j.heares.2010.09.001.Google Scholar
Rauschecker, J. P. & Scott, S. K. (2009) Maps and streams in the auditory cortex: Nonhuman primates illuminate human speech processing. Nature Neuroscience 12(6):718–24. DOI:10.1038/nn.2331.Google Scholar
Rauschecker, J. P. & Tian, B. (2000) Mechanisms and streams for processing of “what” and “where” in auditory cortex. Proceedings of the National Academy of Sciences 97(22):11800.Google Scholar
Rogalsky, C. & Hickok, G. (2011) The role of Broca's area in sentence comprehension. Journal of Cognitive Neuroscience 23(7):1664–80.Google Scholar
Sato, M., Buccino, G., Gentilucci, M. & Cattaneo, L. (2010) On the tip of the tongue: Modulation of the primary motor cortex during audiovisual speech perception. Speech Communication 52(6):533–41.CrossRefGoogle Scholar
Sato, M., Tremblay, P. & Gracco, V. L. (2009) A mediating role of the premotor cortex in phoneme segmentation. Brain and Language 111(1):17. DOI:10.1016/j.bandl.2009.03.002.CrossRefGoogle ScholarPubMed
Skipper, J. I., Goldin-Meadow, S., Nusbaum, H. C. & Small, S. L. (2007a) Speech-associated gestures, Broca's area, and the human mirror system. Brain and Language 101(3):260–77.Google Scholar
Skipper, J. I., Goldin-Meadow, S., Nusbaum, H. C. & Small, S. L. (2009) Gestures orchestrate brain networks for language understanding. Current Biology 19:17.Google Scholar
Skipper, J. I., Nusbaum, H. C. & Small, S. L. (2005) Listening to talking faces: Motor cortical activation during speech perception. NeuroImage 25(1):7689.Google Scholar
Skipper, J. I., van Wassenhove, V., Nusbaum, H.C., & Small, S. L. (2007b) Hearing lips and seeing voices: How cortical areas supporting speech production mediate audiovisual speech perception. Cerebral Cortex 17:2387–99.CrossRefGoogle ScholarPubMed
Straube, B., Green, A., Bromberger, B. & Kircher, T. (2011) The differentiation of iconic and metaphoric gestures: Common and unique integration processes. Human Brain Mapping 32(4):520–33. DOI: 10.1002/hbm.21041.Google Scholar
Tremblay, P., Sato, M. & Small, S. L. (2012) TMS-induced modulation of action sentence priming in the ventral premotor cortex. Neuropsychologia 50(2):319–26. DOI:10.1016/j.neuropsychologia.2011.12.00.Google Scholar
Tremblay, P. & Small, S. L. (2011) On the context-dependent nature of the contribution of the ventral premotor cortex to speech perception, NeuroImage 57(4):1561–71.Google Scholar
Ungerleider, L. G. & Haxby, J. V. (1994) “What” and “where” in the human brain. Current Opinion in Neurobiology 4(2):157–65.Google Scholar
van Wassenhove, V., Grant, K. W. & Poeppel, D. (2005) Visual speech speeds up the neural processing of auditory speech. Procedings of the National Academy of Sciences 102(4):1181–86.CrossRefGoogle ScholarPubMed
Vigneau, M., Beaucousin, V., Hervé, P. Y., Duffau, H., Crivello, F., Houdé, O., Mazoyer, B. & Tzourio-Mazoyer, N. (2006) Meta-analyzing left hemisphere language areas: Phonology, semantics, and sentence processing. NeuroImage 30(4):1414–32. DOI: 10.1016/j.neuroimage.2005.11.002.Google Scholar
Watkins, K., Strafella, A. P. & Paus, T. (2003) Seeing and hearing speech excites the motor system involved in speech production. Neuropsychologia 41:989–94.Google Scholar
Willems, R. M., Özyürek, A. & Hagoort, P. (2007) When language meets action: The neural integration of gesture and speech. Cerebral Cortex 17(10):2322.Google Scholar
Willems, R. M., Özyürek, A. & Hagoort, P. (2009) Differential roles for left inferior frontal and superior temporal cortex in multimodal integration of action and language. NeuroImage 47:19922004.CrossRefGoogle ScholarPubMed
Wilson, S. M. & Iacoboni, M. (2006) Neural responses to non-native phonemes varying in producibility: Evidence for the sensorimotor nature of speech perception. Neuroimage 33(1):316–25.Google Scholar
Wilson, S. M., Saygin, A. P., Sereno, M. I. & Iacoboni, M. (2004) Listening to speech activates motor areas involved in speech production. Nature Neuroscience 7(7):701702.Google Scholar
Xu, J., Gannon, P. J., Emmorey, K., Smith, J. F. & Braun, A. R. (2009) Symbolic gestures and spoken language are processed by a common neural system. Proceedings of the National Academy of Sciences 106(49):20664–69.Google Scholar