Crossref Citations
This article has been cited by the following publications. This list is generated based on data provided by
Crossref.
Fine, Alex B.
Jaeger, T. Florian
Farmer, Thomas A.
Qian, Ting
and
Paterson, Kevin
2013.
Rapid Expectation Adaptation during Syntactic Comprehension.
PLoS ONE,
Vol. 8,
Issue. 10,
p.
e77661.
Wedel, Andrew
Jackson, Scott
and
Kaplan, Abby
2013.
Functional Load and the Lexicon: Evidence that Syntactic Category and Frequency Relationships in Minimal Lemma Pairs Predict the Loss of Phoneme contrasts in Language Change.
Language and Speech,
Vol. 56,
Issue. 3,
p.
395.
Seyfarth, Scott
2014.
Word informativity influences acoustic duration: Effects of contextual predictability on lexical representation.
Cognition,
Vol. 133,
Issue. 1,
p.
140.
Kurumada, Chigusa
and
Jaeger, T. Florian
2015.
Communicative efficiency in language production: Optional case-marking in Japanese.
Journal of Memory and Language,
Vol. 83,
Issue. ,
p.
152.
Brown, Meredith
and
Kuperberg, Gina R.
2015.
A Hierarchical Generative Framework of Language Processing: Linking Language Perception, Interpretation, and Production Abnormalities in Schizophrenia.
Frontiers in Human Neuroscience,
Vol. 9,
Issue. ,
Pate, John K
and
Goldwater, Sharon
2015.
Talkers account for listener and channel characteristics to communicate efficiently.
Journal of Memory and Language,
Vol. 78,
Issue. ,
p.
1.
Dell, Gary S.
and
Jacobs, Cassandra L.
2016.
Neurobiology of Language.
p.
209.
Kuperberg, Gina R.
and
Jaeger, T. Florian
2016.
What do we mean by prediction in language comprehension?.
Language, Cognition and Neuroscience,
Vol. 31,
Issue. 1,
p.
32.
NORCLIFFE, ELISABETH
and
JAEGER, T. FLORIAN
2016.
Predicting head-marking variability in Yucatec Maya relative clause production.
Language and Cognition,
Vol. 8,
Issue. 2,
p.
167.
Buz, Esteban
Tanenhaus, Michael K.
and
Jaeger, T. Florian
2016.
Dynamically adapted context-specific hyper-articulation: Feedback from interlocutors affects speakers’ subsequent pronunciations.
Journal of Memory and Language,
Vol. 89,
Issue. ,
p.
68.
Jaeger, T. Florian
and
Buz, Esteban
2017.
The Handbook of Psycholinguistics.
p.
38.
Hall, Kathleen Currie
Hume, Elizabeth
Jaeger, T. Florian
and
Wedel, Andrew
2018.
The role of predictability in shaping phonological patterns.
Linguistics Vanguard,
Vol. 4,
Issue. s2,
Larsen-Freeman, Diane
2018.
Learning Language through Task Repetition.
Vol. 11,
Issue. ,
p.
311.
Ferreira, Victor S.
2019.
A Mechanistic Framework for Explaining Audience Design in Language Production.
Annual Review of Psychology,
Vol. 70,
Issue. 1,
p.
29.
Long, Madeleine
Moore, Isabelle
Mollica, Francis
and
Rubio-Fernandez, Paula
2021.
Contrast perception as a visual heuristic in the formulation of referential expressions.
Cognition,
Vol. 217,
Issue. ,
p.
104879.
Kuperberg, Gina R.
2021.
Tea With Milk? A Hierarchical Generative Framework of Sequential Event Comprehension.
Topics in Cognitive Science,
Vol. 13,
Issue. 1,
p.
256.
Rubio-Fernandez, Paula
2021.
Color discriminability makes over-specification efficient: Theoretical analysis and empirical evidence.
Humanities and Social Sciences Communications,
Vol. 8,
Issue. 1,
Morgan, Adam M.
and
Ferreira, Victor S.
2022.
Still no evidence for audience design in syntax: Resumptive pronouns are not the exception.
Journal of Memory and Language,
Vol. 127,
Issue. ,
p.
104368.
Pickering & Garrod (P&G) are not the first to propose that comprehension is a predictive process (e.g., Hale Reference Hale2001; Levy Reference Levy2008; Ramscar et al. Reference Ramscar, Yarlett, Dye, Denny and Thorpe2010). Similarly, recent work has found that language production is sensitive to prediction in ways closely resembling comprehension (e.g., Aylett & Turk Reference Aylett and Turk2004; Jaeger Reference Jaeger2010). We believe that forward models (1) offer an elegant account of prediction effects and (2) provide a framework that could generate novel predictions and guide future work. However, in our view, the proposal by P&G fails to advance either goal because it does not take into account two important properties of forward models. The first is learning; the second is the nature of the prediction error that the forward model is minimizing.
Learning
Forward models have been a successful framework for motor control in large part because they provide a unifying framework, not only for prediction, but also for learning. Since their inception, forward models have been used to study learning–both acquisition and adaptation throughout life. However, except for a brief mention of “tuning” (target article, sect. 3.1, para. 15), P&G do not discuss what predictions their framework makes for implicit learning during language production, despite the fact that construing language processing as prediction in the context of learning readily explains otherwise puzzling findings from production (e.g., Roche et al. Reference Roche, Dale, Kreuz and Jaeger2013; Warker & Dell Reference Warker and Dell2006), comprehension (e.g., Clayards et al. Reference Clayards, Tanenhaus, Aslin and Jacobs2008; Farmer et al. Reference Farmer, Brown and Tanenhaus2013; Kleinschmidt et al. Reference Kleinschmidt, Fine and Jaeger2012) and acquisition (Ramscar et al. Reference Ramscar, Yarlett, Dye, Denny and Thorpe2010). If connected to learning, forward models can explain how we learn to align our predictions during dialogue (i.e., learning in order to reduce future prediction errors, Fine et al., submitted; Jaeger & Snider Reference Jaeger and Snider2013; for related ideas, see also Chang et al. Reference Chang, Dell and Bock2006; Fine & Jaeger Reference Fine and Jaeger2013; Kleinschmidt & Jaeger Reference Kleinschmidt and Jaeger2011; Sonderegger & Yu Reference Sonderegger and Yu2010).
Prediction errors
Deriving testable predictions from forward models is integrally tied to the nature of the prediction error that the system is meant to minimize during self- and other-monitoring (i.e., the function of the model, cf. Guenther et al. Reference Guenther, Hampson and Johnson1998). P&G do not explicitly address this. They do, however, propose separate forward models at all levels of linguistic representations. These forward models seem to have just one function, to predict the perceived linguistic unit at each level. For example, the syntactic forward model predicts the “syntactic percept,” which is used to decide whether the production plan needs to be adjusted (how this comparison proceeds and what determines its outcome is left unspecified).
Minimizing communication error: A proposal
If one of the goals of language production is to be understood–or even to communicate the intended message both robustly and efficiently (Jaeger Reference Jaeger2010; Lindblom Reference Lindblom1990)–correctly predicting the intended linguistic units should only be relevant to the extent that not doing so impedes being understood. Therefore, the prediction error that forward models in production should aim to minimize is not the perception of linguistic units, but the outcome of the entire inference process that constitutes comprehension. Support for this alternative view comes from work on motor control, work on articulation, and cross-linguistic properties of language.
For example, if the speaker produces an underspecified referential expression but is understood, there is no need to self-correct (as observed in research on conceptual pacts, Brennan & Clark Reference Brennan and Clark1996). This view would explain why only reductions of words with low confusability tend to enter the lexicon (e.g., “strodny,” rather than “extrary,” for “extraordinary”). If, however, the function of the forward model is to predict linguistic units, as P&G propose, no such generalization is expected. Rather, any deviation from the target phonology will cause a prediction error, regardless of whether it affects the likelihood of being understood. Similar reasoning applies to the reduction of morphosyntactic units, which often is blocked when it would cause systemic ambiguity (e.g., differential or optional case-marking, Fedzechkina et al. Reference Fedzechkina, Jaeger and Newport2012; see also Ferreira Reference Ferreira2008).
Research on motor control finds that not all prediction errors are created equal: Stronger adaptation effects are found after task-relevant errors (Wei & Körding Reference Wei and Körding2009). Indeed, in a recent perturbation study on production, Frank (Reference Frank2011) found that speakers exhibit stronger error correction if the perceived deviation from the intended acoustics makes the actual production more similar to an existing word (see also Perkell et al. Reference Perkell, Guenther, Lane, Matthies, Stockmann, Tiede and Zandipour2004).
This view also addresses another shortcoming of P&G's proposal. At several points, P&G state that the forward models make impoverished predictions. Perhaps predictions are impoverished only in that they map the efference copy directly onto the predicted meaning (rather than the intermediate linguistic units).
Of course, the goal of reducing the prediction error for efficient information transfer is achieved by reducing the prediction error at the levels assumed by P&G. In this case, the architecture assumed by P&G would follow from the more general principle described here. However, in a truly predictive learning framework (Clark Reference Clark2013), there is no guarantee that the levels of representation that such models would learn in order to minimize prediction errors would neatly map onto those traditionally assumed (cf. Baayen et al. Reference Baayen, Milin, Filipović Durdević, Hendrix and Marelli2011).
Finally, we note that, in the architecture proposed by P&G, the production forward model seems to serve no purpose but to be the input of the comprehension forward model (sect. 3.1, Fig. 5; sect. 3.2, Fig. 6). Presumably, the output of, for example, the syntactic production forward model will be a syntactic plan. Hence, the syntactic comprehension forward model takes syntactic plans as input. The output of that comprehension forward model must be akin to a parse, as it is compared to the output of the actual comprehension model. Neither of these components seems to fulfill any independent purpose. Why not map straight from the syntactic efference copy to the predicted “syntactic percept”? If forward models are used as a computational framework, rather than as metaphor, one of their strengths is that they can map efference copies directly onto the reference frame that is required for effective learning and minimization of the relevant prediction error (cf. Guenther et al. Reference Guenther, Hampson and Johnson1998).