The importance of inferencing skills
To make sense of language, children must make inferences. For example, they may need to infer what a pronoun (e.g., it) refers to, or why a protagonist in a story acted in a certain way, based on information distributed through the discourse. Inferencing skills are crucial for language comprehension because speakers and writers leave much of the content of their messages implicit. For example, on hearing that a character entered a room and turned on her torch, good comprehenders readily infer that the room was dark. In this case the missing information is provided by general world knowledge, which is integrated with information from the discourse as it unfolds. In this study, we evaluate the efficacy of a novel language intervention intended to promote four-year-olds’ oral inferencing skills. This parent-delivered intervention was designed to prompt children's inferential thinking by giving them practice in answering inferencing questions during shared book-reading.
Inferencing skills allow comprehenders to construct a full and accurate representation of texts by linking events and working out causes and consequences of actions to create a coherent mental representation. Without good inferencing skills that draw from knowledge removed from the here-and-now to fill in implicit information, we cannot make sense of extended discourses such as narratives or instructions. When children start school, they face a sharp increase in the amount and range of the decontextualized language they hear (Hindman, Connor, Jewkes, & Morrison, Reference Hindman, Connor, Jewkes and Morrison2008; Rowe, Reference Rowe2013), meaning that inferencing skills are in greater demand. Given the importance of good oral language at school, improving inferential language during the preschool years is likely to benefit school readiness when children start formal primary education. Strong inferencing skills across oral and written modalities can help enable a child to fully access the curriculum. More broadly, current educational policy emphasises the need for greater language comprehension skills (Law et al., Reference Law, Charlton, Dockrell, Gascoigne, McKean and Theakston2017; Oxford University Press, 2018), so inferencing remains a priority in primary education.
As children progress through the primary school years, inferencing becomes particularly important for reading comprehension and for related academic success. Indeed, much of the literature on inference-making comes from studies of reading comprehension (e.g., Cain & Oakhill, Reference Cain and Oakhill1999; McGee & Johnson, Reference McGee and Johnson2003; Silva & Cain, Reference Silva and Cain2015; Yuill & Oakhill, Reference Yuill and Oakhill1988; see Elleman, Reference Elleman2017, for a meta-analytic review). Several studies have found that children with poor reading comprehension are less likely to make inferences when reading than those with good comprehension (Cain & Oakhill, Reference Cain and Oakhill1999; Oakhill, Reference Oakhill1984), and a range of approaches including classroom-intervention and individual differences methodologies have reinforced the link between inferencing and reading comprehension. Text-based inference training has been effective in enhancing comprehenders’ reading abilities (Bos, De Koning, Wassenburg, & van der Schoot, Reference Bos, De Koning, Wassenburg and van der Schoot2016; McGee & Johnson, Reference McGee and Johnson2003; Yuill & Oakhill, Reference Yuill and Oakhill1988; though note that these studies are not randomised controlled trials), and latent inferencing skill has been found to predict reading comprehension (Cain & Oakhill, Reference Cain and Oakhill1999; Oakhill & Cain, Reference Oakhill and Cain2012; Silva & Cain, Reference Silva and Cain2015). More broadly, higher-level comprehension processes including inferencing account for unique variance in reading comprehension (Language and Reading Research Consortium [LARRC], 2017). Thus, there is a good evidence base showing that good text-based inferencing abilities provide a firm basis for later reading success.
Training inferencing skills
Although several studies have shown that it is possible to train inferencing in school-aged children to improve reading comprehension (Bos et al., Reference Bos, De Koning, Wassenburg and van der Schoot2016; Clarke, Snowling, Truelove, & Hulme, Reference Clarke, Snowling, Truelove and Hulme2010; McGee & Johnson, Reference McGee and Johnson2003, Yuill & Oakhill, Reference Yuill and Oakhill1988), very little is known about whether and how inferencing can be supported earlier on. Given the strong link between inferencing and reading skills, oral inferencing in the preschool years should help with the later demands of formal literacy education. Inferencing practice may also provide early protection to children at risk of becoming poor comprehenders, since a proportion of children may start school at risk of reading difficulties “not because they have problems with decoding or literal comprehension (although they may have these difficulties, too), but because they have not had extensive exposure to text inferencing that supports later, higher levels of literacy” (van Kleeck, Reference van Kleeck and van Kleeck2006, p. 279). Although there have been many arguments in favour of promoting language skills in the preschool years, and successful interventions for doing so (e.g., Burgoyne, Gardner, Whiteley, Snowling, & Hulme, Reference Burgoyne, Gardner, Whiteley, Snowling and Hulme2018), and although inferencing skills for oral language have been monitored in the preschool years (e.g., Das Gupta & Bryant, Reference Das Gupta and Bryant1989; Filiatrault-Veilleux, Bouchard, Trudeau, & Desmarais, Reference Filiatrault-Veilleux, Bouchard, Trudeau and Desmarais2016; Pyykkönen, Matthews, & Järvikivi, Reference Pyykkönen, Matthews and Järvikivi2010; Schulze, Grassmann, & Tomasello, Reference Schulze, Grassmann and Tomasello2013), how those skills can be strengthened before children learn to read is currently unknown.
In a cross-sectional study with four- to six-year-olds, Florit, Roch, and Levorato (Reference Florit, Roch and Levorato2011) found that inferencing skills play a specific role in oral language comprehension. In preschoolers, only three studies to our knowledge have explored whether it is possible to train inferencing skills – one educational and two clinical. First, in a 3-year, quasi-experimental (i.e., non-randomised) study beginning when a large sample of children were almost four years of age, Bianco et al. (Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010) found improved oral comprehension as a result of regular, long-term, explicit, well-defined, comprehension-focused activities including inferencing. However, this study in preschools had a broader focus on comprehension skills more generally and thus we do not yet know which activities specifically supported inference-making. Second, in an 8-week oral inferencing training programme with preschoolers with language impairment, van Kleeck, Vander Woude, and Hammett (Reference van Kleeck, Vander Woude and Hammett2006) reported that their training group outperformed non-intervention controls on receptive and inferential language (though with a small sample size of 15 children in each of the two groups). More recently, in a small-group book-sharing intervention with Australian pre-primary five- to six-year-olds with developmental language disorder, a randomised controlled trial found that children who had undergone oral inferential comprehension training (n = 19) showed an increase in inferential comprehension scores immediately after the 8-week intervention, maintained 8 weeks later. This group also scored higher than the control group for inferential comprehension on a post-intervention assessment of their ability to generalise inferential skills to new narrative contexts (Dawes, Leitão, Claessen, & Kane, Reference Dawes, Leitão, Claessen and Kane2019).
Although these three studies provide tentative evidence that building inferencing skills can improve oral language comprehension, and that inferencing ability can be trained under certain conditions during the preschool years, it is difficult to draw conclusions due to the diverse nature of the populations and the methodologies used. Until now, no study has investigated whether focused practice in inferencing, delivered as part of typically developing preschoolers’ regular activities at home, will lead to improved inferencing skill in a large and diverse sample. Evidence from such a study would have clear implications for the way that inferencing is supported in the preschool years.
The studies that have found improvements in inferencing ability in school-aged children have used a wide range of instruction methods, from explicit teaching (Bos et al., Reference Bos, De Koning, Wassenburg and van der Schoot2016; Clarke et al., Reference Clarke, Snowling, Truelove and Hulme2010; McGee & Johnson, Reference McGee and Johnson2003) to more implicit practice, e.g., asking comprehension questions about texts and allowing children to naturally discuss their answers with their peers (Yuill & Oakhill, Reference Yuill and Oakhill1988). In line with the literature, we define explicit vs. implicit instructional methods, respectively, as (i) guided activities that focus a child's attention explicitly on the pieces of information required for making an inference, and on the process of integrating them; and (ii) activities that elicit inferencing processes from the child incidentally through comprehension questions (Connor, Morrison, & Katch, Reference Connor, Morrison and Katch2004; Snow, Reference Snow2001). As defined, the intervention reported in the current study uses implicit methods.
As we turn our focus to inferencing training in preschoolers, the range of suitable training methods narrows, since many forms of explicit instruction require an explicit understanding of the components of the skill being taught, e.g., the separability of discrete chunks of information, or of information sources. However, a wide range of implicit methods remains open for this age range: it has been suggested that younger children would particularly benefit from supportive dyadic contexts for inferencing, where they are encouraged to demonstrate their inferencing abilities via narratives rather than undergoing formal question-and-answer tests (van Kleeck, Reference van Kleeck and van Kleeck2006, p. 292). Similarly, van Kleeck (Reference van Kleeck2008) has suggested that one of the best ways to promote inferencing ability in younger children is to give them practice in making inferences by responding to questions about a story and then discussing answers. To our knowledge, these recommendations have yet to be taken up in a rigorous trial. Here we explore the value of parent–child book-reading as a basis for this kind of practice. Specifically, we test whether practising making inferences in order to respond to caregiver questions during shared book-reading promotes four-year-olds’ inferencing ability.
Intervention approach
Strengthening inferencing skills in the early years is likely to have advantages for oral language comprehension and later reading ability. Despite evidence showing that inferences can be trained in school-aged children using a range of methods, from answering inferencing questions to formal explicit teaching, we do not yet know: (i) whether training inferencing in the preschool years is possible; and (ii) if so, whether it is possible in this age group via implicit methods. This is particularly important, since formal, explicit instruction methods rely on an understanding of the subcomponents of inferencing so may not be easily accessible for this age group. To address this gap, we report the results of a randomised controlled trial (RCT) to test the effect of a parent-delivered intervention that gave four-year-old children practice in responding to inferencing questions during book sharing. To provide the best chance of success, our design combined elements of successful inferencing interventions for older children with recommendations for scaffolding inferencing in preschoolers. It follows Yuill and Oakhill's (Reference Yuill and Oakhill1988) finding that comprehension questions improve inferencing ability, as well as van Kleeck's (Reference van Kleeck2008) evidence-based recommendations for fostering inferential language in preschoolers, e.g., embedding scripted inferencing questions, prompts, and feedback in shared reading materials (Ard & Beverly, Reference Ard and Beverly2004; Karweit, Reference Karweit1989; van Kleeck et al., Reference van Kleeck and van Kleeck2006). Inserting these prompts ahead of time increases the amount of ‘thinking aloud’ between dyads (Kucan & Beck, Reference Kucan and Beck1997), and improves fidelity.
Our intervention was designed to prompt younger children in their inferential thinking. Although preschoolers are able to engage in inferencing, evidence suggests that they are less likely than their older peers to do so spontaneously (Florit et al., Reference Florit, Roch and Levorato2011). Through naturalistic questioning (based on evidence showing that some parents naturally engage in literal and inferential talk during shared book-reading; Hammett, van Kleeck, & Huberty, Reference Hammett, van Kleeck and Huberty2003), our training highlighted the fact that there is information to be had that is not explicitly stated, and encouraged children to fill in the gaps using clues provided in the text or from their prior knowledge. By raising their awareness of these gaps, children were alerted that an inference needed to be made, and encouraged to strive for coherence (Cain & Oakhill, Reference Cain and Oakhill1999, p. 501). Further, unlike studies involving classroom-based, group training sessions that use explicit training methods to highlight textual cues to implicit meaning (Bianco et al., Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010; McGee & Johnson, Reference McGee and Johnson2003; Yuill & Oakhill, Reference Yuill and Oakhill1988; Zucker, Justice, Piasta, & Kaderavek, Reference Zucker, Justice, Piasta and Kaderavek2010), the intervention was run at home by parents, meaning that, if successful, the programme could be adopted without the need for specialist training.
Shared book-reading was chosen as the medium for the intervention for several reasons. Children who read regularly with an adult in the preschool years learn language faster, enter school with a larger vocabulary, and become more successful readers in school (Mol, Bus, de Jong, & Smeets, Reference Mol, Bus, de Jong and Smeets2008). Shared book-reading facilitates more complex talk than traditional caretaking or play activities (Snow, Reference Snow, Damon and Daiute1993), and exposes children to vocabulary and syntactic structures beyond what they would hear in everyday speech (Cameron-Faulkner & Noble, Reference Cameron-Faulkner and Noble2013; Hoff-Ginsberg, Reference Hoff-Ginsberg1991). Thus, shared book-reading is a potentially powerful tool for supporting the development of vocabulary, narrative and conversational skills, complex syntax, and other literacy practices such as print and phonological awareness (Burgess, Reference Burgess2010; Ezell & Justice, Reference Ezell and Justice2000). More specifically, our intervention asked open-ended questions; a technique from dialogic reading interventions. Dialogic reading encourages caregivers to be more responsive to the child during shared book-reading, and in general has been shown to have a positive impact on a child's oral language development (Baker & Nelson, Reference Baker and Nelson1984; Cleave, Becker, Curran, Van Horne, & Fey, Reference Cleave, Becker, Curran, Van Horne and Fey2015; Farrar, Reference Farrar1990; Girolametto & Weitzman, Reference Girolametto and Weitzman2002; Huttenlocher, Waterfall, Vasilyeva, Vevea, & Hedges, Reference Huttenlocher, Waterfall, Vasilyeva, Vevea and Hedges2010; Valdez-Menchaca & Whitehurst, Reference Valdez-Menchaca and Whitehurst1992; though see a meta-analysis by Noble et al. (Reference Noble, Sala, Peter, Lingwood, Rowland, Gobet and Pine2019), and a randomised controlled trial by Noble et al. (unpublished observations) for evidence of no effect or small effects of dialogic reading on children's language skills). Most pertinently for our intervention, shared book-reading is a good medium for linking social conversations (e.g., about personal events and real-world knowledge) and text inferencing skills for two reasons. First, because some caregivers naturally ask their children questions about the shared story that require them to make inferences about the text, they model the kinds of information that support text comprehension, and then support the child in answering via various types of scaffolding (van Kleeck, Reference van Kleeck and van Kleeck2006, Reference van Kleeck2008). Second, oral inferencing practice is particularly suited to shared book-reading because it takes place within the same activity that it will later be applied in when reading, i.e., generating meaning from information presented in books (van Kleeck, Reference van Kleeck and van Kleeck2006, p. 275). Thus, we use shared book-reading as an activity that will provide the natural apprenticeship for later independent inferencing.
Design and hypotheses
The aim of this RCT was to test whether training parents to ask their children inference-eliciting questions during shared book-reading (and supporting them to do so with in-text questions) is effective for promoting inferencing ability in four-year-olds. The primary outcome measure was children's ability to answer inferencing questions after completing the 4-week intervention (controlling for baseline ability). The inferencing training group was compared with an active control group of children who spent the intervention period working through a maths exercise book with their caregiver. We hypothesised that the training group would make significantly greater gains in inferencing ability than the control group. The secondary outcome measure was the change in children's NFER Baseline Reception Assessment Language and Communication scale (National Foundation for Educational Research, 2015) (NFERL); a standardised assessment frequently used in British primary schools to gauge children's language ability upon school entry (aged four to five years). We did not have a hypothesis regarding potential effects on the NFERL scale as transfer is not often seen in response to cognitive training programmes (Sala & Gobet, Reference Sala and Gobet2017), but we were interested to assess this all the same.
MethodFootnote 1
This educational intervention was preregistered at <https://clinicaltrials.gov/ct2/home> (NCT02854462, ‘Appendix A’). Ethical approval was granted by the Psychology Ethics sub-committee at the University of Sheffield.
Participants
The Consolidated Standards of Reporting Trials (CONSORT) diagram is reported in Figure 1 and the checklist appears as ‘Appendix B’ (Schulz, Altman, & Moher, Reference Schulz, Altman and Moher2010). One hundred four-year-olds (53 female) were recruited in the north of England from a volunteer database at the University of Sheffield's Department of Psychology (Mean age at Baseline = 50.3 months; Median = 50 months: Range: 48 to 56 months; Mean age at Post-test = 51.5 months; Median = 51 months: Range: 49 to 58 months). Eighty-three caregivers and their children had previously taken part in a separate randomised controlled trial investigating the role of caregiver contingent talk on early language development (McGillion, Pine, Herbert, & Matthews, Reference McGillion, Pine, Herbert and Matthews2017). These children did not differ on any measures collected at baseline from those who had not been involved in the previous study (n = 17). Participants were specifically recruited to be representative of the UK population in terms of SES: forty-five percent of households were not educated to degree level. Eighty-nine caregivers gave permission for their data to be uploaded to the UK Data Archive (UK Data Service. 10.5255/UKDA-SN-853233).
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200327171122418-0547:S0305000919000801:S0305000919000801_fig1.png?pub-status=live)
Figure 1. CONSORT flow diagram.
Inclusion criteria
Children were first-born (to control for potential birth-order effects), full-term (i.e., born no more than 3 weeks prematurely), with birth weight over 2.5 kg, and were monolingual English speakers (to allow for the administration of standardised language assessments). Exclusion criterion: neither caregivers nor children had any significant known physical, mental, or learning disability.
At baseline visits, families were given a cuddly toy and the materials required to complete the intervention, and a second cuddly toy and a £40 gift voucher on completing the post-test visit.
Materials
Intervention videos
A short video was used to deliver the Inferencing Training Intervention to caregivers. The script was developed by the authors to explain in lay terms what inferencing is and why it might be important for language and reading comprehension and, by extension, success in school (‘Appendix C’). Stills and video clips, collected during piloting for this study, were used to illustrate how caregivers and their children might engage in inference-eliciting dialogues while reading books. This method of intervention administration has been used successfully in previous studies of language development (McGillion et al., Reference McGillion, Pine, Herbert and Matthews2017), and was chosen for its consistency. Qualitative feedback in the exit questionnaire suggested that caregivers had enjoyed the reading comprehension video and found it useful in explaining the theory behind the study.
A second video (matched in length, format, production, and aims to the training condition) introduced caregivers to the Mathematical Control Intervention.
Intervention support materials
Children in the inferencing training condition were given 10 books. Inference-making questions were pasted alongside the text of these books to elicit inferencing during shared book-reading. Each question label included a picture of a tiger, who was introduced on the front cover of every book. Caregivers explained to the children that the tiger might need some help to understand the story, and that they could do this by answering the questions beside his picture throughout the story. Caregivers were encouraged to provide supportive individual feedback for correct responses (see information leaflet in ‘Appendix D’). For questions where the child did not respond, or responded incorrectly, model answers and feedback were included on the question labels, e.g.:
– Why does Percy need an extra blanket tonight?
– Perhaps he is trying to get warm.
– Can you remember what the weather was like outside? It was very cold!
One Snowy Night (Butterworth, Reference Butterworth2011, p. 4).
See ‘Appendix E’ for book titles, inferencing questions, and model responses.
Video analysis during piloting confirmed that caregivers understood these instructions and were able to incorporate the question prompts and feedback into their usual book-reading routine (this was also endorsed in their oral feedback to us). Children in the control condition were given the commercially available maths workbook At Home with Counting (Ackland, Reference Ackland2012). This book introduces simple number knowledge (e.g., learning the numbers 1–10) and skills (e.g., sequencing, adding, more/less than) through matching drawing and colouring activities. Each page contained instructions for caregivers to encourage their child's participation and support learning. Families in both conditions were given an intervention diary to record each time they read a particular book (inferencing training condition) or completed a page in the maths workbook (maths control condition) and to comment on their experience of taking part in these activities. Qualitative analysis of these comments after the intervention was complete suggested that caregivers understood what was expected of them during the intervention period.
Measures of inference-making
Age-appropriate story vignettes and questions were used to measure children's inference-making ability at baseline and post-test. Inferencing vignettes for preschoolers taken from the Language and Reading Research Consortium (LARRC; see Currie & Cain, Reference Currie and Cain2015; Language and Reading Research Consortium, 2015, for details of their construction and validation) were administered at baseline (Birthday) and at post-test (A New Pet; A Family Day Out Part 1). Additional author-designed vignettes followed the LARRC template (baseline n = 1, Rover the Dog (see Table 1); post-test n = 1; Jessie's Birthday Party, ‘Appendix F’), and were designed to portray familiar scenarios that tapped into four-year-olds’ world knowledge.
Table 1. Baseline vignette Rover the Dog with inferencing questions
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200327171122418-0547:S0305000919000801:S0305000919000801_tab1.png?pub-status=live)
To demonstrate comparability between the author-designed vignettes and those from the LARRC (2015) materials, the number of utterances, number of morphemes, word tokens (i.e., the total number of words including repetitions of the same word), and word types (i.e., number of different words) were computed using CLAN (Computerized Language Analysis; MacWhinney, Reference MacWhinney2000). Two measures of linguistic richness were used: global syntactic complexity (indexed by the mean length of utterances in morphemes) and lexical diversity (type:token ratio). These analyses suggest that the LARRC and author-designed vignettes were of a comparable level of difficulty.
Each short story was read aloud by the experimenter and was followed by between four and eight questions to assess inferencing ability. Questions followed the order in which the information was presented in each story vignette and required the child to integrate information across the text and/or their world knowledge to, for example, infer character motivations (e.g., Table 1, questions 1 and 5), emotions (question 8), and semantic (question 2) and anaphoric relationships (questions 4 and 6). We tested a range of inference types so that our results could inform interventions that would comprehensively promote the range of inferences that children face during oral comprehension. In this respect, our materials are in line with standardised measures of reading comprehension that frame inferencing as a broad construct, e.g., the Neale Analysis of Reading Ability (NARA II; Neale, Reference Neale1989) and the York Assessment for Reading Comprehension (YARC; Hulme et al., Reference Hulme, Stothard, Clarke, Bowyer-Crane, Harrington, Truelove and Snowling2009). Inferencing questions by type, with by-group scores, are in ‘Appendix G’. Author-designed vignettes and questions were administered first at both time-points. These stories were presented in two parts to minimise memory demands, and included pictorial supports to illustrate characters in the story.
Procedure
Families meeting the eligibility criteria were invited to take part in a study investigating factors that impact on school readiness. Prior to this appointment, participating caregivers completed a Family Questionnaire to measure demographic information, e.g., caregiver education and household income (see Alcock, Meints, & Rowland, Reference Alcock, Meints and Rowland2017, for details of its construction), and a Home Life questionnaire to collect information about literacy-related behaviours and attitudes e.g., how often someone read with the child in a typical week.
Randomisation
Dyads were randomised to either the inferencing training or maths control condition according to CONSORT 2010 guidelines (Schulz et al., Reference Schulz, Altman and Moher2010). Randomisation was conducted by an independent statistician at the University of Sheffield. Randomisation was stratified by household education (degree or no degree) and the condition to which dyads had been allocated in a previous intervention study if they had taken part in it (McGillion et al., Reference McGillion, Pine, Herbert and Matthews2017). If the family had not taken part in the prior intervention (n = 17), they were allocated a condition envelope for a family that had taken part in the prior intervention but who had declined to take part in this study, matching for SES. For each participant number, condition allocations were placed in a sealed envelope, identified only by participant number, by a research assistant not involved in any other aspect of the project. Another researcher who administered the baseline measures and the intervention became aware of condition allocation by opening the envelope during the baseline visit, and only once the final baseline measure had been collected. This ensured that baseline measures were collected blind to condition allocation. This research assistant, having opened the envelope with the appropriate participant number to find out which condition the dyad had been randomised to, administered the relevant intervention. Intervention groups did not differ as a function of child age, gender, or SES.
Baseline data collection
Caregivers and their children completed two baseline visits. On the first visit, at the university, children completed several measures of mathematical ability as part of a separate study on mathematical development (Yanez Diaz Barriga, Reference Yanez Diaz Barriga2018). The second visit took place in the family home. After two cameras had been set up (Sony HDR-PJ810E and Sony HDR-PJ220E) and turned on, caregivers and their child spent approximately 10 minutes completing a book-reading session as a warm-up activity before baseline data collection began. First, this involved collecting a measure of child inferencing ability. The researcher read two story vignettes (one author-designed; Rover the Dog (Table 1), the second from the LARRC; Birthday). Children were asked to listen carefully to the stories so that they could answer the inference-eliciting questions that followed each story. Aside from general encouragement, no other feedback was given. Then, child language and communication was measured by the researcher (secondary outcome) using the NFER Baseline Reception Assessment Language and Communication scale (NFERL) (National Foundation for Educational Research, 2015) and the Language Content index of the Clinical Evaluation of Language Fundamentals Preschool 2 UK (CELF) (Wiig, Second, & Semel, Reference Wiig, Second and Semel2006). The NFERL assesses phonics, picture sequencing, story prediction, word reading, simple sentence reading, and name writing. The Language Content index of the CELF is a measure of vocabulary breadth, concept development, comprehension of simple and complex sentences, and comprehension of associations and relationships among words.
The intervention
After all baseline measures were collected, the researcher opened the envelope containing the dyad's condition allocation and administered the appropriate intervention.
Inference Training Condition
The researcher explained that the study was investigating whether asking questions during shared book-reading could help language comprehension before children start school. Caregivers were shown the intervention materials, watched the intervention video, and were asked to read each of the 10 books (with inferencing questions included) at least twice over the course of the following month (i.e., a minimum of 20 sessions), and were given a leaflet summarising the main intervention message (‘Appendix D’).
Mathematical Control Condition
The researcher explained that the study was investigating whether completing daily maths activities could help children get ready for school. Caregivers watched a video explaining what the intervention involved, were shown the maths workbook, and were asked to complete one or two pages a day over the course of the following month (i.e., a maximum of 20 sessions).
Caregivers in both conditions were given an intervention diary to record how often they completed the relevant intervention activities and their impressions of having done so.
Post-test data collection
Approximately one month later, caregivers and children visited the university for post-test data collection. A version of the Home Life questionnaire (adapted to include questions about the activities completed over the past month), and an exit questionnaire about the general experience of taking part in the study were posted to caregivers in advance of this visit. Caregivers were asked to complete these questionnaires and to bring it with their completed intervention diary to the university in a sealed envelope.
The researcher read three different short vignettes following the protocol established at baseline (see ‘Materials’ section: one author-designed; Jessie's Birthday Party, and two from the LARRC; A Family Day Out Part 1; A New Pet), see ‘Appendix F’). These vignettes were of equivalent total length to those administered at baseline and were matched for story theme. After each story, the researcher asked a series of questions designed to measure the child's inference-making ability (primary outcome). The Communication Language and Literacy and Mathematical Literacy components of the NFER Baseline Reception Assessment were administered to measure child language (secondary outcome) and mathematical ability.
Debrief
In accordance with ethical guidelines laid down by the University of Sheffield ethics committee, all caregivers were fully debriefed by email after all children had completed the final outcome visit.
Coding and measures
Inferencing ability at baseline and post-test was measured by child responses to the inference eliciting questions following story vignettes. Responses to each inference question were scored from video-recordings by a researcher blind to condition allocation. Correct responses that demonstrated full inference-making were awarded 2 points. Partially correct answers that lacked full inferencing scored 1 point. Unintelligible responses, “I don't know”, or incorrect responses scored 0. For example, for question 5 in Table 1; “Where did they walk to with Rover?”, “The park / the playground” scored 2 points, “swings and slides / roundabout” scored 1, and “shopping / for a walk” scored 0. For any response that was scored as partially correct or incorrect, the researcher asked a background question(s) or prompted the child in line with the rubric to help the child follow the narrative, and so that subsequent questions could be administered. Complete scoring schemes are presented in ‘Appendix F’. Scores for individual questions were summed to produce an overall inferencing score at baseline (out of a maximum score of 32 for the 16 questions at baseline) and post-test (out of a maximum score of 40 for the 20 questions at post-test). Rare instances of missing data were replaced with the sample mean for the particular item. At baseline, five participants had at least one missing datapoint on measures of inferencing ability, each with an average of 2.4 items missed out of the 16 items on this scale (totalling less than 1% of data on this measure). Four participants had at least one missing datapoint on post-test measures of inferencing ability, each with an average of 3 missed items out of the 20 items on this scale (totalling less than 1% of data on this measure). Incidences of missing data due to experimenter error or caregiver interference were replaced with the sample mean for the item in question.
Child language and communication ability was measured using the NFER Baseline Reception Assessment Language (baseline and post-test) and the CELF Language Content Index (baseline). These were scored from video-recordings by a researcher blind to condition allocation. A raw frequency score was calculated for each test according to individual assessment guidelines. Incidences of missing data due to experimenter error or caregiver interference were replaced with the sample mean for the item in question. Eight participants had at least one missing datapoint on the CELF, each with an average of 2.7 missed items out of the 59 items on this test (totalling less than 0.5% of data across the dataset on this measure). Nine participants had at least one missing datapoint on the baseline NFER, each with an average of 2.8 missed items out of the 43 items on this test (totalling 1% of data across the dataset on this measure). Twelve participants had at least one missing datapoint on the post-test NFER, each with an average of 1.8 missed items out of the 43 items on this test (totalling 0.5% of data across the dataset on this measure). Results for primary and secondary outcomes (as well as for the post-hoc analyses reported below) did not change when statistical models were run without the imputed data.
Reliabilities
Ten percent of responses to questions measuring inferencing ability, randomly selected, were double coded by a researcher blind to condition allocation at baseline (n = 10) and post-test (n = 10). Correlations between scorers indicated high levels of agreement at baseline (r = 0.97). There was 100% agreement at post-test. The CELF and NFER Language were coded live, using the standardised tests stopping rules. A second researcher blind to condition allocation recoded these tests from the video-recording to check the accuracy of the test administration and scoring. Internal consistency was acceptable for our main measure of inferencing ability at post-test (α = .76). Baseline tests of inferencing ability had a Cronbach's alpha of .65. As a measure of the predictive validity of the inferencing measures, baseline and post-tests of inferencing were found to be positively correlated in the control group (r = 0.59).
Sample size and statistical methods
Sample size was calculated to detect a medium effect size for the primary outcome measure (inferencing ability) with 80% power at the 5% level of significance, allowing for up to 20% attrition rate. To compare primary and secondary outcomes across intervention groups, we fitted separate linear regression models to each outcome measure with condition and an equivalent baseline measure as predictors. All analyses were conducted using R version 3.3.1 (R Core Team, 2013) and RStudio Version 1.1.419 (RStudio Team, 2015). In two final post-hoc analyses we used correlation to explore individual differences between language (at baseline and post-test) and inferencing ability.
Results
Children were generally able to engage with the inferencing task, and scores on the baseline and post-test were normally distributed. Table 2 reports descriptive statistics for all baseline and post-test measures along with the number of sessions parents reported having completed at home for each condition. Percentage scores for the baseline and post-tests of inferencing were calculated by dividing total scores by the maximum possible score (32 at baseline and 40 at post-test), and are shown in Table 2 to facilitate interpretation: results of statistical analysis are reported for the raw scores only. Children in the maths control condition scored slightly but significantly higher on the baseline test of inferencing than children in the inferencing training condition (t(98) = 2.23, p = .03). In line with our statistical analysis plan, these baseline scores are controlled for in the analysis of the effect of the intervention below (Table 3). There was no significant difference between groups on either measure of language or communication collected at baseline (NFER Language t(98) = 0.48, p = .63; CELF t(93) = 1.48, p = .14). Both groups scored within expected norms on the CELF Language Content Index. Children in the inferencing training condition made bigger pre- to post-test numerical gains on the inference tests (45% at baseline and 50% at post-test) than children in the maths control condition (53% at baseline and 52% at post-test). Parents in the inferencing condition reported completing sessions at home on more days than parents in the maths condition. This simply reflects the fact that, once the maths book had been completed, parents were less likely to return to it again, whereas the story-books for the inferencing training were often shared multiple times.
Table 2. Means (SD) scores for Inference, CELF, and NFER language tests at baseline and post-test as a function of condition
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200327171122418-0547:S0305000919000801:S0305000919000801_tab2.png?pub-status=live)
Table 3. Regression model fitting condition and baseline inferencing to post-test inferencing (n = 95)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200327171122418-0547:S0305000919000801:S0305000919000801_tab3.png?pub-status=live)
Note. R2 = .24.
Effect of the intervention on primary and secondary outcome measures
To test for an effect of the intervention on inferencing scores (our primary outcome measure), we built a linear regression model with inferencing at post-test as the outcome variable and intervention condition plus inferencing at baseline (scaled and grand mean centered) as predictors. The model is reported in Table 3. Controlling for baseline, there was no statistically significant effect of condition on inferencing outcomes (Hedge's g = 0.14). Hedges’ g (calculated using the R package metanalytic: Xiao, Kasim, & Higgins, Reference Xiao, Kasim and Higgins2016) is a corrected measure of effect size for continuous variables in smaller samples. It is interpreted in the same way as Cohen's d, i.e., 0.2 is a small effect, 0.5 is medium, and 0.8 is a large effect.
To test for an effect of the intervention on NFER language scores (our secondary outcome measure), we built a linear regression model with NFERL at post-test as the outcome variable and intervention condition plus NFERL at baseline (scaled and grand mean centred) as predictors. The model is reported in Table 4. Controlling for baseline, there was no statistically significant effect of condition on language outcomes (Hedge's g = 0.18).
Table 4. Regression model fitting condition and NFER language baseline to post-test NFER language (n = 98)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200327171122418-0547:S0305000919000801:S0305000919000801_tab4.png?pub-status=live)
Note. R2 = .70.
To further explore these null effects, we ran equivalence tests on our primary (inferencing ability) and secondary (NFER language scores) outcome measures (Lakens, Scheel, & Isager, Reference Lakens, Scheel and Isager2018). Equivalence testing is a variant of hypothesis testing that examines whether the difference between groups is more or less extreme than the smallest effect size of interest, i.e., are groups significantly equivalent. We used the two one-sided test procedure (TOST from R package TOSTER: Lakens, McLatchie, Isager, Scheel, & Dienes, Reference Lakens, McLatchie, Isager, Scheel and Dienes2018), setting the minimal effect size of interest at 0.5 (Colmar, Reference Colmar2014; Noble et al., unpublished observations). These analyses showed that, at post-test, there was no meaningful difference between participants in the control and intervention groups with respect to both inferencing (t(91) = –1.847, p = .034) and language ability (t(91) = 2.373, p = .001).
Individual differences in inferencing and language ability
As an exploratory post-hoc analysis, we tested whether individual differences in inferencing were associated with language ability as measured by the CELF Language Content Index. Recall that the CELF measures vocabulary, concept development, and comprehension of sentences and of relationships between words – all key components of inferencing. A positive correlation would indicate that the inferencing is related to general language skill. Considering all participants at baseline, inferencing ability was significantly associated with performance on the CELF Language Content Index (r = 0.47, p < .01). At post-test, considering only children in the maths control condition (for whom there could have been no effect, however small, of the intervention), here too this measure of inferencing was highly correlated with children's baseline CELF Language Content Index (r = 0.63). These correlations are about the same size as the correlation between baseline and post-test measures of inferencing themselves (r = 0.59, again considering children in the control condition only, although correlations are of similar size for the full sample; r = 0.49). This suggests that our measure of inferencing is related to more general language ability.
Parents’ responses to the intervention
Parents’ qualitative comments about the inferencing training were extracted from the intervention diaries. Here we summarise the major trends, together with more general comments about study participation from an exit questionnaire.
Thirty-eight diaries were returned from the 47 families taking part in the inferencing training. These indicated that a mean of 23.9 sessions were completed over the duration of the intervention (SD = 7.8; Range 11–49). Recall that families were instructed to read each book at least twice during the month, so the minimum expected number of sessions was 20. Thus, we take the level of uptake and engagement in the training as moderate to high. Most parents saw value in preparing their children for the transition to school and reported that the activities were enjoyable, though they reported some difficulty fitting the sessions around other daily activities. Although parents commented that their children enjoyed certain books more than others, there was no difference in how often each of the 10 books was read (the mean number of reading sessions per book ranged from 2.3–2.8). Regarding children's levels of concentration during the shared reading sessions, some parents reported that their children enjoyed answering the questions, whereas others were distracted from the story by looking for the tiger stickers. Time of day was also cited as a factor in levels of tiredness and concentration. The repetition of books and questions elicited both positive and negative comments. Many parents were keen to report that the repetition strengthened their children's confidence and understanding of what was happening in the stories and, at least for the books that their children enjoyed, they were enthusiastic about repeated reading. However, a few parents also reported that children were frustrated by being asked the same questions. The main implication for our intervention is that children may not have been engaging in inferencing on subsequent sessions and instead either refusing to answer the question or rote responding from memory.
Of the range of inference types in the training materials, parents reported that their children found some harder than others. For example, inferences about why characters were feeling a certain way were challenging for some children, as were predictions. Inferencing was also sometimes hindered by a lack of world knowledge. For example, some children needed an explanation of the meanings of mustard or carsick in order to attempt the relevant inferencing question.
An unanticipated advantage of the inferencing training was its ability to give parents a means of explicitly assessing how much their children understood from a shared story. Many appreciated the chance to learn about their child's abilities. Some had underestimated how much their children understood but had revised their assessment from their child's responses to the inferencing questions. Together with the observations of spontaneous shared reading during the pilot study, these comments also suggest that the intervention went beyond parents’ usual practices when reading with their children. The fact that parents would not typically ask this number and type of questions during shared book-reading means that our intervention was qualitatively different to business-as-usual for the majority of families in our sample. Although some parents (particularly those with higher levels of literacy; Bus, Leseman, & Keultjes, Reference Bus, Leseman and Keultjes2000; Heath, Reference Heath1983) may engage in a lot of extra-textual talk when reading some genres (particularly information books; Anderson, Anderson, Lynch, Shapiro, & Kim, Reference Anderson, Anderson, Lynch, Shapiro and Kim2012; Pellegrini, Perlmutter, Galda, & Brody, Reference Pellegrini, Perlmutter, Galda and Brody1990; Potter & Haynes, Reference Potter and Haynes2000; Price, van Kleeck, & Huberty. Reference Price, van Kleeck and Huberty2009), many do not. The socioeconomic diversity of our sample and the use of story-books in our study means that substantial differences are likely between the reading style imposed by our intervention and what the majority of parents in our sample would normally do.
Discussion
This randomised controlled trial evaluated a language intervention intended to boost inferencing skills using implicit training, delivered by parents to their preschool children. The training was designed to prompt four-year-olds’ inferential thinking by giving them practice in answering inferencing questions during shared book-reading. The training had no significant effect on either inferencing skills or on language and communication skills. Despite good theoretical justification, high levels of engagement by the participating families, and a rigorous RCT design, our intervention did not effect significant change. Based on methods used in previous inferencing interventions that successfully improved comprehension in school-aged children (Bianco et al., Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010; McGee & Johnson, Reference McGee and Johnson2003; van Kleeck et al., Reference van Kleeck, Vander Woude and Hammett2006; Yuill & Oakhill, Reference Yuill and Oakhill1988), our design focused on asking children inferencing questions while they listened to stories. It also closely followed van Kleeck's (Reference van Kleeck2008) evidence-based recommendations for fostering inferential language in younger children, e.g., targeted questions and scripted feedback in a shared reading context.
Our findings have several important implications for the field. First, having used gold-standard methods to test the efficacy of supporting preschoolers’ inferencing skills using implicit methods at home, the evidence base for this type of training remains negligible. Future interventions should offer more support for children of this age by using direct teaching methods, and should scrutinise the potential benefits of professionally implemented interventions that use explicit, well-defined, comprehension-focused activities. Second, our findings highlight the link between inferencing and general language ability. Specifically, we would like to promote strategies that strengthen vocabulary to provide a solid foundation for inferencing. The results of this rigorous RCT will benefit researchers engaged in theory-building and testing as well as practitioners choosing how to allocate resources.
Quality of parent delivery in the intervention was high. Videos of intervention sessions with pilot caregiver–child dyads showed that the training was accessible and implemented as intended. We designed the training to be consistent and easy to follow, and gave clear instructions in the support materials. Qualitative and quantitative comments from the intervention diaries and exit questionnaires showed that engagement was generally good, with a mean dosage of 24 sessions out of a recommended 20 over the month-long intervention. Responses to the Home Life questionnaire revealed that, for the vast majority of our inferencing training group, reading is a frequent and enjoyable activity. Ninety-five percent of returned questionnaires (n = 44) stated that someone reads or looks at books with their child daily (84%) or more than three times per week (11%), and 93% of parents who returned questionnaires agreed or strongly agreed that they found reading on their own enjoyable. While these caregivers are not necessarily representative of the general population, it suggests the format of the training was familiar and pleasurable, raising the likelihood of good quality implementation.
In addition to sound theoretical foundations and good treatment fidelity (according to our measures), our study used an RCT as the gold standard for testing the effectiveness of an intervention. Despite having used these three core strategies for maximising success, we are left with the question of why the training did not have reliable effects on our primary outcome of inferencing skills.
Recall that our original aims were to investigate: (i) whether training inferencing in the preschool years is possible; and (ii) if so, whether it is possible in this age group using implicit methods. Our results suggest not on both counts, at least in an intervention of this length. First, the children that our intervention targeted may not be developmentally ready to benefit from this kind of implicit inferencing training (where a parent asks an open question with basic scaffolding in the case of incorrect responses). This is supported by another intervention study with preschoolers, which found that, although mothers’ inferential yes/no questions and statements predicted children's receptive vocabulary growth over six months, mothers’ inferential wh-questions did not (Tompkins, Bengochea, Nicol, & Justice, Reference Tompkins, Bengochea, Nicol and Justice2017). The authors suggest that since open-ended wh-questions (similar to those used in our study) do not provide the child with the correct information (in contrast to closed questions and statements), preschoolers may need inferences to be made more explicit for them to facilitate language development (see also Carmiol, Matthews, & Rodríguez-Villagra, Reference Carmiol, Matthews and Rodríguez-Villagra2018).
Our approach was novel in its focus on the oral language of children in the preschool years. On the whole, comparable successful interventions have targeted children ranging from six to nine years old (Clarke et al., Reference Clarke, Snowling, Truelove and Hulme2010; McGee & Johnson, Reference McGee and Johnson2003; Yuill & Oakhill, Reference Yuill and Oakhill1988), due to their focus on reading comprehension as an outcome measure. Although inferences are within reach of children from three to four years old (Filiatrault-Veilleux et al., Reference Filiatrault-Veilleux, Bouchard, Trudeau and Desmarais2016, and as shown by the distributions of scores on our tests of inferencing), evidence that the same skills can be trained in preschoolers is scant. To the best of our knowledge, a single study has shown that five-year-old children showed improvements in oral comprehension (including inferencing) after explicit training activities that spanned seven months (Bianco et al., Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010). Crucially, to be effective, the training in that study had to comprise explicit, well-defined, comprehension-focused activities, i.e., not shared reading and discussion alone – a point we will return to below. Therefore, despite showing competence in inferencing and engaging with the training material, under-fives may not be able to transfer the skills they practised during the shared reading activities to a test situation.
The reasons for this apparent age threshold cannot be conclusively answered by our data, but one potential factor could be four-year-olds’ immature executive function skills. The working memory (WM) demands of the inferencing task may have prevented children from responding even if in principle they could make relevant inferences. While some of the vignettes were presented in two halves and with picture prompts (Rover the Dog; Jessie's Birthday Party), others were presented without a break and without visual support (Birthday; A New Pet). The latter two vignettes were therefore quite long (211 and 161 words, respectively), and were also administered later in the session so fatigue effects are likely to have been at play. The lower mean scores for these particular vignettes relative to other vignettes administered at the same time-point suggest that WM demands may have impeded children's inferencing performance (see also Freed & Cain, Reference Freed and Cain2017, for evidence that younger children benefit from a segmented format when being tested on inference-making). While many real-world inferences necessitate the retention in memory of large blocks of texts, future studies might explore reducing these demands with preschool children.
Returning to the second consideration of whether inference training is possible in preschoolers using implicit methods, the indirect nature of the instruction provided may also explain the null results. Our training was focused on parents asking inferencing questions and children answering them, with parents responding to incorrect answers using minimal, prompted, item-specific feedback. Although the shared reading materials were designed to highlight gaps in the text, relying on children to realise that these gaps existed and then make the required inference without more explicit feedback, may have overestimated their capabilities at this age: learning opportunities may have been too subtle to effect the hypothesised change. One reason for adopting this implicit approach was that explicit instruction is not easily accessible by four-year-olds, yet to begin formal education. Another was the challenge of training parents in explicit methods. The current evidence base for the effectiveness of explicit parent-delivered interventions is small (Burgoyne et al., Reference Burgoyne, Gardner, Whiteley, Snowling and Hulme2018; Huat See & Gorard, Reference Huat See and Gorard2013) relative to the more substantial literature on the success of interventions by trained professionals (Bianco et al., Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010; Fricke, Bowyer-Crane, Haley, Hulme, & Snowling, Reference Fricke, Bowyer-Crane, Haley, Hulme and Snowling2013; Fricke et al., Reference Fricke, Burgoyne, Bowyer-Crane, Kyriacou, Zosimidou, Maxwell, Snowling and Hulme2017; Rogde, Melby Lervag, & Lervag, Reference Rogde, Melby-Lervag and Lervag2016). Thus, inferencing interventions may be more effective if delivered by early years professionals who could adapt some of the explicit methods used in the classroom with older children, e.g., giving practice in text prediction or in lexical inference (McGee & Johnson, Reference McGee and Johnson2003; Yuill & Oakhill, Reference Yuill and Oakhill1988). Indeed, a single study has shown that these explicit approaches can be effective for four-year-olds’ inferencing abilities over a longer period (Bianco et al., Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010). This is not to say that the medium of shared book-reading is problematic in itself; explicit feedback can be integrated into natural book-sharing interactions via adult modelling (van Kleeck, Reference van Kleeck2008, p. 638). Indeed, the discussions between caregiver and child resulting from adult feedback is likely to be beneficial for inferencing training, compared to simply answering comprehension questions (Yuill & Oakhill, Reference Yuill and Oakhill1988).
Our results raise the more general question of whether it makes sense to train inferencing as an isolated skill in preschoolers, or to instead concentrate on other aspects of language such as vocabulary. We found that both baseline and post-test measures of inferencing were highly correlated with children's baseline language ability as measured by the CELF Language Content Index (tapping vocabulary breadth, concept development, sentence comprehension, and comprehension of lexical relationships). This suggests that our measure of inferencing – and indeed inferencing ability in general – might reflect general language ability, and we would welcome studies that further analyse the nature of this association. Language skill (or more specifically the vocabulary component) may be a more powerful determinant of inferencing ability than the type of inferencing training we administered. This explanation is in line with evidence from individual differences and longitudinal studies showing vocabulary knowledge to be a key predictor of inferencing (Currie & Cain, Reference Currie and Cain2015; Language and Reading Research Consortium, Currie, & Muijselaar, Reference Currie and Muijselaar2019; Lucas & Norbury, Reference Lucas and Norbury2015; Silva & Cain, Reference Silva and Cain2015), and is also supported by the lexical quality hypothesis, which predicts that more precise knowledge of words promotes efficient text comprehension (Perfetti, Reference Perfetti2007). If a child doesn't yet know a word (i.e., developing vocabulary breadth) or have a sufficiently rich representation of its meaning (i.e., developing vocabulary depth), they are less likely to integrate the word into the situation model to make the required inference during comprehension. In the case of semantic inferences, for example, a rich and robust knowledge of word meanings is required to map between a word and its synonyms, co-hyponyms, or superordinates (e.g., knowing that apples and bananas are types of fruit), thus greater vocabulary depth and richer semantic networks facilitate more efficient and more complex semantic inferences. In a recent study that analysed the concurrent and longitudinal relations between inference-making, vocabulary, and verbal working memory in four- to five-year-olds to eight- to nine-year-olds, both vocabulary breadth and (to a lesser extent) depth explained inference-making skill in the early grades, i.e., at the same age as our sample (Language and Reading Research Consortium et al., Reference Currie and Muijselaar2019).
Furthermore, vocabulary can boost the memory processes recruited during inferencing. Robust word representations can support the maintenance of semantic information in verbal working memory (Nation, Adams, Bowyer-Crane, & Snowling, Reference Nation, Adams, Bowyer-Crane and Snowling1999; Walker & Hulme, Reference Walker and Hulme1999), and efficient maintenance of word meaning is necessary for integrating information distributed throughout the discourse. Thus, good vocabulary supports inferencing in (at least) two distinct ways. A more effective approach to scaffolding inferencing might be to focus on boosting vocabulary breadth, depth, and conceptual knowledge, all of which can be used in making inferences. As vocabulary has been implicated in the development of multiple aspects of comprehension (Marulis & Neuman, Reference Marulis and Neuman2010, Reference Marulis and Neuman2013; Ouellette, Reference Ouellette2006; Wright & Cervetti, Reference Wright and Cervetti2017), training word learning could be a powerful tool to benefit language across the board. Accruing a greater vocabulary size and a richer knowledge of word meaning may be of particular benefit to inferencing. Interestingly, there is emerging evidence for a reciprocal relationship between vocabulary and inferencing skill (Language and Reading Research Consortium et al., Reference Currie and Muijselaar2019). This highlights the importance of practising both skills in the classroom to benefit not only the discrete skills, but also the ways that each can support the other.
Another reason for favouring a focus on lower-level language skills such as vocabulary and lexical relationships is the apparent lack of transfer in our data. That is, the lack of transfer between inferencing skills practised during the training to those required at post-test could be taken to suggest a lack of generalisability between semantic domains. For example, if a child can make the inference that sitting on sand means being at the beach, this does not guarantee that they can make a different type of inference, say about a character's motivation or the consequences of their actions. Although learning about the sand/beach connection would increase a child's knowledge about that specific domain, it may not be useful for higher-level, general-purpose inferencing ability (if such a thing exists). This hypothesis is also in line with the modest correlations between baseline and post-tests of inferencing (r = 0.59 in the control group and r = 0.4 in the training group). That is, it may not be possible to train ‘general purpose’ inferencing. If this is the case then a more fruitful approach to boosting inferencing may be to focus on vocabulary and the development of richly connected semantic networks.
Summary and recommendations
There are several reasons – separately or in combination – which might explain why our training was not effective in improving inferencing. Age of the children, use of implicit-exposure training, lack of transfer between inferencing domains, and the tentative link between inferencing and underlying language ability could all have limited its potential to effect change. Nonetheless, due to the firm evidence base suggesting that the development of inferencing can be supported (albeit in older children and/or using more explicit methods), and our use of a robust RCT, we had good reason to believe that the children might learn from the training. The fact that they did not means that the evidence for inferencing training using implicit methods with younger children remains negligible. Future interventions would need to offer more support for children of this age. Our findings suggest the following priorities for future research and practice.
First, the association that we found between language and inferencing skills suggests that a more effective approach to scaffolding inferencing in the preschool years might be to focus on promoting vocabulary to develop the broad, deep, and rapidly accessed semantic knowledge necessary to make inferences viable. This should be preceded by in-depth analysis of the links between inferencing and language ability.
Second, interventions using more explicit inferencing training could shed light on whether the implicit nature of the current intervention was the limiting factor for preschoolers. To do this, materials could be adapted from successful interventions for older children (e.g., Bianco et al., Reference Bianco, Bressoux, Doyen, Lambert, Lima, Pellenq and Zorman2010; McGee & Johnson, Reference McGee and Johnson2003; Yuill & Oakhill, Reference Yuill and Oakhill1988) to determine whether explicit inferencing training could be accessible and effective for preschoolers. It may be that this type of training is best delivered professionally, i.e., at preschool. As reviewed above, the most successful inferencing interventions have used direct teaching methods, with frequent, explicit focus on the target skill or structures, and have often used group-based delivery methods (Elleman, Reference Elleman2017). These specialist skills that might allow children to gain insights into inference-making are likely best found in well-trained teachers.
Third, we would support future lab-based experimental studies to unpick the components of specific types of inferences, the inferencing-making process, and the associated cognitive resources (e.g., working memory, background knowledge, vocabulary – including its speed of access; Cain, Lemmon, & Oakhill, Reference Cain, Lemmon and Oakhill2004; Freed & Cain, Reference Freed and Cain2017; Language and Reading Research Consortium, 2018; Oakhill, Cain, & McCarthy, Reference Oakhill, Cain, McCarthy, O'Brien, Cook and Lorch2015). While the current study purposely chose to test a wide range of inference types, future studies might select from a more restricted range to investigate how different inference types vary in their developmental trajectories, how responsive they are to training, and how they are underpinned by different cognitive resources. This kind of research programme would provide new insights into limiting factors in early development and how best to support inference development at different points of development. This knowledge could also inform more effective tests for assessing inferencing in the preschool years, e.g., exploring the use of graphic organisers to support memory demands and organise key ideas when inferencing (Nesbit & Adesope, Reference Nesbit and Adesope2006), and using the most supportive dyadic contexts for inferencing, e.g., allowing children to demonstrate their inferencing abilities via narratives rather than undergoing formal question-and-answer tests (van Kleeck, Reference van Kleeck and van Kleeck2006, p. 292). There is also scope for future studies to adopt a joint enquiry approach in which parents and children collaboratively answer questions, giving parents the opportunity to be reading role models and model their own inferencing and deduction processes.
Although our intervention was designed to maximise fidelity – indeed we have no reason to infer that parents did not administer it as intended – future studies should monitor implementation directly, e.g., by asking families to record their intervention sessions and then analysing a proportion of these against the protocols (e.g., Noble et al., unpublished observations). In studies which find no effect of the chosen intervention, direct monitoring would provide specific information about why parent-delivered interventions are not effective. Prior to further RCTs, feasibility studies are essential for clarifying the acceptability of proposed interventions to stakeholders.
Finally, we would like to reiterate the importance of reporting and publishing null results. Our findings show the usefulness of rigorously evaluating well-founded interventions to inform future work, and of disseminating the findings to practitioners who can use the emerging evidence in identifying and developing effective practices.
Acknowledgements
This research was supported by ESRC grant ES/M003752/1. The authors thank the families for their participation, and Kiera Solaiman, Charlotte Rowley, Annalise Guild, Lauren Lofthouse, Katie Gascoigne, Hannah Putrus, Rachael Staunton, and Jonathan Turner for assistance with collecting and coding data. They also thank Dr Tim Heaton for conducting the randomisation, Dr Jamie Lingwood for assistance with CLAN, Dr Paula Clarke for comments on an earlier draft, and members of the shared book-reading grant team for advice and support. CR is also supported by the ESRC International Centre for Language and Communicative Development (LuCiD; ES/L008955/1). The authors have declared that they have no competing or potential conflicts of interest in relation to this paper.
AppendicesFootnote 2
Appendix A: Trial registration
Appendix B: CONSORT 2010 checklist
Appendix C: Inferencing training video script
Appendix D: Inferencing intervention support leaflet
Appendix E: Inferencing intervention materials: book titles, inference-eliciting questions, and model responses
Appendix F: Inferencing test materials: bespoke inferencing vignettes, comprehension questions, and coding scheme
Appendix G: Scores by inference question and inference type