1. Introduction
The early childhood period, from conception to age 5, is recognised as an important period for children's development (see e.g. Reference Phillips and ShonkoffPhillips and Shonkoff, 2000; Reference Shonkoff, Boyce and McEwenShonkoff et al., 2009, Reference Shonkoff, Garner, Siegel, Dobbins, Earls and McGuinn2012; Reference HeckmanHeckman, 2008). Children's relationships and interactions with parents and caregivers are a key mechanism through which skills, behaviours, and other developmental milestones are achieved (Reference PiagetPiaget, 2013; Reference Asmussen, Feinstein, Martin and ChowdryAsmussen et al., 2016). It is also widely established that events, developments and shocks occurring in the first five years of a child's life can be expected to have effects that last into adulthood (Reference Almond, Currie, Ashenfelter and CardAlmond and Currie, 2011). Thus, the broad case that early life matters to outcomes is well established.Footnote 1 Moreover, well-regarded direct (i.e. retrospective) cost-benefit analyses are available for a small number of highly stylised programmes, many of which were implemented in small, rigorously controlled and evaluated experiments in the past (e.g. Reference Eckenrode, Campa, Luckey, Henderson, Cole, Kitzman, Anson, Sidora-Arcoleo, Powers and OldsEckenrode et al., 2010; Reference Cunha, Heckman and SchennachHeckman et al., 2010; Reference Kitzman, Olds, Henderson, Hanks, Cole, Tatelbaum, McConnochie, Sidora, Luckey, Shaver, Englehardt, James and BarnardKitzman et al., 1997, Reference Campbell, Pungello, Burchinal, Kainz, Pan, Wasik, Barbarin, Sparling and RameyCampbell et al., 2012). From this evidence we learn that some forms of intervention, with sufficient specificity and intensity, can potentially make a substantive difference to child development in certain contexts. As a result, policies to promote improved parenting and home environments during this period have become of significant interest to local and national policymakers (Reference AllenAllen, 2011; National Audit Office, 2013), as well as the research and scientific communities. This includes activities known as early intervention, for babies, children and adolescents. We define early intervention as targeted activity (in response to emerging signals of additional need) which addresses problems before they become entrenched, thereby preventing costly outcomes to individuals, society and the state.
The Early Intervention Foundation (EIF) was set up in England by the coalition government of 2010–15, to draw together the evidence on this sort of intervention and support local authorities and other agencies in translating this into action. The EIF was established as a ‘What Works Centre’, one of a growing number of independent organisations seeking to review and synthesise evidence on the effectiveness of policy and practice interventions and to share this with policymakers and practitioners in accessible ways. Other What Works Centres have tended to focus on specific domains of practice. Perhaps the best known example is the Education Endowment Foundation (EEF) which has centred on reviewing and developing the evidence on the cost and effectiveness of teaching practice in the classroom. The EEF's Teaching and Learning ToolkitFootnote 2 assesses different teaching and learning methods, using ratings on a 1–5 scale to represent cost of delivery, quality of evidence and scale of impact. This approach is underpinned by three key features and assumptions: i) an emphasis on evidence that can establish causality through the use of high quality, randomised control trial designs that minimise the biases that result from observational approaches with weak, irrelevant; ii) a focus on the single outcome of progress in literacy or numeracy scores, expressed in a single currency of months of learning; iii) the assessment of cost in terms of staff time or other inputs.
This model, rooted in recognition of economic principles of cost effectiveness and a concern for valid causal inferences, provides an important benchmark for other What Works Centres such as on Local Economic GrowthFootnote 3 or Crime Reduction.Footnote 4 This has stimulated a lively debate about the validity of RCTs as opposed to other forms of evaluation such as realist approaches or process evaluation (Reference CartwrightCartwright, 2011; Reference Bonell, Fletcher, Morton, Lorenc and MooreBonell et al., 2012; Reference Marchal, Westhorp, Wong, Van Belle, Greenhalgh, Kegels and PawsonMarchal et al., 2013; Reference PawsonPawson, 2013). Methodologists and practitioners have been inspired to engage in new ways in a fascinating example of the messy frontier between technical issues of methodology and the realities of practice.
This paper describes some of the challenges in applying this model to the field of early intervention, and sets out findings from two substantial studies the EIF has conducted that make small steps towards resolving these conundrums. Section 2 describes some of the issues and difficulties in applying the EEF model to early intervention, with a focus on the challenge of considering the fiscal benefits. Section 3 summarises recent EIF evidence on currently delivered early childhood programmes in the UK, and their levels of evidence of effectiveness. Section 4 reports on a new way of measuring the fiscal costs of late intervention. Section 5 concludes with a discussion of the challenges of integrating findings from sections 3 and 4 and makes a small number of recommendations for future research and practice in this field.
Carefully targeted early intervention and prevention activities are widely recognised to provide benefits for children and their families and wider society and government (Reference AllenAllen, 2011; Department of Health, 2013). However, local agencies often rightly seek evidence in the form of cost-benefit analysis demonstrating credible financial or social returns fit to their populations, location and investments. As will be seen, the current state of the evidence is some way from having this breadth and specificity of information. Therefore, we do not present here a model or assessment of long-term impact; instead, we seek to explain why the demand for simple cost-benefit numbers cannot yet be met meaningfully. We offer new ways of thinking about (a) what is now known about the effectiveness of available programmes that are currently widely commissioned, and (b) what is known about the fiscal costs of the ‘late intervention’ problems which those programmes seek to prevent. We hope in this way to mitigate against the kind of “magical thinking” (Reference Brooks-GunnBrooks-Gunn, 2003) that can sometimes become embedded in policy debate.
2. Challenges in modelling and measuring the impact of early intervention
A first key challenge in estimating the impact of early intervention is the diversity of what is meant by early intervention and of approaches to delivering it. Early intervention as defined by the EIF is not just policy for the early years, but as action that is preventive, corresponding to the notion of secondary intervention in public health (e.g. Reference Durlak and WellsDurlak and Wells, 1998), supportive of a reduction in harm where there is evidence that risk is above that for a population norm. The NAO landscape review (2013, op. cit.) emphasises the importance of this form of policy and practice, that is situated between universal services and the wide array of statutory and acute services that are necessarily involved when public harms are so intense for individuals, their families or communities that the state must intervene through, for example, child protection, criminal justice or mental health services. Early intervention can usefully be thought of as activity that exists between universal services such as GPs and schools, and these more statutory responses and ‘late’ intervention.Footnote 5 Early intervention is a wide terrain that ranges from early years activities (such as targeted health visiting in low income communities) to support for adolescents (such as early identification and support for mental health difficulties). Hence early intervention refers to a wide array of different sorts of activity, in different settings, for children of different ages, at different levels of need, with need and risk defined in different ways, by different practitioner groups and professionals. Not only is early intervention not one thing, the effectiveness of any form of early intervention will depend on how well it is targeted to meet need and how it fits with other parts of the local portfolio of supports and interventions.
A second issue we would highlight is external validity and the transferability of evidence from the best studies to the realities of local action. As stated above there is evidence from rigorous trials (such as Reference Eckenrode, Campa, Luckey, Henderson, Cole, Kitzman, Anson, Sidora-Arcoleo, Powers and OldsEckenrode et al., 2010) of very specific forms of early intervention but this often describes activities that are very different from many local forms of early intervention. The issue of external validity has been highlighted recently by the attempt to transfer the success of the US programme Nurse Family Partnership to the UK as the Family Nurse Partnership (FNP). This form of evidence based policymaking has seen the attempt to replicate the success of a rigorously evaluated US programme in the UK, by adapting what has been shown to work in the US with a carefully managed roll out across multiple parts of England and a high quality randomised controlled trial (Reference Robling, Bekkers, Bell, Butler, Cannings-John, Channon, Martin, Gregory, Hood, Kemp, Kenkre, Montgomery, Moody, Owen-Jones, Pickett, Richardson, Roberts, Ronaldson, Sanders, Stamuli and TorgersonRobling et al., 2016), following an earlier pilot study (Reference Barnes, Ball, Meadows and BelskyBarnes et al., 2009). Although this has been a careful and valuable trial, the early findings at 24 months have been disappointing for advocates of this translation approach.
We referred above to the diversity implicit in the term early intervention. There is also considerable variation in how each type of early intervention is delivered. The recently emerging field of implementation science has emphasised the difficulty for narrow definitions of evidence based policy that for human services interventions in part ‘the practitioner is the intervention’.Footnote 6 In practice, the child and family are also part of the intervention. This makes each local implementation subject to complex, individual level heterogeneity. The role of science and evidence in this approach is much more to support the quality of practice than to develop rigorous evaluations of gold standard products that can then be easily ‘rolled out’.
As described in Section 3 the EIF has taken a different approach that has recognised more the diversity of existing practice and sought to develop more evidence based approaches by supporting existing local practice to improve its evidence, rather than by importing wholesale the most evidence based approaches.
The third issue that we highlight here is the multifinality and long-term nature of the impacts of early intervention (see e.g. Reference Cicchetti and RogoschCicchetti and Rogosch, 1996), i.e. that human development allows a wide diversity of possible pathways from variable starting conditions to variable outcomes, which makes specific impacts hard to predict accurately across multiple domains of outcome.
There are no large sample studies that monitor the impact of programmes through from the early years to later life in ways that can capture all of the hypothesised benefits in different domains of outcome. This adds to the difficulty of construction of a single currency, although in principle cost-benefit analysis is a way to evaluate various types of benefit in comparable forms – such as quality adjusted life years (QALYs), social value and hedonic pricing. In the case of early intervention, the quality of estimate of benefits depends on the degree to which short-run impacts can be linked to subsequent outcomes upon which an economic valuation can be placed.
Important work of this sort has been conducted elsewhere. For example, the Washington Institute for Public Policy (WSIPP)Footnote 7 has combined meta-analyses of programme effects with other longitudinal research on the links between childhood circumstances and future adolescent and adult outcomes. In this way WSIPP has been able to produce hypothetical estimates of long-run economic and fiscal benefits – including net present values, cost-benefit ratios, and benefits by public sector agency – for a range of programmes.Footnote 8 This form of appraisal has expanded the pool of programmes for which cost-benefit estimates are available. However, it focuses on a select group of branded, well-evidenced programmes, using evidence and modelling that is more relevant to US contextFootnote 9 and makes considerable assumptions about the relationships between early impact and long-term development. The lag between immediate impact and fiscal savings introduces unavoidable uncertainty over impacts that increase with the length of the lag. Moreover, ultimate impacts may be diffuse and hard to predict because of multifinality.
For the current crop of provision, especially early intervention and early years programmes delivered in the UK, there remains a lack of information on the expected potential returns on investment, in large part because there has not been enough modelling of local population trajectories in the absence of early intervention, linked to estimates of impact that can address the issues of long lags and multifinality. This serves as a barrier to the funding and commissioning of programmes, since local and national commissioners of services often demand credible, rigorous evidence of future budgetary savings before committing to invest in early intervention. Commissioners and policymakers are interested not just in the general scientific evidence that early intervention might work in theory; they also demand evidence that it will work in practice when they commission it for their populations. Moreover, where cost-benefit analysis is based only on projections from the best-evidenced programmes, it is likely to be an overestimate of what may be achieved from more standard practice. A lack of evidence relevant to the majority of early intervention practice – as delivered ‘on the ground’ – therefore inhibits opportunities to improve outcomes and life chances for children.
3. Effective support for parent-child interactions in the UK
The Early Intervention Foundation was set up in part to begin to address some of these gaps in evidence. Because of the diversity of types of early intervention referred to above, a series of reviews have been established which consider the evidence for different forms of early intervention. The review we summarise here (Reference Asmussen, Feinstein, Martin and ChowdryAsmussen et al., 2016) provides the most extensive summary of what is currently known about the effectiveness of programmes that aim to improve developmental outcomes of children up to age 5 through improved parent-child interactions.Footnote 10 The definition of developmental outcomes included parent-child attachment, behaviour or self-regulation, and cognitive/language development. We reiterate that this particular theme – parent-child interactions in the early years – is just one type of early intervention. Other types of early intervention currently under EIF review include activities to support the relationship between parents, to support social and emotional skills in adolescence, and to support vulnerable groups such as children in gangs. In this paper, we focus on what was learnt about scale and nature of impact and provide a brief description of methods and findings, emphasising uncertainty and the need for innovation as well as the importance of rigour in evaluation of scale of impact.
A full description of the sample frame, methods of identification, methods of assessment and findings is provided on the EIF website.Footnote 11 This includes an up to date assessment through the EIF Guidebook of those programmes with the best evidence of impact.
The review assessed the research and evidence underpinning 75 programmes available in the UK. For each programme, the EIF assessment process generated a rating summarising the strength of evidence for impact on a child outcome, based on detailed and transparent standards of evidence.Footnote 12 The higher levels of this scale are based on the Maryland Scientific Methods Scale (Reference Sherman, Gottfredson, MacKenzie, Eck, Reuter and BushwaySherman et al., 1998) and Reference Flay, Biglan, Boruch, Castro, Gottfredson, Kellam, Moscicki, Schinke, Valentine and JiFlay et al. (2005). The standards also include more formative types of evidence in order to recognise a wider range of activities than those with large sample trials. The ratings summarise the confidence with which one can state that a programme has demonstrably (and causally) led to an improvement in a child outcome; this combines a consideration of the quality of the research design with a binary judgement about whether the balance of the findings is consistent with a positive and statistically significant benefit. There are five levels to this EIF scale: Level 3 on this scale is equivalent to what is commonly considered as efficacy (Reference RushRush, 2009), or proof of concept under supportive circumstances; Level 4 equates to effectiveness at scale, with replicated impacts under multiple trials including adaptations. Both of these levels involve evidence where a causal link between exposure to an intervention and subsequent outcomes would be accepted as credible, subject to good implementation of the study and accurate data collection. By contrast, Level 2 would not count as causal evidence of impact in standard evidence-based policymaking frameworks such as the UK government's Magenta Book (HM Treasury, 2011), but would be recognised as important steps in developing a sound understanding of the mechanism by which impacts might be achieved and how to measure them. The rating ‘NL2’ was used for programmes that do not have evidence from which even preliminary assessments about impact upon a child outcome could be made. Such programmes may only have examined effects on a parent outcome, or may have had qualitative or process evaluations. Programmes assessed as ‘NE’ or ‘No effect’ have evidence from a high quality study with null findings in relation to child developmental outcomes.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table1.png?pub-status=live)
It must be emphasised that the ratings reflected the evidence from the balance of studies conducted on each programme and that the most rigorous evidence was always privileged in the rating system. It must also be emphasised in line with the discussion of Section 1 that the intention was not to determine a set of proven programmes that should then be rolled out nationally, but to provide local decision makers with findings about a range of programmes that provided detailed information about how much evidence was available for each, together with guidance on how to interpret it in the light of local conditions, assets and needs.
It is an important feature of this approach that it assesses the evidence that a given programme has achieved in empirical studies a hypothesised set of impacts, rather than asking whether the programme is based on activities that have a sound scientific basis, or that have been shown to have impact when delivered in other forms by different organisations. We address the question of whether specific programmes have demonstrated effectiveness, not whether they are rooted in evidence, and not the general science of what might work without regard to implementation and delivery. We also took the approach of seeking to assess in a formative as well as summative sense, providing information on how programmes might develop their evidence as well as on what is currently well evidenced. This is important in a field for which much activity is voluntary or developed locally and participants may be hostile or unsupportive to the parachuting in of well-established programmes with higher levels of evidence.
The detailed findings on individual programmes are best reviewed on the EIF website where the information can be kept up to date as the evidence changes and programmes change. We focus here on the general findings of the review considered in terms of aggregate findings across the 75 programmes of this type. Briefly, we found that there were a good number of well-evidenced programmes that if carefully commissioned are likely to be effective. We also found many promising approaches that are based on firm scientific principles but have not yet been proven. The evidence from this review was strongest for programmes that target based on early signals of risk in child development, although universal programmes or programmes that target on the basis of demographic factors can in principle be effective. Evidence was also strongest for programmes that aimed to promote early behavioural self-regulation rather than attachment or cognitive development. This does not necessarily mean that attachment or cognitive development programmes are ineffective; more and better evaluation is required. Nonetheless, evidence-based approaches to improving behaviour are better prepared for replication and taking to scale.
A total of 17 programmes were assessed with an evidence rating of ‘3’ or ‘4’. We refer to these programmes as ‘evidence-based’, which for the purposes of this paper is a shorthand label for having demonstrated statistically significant effectiveness in at least one rigorously controlled study.13 The 17 evidence-based programmes are listed in table 2 above together with their primary outcome domain.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table2.png?pub-status=live)
The review also assessed the implementation costs for these programmes, providing a rating which indicated a broad estimate of implementation cost per child treated. This cost assessment is not the same as the actual market price that a commissioner would pay for an intervention (although both may be correlated, and the latter may influence the former). The actual market price will in practice need to be negotiated between provider and commissioner, taking into account commercial policies, trading conditions and local resources and circumstances. Typically, therefore, information on market prices or unit costs is either unavailable or commercially sensitive, so we did not have access to it in a robust, consistent and comparable way. We therefore developed a new approach that would provide us with comparable data for each programme. We surveyed programme providers and developers to obtain information on their programme's implementation inputs and activities, including costs but also other features of delivery:
• Training fees
• Training time for each practitioner
• Whether booster training is required
• Costs of programme material (initial and ongoing)
• Programme delivery hours for each practitioner involved
• Qualification level of each practitioner involved
• Whether internal and external supervision are required
• Qualifications of internal and external supervisors (if applicable)
• Whether a licence is required
• Typical size of intervention group.
Using statistical and econometric methods, we then weighted and combined this information into a single scale (see Annex 3 of Reference Asmussen, Feinstein, Martin and ChowdryAsmussen et al., 2016 for further detail). We then discretised the scale into the following estimated unit cost bins:
• Low (< £100 per child)
• Medium-Low (£100–£499 per child)
• Medium (£500–£999 per child)
• Medium-High (£1,000–£1,999 per child)
• High (≥ £2,000 per child).
These cost ratings permit one programme to be compared to another in broad terms on the basis of the total amount of resources that one would normally expect are required to deliver it, normalised according to the number of children that would receive the treatment simultaneously. It should be noted that our methodology is an attempt to arrive at a high-level cost for a number of programmes, in a relatively scalable and parsimonious manner. Other approaches exist which produce a more detailed estimate for an individual programme or a small group of related programmes, e.g. Reference Charles, Edwards, Bywater and HutchingsCharles et al. (2013).
The distribution of implementation cost ratings for the evidence-based programmes is shown in figure 1. By and large these are inexpensive programmes: we estimate that many could be delivered at a cost of less than £500 per child. However, this set also includes two attachment programmes with a significantly higher estimated implementation cost of over £2,000 per child.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-fig1.png?pub-status=live)
Figure 1. Distribution of estimated programme implementation costs (per child)
For each of the review's evidence-based programmes, data were also collected on scale of impact and effect size from its most rigorous studies. Across the 17 programmes above, we identified 53 such studies containing 762 effects, of which 304 related to relevant child outcomes – attachment, behaviour, or cognitive development.Footnote 14 Of these, 155 were genuine effect size parameters; in the other cases, an effect size was not actually reported in the original study (the most common alternative was a p-value). We focus our attention on the cases where an actual Cohen's D or equivalent parameter was reported, and in particular the 82 effect sizes that were statistically significant. These are shown in table 3. Effects are coded such that positive numbers represent improvements in a child outcome, and higher numbers represent larger improvements. For full details on the effect sizes for each of the 17 programmes, see table A2 in the Appendix.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table3.png?pub-status=live)
Overall, effect sizes appear to be larger for the evidence-based programmes that target child behaviour, and smaller for evidence-based programmes that target attachment – although the differences would not be statistically significant.
Cost-benefit analysis of early intervention programmes requires knowledge of, or assumptions about, potential long-term effects, in order to make statements about economic and fiscal implications. Table 4 investigates this by breaking down the 82 effect sizes according to the point at which effect sizes were measured. As can be seen, the majority of these effects were measured at post-test, i.e. at the end of the intervention, with only a small number of long-term follow-up measures of impact. This starkly illustrates a key challenge for carrying out cost-benefit analysis of early years parenting programmes – despite our focus on the most rigorous available evidence. The outcomes of these programmes are not generally measured over sufficient duration to permit direct, non-speculative cost-benefit analysis. Furthermore, in the case of behaviour programmes, there appears to be a downward trend whereby effect sizes measured in long-term follow-ups are smaller than effect sizes measured at post-test.Footnote 15
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table4.png?pub-status=live)
When only short-term and immediate shifts in skills, development or human capital are observed, auxiliary methods linking aspects of early and later development are required in order to make statements about potential longer-term outcomes. As highlighted in Section 2, we note that other evidence bodies engaged in the practice of cost-benefit analysis have combined data on programme effects with other research and data on longitudinal trajectories between child outcomes and adult outcomes. There is a rich literature on quantitative approaches for estimating the link between early and later measures of skill or development. These vary from structural models of the dynamics of skills and human capital (Reference Cunha and HeckmanCunha and Heckman, 2008; Reference Cunha, Heckman and SchennachCunha et al., 2010) and of the effects of early-life shocks (Reference Almond, Currie and DuqueAlmond et al., 2017), to more reduced-form approaches involving regressions of adolescent or adult outcomes on early childhood conditions (Reference Smith, Monica and SmithSmith et al., 2010; Reference Goodman, Joyce and SmithGoodman et al., 2011, Reference Goodman, Joshi, Nasim and Tyler2015).
Of course, there are a number of challenges that make this difficult to execute in ways that are not open to criticism. First, programme effects may dissipate or fade out over time – see table 4 – in ways that can be difficult to pin down without direct observation. Hence, forward projections based on short-term or immediate effect sizes risk overestimating the actual long-term effect. Second, even though similar constructs may be measured in evaluation trials and longitudinal cohort studies such that they can be linked, they may be structurally different in important ways. Programme effects from an RCT are not the same as naturally occurring cross-sectional differences in outcomes. Individual heterogeneity, both observed and unobserved, may be responsible for much of the latter; this may mean that long-term benefits may not materialise in the manner hypothesised through the application of ex ante arithmetic. As an example of this, we note evidence that programmes might improve parenting without a corresponding longer-term improvement in child outcomes (e.g. Reference Kalinauskiene, Cekuoliene, VanIJzendoorn, Bakermans-Kranenburg, Juffer and KusakovskajaKalinauskiene et al., 2009), even though the correlation between parenting and child outcomes is well-established in observational studies (Reference UttingUtting, 2007; Reference Dearden, Sibieta and SylvaDearden et al., 2011; Reference Bono, Francesconi, Kelly and SackerBono et al., 2016).
Given these considerable modelling challenges, we are not able to – and do not attempt to – make cost-benefit statements in relation to the programmes considered here. In the next section, we consider evidence from the other end of the spectrum on known or accepted costs of adolescent and adult outcomes but we do not attempt to link the two formally.
4. Evidence on financial and fiscal costs
An increasingly popular way of evaluating the financial case for early intervention is to explore the financial costs of failing to intervene. For example, in the case of behaviour programmes, the costs of conduct disorder provide information and evidence on the case for intervening earlier. Important work has been done in this area by Reference Scott, Knapp, Henderson and MaughanScott et al. (2001), Reference Friedli and ParsonageFriedli and Parsonage (2007), Sainsbury Centre for Mental Health (2009), Reference Bonin, Stevens, Beecham, Byford, Parsonage, Knapp, McDaid and ParsonageBonin et al. (2011), and most recently NICE (2013). However there is considerable variation among these sources in the types of cost, types of benefit, types of analysis and time periods considered. Our work relates more directly to, and builds on, existing evidence on the unit costs of public services relevant to targeted interventions, most notably the Unit Cost Database (produced by New Economy)Footnote 16 and the Unit Costs of Health and Social Care (produced by the Personal Social Services Research Unit).Footnote 17
EIF has recently added to this evidence with estimates of the annual cost of ‘late intervention’. We define ‘late intervention’ as the acute or statutory services and benefits – such as social care, hospital, police and criminal justice services – that may be required if problems are not addressed. These services, while necessary, are costly to individuals, society and state, and tend to be aimed more at management or containment of problems rather than remedying them. The associated costs and harms may, therefore, extend into the future as problems occur repeatedly, although that is not the focus of our estimates. A fuller description of what is captured and included within the EIF estimate of the fiscal costs of late intervention is shown in table 5.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table5.png?pub-status=live)
An important feature of this estimated cost of late intervention is that it is a current, short-term annual, fiscal cost – the fiscal cost this year – rather than a projected cost cumulated over years or decades. Also, as it relates only to known costs to taxpayers, it is not an estimate of social costs (e.g. costs to others from the experience of crime) or economic costs (such as in lost output or lost productivity). Although limited in this way, expressing the cost of late intervention as a short-run fiscal cost offers a number of benefits. First, it is more comparable to the current costs of investing in early intervention and prevention. Second, it is subject to less uncertainty than estimated lifetime costs or costs over a generation, important though those may be (e.g. Action for Children, 2013). Third, our focus on the current annual cost of late intervention sits more easily within budgetary or political cycles relevant to decision-makers – unlike cumulative estimates over a longer period. Focussing on current annual government spending on late intervention for children and young people – while they are still children and young people – enables a clearer discussion of the size and scope of resources currently expended which might become available to use in other ways if early intervention is employed in effective and appropriate ways. It also focusses attention on where the costs of late intervention currently fall between different parts of government, in ways that provide greater impetus for policy response.
The EIF estimate of late intervention complements other estimates of the annual costs of specific problems: youth crime and youth unemployment have both been estimated to cost over £1 billion a year (The Prince's Trust, 2010), and the cost of dealing with child behavioural disorders is estimated at £1.6 billion a year.Footnote 18 In relation to healthcare, National Audit Office (2013) estimated that the NHS spent nearly £10 billion in 2011–12 on the costs of obesity, alcohol misuse and smoking-related illness. In comparison to these estimates, the EIF late intervention estimate is a more ‘global’ summary of costs in that it aggregates across all key issues rather than focussing on one. The EIF estimate is also a ‘bottom-up’ estimate, rooted in actual data on children and young people and the services they use, within each local area and for the country as a whole.
Our general approach for arriving at the annual fiscal cost of late intervention is to take total demand or caseloads for each form of intervention shown in table 5 – obtained from published statistics – and combine that with an estimated ‘unit cost’ of each case that provision. This has the advantage of being directly linked to what we know about demand for late intervention services for children and young people. We have made use of existing unit cost dataFootnote 19 that are widely accepted, including by HM Treasury, as a basis for financial analysis and cost-benefit analysis. Information on unit costs tends to be available at a national level only, even though the true unit cost may vary from one local area to another. Therefore, where it would lead to more robust results, we have also used published data on actual local authority spending on particular acute services.
In summarising this work we are mindful of its key limitations. First, judgements have been made about which items to include in the analysis. Table 5 is not a list of every activity that counts as late intervention, and the items in the table do not add up to the true amount of total spending on late intervention. The table is better interpreted as setting out some key types of late intervention for which national and local data on caseloads are available, along with information on total spend or unit costs. For this reason our estimate is necessarily an underestimate of the true current annual fiscal cost of late intervention. Second, late intervention relates to services not wellbeing. The data we have used will reflect reporting and recording of incidents, local and national decisions about the availability, referral and resourcing of acute services, as well as the broader fiscal climate. While such data are useful for understanding the fiscal implications of policy decisions, and for informing discussions about spending, these data do not capture the underlying outcomes and wellbeing of children and families.
We estimate that, annually, £17 billion per year is spent by the public sector in England and Wales on the costs of providing late intervention for children and young people (Reference Chowdry and FitzsimonsChowdry and Fitzsimons, 2016a). Further details behind the total overall cost are shown in table 6. The single largest components of this are the fiscal costs of responding to incidents of domestic violence and abuse, and spending on looked after children.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table6.png?pub-status=live)
Figure 2 presents a visual summary in terms of the social issues that drive costs. Children's social care – including spending on children in need, child protection and looked after children – represents the issue that constitutes the largest portion of the £17 billion total. The second largest issue relates to crime and anti-social behaviour.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-fig2.png?pub-status=live)
Figure 2. Fiscal cost of late intervention by social issue
The data can also be considered in terms of who pays, as shown in figure 3. Nearly 40 per cent of the cost of late intervention, as currently defined, is borne by local government. Importantly, however, the majority is borne elsewhere. Given that most early intervention programmes and provision are currently funded and commissioned by local authority services, this sheds light on another major economic and financial challenge: even if benefits from early intervention do materialise, they are likely to accrue to other organisations and agencies – for example, police and health services, or the national exchequer. This externality reduces the incentives for a single agency to invest in early intervention and prevention. However, this reflects institutional shortcomings and failures of governance rather than of early intervention itself. We believe that analysis such as figure 3 has the potential to drive the practice and improvement of joint investment and commissioning arrangements, ensuring that investment and benefit are distributed across all relevant public bodies.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-fig3.png?pub-status=live)
Figure 3. Fiscal cost of late intervention by budget area
Finally, because the cost of late intervention is a bottom-up estimate computed by aggregating local data on late intervention caseloads and spending, late intervention costs can also be computed for any local authority in England. The analyses shown in table 6, figure 2 and figure 3 can all be replicated for any English local authority. For reference local authority levels of late intervention cost, normalised on a per-capita basis, range from a minimum of £164 to a maximum of £531. The 25th percentile is £248; the 75th percentile £337.
We stop short of implying that the fiscal amounts presented above represent achievable fiscal savings as a result of early intervention. First, some amount of late intervention will be inevitable and unavoidable – the cost cannot be brought down to zero, and we do not have evidence to prove that it is far from a natural limit. Second, unit costs are usually defined as the ratio between total budgetary spend and total caseloads, hence they may not be equal to the marginal saving that would result from incremental reductions in demand for such services. Third, in reality, the services mentioned in table 4 may be de facto rationed, in the sense that there may be thresholds or waiting lists. Even if early intervention programmes successfully improve outcomes, the result may be simply freeing up of capacity within these services and the reallocation of places, rather than a financial saving from reduced total service use. Even if total service use can be reduced, these services may be governed by fixed budgets in the short run. If so, service expenditure will not be directly linearly related to caseloads and savings will not be proportional to the change in caseloads. To be clear, this is a limitation of the structure and organisation of services, not of early intervention. Nevertheless, this caveat should be borne in mind when assessing the capability of early intervention to deliver financial savings, and so we are careful not to treat the costs of late intervention interchangeably as the benefits of early intervention.
While our estimates are not an attempt at cost-benefit analysis, they do provide important data and evidence which could support it. Further analysis to get closer to cost-benefit estimates, building on this work, would have to model the relationship between the presence or prevalence of child problems and the resulting likelihood of use of, or volume of demand for the late intervention services considered here. Specific cost items, such as antisocial behaviour, youth justice, persistent absenteeism and school exclusions, may be the most fertile ground for savings as a result of improvements to child behaviour, for example.
5. Conclusion and discussion
We summarise our headline findings and messages in the following key points:
1. The scientific basis for early intervention is strong. The importance of promoting improved early childhood environments and child development is strongly supported by research across a range of disciplines. The rationale for policies that take a preventative approach is also sound.
2. Early intervention is a broad and diverse concept. It can vary by target group, type of risk, level of need, type of intervention and type of outcome. The Early Intervention Foundation has conducted substantial reviews of precisely defined types of early intervention; specific provision to achieve specific outcomes for specific target groups. As an example, we focus in this paper on programmes that aim to promote early child development via improved parent-child interactions (up to a child age of 5 years).
3. Cost-benefit analysis for commissioners should reflect what is available, not just what has the best evidence. We reviewed 75 such programmes available in the UK, and found that 1 in 4 have shown a statistically significant improvement in child attachment, behaviour or cognitive development in a well-conducted and rigorous trial with a good counterfactual. Most of these programmes can be delivered at a low or moderate cost per child. Other programmes have less rigorous evidence, but may still be of interest to commissioners.
4. Establishing that a programme is effective is important, but not sufficient, for cost-benefit analysis. Even among programmes with the most rigorous evidence, data on scale and longevity of impact are relatively weak. We find a plethora of different outcome measures, limited reporting of effect sizes, and a lack of long-term follow-up.
5. Future research should focus on the future economic and fiscal value that results from initial programme effectiveness. We support this approach by presenting data on the immediate fiscal costs of late intervention for children and young people. Further evaluation of programmes is required to improve what is known about their impacts on long-run outcomes in adolescence and adulthood, and on demand for late intervention services.
6. Cost-benefit analysis should be applicable to the contexts in which commissioners operate. Commissioners seek to understand how evaluation findings will translate into benefits for their local population and their local system (including workforces, referral pathways and other provision). Commissioners also seek to understand how benefits are distributed across agencies, as well as the extent to which these benefits might lead to tangible savings.
Discussion
While it is important for commissioners and policymakers to have clear information on effectiveness and value for money to guide decision-making, there are important aspects of the reality of implementation of early intervention that should not be ignored when decisions are made. For these reasons the Early Intervention Foundation in its early work has sought to map out the evidence for early intervention programmes recognising the diversity of early intervention activities, the importance of innovation, and the crucial role of implementation. There is a danger that the recognition of wider complexity and uncertainty falsely undermines confidence in the well-established science of early intervention and of evaluation. However, we think the rewards of addressing gaps in evidence makes this risk worthwhile.
The work summarised here assessed programmes that aimed to improve child development by supporting the quality of parent-child interactions in early childhood. We have established that there are programmes available in the UK which have demonstrated the capability to positively shift child outcomes in rigorously controlled environments. There are also many other approaches, not mentioned in this paper, that may be promising but do not yet have rigorous evidence of impact. We find that programmes that focus on behaviour tend to have the strongest evidence of effectiveness. In our view, further innovation, testing and learning is required. Attachment represents another important area of innovation that has much potential, but where few programmes have yet demonstrated that the underpinning science can be translated into effective practice. Importantly, there is a paucity of evidence on long-run effectiveness several years from the point at which interventions are delivered.
We note that the evidence we summarise here is largely academic in nature. More work is required to understand how to translate this evidence and best apply it in the context of a particular population and system. The effectiveness of programmes in the real world will depend on features of the broader system, including implementation, workforces, assessments, referrals and monitoring. This requires an integration of implementation science with economic thinking and other sciences, and new ways of understanding the messy frontier between practice, policy and evidence.
We hope that some gaps in knowledge will be addressed over time by improvements to the evidence base, with more studies over longer duration that make a clearer connection between ‘what works’ and ‘for whom’. This requires more evaluation, not just with the necessary standards of rigour for causality to be ascertained, but also with the types of data (e.g. linked administrative records) that facilitate long-run tracking of economic and fiscal outcomes in an affordable manner. While we have emphasised the importance of linking to fiscal costs for any discussion of benefits, it is important that estimates of potential benefits of early intervention are aware of the challenges involved in quantifying and ultimately achieving cashable fiscal benefits – tangible, in-year financial savings. Too often it is assumed that if children benefit then savings will result; much more sophisticated analysis is required.
It is also important to recognise that a diversity of outcomes generally implies a diversity of beneficiaries. We present here an attempt to move towards bridging the gap, showing the degree to which the costs of late intervention fall across different public agencies. In so doing, we hope to offer further stimulus to those agencies to come together in developing approaches that better match services to populations and reduce the collective costs of failure. Devolution is one such opportunity, but despite some promising pilots, as in Greater Manchester, there is not substantive evidence yet that the opportunity to reduce the mismatch of agencies, budgets and populations has been grasped.
Returning to the title of this paper, we have shown that the current state of the evidence in relation to these programmes is not yet mature and established enough to enable rigorous, place-specific cost-benefit analysis on a wide scale. Even the ‘gold-star’ evidence containing rigorous causal estimates of effectiveness, while a useful and important point of departure, provides an insufficient basis for estimation and modelling. In order to make a firm and credible connection to either of these, much more evaluation, modelling and analysis is required; we hope this stimulates further work by government, researchers and the EIF over the coming years.
Appendix
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table7.png?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200302061626415-0803:S0027950100003768:S0027950100003768-table8.png?pub-status=live)