Hostname: page-component-7b9c58cd5d-g9frx Total loading time: 0 Render date: 2025-03-14T02:09:51.489Z Has data issue: false hasContentIssue false

Screening Tools for Cognitive Impairment in Adults with Substance Use Disorders: A Systematic Review

Published online by Cambridge University Press:  26 August 2021

Katherine Y. Ko
Affiliation:
Department of Psychology, Macquarie University, Sydney, New South Wales, Australia
Nicole Ridley
Affiliation:
Drug and Alcohol Services, South Eastern Sydney Local Health District, Sydney, New South Wales, Australia
Shayden D. Bryce
Affiliation:
Centre for Youth Mental Health, The University of Melbourne, Parkville, Victoria, Australia Orygen, Parkville, Victoria, Australia
Kelly Allott
Affiliation:
Centre for Youth Mental Health, The University of Melbourne, Parkville, Victoria, Australia Orygen, Parkville, Victoria, Australia
Angela Smith
Affiliation:
Hunter New England Health Libraries, Hunter New England Local Health District, Newcastle, New South Wales, Australia
Jody Kamminga*
Affiliation:
Drug and Alcohol Clinical Services, Hunter New England Local Health District, Newcastle, New South Wales, Australia
*
*Correspondence and reprint requests to: Jody Kamminga, Drug and Alcohol Clinical Services, Hunter New England Local Health District, Level 3, 670 Hunter Street, Newcastle West, NSW 2304, Australia. E-mail: Jody.Kamminga@health.nsw.gov.au
Rights & Permissions [Opens in a new window]

Abstract

Objectives:

Cognitive impairment is common in individuals with substance use disorders (SUDs), yet no evidence-based guidelines exist regarding the most appropriate screening measure for use in this population. This systematic review aimed to (1) describe different cognitive screening measures used in adults with SUDs, (2) identify substance use populations and contexts these tools are utilised in, (3) review diagnostic accuracy of these screening measures versus an accepted objective reference standard, and (4) evaluate methodology of included studies for risk of bias.

Methods:

Online databases (PsycINFO, MEDLINE, Embase, and CINAHL) were searched for relevant studies according to pre-determined criteria, and risk of bias and applicability was assessed using the Quality Assessment of Diagnostic Accuracy Studies–2 (QUADAS–2). At each review phase, dual screening, extraction, and quality ratings were performed.

Results:

Fourteen studies met inclusion, identifying 10 unique cognitive screening tools. The Montreal Cognitive Assessment (MoCA) was the most common, and two novel screening tools (Brief Evaluation of Alcohol-Related Neuropsychological Impairments [BEARNI] and Brief Executive Function Assessment Tool [BEAT]) were specifically developed for use within SUD populations. Twelve studies reported on classification accuracy and relevant psychometric parameters (e.g., sensitivity and specificity). While several tools yielded acceptable to outstanding classification accuracy, there was poor adherence to the Standards for Reporting Diagnostic Accuracy Studies (STARD) across all studies, with high or unclear risk of methodological bias.

Conclusions:

While some screening tools exhibit promise for use within SUD populations, further evaluation with stronger methodological design and reporting is required. Clinical recommendations and future directions for research are discussed.

Type
Critical Review
Copyright
Copyright © INS. Published by Cambridge University Press, 2021

Substance use disorder (SUD) is a pathological cluster of cognitive, behavioural, and physiological symptoms stemming from compulsive use of a substance despite adverse effects on psychosocial functioning (American Psychological Association, 2013). Between 30% and 80% of individuals presenting for the treatment of SUDs experience some form of cognitive impairment (CI; Copersino et al., Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009), with neurophysiological changes occurring as a result of intoxication, withdrawal, or chronic use. Substance users frequently demonstrate deficits in attention, memory, executive functions, and decision-making (Ramey & Regier, 2020), with nature and chronicity of impairment differing between types of substances, quantity, and frequency of use (Bruijnen et al., Reference Bruijnen, Jansen, Dijkstra, Walvoort, Lugtmeijer, Markus and Kessels2019; Chen, Strain, Crum, & Mojtabai, Reference Chen, Strain, Crum and Mojtabai2013).

CI may also result from other comorbid factors. Psychiatric illness is highly prevalent in individuals with SUD (Gould, Reference Gould2010; Kelly & Daley, Reference Kelly and Daley2013) and can increase severity of health issues and complicate recovery (Daley & Moss, 2002; Grella, Hser, Joshi, & Rounds-Bryant, Reference Grella, Hser, Joshi and Rounds-Bryant2001). Other common comorbidities that can impact cognition include physical illness, such as vascular conditions (Volkow, Baler, Compton, & Weiss, Reference Volkow, Baler, Compton and Weiss2014), respiratory problems (Owen, Sutter & Albertson, Reference Owen, Sutter and Albertson2014), and liver damage (Brodersen et al., Reference Brodersen, Koen, Ponte, Sánchez, Segal, Chiapella and Lemberg2014). Between 37% and 66% of individuals with an acquired brain injury (ABI, e.g., due to falls, head trauma, overdose, or chronic substance use; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018) misuse alcohol, and 10%–44% use illicit substances (Parry-Jones, Vaughan, & Miles Cox, Reference Parry-Jones, Vaughan and Miles Cox2006), suggesting potential cumulative cognitive dysfunction. Developmental disorders (e.g., intellectual or learning disability, attention-deficit hyperactivity disorder) also co-occur with SUDs, with rates varying (between 1% and 45%; Levin, Reference Levin2007; Lin et al., Reference Lin, Balogh, McGarry, Selick, Dobranowski, Wilton and Lunsky2016; Salavert et al., Reference Salavert, Clarabuch, Fernández-Gómez, Barrau, Giráldez and Borràs2018). Many developmental disorders may remain undiagnosed at treatment (Gooden et al., Reference Gooden, Cox, Petersen, Curtis, Manning and Lubman2020), and it is possible that similar factors that form barriers to diagnosis and support for individuals with developmental disorders (i.e., social and environmental disadvantages; Carroll Chapman & Wu, Reference Carroll Chapman and Wu2012) are also risk factors for developing SUDs.

Regardless of aetiology, CI in individuals with SUDs can severely impede everyday functioning. The degree of executive dysfunction could determine how effectively individuals with SUDs are able to cope with day-to-day demands (see Verdejo-Garcia, Garcia-Fernandez, & Dom, Reference Verdejo-Garcia, Garcia-Fernandez and Dom2019). Impaired inhibitory control can contribute to behavioural difficulties, including increased impulsivity and poor decision-making (Ramey & Regier, 2020). CI also has impacts on the ability to engage successfully with SUD treatment (Bates, Bowden, & Barry, Reference Bates, Bowden and Barry2002), with individuals less likely to benefit from the therapeutic process or acquire effective strategies, related to diminished insight and motivation (Aharonovich, Amrhein, Bisaga, Nunes, & Hasin, Reference Aharonovich, Amrhein, Bisaga, Nunes and Hasin2008). CI is also associated with poorer treatment compliance (Copersino et al., Reference Copersino, Schretlen, Fitzmaurice, Lukas, Faberman, Sokoloff and Weiss2012). However, individuals with CI can benefit from treatment, particularly when treatment is appropriately matched to cognitive level (Bates, Buckman, & Nguyen, Reference Bates, Buckman and Nguyen2013). For instance, cognitive-behavioural approaches are geared to enhance higher-order cognitive processes related to self-awareness and decision-making. These cognitive skills may be fundamentally reduced in individuals with CI (Hagen et al., Reference Hagen, Erga, Hagen, Nesvåg, McKay, Lundervold and Walderhaug2016), who may benefit more from interactional programmes with reduced cognitive demands (Secades-Villa, García-Rodríguez, & Fernández-Hermida, Reference Secades-Villa, García-Rodríguez and Fernández-Hermida2015) or increased involvement of social supports (Buckman, Bates, & Cisler, Reference Buckman, Bates and Cisler2007; Buckman, Bates, & Morgenstern, Reference Buckman, Bates and Morgenstern2008).

Given the impact of CI on SUD treatment, accurate and timely detection of CI is essential. The wide range of factors contributing to CI in individuals presenting for SUD treatment, however, results in heterogeneous cognitive profiles that vary in their clinical presentation, making CI difficult to detect (Horner, Harvey, & Denier, Reference Horner, Harvey and Denier1999). Consequently, individuals with CI presenting for SUD treatment might not receive the most appropriate support (Braatveit, Torsheim, & Hove, Reference Braatveit, Torsheim and Hove2018).

While neuropsychological assessment is the gold standard in evaluating cognitive functioning (Roebuck-Spencer et al., Reference Roebuck-Spencer, Glen, Puente, Denney, Ruff, Hostetter and Bianchini2017), this is not always feasible as it is time-consuming (taking up to several hours) and requires extensive training in administration and interpretation (Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018; Roebuck-Spencer et al., Reference Roebuck-Spencer, Glen, Puente, Denney, Ruff, Hostetter and Bianchini2017). According to the 2014 American Psychological Association Working Group on Screening and Assessment (APA–WGSA) guidelines, cognitive screening tools are used for early identification of individuals at high risk for a specific disorder, are generally brief, monitor treatment progress or change in symptoms over time, can be administered by clinicians with appropriate training, and are not definitively diagnostic (APA–WGSA, 2014). Cognitive screening, therefore, represents a critical preliminary stage of triaging those in need of more comprehensive neuropsychological assessment (Roebuck-Spencer et al., Reference Roebuck-Spencer, Glen, Puente, Denney, Ruff, Hostetter and Bianchini2017).

Cognitive screening is routinely practiced in primary and tertiary settings in the assessment of adults with mild CI, dementia, ABI, and psychiatric illness (Roebuck-Spencer et al., Reference Roebuck-Spencer, Glen, Puente, Denney, Ruff, Hostetter and Bianchini2017); however, it is much less established as a standard clinical process for individuals with SUD. Cognitive screening tools should be chosen carefully, with reference to both clinical utility and psychometric properties for the population in which it will be used (Slater & Young, Reference Slater and Young2013). The measure should have high sensitivity (to correctly detect those with CI) and good specificity (to accurately rule out those without CI; Bunnage, Reference Bunnage and Bowden2017). If specificity is too low, clients may inaccurately be perceived as needing full assessment, with potential for resource overload as well as unnecessary alarm for the individual. Yet, if sensitivity is too low, clients that would benefit from full assessment may not receive this, leading to a potential mismatch in treatment provision.

There exists limited evidence to guide decision-making on which cognitive screening tool is appropriate for use with individuals presenting for the treatment of SUDs. The most widely used cognitive screening tools, the Mini-Mental State Examination (MMSE; Folstein, Folstein, & McHugh, Reference Folstein, Folstein and McHugh1975) and the Montreal Cognitive Assessment (MoCA; Nasreddine et al., Reference Nasreddine, Phillips, Bédirian, Charbonneau, Whitehead, Collin and Chertkow2005), have most consistently been used in older adult populations (e.g., for stroke and dementia screening). It remains unclear how well these measures can detect CI in individuals with SUDs, or whether there are more appropriate screening tools for use in this population. Given the impact of CI on SUD treatment outcomes, there is a clear need for evidence-based review and recommendations around the use of cognitive screening tools in adults engaging in SUD treatment.

Objective

The aims of this systematic review were to: (1) describe types of cognitive screening measures used in adults with SUDs; (2) identify specific substance use populations and settings these tools are utilised in; (3) review diagnostic accuracy of these screening measures in detecting CI versus an accepted objective reference standard; and (4) evaluate methodology of these studies for risk of bias. It was hoped this review would yield clinical recommendations for use of cognitive screening in individuals presenting for the treatment of SUDs in clinical settings, as well as identify areas for future research in this field.

METHOD

Search Strategy

The Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) statement guided the reporting standards of this review (Moher, Liberati, Tetzlaff, Altman, & The PRISMA Group, Reference Moher, Liberati, Tetzlaff and Altman2009). Results of systematic review registry searches (e.g., PROSPERO) did not reveal similar systematic reviews as of May 2020. The review protocol was registered with PROSPERO on 7 November 2020 (ID: CRD42020185902; www.crd.york.ac.uk/prospero).

A systematic search of the Medline, Embase, PsycINFO (OVID), and CINAHL (EBSCO) databases was undertaken in June 2019 and updated in May 2020. The search strategy was constructed around the aims of the review. Subject headings, keywords, and keyword phrases were developed for each of the search concepts, substance use disorder, cognition, and screening or assessment. Subject headings and keywords for each concept were combined using the Boolean operator ‘OR’ before the concepts were bought together using the ‘AND’ operator. The search strategy was developed by a senior research librarian (AS) and was designed in Medline before being translated to the other databases. Searches were limited to adult populations and studies in English. No date limits were applied. Comments, editorials, letters, news, book, and book chapters as well as conference abstracts were excluded (see Supplementary Material 1 for full search strategies). In addition, supplementary searching was undertaken in Google Scholar using a combination of search terms for the first 200 results. The reference lists of all papers that met the final inclusion criteria were screened for any additional relevant citations not captured in the database searches. Screening for articles prepared for publication but not yet in print was performed through a search of above search terms of the Psychology Preprint Archive (psyarxiv.com); an author known to be developing a measure in this field was also contacted for potential preprints. A comprehensive grey literature search was not conducted. Results were exported to EndNote and duplicates removed. Studies were then exported to Covidence (Veritas Health Innovation; www.covidence.org) for screening.

Inclusion Criteria

The following inclusion criteria were set: (1) original empirical article, published or prepared for publication (i.e., under review); (2) in English; (3) reported on a tool designed to screen for cognitive functioning, or the tool was used for this purpose (in isolation or embedded as part of a broader screening process); (4) examined or provided data that supported inferences regarding criterion validity of the screening tool compared to an accepted standard criterion of cognitive functioning (e.g., neuropsychological assessment); (5) participants were adults (aged 18 years and above); and (6) participants were reported to have some form of SUD (inclusive of alcohol-use or other specific disorder types). The standards described by the 2014 APA–WGSA statement guided screening tool eligibility, specifying they must: (1) be brief to administer; (2) collect objective test data with an established protocol for scoring; (3) require minimal, non-specialist training; and (4) not used for the purpose of definitive diagnosis. To maximise eligible studies, screening tools were not excluded if they also collected functional or self-report information, or focused on one cognitive domain. No restriction on type of criterion validity analysis was applied (e.g., correlations, regression, receiver operating curve).

Screening, Extraction, and Risk of Bias

Two authors (JK and KK) independently screened titles and abstracts from the database searches for articles meeting inclusion criteria (Phase 1). JK and KK independently reviewed full-text articles for eligibility using the same review criteria (Phase 2). If there was disagreement between authors on eligibility for an article, a third author (NR) was consulted and decisions were resolved by consensus. Additional studies retrieved from correspondence or Google Scholar searches were included if author consensus (JK, KK, and NR) indicated they met inclusion criteria.

Data extraction from articles meeting criteria was conducted by JK and KK, with a check by author NR. Variables of interest included target population, study design, recruitment method, participant demographics, index test, reference standard, and relevant psychometric data (e.g., sensitivity, specificity, likelihood ratios, and correlation coefficients).

Included studies were also assessed for risk of bias and applicability (JK and KK) using the revised Quality Assessment of Diagnostic Accuracy Studies (QUADAS–2) tool according to scoring guidelines presented in Whiting and colleagues (Reference Whiting, Rutjes, Westwood, Mallett, Deeks and Reitsma2011). Data extraction, risk of bias, and applicability ratings followed the same dual, independent examination and consensus procedures as described for article screening. NR was not involved in reviewing her respective paper (Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018); instead, SB was consulted for consensus decisions.

Risk of bias assessment using QUADAS–2

The QUADAS–2 assessment tool (Whiting et al., Reference Whiting, Rutjes, Westwood, Mallett, Deeks and Reitsma2011) evaluated risk of bias and applicability of diagnostic accuracy studies (www.quadas.org). Guided by the Standards for Reporting Diagnostic Accuracy Studies (STARD) criteria, this scale was developed to allow more transparency in reporting (Cohen et al., Reference Cohen, Korevaar, Altman, Bruns, Gatsonis, Hooft and Bossuyt2016; Whiting et al., Reference Whiting, Rutjes, Westwood, Mallett, Deeks and Reitsma2011) and is recommended by the Cochrane Collaboration (www.cochrane.org) and National Institute for Health and Clinical Excellence (NICE; www.nice.org.uk).

The QUADAS–2 rating process has four phases: (1) summarise review question; (2) tailor the tool to the review; (3) construct a flow diagram for the study; and (4) assess risk of bias and applicability concerns. The tool consists of four key domains: (1) patient selection; (2) index test; (3) reference standard; and (4) flow and timing (i.e., flow of participants through the study, timing of administration of index test and reference standard). Each domain is assessed for risk of bias, the first three also for applicability concerns. Signalling questions guided decision-making on high or low risk of bias (Schueler, Schuetz, & Dewey, Reference Schueler, Schuetz and Dewey2012), rated as ‘Yes’, ‘No’, or ‘Unclear’. Risk of bias is deemed ‘Low’ if all relevant signalling questions are rated ‘Yes’; however, bias may exist if these questions are rated as ‘No’ or ‘Unclear’. If questions relating to concerns regarding applicability are rated as ‘High’ or ‘Unclear’, aspects of the study may not match the review question.

Statistical Interpretation

Sensitivity, the ability of a screening tool to correctly identify individuals with a condition of interest, and specificity, the ability of a screening tool to detect individuals who do not have a condition, are relative to a reference standard. A receiver operating characteristic (ROC) curve can be plotted by using sensitivity against the false-positive rate (1–specificity) for given cut-off points of a screening test. An area under ROC curve (AUC) value summarises overall test diagnostic accuracy (Trevethan, Reference Trevethan2017), ranging from 0 to 1, where 0 and 1 represent perfectly inaccurate or accurate tests, respectively. As a guide, AUC values of 0.5 indicate no meaningful discrimination ability, 0.7 to 0.8 acceptable, 0.8 to 0.9 excellent, and >0.9 outstanding (Mandrekar, Reference Mandrekar2010).

Sensitivity and specificity are used to obtain predictive power and likelihood ratios. Positive predictive power describes the probability that an individual with a positive test result is a true positive; negative predictive power indicates the probability an individual who obtains a negative test result is a true negative. Both are dependent on the prevalence of the target condition in each population sample (Bunnage, Reference Bunnage and Bowden2017; Trevethan, Reference Trevethan2017).

A positive likelihood ratio (LR+) refers to the ratio between the probability of correctly classifying an individual as having CI and the probability of incorrectly classifying an individual who does not have CI (i.e., sensitivity/100–specificity). A LR+ greater than 10 indicates CI is highly likely when a positive test result is yielded. A negative likelihood ratio (LR–) is the ratio between the probability of incorrectly classifying an individual as having CI and the probability of correctly classifying an individual who does not have CI (i.e., 100–sensitivity/specificity). A LR– below 0.1 indicates that impairment is highly unlikely when a negative test result is obtained (McGee, Reference McGee2002).

For studies that examined correlations between cognitive screening tools and reference tests, coefficients of 0.1, 0.3, and 0.5 indicated small, medium, and large relationships, respectively (Cohen, Reference Cohen1988).

RESULTS

Search Results

Figure 1 presents a flow diagram of search results. Searches of electronic databases of published articles identified 5,135 candidate studies. Removal of 1,672 duplicates resulted in 3,463 studies. An additional study identified from a Google Scholar search and another through correspondence with an author brought the total to 3,465 studies. A search of the Psychology Preprint Archive yielded no additional studies. After screening the titles and abstracts (Phase 1), the number of studies was reduced to 35. Examination of full-text versions of eligible studies identified 14 relevant papers. Full-text articles were excluded for not meeting the following criteria: not appropriate study design (e.g., did not evaluate screening tool against a standard criterion of cognitive functioning); did not meet 2014 APA–WGSA screening criteria; and did not include participants with SUDs. One duplicate was also removed during this phase.

Fig. 1. PRISMA flow diagram.

Study Characteristics

Tables 13 present characteristics of the 14 included studies and an overview of the screening measures and reference standards. Studies were conducted in the United States of America (n = 3), France (n = 3), the Netherlands (n = 2), Spain (n = 1), Norway (n = 1), Brazil (n = 1), Belgium (n = 1), and Australia (n = 2). Across all 14 studies, the average clinical sample size was 88.33 (SD = 116.75); however, the range between studies varied widely from 20 to 501 participants. The majority of studies (n = 11; 78.57%) made reference to formal diagnostic criteria (e.g., Diagnostic and Statistical Manual of Mental Disorders) to operationalise SUD. Over a quarter (n = 4) included a sample with a single problematic substance (Ewert et al., Reference Ewert, Pelletier, Alarcon, Nalpas, Donnadieu-Rigole, Trouillet and Perney2018; Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013). All studies included participants with alcohol use disorder, nearly two-thirds also included those with cannabis use (n = 9; 64.29%), followed by cocaine use (n = 7; 50.00%), then opioids (n = 6; 42.86%). Only one study failed to report the specific substance of misuse (To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015). Over two-thirds (n = 8; 57.14%) required a period of abstinence before engaging in cognitive assessment, while three studies did not require any period of abstinence (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018; Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017) and one study did not specify abstinence status (Wester et al., Reference Wester, Westhoff, Kessels and Egger2013). Nearly, half of studies (n = 6; 46.15%) were in an inpatient setting, two (15.38%) outpatient, four (30.77%) residential rehabilitation, and two (14.29%) did not report the setting.

Table 1. Study characteristics of articles included for review

Note: Only clinical sample characteristics are presented here. As Ridley et al. (Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018) and Wester et al. (Reference Wester, Westhoff, Kessels and Egger2013) combined clinical and control data when calculating accuracy statistics, both groups are presented in this table. DSM–III–R, Diagnostic and Statistical Manual–Third Edition–Revised; DSM–IV, Diagnostic and Statistical Manual–Fourth Edition; DSM–IV–TR, Diagnostic and Statistical Manual–Fourth Edition–Text-Revised; DSM–5, Diagnostic and Statistical Manual–Fifth Edition; ICD–10, International Statistical Classification of Diseases and Related Health Problems–10th Revision; NR, not reported; SD, standard deviation; WAIS–III, Wechsler Adult Intelligence Scale–Third Edition; WAIS–IV, Wechsler Adult Intelligence Scale–Fourth Edition.

Table 2. Overview of cognitive screening measures

M, mean; SD, standard deviation; SUD, substance use disorder.

Table 3. Overview of screening measures and reference standards for included studies

ACE–R, Addenbrooke’s Cognitive Examination–Revised; ACLS–5, Allen Cognitive Level Screen–Fifth Edition; BCSE, Brief Cognitive Status Examination; BEARNI, Brief Evaluation of Alcohol-Related Neuropsychological Impairments; BEAT, Brief Executive Function Assessment Tool; CI, cognitive impairment; CST–14, Cognitive Screening Test–14; CVLT, California Verbal Learning Test; CVLT–II, California Verbal Learning Test–Second Edition; DART, Dutch Adult Reading Test; D–KEFS TMT, Delis–Kaplan Executive Function System Trail Making Test; DSB, Digit Span Backward; DSF, Digit Span Forward; DSS, Digit Span Sequencing; FAB, Frontal Assessment Battery; FCSRT, Free and Cued Selective Reminding Test; FPT, Five-Point Test; FSIQ, Full Scale Intelligence Quotient; GMI, General Memory Index; HASI, Hayes Ability Screening Index; LNS, letter–number sequencing; LOTCA, Loewenstein Occupational Therapy Cognitive Assessment; MAE, Multilingual Aphasia Exam; MMSE, Mini-Mental State Examination; MoCA, Montreal Cognitive Assessment; NAB–SM, Neuropsychological Assessment Battery–Screening Module; NART, National Adult Reading Test; NCSE, Neurobehavioural Cognitive Status Examination; NPA, neuropsychological assessment; NR, not reported; NSB, Neuropsychological Screening Battery; RAVLT, Rey Auditory Verbal Learning Test; RBANS, Repeatable Battery for the Assessment of Neuropsychological Status; RBMT–3, Rivermead Behavioural Memory Test–Third Edition; RFFT, Ruff Figural Fluency Test; ROCF, Rey–Osterrieth Complex Figure; SCWT, Stroop and Colour Word Test; SD, standard deviation; SDMT, Symbol Digit Modalities Test; SUD, substance use disorder; TMT, Trail Making Test; TMT A, Trail Making Test-Part A; TMT B, Trail Making Test-Part B; TMT B–A, Trail Making Test Part A minus Part B; Vineland II, Vineland Adaptive Behaviour Scales–Second Edition; WAIS–R, Wechsler Adult Intelligence Scale–Revised; WAIS–III, Wechsler Adult Intelligence Scale–Third Edition; WAIS–IV, Wechsler Adult Intelligence Scale–Fourth Edition; WAIS–IV–NL, Wechsler Adult Intelligence Scale–Fourth Edition (Dutch version); WCST, Wisconsin Card Sorting Test; WMS, Wechsler Memory Scale; WMS–IV, Wechsler Memory Scale–Fourth Edition.

As per inclusion criteria, each study that was included compared an index test (i.e., the cognitive screening measure) to a reference standard (i.e., neuropsychological assessment). In over a quarter of studies (28.57%), a particular syndrome of interest (e.g., Korsakoff’s syndrome, intellectual disability) influenced test selection for the reference standard. The index tests and reference standards are presented in Tables 2 and 3.

The majority of studies (n = 12; 85.57%) examined only one screening tool. Across all 14 studies, 10 different index measures were used, with the most common being the MoCA (n = 6; 46.15%). Most (n = 8; 80.00%) were developed for use within other populations (e.g., dementia, intellectual disability, brain injury); however, two were developed specifically for use within a SUD population (Brief Evaluation of Alcohol-Related Neuropsychological Impairments [BEARNI] and Brief Executive Function Assessment Tool [BEAT]). Length of administration for cognitive screening measures varied between 10 and 45 minutes. Almost all index measures targeted multiple cognitive domains, including: orientation (n = 5; 35.71%), attention (n = 6; 42.86%), language (n = 6; 42.86%), visual-spatial abilities (n = 8; 57.14%), memory (n = 7; 50%), motor ability (n = 3; 21.43%), and executive functions (n = 6; 42.86%). The Frontal Assessment Battery (FAB) focused on one cognitive domain (executive functioning).

Across all 14 studies, reference standards varied in cognitive domains and specific neuropsychological tests. While the reference standard largely encompassed a range of cognitive domains, two studies (14.28%) utilised a reference standard focusing on one cognitive domain (i.e., executive functioning, memory; Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013). Another study (Cunha et al., Reference Cunha, Nicastri, de Andrade and Bolla2010) restricted the reference standard to attention, working memory, and executive function. Other variations across studies included time between administration of index and reference standard (ranging from the same day to three months later), and operationalisation of ‘impairment’ for the reference standard.

Screening Measure Utility

Diagnostic accuracy

Table 4 displays psychometric data provided by 12 studies that reported classification accuracy (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Braatveit et al., Reference Braatveit, Torsheim and Hove2018; Bruijnen et al., Reference Bruijnen, Jansen, Dijkstra, Walvoort, Lugtmeijer, Markus and Kessels2019; Copersino et al., Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009; Ewert et al., Reference Ewert, Pelletier, Alarcon, Nalpas, Donnadieu-Rigole, Trouillet and Perney2018; Fals-Stewart, Reference Fals-Stewart1997; Grohman & Fals-Stewart, Reference Grohman and Fals-Stewart2004; Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013). While all of these studies provided sensitivity and specificity data for at least one defined cut-point, two did not provide overall AUC data (Fals-Stewart, Reference Fals-Stewart1997; Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018). Although several studies included control groups in the overall study, two included control groups in diagnostic accuracy data (Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013).

Table 4. Psychometric data retrieved from primary classification accuracy studies

ACE–R, Addenbrooke’s Cognitive Examination–Revised; BEARNI, Brief Evaluation of Alcohol-Related Neuropsychological Impairments; BEAT, Brief Executive Function Assessment Tool; FAB, Frontal Assessment Battery; GMI, General Memory Index; HASI, Hayes Ability Screening Index; IQ, intellectual quotient; LOTCA, Loewenstein Occupational Therapy Cognitive Assessment; MMSE, Mini-Mental State Examination; MoCA, Montreal Cognitive Assessment; MoCA–D, Montreal Cognitive Assessment (Dutch version); NAB–SM, Neuropsychological Assessment Battery–Screening Module; NCSE, Neurobehavioural Cognitive Status Examination; NPA, europsychological assessment; NR, not reported; NSB, Neuropsychological Screening Battery; WAIS–III, Wechsler Adult Intelligence Scale–Third Edition; WAIS–IV, Wechsler Adult Intelligence Scale–Fourth Edition.

Of the studies that provided overall AUC data, the MoCA ranged from acceptable to outstanding in the detection of CI (range: 0.75–0.96). The single study examining the Addenbrooke’s Cognitive Examination–Revised (ACE–R; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018) revealed excellent discriminative ability (AUC = 0.85), as did the Hayes Ability Screening Index (HASI; AUC = 0.94–0.95) across two studies (Braatveit et al., Reference Braatveit, Torsheim and Hove2018; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015), and the Neuropsychological Assessment Battery–Screening Module (NAB–SM; AUC = 0.88; Fals-Stewart, Reference Fals-Stewart1997). The MMSE by comparison had lower accuracy that ranged from no discrimination ability to acceptable (AUC = 0.21–0.79) across two studies (Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015). For index measures specifically developed for SUD populations (BEAT and BEARNI), at least excellent accuracy in detecting CI was reported (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015).

Half of studies that reported diagnostic accuracy data (n = 6) included Positive Predictive Power and Negative Predictive Power; only two studies reported positive and negative likelihood ratios (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Copersino et al., Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009). Berry and colleagues (Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021) was the sole study that reported classification accuracy as well as predictive power and likelihood ratios. Using a LR+ of ≥10 and LR– of ≤0.1 as threshold (McGee, Reference McGee2002), a score of ≤14 on the BEAT (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021) was needed to rule in likely impairment, while a score of ≥31 was needed to likely rule out impairment. Copersino et al. (Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009) indicated that a cut-score of ≤22 on the MoCA had strong evidence for ruling in CI and a score of ≥27 had strong evidence.

Correlational data

More than half of the 14 studies (n = 8) examined correlations between index tests and neuropsychological measures, with most (n = 6) presenting this in addition to classification accuracy data. Two studies (Cunha et al., Reference Cunha, Nicastri, de Andrade and Bolla2010; Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017) provided correlational data only. As outlined in Table 5, various correlations were performed, including: index total summary scores as compared to reference standard total summary score; index domain score as compared to corresponding reference standard domain scores; and total index score versus specific test outcome on the reference standard. One study (Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018) combined control and clinical groups to produce correlational data, while all others used clinical samples only. The largest correlations reported were between total summary scores derived from a reference standard (i.e., the Repeatable Battery for the Assessment of Neuropsychological Status; RBANS) and total scores on the MoCA (r = 0.75) and the ACE–R (r = 0.81; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018). Similarly, large correlations were observed between HASI total score and Wechsler Adult Intelligence Scale–Third Edition Full Scale IQ (r = 0.69–0.70; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015). For the MoCA, medium to large correlations between MoCA total score and neuropsychological tasks assessing executive function (r = 0.32–0.46), learning and memory (r = 0.45–0.56), working memory (r = 0.27–0.29), and fluency (r = 0.37–0.53) were found (Ewert et al., Reference Ewert, Pelletier, Alarcon, Nalpas, Donnadieu-Rigole, Trouillet and Perney2018). MoCA domain scores also yielded medium to large correlations to corresponding neuropsychological domains (Bruijnen et al., Reference Bruijnen, Jansen, Dijkstra, Walvoort, Lugtmeijer, Markus and Kessels2019; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018), with the exception of the domains of language and attention/orientation items (Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018). For the ACE–R, Ridley and colleagues (Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018) reported medium to large correlations between reference domains and relevant ACE–R subscales, with the exception of language. For the FAB, Cuhna and colleagues (Reference Cunha, Nicastri, de Andrade and Bolla2010) reported medium to large correlations (r = 0.41–0.57) between FAB tasks and performance on specific executive function measures; however, not all correlations were significant (r = 0.03–0.27). For the Loewenstein Occupational Therapy Cognitive Assessment (LOTCA), medium to large correlations between index and reference tasks were found for measures of visual-perceptual abilities and psychomotor abilities (Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017). However, reference measures assessing executive functioning and memory did not correspond consistently with LOTCA subscales. For the BEAT, there were medium to large correlations between overall score and performance on reference tasks measuring executive functioning (r = 0.43–0.57; Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021).

Table 5. Summary of studies examining correlations between cognitive screening tools and comprehensive cognitive assessment

* Included if at least significant at the .05 level or higher.

ACE–R, Addenbrooke’s Cognitive Examination–Revised; ACLS–5, Allen Cognitive Level Screen–Fifth Edition; BCSE, Brief Cognitive Status Examination; BEAT, Brief Executive Function Assessment Tool; DSF, Digit Span Forward; DSS, Digit Span Sequencing; FAB, Frontal Assessment Battery; FDT, Five-Digit Test; FPT, Five-Point Test; FSIQ, Full Scale Intelligence Quotient; HASI, Hayes Ability Screening Index; LOTCA, Loewenstein Occupational Therapy Cognitive Assessment; MMSE, Mini-Mental State Examination; MoCA, Montreal Cognitive Assessment; NAB–SM, Neuropsychological Assessment Battery–Screening Module; NPA, neuropsychological assessment; PSI, Processing Speed Index; RBANS, Repeatable Battery for the Assessment of Neuropsychological Status; ROCF, Rey-Osterrieth Complex Figure; TMT, Trail Making Test; SCWT, Stroop and Colour Word Test; WAIS–III, Wechsler Adult Intelligence Scale–Third Edition; WCST, Wisconsin Card Sorting Test; WMS–IV, Wechsler Memory Scale–Fourth Edition.

Quality

Ratings of risk of bias, applicability, and quality of study design are presented in Table 6. Almost all studies were judged to have a high or unclear risk of bias in terms of patient selection (QUADAS–2 Domain 1;92.86%) due to convenience or unclear sampling methods (57.14%; Braatveit et al., Reference Braatveit, Torsheim and Hove2018; Bruijnen et al., Reference Bruijnen, Jansen, Dijkstra, Walvoort, Lugtmeijer, Markus and Kessels2019; Cunha et al., Reference Cunha, Nicastri, de Andrade and Bolla2010; Copersino et al., Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009; Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015, To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013); inappropriate or unclear exclusions (57.14%; Cunha et al., Reference Cunha, Nicastri, de Andrade and Bolla2010; Ewert et al., Reference Ewert, Pelletier, Alarcon, Nalpas, Donnadieu-Rigole, Trouillet and Perney2018; Fals-Stewart, Reference Fals-Stewart1997; Grohman & Fals-Stewart, Reference Grohman and Fals-Stewart2004; Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015; Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017, Wester et al., Reference Wester, Westhoff, Kessels and Egger2013); inclusion of a control group in analyses (14.29%; Ridley et al., Reference Ridley, Batchelor, Draper, Demirkol, Lintzeris and Withall2018; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015); or a combination of these factors. Nearly half of the included studies (n = 6;42.86%) had the main aim of validating the index measure within a SUD population (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Braatveit et al., Reference Braatveit, Torsheim and Hove2018; Copersino et al., Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015; Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015). Half of the studies had high or unclear risk of bias for the index test (QUADAS–2 Domain 2; 50.00%), while a greater number had a high or unclear risk of bias for the reference standard (QUADAS–2 Domain 3; 71.43%). The high risk was largely due to the non-blinding of index or reference test results; only four studies (Ewert et al., Reference Ewert, Pelletier, Alarcon, Nalpas, Donnadieu-Rigole, Trouillet and Perney2018; Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018; Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013) used different individuals to administer the index and reference measures, while in another seven the order of administration or prior knowledge of test results by the tester was unclear (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Braatveit et al., Reference Braatveit, Torsheim and Hove2018; Bruijnen et al., Reference Bruijnen, Jansen, Dijkstra, Walvoort, Lugtmeijer, Markus and Kessels2019; Cunha et al., Reference Cunha, Nicastri, de Andrade and Bolla2010; Fals-Stewart, Reference Fals-Stewart1997; Grohman & Fals-Stewart, Reference Grohman and Fals-Stewart2004; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015). For participant flow and timing (QUADAS–2 Domain 4), nearly half of included studies had either high or unclear risk of bias (42.86%) due to inappropriate or unclear time interval between the administration of the index test and the reference standard (n = 2), participants not all receiving the reference standard (n = 3), or not accounting for missing participant data, either clinically or statistically. Overall, quality rating results based on QUADAS–2 indicated generally poor adherence to STARD criteria across the included studies. Two studies (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Bruijnen et al., Reference Bruijnen, Jansen, Dijkstra, Walvoort, Lugtmeijer, Markus and Kessels2019) were not rated as high risk across any domains; however, unclear ratings meant that risk was unable to be fully assessed.

Table 6. QUADAS–2 ratings for studies included in the review

QUADAS–2, Quality Assessment of Diagnostic Accuracy Studies – Version 2; HR, high risk; LR, low risk; ? = unclear risk.

By contrast, the majority of studies met applicability criteria for the current review. The exception was in relation to the reference standard (QUADAS–2 Domain 3) across two studies (Braatveit et al., Reference Braatveit, Torsheim and Hove2018; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015). These studies sought to evaluate presence of intellectual disability within SUD populations, and consequently, it is possible the reference standards could fail to detect CI specifically related to SUD, or CI more broadly in this population.

DISCUSSION

This review had four primary aims: (1) describe types of cognitive screening measures used in adults with SUDs; (2) identify specific SUD population and settings in which these tools are utilised; (3) review diagnostic accuracy of these screening measures in relation to detecting CI versus an accepted objective reference standard; and (4) evaluate methodology of these studies to determine risk of bias.

Aims 1 and 2: Overview of Cognitive Screening Measures in Adults with SUDs

The current review, to the authors’ knowledge, is the first systematic overview of cognitive screening tools utilised in adult SUD populations as compared to a reference standard (e.g., neuropsychological assessment). Overall, 14 studies were identified which broadly aligned with definitions of cognitive screening as proposed by Roebuck-Spencer and colleagues (Reference Roebuck-Spencer, Glen, Puente, Denney, Ruff, Hostetter and Bianchini2017). In summary, more than three-quarters of studies made reference to specific SUD diagnostic criteria. The majority included individuals with poly-substance use, while the remaining specifically examined alcohol use. Alcohol was the most common substance used by participants, with a period of abstinence also frequently endorsed, and nearly half of studies were conducted in an inpatient setting. However, a number of studies failed to report key study details including study setting, specific substance of misuse, and abstinence status. Sample sizes were generally small, with the exception of Berry and colleagues (Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021), who commendably captured a large sample.

Across the 14 included studies, 10 different cognitive screening tools were evaluated. The majority of studies examined only one measure, and all were paper and pen tasks. The range of cognitive domains and tasks assessed by the screening tool and length of administration varied widely across studies. Most studies sought to validate existing cognitive screening tools in this new population, primarily the MoCA, with two seeking to validate new measures for specific use in SUD populations (BEAT and BEARNI). In over a quarter of studies, a particular syndrome of interest was evaluated within a SUD sample, with all other studies seeking to identify CI more generally.

Aim 3: Classification Accuracy

Out of the 14 retrieved studies, 12 reported on classification accuracy data. All reported sensitivity and specificity that varied widely (range: 36%–100%, 2%–100%, respectively). A further two used correlational data to assess associations between screening tools and reference measures. Of the 10 studies that provided overall AUC data, the six that related to the MoCA indicated an acceptable to outstanding ability to detect CI versus respective reference standards. By contrast, the MMSE, another well-known cognitive screening measure, yielded much lower accuracy that ranged from no discrimination to acceptable across two studies.

Although the HASI exhibited excellent discriminant ability (Braatveit et al., Reference Braatveit, Torsheim and Hove2018; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015), this measure was designed to screen specifically for intellectual disability and the reference standards in both instances were restricted to an intellectual disability diagnosis or intellectual cognitive domains. Therefore, classification accuracy of the HASI for detecting CI more broadly in individuals with SUDs remains unknown.

Berry and colleagues (Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021) reported the most extensive psychometric data on the BEAT, a measure specifically developed for use in those with SUDs. The BEAT demonstrated excellent discriminative ability and good sensitivity and specificity at the recommended cut-score of ≤24. However, low prevalence of CI in this group (7.8%) contributed to a low positive predictive power (Abu-Akel, Bousman, Skafidas, & Pantelis, Reference Abu-Akel, Bousman, Skafidas and Pantelis2018). The authors suggest this was due to using a conservative reference standard (i.e., impairment across three tests); however, it is also possible that this relates to sample characteristics (residential participants only), as well as choice of the reference battery (executive tasks only). Given that the prevalence rate in this study is much lower than that in other treatment groups and the SUD literature more broadly (e.g., Copersino et al., Reference Copersino, Fals-Stewart, Fitzmaurice, Schretlen, Sokoloff and Weiss2009), further evaluation of the psychometric properties of the BEAT in broader SUD populations is necessary.

Another cognitive screening tool specifically designed for SUDs, the BEARNI, exhibited good discriminative abilities in the original study but a follow-up analysis demonstrated that despite good sensitivity, the specificity of the tool was poor (Pelletier et al., Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018). To this end and given superior MoCA data, Pelletier and colleagues (Reference Pelletier, Alarcon, Ewert, Forest, Nalpas and Perney2018) suggest that the MoCA may be a more suitable choice for routine use.

Less than half of studies provided predictive value data for their screening tools. These data take into account the prevalence of CI in the population of interest. Including this information is crucial as in the event of a mismatch between the prevalence of CI in the screening sample and in individuals attending SUD treatment, classification data may misrepresent a tool’s screening ability (Slater & Young, Reference Slater and Young2013). Of the studies that provided predictive data, the rate of CI across samples varied widely (8%–86%), which is unsurprising given the broad range of study settings and samples. Clinicians must be mindful of the expected prevalence of CI in their particular cohort, as screening tools examined in groups with different prevalence rates may differentially capture CI in their cohort (Molinaro, Reference Molinaro2015).

Two studies relied on correlational data to examine screening measures in individuals with SUD. Medium to large correlations were reported between index and reference measures on the FAB (Cunha et al., Reference Cunha, Nicastri, de Andrade and Bolla2010) and LOTCA (Rojo-Mota et al., Reference Rojo-Mota, Pedrero-Pérez, Ruiz-Sánchez de León, León-Frade, Aldea-Poyo, Alonso-Rodríguez and Morales-Alonso2017). However, not all domains corresponded, and no further classification data were provided to support these associations, resulting in only indirect inferences regarding criterion validity. As such, further evaluation of these screening measures is required. There were more consistent medium to large correlations between MoCA scores (total and domain) and reference measures, providing initial evidence for construct validity. The ACE–R, BEAT, and HASI also had promising correlational data between index and reference measures; however, these findings need to be replicated.

Aim 4: Risk of Bias

This review represents the first time that the methodological quality of studies evaluating cognitive screening tools in SUDs have been systematically reviewed. The studies included in this review were assessed overall to have high or unclear risk of bias according to the QUADAS–2, suggesting significant methodological limitations or insufficient reporting. Patient selection represented the most problematic domain with high or unclear risk of bias occurring in 85.71% of studies. This was primarily due to the use of convenience sampling methods and overly stringent exclusion criteria. This included exclusion of individuals who had a history of psychiatric illness, had head injuries, or were on opioid treatment programmes. As individuals with SUDs present clinically with a range of comorbid and co-occurring conditions (see Gooden et al., Reference Gooden, Cox, Petersen, Curtis, Manning and Lubman2020), excluding participants based on these factors has the potential to significantly limit translation of study findings to applied clinical settings.

An additional source of high or unclear risk of bias related to whether index measures were interpreted without the knowledge of the reference standard (i.e., blind ratings), or vice versa, an issue in almost three-quarters of studies. Administration of both measures by the same individual is problematic, as it can inadvertently lead to bias in administration and scoring of measures. This is a clear area in need of improvement in future studies. There were also issues with flow and timing, with nearly half of studies classified as having high or unclear risk of bias. This was primarily due to inappropriate or unclear time intervals between the administration of the index test and reference standard. In a SUD population, administration of both the index and reference standard at a close time interval is critical given the potential of ongoing cognitive recovery with continued abstinence (Stavro, Pelletier, & Potvin, Reference Stavro, Pelletier and Potvin2013). Other issues with flow and timing included not all participants receiving the reference standard, or not accounting for missing participant data, creating an unknown amount of potential bias. It is possible that participants who are more cognitively impaired may drop out more readily due to increased complexity of tasks. It is essential that all aspects of participant flow are tracked, including those invited, those that accept, and those that complete the study, in order to establish whether the population sample reflects the population at large.

The selection of the reference standard was also a source of discrepancy across studies. Given the aim of the study was to include all studies examining CI in general amongst individuals with SUD, three studies focused on specific sub-populations (intellectual disability, Korsakoff syndrome; Braatveit et al., Reference Braatveit, Torsheim and Hove2018; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013; To et al., Reference To, Vanheule, Vanderplasschen, Audenaert and Vandevelde2015). For this reason, the reference standard used for classification was specific and focused, raising concerns on the applicability of identifying CI more broadly in individuals with SUDs.

Another concern regarding reference standards was inconsistency in impairment ratings. Some studies operationalised impairment as ranging from −1 to −2 standard deviations below normative data (the equivalent to a difference of 15th percentile versus 2nd percentile). Other studies required impairment on a certain number of tasks, by domain, or an average overall score. What constitutes psychometric impairment is a source of debate in the neuropsychological community (Schoenberg & Rum, Reference Schoenberg and Rum2017), and this inconsistency is reflected in the studies included. Interestingly, Godefroy et al. (Reference Godefroy, Gibbons, Diouf, Nyenhuis, Roussel, Black and Bugnicourt2014) analysed various methods of classifying impairment and suggested that a cut-off of the 5th percentile is most appropriate. A consensus on how to characterise impairment is needed if analyses of screening measures are to be accurately compared.

Two studies (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Wester et al., Reference Wester, Westhoff, Kessels and Egger2013) presented a unique methodological challenge that was not captured via the QUADAS–2. In both studies the reference standard focused on a limited area of cognitive functioning (executive functioning and memory, respectively), while the index measures evaluated cognitive function more broadly. In both instances, despite the apparent differences in screening and reference tasks, screening measures were able to detect individuals impaired on the reference standard at an acceptable level (AUC ≥75). This suggests that the BEAT and MoCA were able to adequately detect executive functioning and memory impairments, respectively, despite assessing a broader range of skills. It remains unclear how accurately these screening measures detect impairment in other domains. Use of a comprehensive reference standard that includes a range of cognitive domains would allow evaluation of both how well these screens pick up on a range of impairments other than the domain of interest and establish further construct validity.

Overall, findings suggest that clear adherence to STARD criteria was generally poor in studies evaluating the utility of cognitive screening tools in adults with SUDs (Bossuyt et al., Reference Bossuyt, Reitsma, Bruns, Gatsonis, Glasziou, Irwig and Cohen2015; Bossuyt et al., Reference Bossuyt, Reitsma, Bruns, Gatsonis, Glasziou, Irwig and Lijmer2003; Cohen et al., Reference Cohen, Korevaar, Altman, Bruns, Gatsonis, Hooft and Bossuyt2016). The current results reflect the methodological limitations of the studies in this area to date; not one study was classified as low risk across all domains. This primarily relates to insufficient reporting of study procedure and data as in many cases domains were rated unclear, resulting in risk being unable to be determined. This is a novel and important finding. Not only does it provide direction for further research, but it highlights caution that should be taken in interpretation of the findings to date, given that studies with risk of bias are likely to over- or under-estimate diagnostic accuracy.

Limitations

The current study intentionally set broad limits on search strategy and inclusion criteria to capture as many studies as possible. A key limitation of applying the QUADAS–2 to evaluate risk of bias in captured studies includes dealing with studies that omit key methodological details versus those which report on these more comprehensively and which may be disproportionately penalised as a result. For example, a number of studies did not report how many individuals were approached versus declined to participate, and the reasons for refusal. This could have potentially introduced bias in the recruitment and exclusion of participants in some studies; however, with omission of this information, it cannot be evaluated.

Another limitation is that a comprehensive grey literature search was not incorporated in the search strategy. This was partially mitigated by searching the Psychology Preprint Archive, as well as through inclusion of a publication that was under review and subsequently published (Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021). A more comprehensive grey literature search would have enriched the evidence base for this review by including findings from a broader range of sources, as well as reduce potential publication bias (Paez, Reference Paez2017).

Research Recommendations

While in a research context a homogeneous sample may be useful for interpreting study findings, this does not reflect the rich diversity and heterogeneity of individuals presenting for SUD treatment in health settings. To maximise ecological validity and translation of study findings to applied clinical practice, future studies seeking to validate cognitive screening tools in adults with SUDs should seek to adhere to STARD criteria (Bossuyt et al., Reference Bossuyt, Reitsma, Bruns, Gatsonis, Glasziou, Irwig and Cohen2015; Bossuyt et al., Reference Bossuyt, Reitsma, Bruns, Gatsonis, Glasziou, Irwig and Lijmer2003), while also maximising ecological validity. This includes keeping exclusion criteria to a minimum, such as instances where reduced capacity to provide fully informed consent is evident, or there are acute medical, psychiatric, or other risks that warrant priority treatment. This ensures the study sample is more representative of individuals presenting for the treatment of SUD and captures the potential wide range of co-occurring conditions that may impact cognition in a real-world setting. Studies should also clearly report on sample characteristics (including study setting, substances of misuse, abstinence status) to assist the clinician in understanding whether the cognitive screening tool may be useful for a particular client group. Additionally, including comprehensive classification accuracy statistics supports a greater understanding of the utility of the particular cognitive screening measure. Future studies should: use blinding to ensure objectivity of results; not integrate control with clinical data (as this can conflate results); and provide a clear description of study flow; accounting for any missing data. Future systematic reviews should also adhere to the recently updated PRISMA guidelines (Page et al., Reference Page, McKenzie, Bossuyt, Boutron, Hoffmann, Mulrow and Shamseer2020), which were not available at the time of the current review. Given the wide number of cognitive screening tools available and critiqued in this review, some of which show good promise in use in a drug and alcohol clinical setting (e.g., the MoCA), comprehensively validating existing tools should be an initial first step.

Clinical Recommendations

CI is frequent in individuals with substance use and, given the negative impact on treatment outcomes, should be screened for on treatment entry and periodically over treatment engagement (Bates, Buckman, & Nguyen, Reference Bates, Buckman and Nguyen2013). A cognitive screen used in this context must also be able to capture a range of cognitive disorders, given that the heterogeneity of contributing factors (e.g., ABI, learning or intellectual disability) forms part of a typical client’s presentation. Of the tools reviewed, the MoCA was the most studied and demonstrated the most consistently adequate diagnostic classification accuracy in this population. Practically, the MoCA is brief to administer, evaluates a broad range of cognitive domains, and is used widely across healthcare settings, which may aid in effective communication of health information. It also has alternate versions available, which facilitate re-testing over short periods of time (e.g., at the beginning and end of a detoxification stay, and then entry into outpatient clinics). While the MMSE is another well-known cognitive screening measure, it failed to demonstrate adequate classification accuracy in SUD populations and therefore cannot be recommended for clinical use in this setting. Other measures (e.g., ACE–R and BEAT) have initial promise for use in this area but require further empirical validation before they can be recommended for wide clinical use.

Regardless of the specific screening tool, a key consideration is the risk of over-reliance on a cut-off score alone to determine presence or absence of CI. A cognitive screen is not diagnostic and can only flag clients who may be in need of further assessment, and understanding the context in which the screen was administered is essential in an applied clinical setting. For example, capturing whether a client who performed below the cut-off score on a particular cognitive screen was acutely intoxicated, distracted, or in an uncharacteristically high state of distress is not an uncommon dilemma, and are all relevant details when deciding on clinical recommendations. Incorporating this information can reduce excessive use of resources but is also heavily reliant on clinical acumen. Given the complexities of clinical presentation and subtlety in interpreting cognitive screening results in adults with SUDs, it is recommended that a clinician skilled in the cognitive field provides oversight of cognitive screening processes within this setting, with clinical neuropsychologists ideally situated to perform such a role. Additionally, the development of clear clinical pathways within a service that provide further assessment and comprehensive support for clients identified as at risk of CI is also essential.

Conclusions

The current review provides the first systematic overview of cognitive screening tools utilised in adult SUD populations as compared to an accepted standard criterion of cognitive functioning (e.g., neuropsychological assessment). A strength of this review was the broad inclusion criteria; that is, that CI from a range of aetiologies and client populations was included. The MoCA was the most frequently reported cognitive screening tool, and it exhibited promising classification accuracy data across studies, suggesting the potential for use within this population. However, based on our search strategy and inclusion criteria, studies examining cognitive screening in adults with SUDs were significantly limited in number and quality. Other tools indicate initial promising classification data (ACE–R and BEAT); however, these also require further validation. Overall, given the very small number of studies, diverse sample characteristics, wide range of index measures and reference standards used, as well as high risk of bias identified amongst studies, further evaluation with stronger and more transparent methodological design and reporting is required. Without a clearly established tool to screen for CI, there is the potential that CI may not be systematically evaluated within SUD services, and this may negatively impact individual treatment outcomes. Crucially, future studies need to focus on adhering to the STARD criteria (Bossuyt et al., Reference Bossuyt, Reitsma, Bruns, Gatsonis, Glasziou, Irwig and Cohen2015; Bossuyt et al., Reference Bossuyt, Reitsma, Bruns, Gatsonis, Glasziou, Irwig and Lijmer2003), while also seeking to maximise ecological validity. Positively, there remains promise in the measures utilised to date such as the MoCA, and it is pleasing to see the development of measures specifically designed for use in the SUD population (e.g., Berry et al., Reference Berry, Shores, Nardo, Sedwell, Lunn, Marceau and Batchelor2021; Ritz et al., Reference Ritz, Lannuzel, Boudehent, Vabret, Bordas, Segobin and Beaunieux2015). Future studies should continue to build on the learnings of existing studies to ensure evidence-based measures are available to direct treatment and care of individuals who have comorbid substance use and cognitive needs.

SUPPLEMENTARY MATERIAL

To view supplementary material for this article, please visit https://doi.org/10.1017/S135561772100103X

ACKNOWLEDGEMENTS

The authors thank Hunter New England and South East Sydney Local Health District Drug and Alcohol Services for their support in producing this review.

AUTHOR CONTRIBUTIONS

JK and NR contributed equally to conception, design, interpretation, and writing of the manuscript. KK was involved in analysis, interpretation, and writing of the manuscript. AS designed and conducted the systematic search strategy. SB and KA contributed to design and revisions of the manuscript. All authors contributed to final manuscript review.

FINANCIAL SUPPORT

KA is supported by a Career Development Fellowship from the National Health and Medical Research Council (NHMRC; APP1141207). KK, NR, SB, AS, and JK have no financial support to disclose.

CONFLICTS OF INTEREST

The authors have nothing to disclose.

ETHICAL STANDARDS

This study did not involve human subject participation.

References

REFERENCES

Abu-Akel, A., Bousman, C., Skafidas, E., & Pantelis, C. (2018). Mind the prevalence rate: overestimating the clinical utility of psychiatric diagnostic classifiers. Psychological Medicine, 48, 12251227. doi: 10.1017/S0033291718000673 CrossRefGoogle ScholarPubMed
Aharonovich, E., Amrhein, P.C., Bisaga, A., Nunes, E.V., & Hasin, D.S. (2008). Cognition, commitment language, and behavioral change among cocaine-dependent patients. Psychology of Addictive Behaviors: Journal of the Society of Psychologists in Addictive Behaviors, 22, 557562. doi: 10.1037/a0012971 CrossRefGoogle ScholarPubMed
American Psychiatric Association. (2013). Diagnostic and Statistical Manual of Mental Disorders (5th ed.). Washington, DC: Author.Google Scholar
American Psychological Association Practice Organization. (2014). Distinguishing between screening and assessment for mental and behavioral health problems: a statement from an American Psychological Association Practice Organization work group on screening and psychological assessment. Retrieved from: https://www.apaservices.org/practice/reimbursement/billing/assessment-screening Google Scholar
Bates, M.E., Bowden, S.C., & Barry, D. (2002). Neurocognitive impairment associated with alcohol use disorders: implications for treatment. Experimental and Clinical Psychopharmacology, 10, 193212. doi: 10.1037//1064-1297.10.3.193 CrossRefGoogle ScholarPubMed
Bates, M.E., Buckman, J.F., & Nguyen, T.T. (2013). A role for cognitive rehabilitation in increasing the effectiveness of treatment for alcohol use disorders. Neuropsychology Review, 23, 2747. doi: 10.1007/s11065-013-9228-3 CrossRefGoogle ScholarPubMed
Berry, J., Shores, E.A., Nardo, T., Sedwell, A., Lunn, J., Marceau, E., …Batchelor, J. (2021). The brief executive-function assessment tool: a new cognitive impairment screening tool for alcohol and other drug services. Applied Neuropsychology: Adult. doi: 10.1080/23279095.2021.1895791 CrossRefGoogle ScholarPubMed
Bossuyt, P.M., Reitsma, J.B., Bruns, D.E., Gatsonis, C.A., Glasziou, P.P., Irwig, L., … Cohen, J.F. (2015). STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. BMJ: British Medical Journal, 351, h5527. doi: 10.1136/bmj.h5527 CrossRefGoogle ScholarPubMed
Bossuyt, P.M., Reitsma, J.B., Bruns, D.E., Gatsonis, C.A., Glasziou, P.P., Irwig, L.M., … Lijmer, J.G. (2003). The STARD statement for reporting studies of diagnostic accuracy: explanation and elaboration. Annals of Internal Medicine, 138, W1W12.CrossRefGoogle ScholarPubMed
Braatveit, K.J., Torsheim, T., & Hove, O. (2018). Screening for intellectual disabilities: a validation of the Hayes Ability Screening Index for in-patients with substance use disorder. Nordic Journal of Psychiatry, 72, 387392. doi: 10.1080/08039488.2018.1468923 CrossRefGoogle ScholarPubMed
Bruijnen, C.J., Jansen, M., Dijkstra, B.A., Walvoort, S.J., Lugtmeijer, S., Markus, W., … & Kessels, R.P. (2019). The Montreal Cognitive Assessment (MoCA) as a cognitive screen in addiction health care: a validation study for clinical practice. Journal of Substance Use, 24, 4754. doi: 10.1080/14659891.2018.1497102 CrossRefGoogle Scholar
Brodersen, C., Koen, E., Ponte, A., Sánchez, S., Segal, E., Chiapella, A., … & Lemberg, A. (2014). Cognitive function in patients with alcoholic and nonalcoholic chronic liver disease. The Journal of Neuropsychiatry and Clinical Neurosciences, 26, 241248. doi: 10.1176/appi.neuropsych.12040091 CrossRefGoogle ScholarPubMed
Buckman, J.F., Bates, M.E., & Cisler, R.A. (2007). Social networks and their influence on drinking behaviors: differences related to cognitive impairment in clients receiving alcoholism treatment. Journal of Studies on Alcohol and Drugs, 68, 738747. doi: 10.15288/jsad.2007.68.738 CrossRefGoogle ScholarPubMed
Buckman, J.F., Bates, M.E., & Morgenstern, J. (2008). Social support and cognitive impairment in clients receiving treatment for alcohol- and drug-use disorders: a replication study. Journal of Studies on Alcohol and Drugs, 69, 738746. doi: 10.15288/jsad.2008.69.738 CrossRefGoogle ScholarPubMed
Bunnage, M. (2017). How do I know when a diagnostic test works? In Bowden, S. (Ed.), Neuropsychological assessment in the age of evidence-based practice: Diagnostic and treatment evaluation (pp. 223238). New York, NY: Oxford University Press.Google Scholar
Carroll Chapman, S.L. & Wu, L.T. (2012). Substance abuse among individuals with intellectual disabilities. Research in Developmental Disabilities, 33, 11471156. doi: 10.1016/j.ridd.2012.02.009 CrossRefGoogle ScholarPubMed
Chen, L.Y., Strain, E.C., Crum, R.M., & Mojtabai, R. (2013). Gender differences in substance abuse treatment and barriers to care among persons with substance use disorders with and without comorbid major depression. Journal of Addiction Medicine, 7, 325334. doi: 10.1097/ADM.0b013e31829b7afe CrossRefGoogle ScholarPubMed
Cohen, J. (1988). Statistical Power Analysis for the Behavioral Sciences (2nd ed.). Hillsdale, NJ: Lawrence Earlbaum Associates.Google Scholar
Cohen, J.F., Korevaar, D.A., Altman, D.G., Bruns, D.E., Gatsonis, C.A., Hooft, L., … Bossuyt, P.M. (2016). STARD 2015 guidelines for reporting diagnostic accuracy studies: explanation and elaboration. BMJ Open, 6, e012799. doi: 10.1136/bmjopen-2016-012799 CrossRefGoogle ScholarPubMed
Copersino, M.L., Fals-Stewart, W., Fitzmaurice, G., Schretlen, D.J., Sokoloff, J., & Weiss, R.D. (2009). Rapid cognitive screening of patients with substance use disorders. Experimental and Clinical Psychopharmacology, 17, 337344. doi: 10.1037/a0017260 CrossRefGoogle ScholarPubMed
Copersino, M.L., Schretlen, D.J., Fitzmaurice, G.M., Lukas, S.E., Faberman, J., Sokoloff, J., & Weiss, R.D. (2012). Effects of cognitive impairment on substance abuse treatment attendance: predictive validation of a brief cognitive screening measure. The American Journal of Drug and Alcohol Abuse, 38, 246250. doi: 10.3109/00952990.2012.670866 CrossRefGoogle ScholarPubMed
Covidence Systematic Review Software. Veritas Health Innovation. Melbourne, Australia. Available at www.covidence.org Google Scholar
Cunha, P.J., Nicastri, S., de Andrade, A.G., & Bolla, K.I. (2010). The frontal assessment battery (FAB) reveals neurocognitive dysfunction in substance-dependent individuals in distinct executive domains: abstract reasoning, motor programming, and cognitive flexibility. Addictive Behaviors, 35, 875881. doi: 10.1016/j.addbeh.2010.05.005 CrossRefGoogle ScholarPubMed
Daley, D.C. & Moss, H.B. (2009). Dual disorders: Counseling clients with Chemical Dependency and Mental Illness. Center City: Simon and Schuster.Google Scholar
Dubois, B., Slachevsky, A., Litvan, I., & Pillon, B. (2000). The FAB: a Frontal assessment battery at bedside. Neurology, 55, 16211626. doi: 10.1212/wnl.55.11.1621 CrossRefGoogle Scholar
Ewert, V., Pelletier, S., Alarcon, R., Nalpas, B., Donnadieu-Rigole, H., Trouillet, R., & Perney, P. (2018). Determination of MoCA cutoff score in patients with alcohol use disorders. Alcoholism: Clinical and Experimental Research, 42, 403412 doi: 10.1111/acer.13547 CrossRefGoogle ScholarPubMed
Fals-Stewart, W. (1997). Detection of neuropsychological impairment among substance-abusing patients: accuracy of the Neurobehavioral Cognitive Status Examination. Experimental and Clinical Psychopharmacology, 5, 269.CrossRefGoogle ScholarPubMed
Folstein, M.F., Folstein, S.E., & McHugh, P.R. (1975). “Mini-mental state”: a practical method for grading the cognitive state of patients for the clinician. Journal of Psychiatric Research, 12, 189198.CrossRefGoogle ScholarPubMed
Godefroy, O., Gibbons, L., Diouf, M., Nyenhuis, D., Roussel, M., Black, S., Bugnicourt, J.M., & GREFEX Study Group. (2014). Validation of an integrated method for determining cognitive ability: implications for routine assessments and clinical trials. Cortex, 54, 5162. doi: 10.1016/j.cortex.2014.01.016 CrossRefGoogle ScholarPubMed
Gooden, J.R., Cox, C.A., Petersen, V., Curtis, A., Manning, V., & Lubman, D.I. (2020). Characterisation of presentations to a community-based specialist addiction neuropsychology service: cognitive profiles, diagnoses and comorbidities. Drug and Alcohol Review. doi: 10.1111/dar.13135 Google ScholarPubMed
Gould, T.J. (2010). Addiction and cognition. Addiction Science & Clinical Practice, 5, 414.Google ScholarPubMed
Grella, C.E., Hser, Y.I., Joshi, V., & Rounds-Bryant, J. (2001). Drug treatment outcomes for adolescents with comorbid mental and substance use disorders. The Journal of Nervous and Mental Disease, 189, 384392.CrossRefGoogle ScholarPubMed
Grohman, K. & Fals-Stewart, W. (2004). The detection of cognitive impairment among substance-abusing patients: the accuracy of the Neuropsychological Assessment Battery–screening module. Experimental and Clinical Psychopharmacology, 12, 200. doi: 10.1037/1064-1297.12.3.200 CrossRefGoogle ScholarPubMed
Hagen, E., Erga, A.H., Hagen, K.P., Nesvåg, S.M., McKay, J.R., Lundervold, A.J., & Walderhaug, E. (2016). Assessment of executive function in patients with substance use disorder: a comparison of inventory- and performance-based assessment. Journal of Substance Abuse Treatment, 66, 18. doi: 10.1016/j.jsat.2016.02.010 CrossRefGoogle ScholarPubMed
Hayes, S.C. (2000). Hayes Ability Screening Index (HASI) Manual. Sydney: Faculty of Medicine, University of Sydney.Google Scholar
Horner, M.D., Harvey, R.T., & Denier, C.A. (1999). Self-report and objective measures of cognitive deficit in patients entering substance abuse treatment. Psychiatry Research, 86, 155161.CrossRefGoogle ScholarPubMed
Katz, N., Itzkovich, M., Averbuch, S., & Elazar, B. (1989). Loewenstein occupational therapy cognitive assessment (LOTCA) battery for brain-injured patients: reliability and validity. American Journal of Occupational Therapy, 43, 184192.CrossRefGoogle ScholarPubMed
Kelly, T.M. & Daley, D.C. (2013). Integrated treatment of substance use and psychiatric disorders. Social Work in Public Health, 28, 388406. doi: 10.1080/19371918.2013.774673 CrossRefGoogle ScholarPubMed
Kiernan, R.J., Mueller, J., Langston, J.W., & Van Dyke, C. (1987). The Neurobehavioral Cognitive Status Examination: a brief but quantitative approach to cognitive assessment. Annals of Internal Medicine, 107(4), 481485. doi: 10.7326/0003-4819-107-4-481 CrossRefGoogle ScholarPubMed
Levin, F.R. (2007). Diagnosing attention-deficit/hyperactivity disorder in patients with substance use disorders. Journal of Clinical Psychiatry, 68, 914.Google ScholarPubMed
Lin, E., Balogh, R., McGarry, C., Selick, A., Dobranowski, K., Wilton, A.S., & Lunsky, Y. (2016). Substance-related and addictive disorders among adults with intellectual and developmental disabilities (IDD): an Ontario population cohort study. BMJ Open, 6, e011638. doi: 10.1136/bmjopen-2016-011638 CrossRefGoogle ScholarPubMed
Mandrekar, J.N. (2010). Receiver operating characteristic curve in diagnostic test assessment. Journal of Thoracic Oncology, 5, 13151316. doi: 10.1097/jto.0b013e3181ec173d CrossRefGoogle ScholarPubMed
McGee, S. (2002). Simplifying likelihood ratios. Journal of General Internal Medicine, 17, 646649. doi: 10.1046/j.1525-1497.2002.10750.x CrossRefGoogle ScholarPubMed
Mioshi, E., Dawson, K., Mitchell, J., Arnold, R., & Hodges, J.R. (2006). The Addenbrooke’s Cognitive Examination Revised (ACE–R): a brief cognitive test battery for dementia screening. International Journal of Geriatric Psychiatry, 21, 10781085. doi: 10.1002/gps.1610 CrossRefGoogle ScholarPubMed
Moher, D., Liberati, A., Tetzlaff, J., Altman, D.G., & The PRISMA Group. (2009). Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLOS Medicine, 6, e1000097. doi: 10.1371/journal.pmed1000097 CrossRefGoogle ScholarPubMed
Molinaro, A.M. (2015). Diagnostic tests: how to estimate the positive predictive value. Neuro-Oncology Practice, 2, 162166. doi: 10.1093/nop/npv030 Google ScholarPubMed
Nasreddine, Z.S., Phillips, N.A., Bédirian, V., Charbonneau, S., Whitehead, V., Collin, I., … & Chertkow, H. (2005). The Montreal cognitive assessment, MoCA: a brief screening tool for mild cognitive impairment. Journal of the American Geriatrics Society, 53, 695699. doi: 10.1111/j.1532-5415.2005.53221.x CrossRefGoogle ScholarPubMed
Owen, K.P., Sutter, M.E. & Albertson, T.E. (2014). Marijuana: respiratory tract effects. Clinical Reviews in Allergy & Immunology, 46, 6581. doi: 10.1007/s12016-013-8374-y CrossRefGoogle ScholarPubMed
Paez, A. (2017). Gray literature: an important resource in systematic reviews. Journal of Evidence-Based Medicine, 10, 233240. doi: 10.1111/jebm.12265 CrossRefGoogle ScholarPubMed
Page, M.J., McKenzie, J.E., Bossuyt, P.M., Boutron, I., Hoffmann, T.C., Mulrow, C.D. … & Shamseer, L. (2020). The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. MetarXiv. https://osf.io/preprints/metaarxiv/v7gm2/ Google Scholar
Parry-Jones, B.L., Vaughan, F.L., & Miles Cox, W. (2006). Traumatic brain injury and substance misuse: a systematic review of prevalence and outcomes research (1994–2004). Neuropsychological Rehabilitation, 16, 537560. doi: 10.1080/09602010500231875 CrossRefGoogle Scholar
Pelletier, S., Alarcon, R., Ewert, V., Forest, M., Nalpas, B., & Perney, P. (2018). Comparison of the MoCA and BEARNI tests for detection of cognitive impairment in in-patients with alcohol use disorders. Drug and Alcohol Dependence, 187, 249253. doi: 10.1016/j.drugalcdep.2018.02.026 CrossRefGoogle ScholarPubMed
Ramey, T. & Regier, P. (2019). Cognitive impairment in substance use disorders. CNS Spectrums, 24, 102113. doi: 10.1017/S1092852918001426 CrossRefGoogle ScholarPubMed
Ridley, N., Batchelor, J., Draper, B., Demirkol, A., Lintzeris, N., & Withall, A. (2018). Cognitive screening in substance users: diagnostic accuracies of the mini-mental state examination, Addenbrooke’s Cognitive Examination–Revised, and Montreal Cognitive Assessment. Journal of Clinical and Experimental Neuropsychology, 40, 107122. doi: 10.1080/13803395.2017.1316970 CrossRefGoogle ScholarPubMed
Ritz, L., Lannuzel, C., Boudehent, C., Vabret, F., Bordas, N., Segobin, S., … & Beaunieux, H. (2015). Validation of a brief screening tool for alcohol-related neuropsychological impairments. Alcoholism: Clinical and Experimental Research, 39, 22492260. doi: 10.1111/acer.12888 Google ScholarPubMed
Roebuck-Spencer, T.M., Glen, T., Puente, A.E., Denney, R.L., Ruff, R.M., Hostetter, G., & Bianchini, K.J. (2017). Cognitive screening tests versus comprehensive neuropsychological test batteries: a National Academy of Neuropsychology Education Paper. Archives of Clinical Neuropsychology, 32, 491498. doi: 10.1093/arclin/acx021 Google Scholar
Rojo-Mota, G., Pedrero-Pérez, E.J., Ruiz-Sánchez de León, J.M., León-Frade, I., Aldea-Poyo, P., Alonso-Rodríguez, M., … & Morales-Alonso, S. (2017). Loewenstein occupational therapy cognitive assessment to evaluate people with addictions. Occupational Therapy International, 17. doi: 10.1155/2017/2750328 CrossRefGoogle ScholarPubMed
Salavert, J., Clarabuch, A., Fernández-Gómez, M.J., Barrau, V., Giráldez, M.P., & Borràs, J. (2018). Substance use disorders in patients with intellectual disability admitted to psychiatric hospitalisation. Journal of Intellectual Disability Research: JIDR, 62, 923930. doi: 10.1111/jir.12514 CrossRefGoogle ScholarPubMed
Schoenberg, M.R. & Rum, R.S. (2017). Towards reporting standards for neuropsychological study results: a proposal to minimize communication errors with standardized qualitative descriptors for normalized test scores. Clinical Neurology and Neurosurgery, 162, 7279. doi: 10.1016/j.clineuro.2017.07.010 CrossRefGoogle ScholarPubMed
Schueler, S., Schuetz, G.M., & Dewey, M. (2012). The revised QUADAS–2 tool. Annals of Internal Medicine, 156, 323324. doi: 10.7326/0003-4819-156-4-201202210-00018 CrossRefGoogle ScholarPubMed
Secades-Villa, R., García-Rodríguez, O., & Fernández-Hermida, J.R. (2015). Contingency management for substance use disorders in Spain: implications for research and practice. Preventive Medicine, 80, 8288. doi: 10.1016/j.ypmed.2015.07.001 CrossRefGoogle ScholarPubMed
Slater, H. & Young, J. (2013). A review of brief cognitive assessment tests. Reviews in Clinical Gerontology, 23, 164176. doi: 10.1017/s0959259813000038 CrossRefGoogle Scholar
Stavro, K., Pelletier, J., & Potvin, S. (2013). Widespread and sustained cognitive deficits in alcoholism: a meta-analysis. Addiction Biology, 18, 203213. doi: 10.1111/j.1369-1600.2011.00418.x CrossRefGoogle ScholarPubMed
Stern, R.A., & White, T. (2003). NAB, Neuropsychological Assessment Battery: Administration, scoring, and interpretation manual. Lutz, FL: Psychological Assessment Resources.Google Scholar
To, W.T., Vanheule, S., Vanderplasschen, W., Audenaert, K., & Vandevelde, S. (2015). Screening for intellectual disability in persons with a substance abuse problem: exploring the validity of the Hayes Ability Screening Index in a Dutch-speaking sample. Research in Developmental Disabilities, 36, 498504. doi: 10.1016/j.ridd.2014.10.046 CrossRefGoogle Scholar
Trevethan, R. (2017). Sensitivity, specificity, and predictive values: foundations, pliabilities, and pitfalls in research and practice. Frontiers in Public Health, 5. doi: 10.3389/fpubh.2017.00307 CrossRefGoogle ScholarPubMed
Verdejo-Garcia, A., Garcia-Fernandez, G., & Dom, G. (2019). Cognition and addiction. Dialogues in Clinical Neuroscience, 21, 281290. doi: 10.31887/DCNS.2019.21.3/gdom CrossRefGoogle ScholarPubMed
Volkow, N.D., Baler, R.D., Compton, W.M., & Weiss, S.R. (2014). Adverse health effects of marijuana use. The New England Journal of Medicine, 370, 22192227. doi: 10.1056/NEJMra1402309 CrossRefGoogle ScholarPubMed
Wester, A.J., Westhoff, J., Kessels, R.P., & Egger, J.I. (2013). The Montreal Cognitive Assessment (MoCA) as a measure of severity of amnesia in patients with alcohol-related cognitive impairments and Korsakoff syndrome. Clinical Neuropsychiatry, 10, 134141.Google Scholar
Whiting, P.F., Rutjes, A.W., Westwood, M.E., Mallett, S., Deeks, J.J., Reitsma, J.B., … QUADAS–2 Group. (2011). QUADAS–2: a revised tool for the quality assessment of diagnostic accuracy studies. Annals of Internal Medicine, 155, 529536. doi: 10.7326/0003-4819-155-8-201110180-00009 CrossRefGoogle ScholarPubMed
Figure 0

Fig. 1. PRISMA flow diagram.

Figure 1

Table 1. Study characteristics of articles included for review

Figure 2

Table 2. Overview of cognitive screening measures

Figure 3

Table 3. Overview of screening measures and reference standards for included studies

Figure 4

Table 4. Psychometric data retrieved from primary classification accuracy studies

Figure 5

Table 5. Summary of studies examining correlations between cognitive screening tools and comprehensive cognitive assessment

Figure 6

Table 6. QUADAS–2 ratings for studies included in the review

Supplementary material: File

Ko et al. supplementary material

Table S1 and S2

Download Ko et al. supplementary material(File)
File 25.1 KB