Symptom validity testing (SVT) has emerged as a major theme of neuropsychological research and clinical practice. Neuropsychological assessment methods and procedures strive for and require the most valid and reliable techniques to assess cognitive and neurobehavioral functioning, to make neuro-psychological inferences and diagnostic conclusions. From the beginnings of neuropsychology as a discipline, issues of test reliability and validity have always been a concern (Filskov & Boll, Reference Filskov and Boll1981; Lezak, Reference Lezak1976). However, neuropsychology's initial focus was mostly on test development, standardization and the psychometric properties of a test and not independent measures of test validity. A variety of SVT methods are now available (Larrabee, Reference Larrabee2007). While contemporary neuropsychological test development has begun to more directly incorporate SVT indicators embedded within the primary neuropsychological instrument (Bender, Martin Garcia, & Barr, Reference Bender, Martin Garcia and Barr2010; Miller et al., Reference Miller, Millis, Rapport, Bashem, Hanks and Axelrod2011; Powell, Locke, Smigielski, & McCrea, Reference Powell, Locke, Smigielski and McCrea2011), traditional neuropsychological test construction and the vast majority of standardized tests currently in use do not. Current practice has been to use what are referred to as “stand-alone” SVT measures that are separately administered during the neuropsychological examination (Sollman & Berry, Reference Sollman and Berry2011). SVT performance is then used to infer validity or lack thereof for the battery of all neuropsychological tests administered during that test session.
The growth in SVT research has been exponential. Using the search words “symptom validity testing” in a National Library of Medicine literature search yielded only one study before 1980, five articles during the 1980s, but hundreds thereafter. SVT research of the last decade has led to important practice conclusions as follows: (1) professional societies endorse SVT use (Bush et al., Reference Bush, Ruff, Troster, Barth, Koffler, Pliskin and Silver2005; Heilbronner, Sweet, Morgan, Larrabee, & Millis, Reference Heilbronner, Sweet, Morgan, Larrabee and Millis2009), (2) passing a SVT infers valid performance, (3) SVT measures have good face validity as cognitive measures, all have components that are easy for healthy controls and even for the majority of neurological patients to pass with few or no errors, and (4) groups that perform below established cut-score levels on a SVT generally exhibit lower neuropsychological test scores. This last observation has been interpreted as demonstrating that SVT performance taps a dimension of effort to perform, where SVT “failure” reflects non-neurological factors that reduce neuropsychological test scores and invalidates findings (West, Curtis, Greve, & Bianchini, Reference West, Curtis, Greve and Bianchini2011). On forced-choice SVTs, the statistical improbability of below chance performance implicates malingering (the examinee knows the correct answer but selects the incorrect to feign impairment).
A quote from Millis (Reference Millis2009) captures the importance of why neuropsychology must address the validity of test performance:
All cognitive tests require that patients give their best effort (italics added) when completing them. Furthermore, cognitive tests do not directly measure cognition: they measure behavior from which we make inferences about cognition. People are able to consciously alter or modify their behavior, including their behavior when performing cognitive tests. Ostensibly poor or “impaired” test scores will be obtained if an examinee withholds effort (e.g., reacting slowly to reaction time tests). There are many reasons why people may fail to give best effort on cognitive testing: financial compensation for personal injury; disability payments; avoiding or escaping formal duty or responsibilities (e.g., prison, military, or public service, or family support payments or other financial obligations); or psychosocial reinforcement for assuming the sick role (Slick, Sherman, & Iverson, Reference Slick, Sherman and Iverson1999). “…. Clinical observation alone cannot reliably differentiate examinees giving best effort from those who are not.” (Millis & Volinsky, Reference Millis and Volinsky2001, p. 2409)
This short review examines key SVT concepts and the “effort” term as used in neuropsychology. “Effort” seems to capture a clinical descriptive of patient test performance that, at first blush, seems straightforward enough. However, effort also has neurobiological underpinnings, a perspective often overlooked in SVT research and clinical application. Furthermore, does the term effort suggest intention, such as “genuine effort” – the patient is trying their best? Or if maximum “effort” is not being applied to test performance or exhibited by the patient, when does it reflect performance that may not be trustworthy? What is meant by effort?
“Effort” – its Multiple Meanings
In the SVT literature when the “effort” term is linked with other nouns, verbs, and/or adjectives such statements appear to infer or explain a patient's mental state, including motivation. For example it is common to see commentaries in the SVT literature indicating something about “cognitive effort,” “mental effort,” “insufficient effort,” “poor effort,” “invalid effort,” or even “faked effort.” Some of these terms suggest that neuropsychological techniques, and in particular the SVT measure itself, are capable of inferring intent? Can they (see discussion by Dressing, Widder, & Foerster, Reference Dressing, Widder and Foerster2010)? There are additional terms often used to describe SVT findings including response bias, disingenuous, dissimulation, non-credible, malingered, or non- or sub-optimal, further complicating SVT nomenclature. There is no agreed upon consensus definition within neuropsychology of what effort means.
Below Cut-Score SVT Performance and Neuropsychological Test Findings: THE ISSUE
An exemplary SVT study has been done by Locke, Smigielski, Powell, and Stevens (Reference Locke, Smigielski, Powell and Stevens2008). This study is singled out for this review because it was performed at an academic medical center (many SVT studies are based on practitioners’ clinical cases), had institutional review board (IRB) approval (most SVT studies do not), examined non-forensic (no case was in litigation although some cases had already been judged disabled and were receiving compensation), and was based on consecutive clinical referrals (independently diagnosed with some type of acquired brain injury [ABI] before the neuropsychological assessment) with all patients being seen for treatment recommendations and/or actual treatment. Figure 1 shows the distribution of pass/fail SVT scores where 21.8% performed below the SVT cut-point [pass ≥ 45/50 items correct on Trial 2 of the Test of Memory Malingering (TOMM), Tombaugh, Reference Tombaugh1996], which Locke et al. defined as constituting a group of ABI patients exhibiting “sub-optimal” effort. Of greater significance for neuropsychology is that the subjects in the failed SVT group, while being matched on all other demographic factors, performed worse and statistically lower across most (but not all) neuropsychological test measures.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160704235124-51140-mediumThumb-S1355617712000252_fig1g.jpg?pub-status=live)
Fig. 1 The distribution of Test of Memory Malingering (TOMM) scores that were above (green) the cut-point of 45 compared to those who performed below (blue-red). Note the bi-modal distribution of those who failed, compared to the peaked performance of 50/50 correct by those who pass. Below chance responding is the most accepted SVT feature indicative of malingering. As indicated by the shading of blue coloration emerging into red, as performance approaches chance, that likely reflects considerable likelihood of malingering. However, recalling that all of these patients had some form of ABI, those in blue, are much closer to the cut-point, raising the question of whether their neurological condition may contribute to their SVT performance.
As seen in Figure 1, the modal response is a perfect or near perfect TOMM score of 50/50 correct, presumably reflective of valid test performance. The fact that all subjects had ABI previously and independently diagnosed before the SVT being administered, and most performed without error, is a testament to the ease of performing an SVT. For those who scored below the cut-point and thereby “failed” the SVT measure, a distinct bi-modal distribution emerges. One group, shown in red (see Figure 1), hovers around chance (clearly an invalid performance) but the majority of others who fail, as shown in blue, hover much closer to, but obviously below, the SVT cut-point.
For the purposes of this review, the distinctly above-chance but below cut-score performing group will be labeled the “Near-Pass” SVT group. It is with this group where the clinician/researcher is confronted with Type I and II statistical errors when attempting to address validity issues of neuropsychological test findings. All current SVT methods acknowledge that certain neurological conditions may influence SVT performance, but few provide much in the way of guidelines as to how this problem should be addressed.
Type II Error and the Patient With a “Near-Pass SVT” performanceFootnote 1
Two cases among many seen in our University-based neuropsychology program are representative of the problems with classifying neuropsychological test performance as invalid in the “Near-Pass” SVT performing patient. Figure 2 depicts the post-temporal lobectomy MRI in a patient with intractable epilepsy who underwent a partial right hippocampectomy. Pre-surgery he passes all SVT measures but post-surgery passes some but not others. Figure 3 shows abnormal medial temporal damage in a patient with herpes simplex encephalitis who fails SVT measures. In both cases, SVT failure is not below chance, both patients have distinct, bona fide and unequivocal structural damage to critical brain regions involved in memory. Is not their SVT performance a reflection of the underlying damage and its disruption of memory performance? If a neuropsychologist interpreted these test data as invalid because of “failed” SVT performance, is that not ignoring the obvious and committing a Type II error? So when is “failed” SVT performance just a reflection of underlying neuropathology? For example, cognitive impairment associated with Alzheimer's disease is sufficient to impair SVT performance resulting in below recommended cut-score levels and therefore constituting a SVT failure (Merten, Bossink, & Schmand, Reference Merten, Bossink and Schmand2007). However, in such a circumstance impaired neuropsychological performance and the low SVT score are both thought to reflect genuinely impaired cognitive ability.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160704235124-51433-mediumThumb-S1355617712000252_fig2g.jpg?pub-status=live)
Fig. 2 MRI showing partial temporal lobectomy. Pre-Surgery Test of Memory and Malingering (TOMM): Trial 1 = 50/50; Trial 2 = 50/50; Test of Neuropsychological Malingering (TNM) = 90% correct (see Hall and Pritchard, Reference Hall and Pritchard1996). Post-Surgery: TOMM Trial 1 = 42/50; Trial 2 = 46/50; Delayed = 44/50; Rey 15-Item = 6/15; Word Memory Test (WMT): IR 67.5%, 30 min delay 75%; Free Recall 5%; Free Recall Delay = 7.5%; Free Recall Long Delay = 0.0.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160704235124-12782-mediumThumb-S1355617712000252_fig3g.jpg?pub-status=live)
Fig. 3 The fluid attenuated inversion recovery (FLAIR) sequence magnetic resonance imaging (MRI) was performed in the sub-acute stage demonstrating marked involvement of the right (arrow) medial temporal lobe region of this patient (see arrow). The patient was attempting to return to college and was being evaluated for special assistance placement. On the Wechsler Memory Scale (WMS-IV) he obtained the following: WMS-IV: Audio Memory = 87; Visual Memory = 87; Visual Word Memory = 73; Immediate Memory = 86; Delayed Memory = 84. Immediate Recall: 77.5% (Fail); Delayed Recall: 72.5% (Fail); Consistency: 65.0% (Fail); Multiple Choice: 50.0% (Warning); Paired Associate: 50.0% (Warning); Free Recall: 47.5%; Test of Memory and Malingering (TOMM): Trial 1: 39, Trial 2: 47.
No Systematic Study Lesion/Localization Studies of SVT performance
Neuropsychological assessment has a long tradition of examining lesion effects (or lack thereof) on neuropsychological test performance (Cipolotti & Warrington, Reference Cipolotti and Warrington1995). There are no systematic studies of lesion effects on SVT performance. Despite the assumption of minimal requirements to perform a SVT task, episodic memory is being tapped where functional neuroimaging studies demonstrate expected medial temporal, cingulate and parietofrontal attentional network activation during SVT performance (Allen, Bigler, Larsen, Goodrich-Hunsaker, & Hopkins, Reference Allen, Bigler, Larsen, Goodrich-Hunsaker and Hopkins2007; Browndyke et al., Reference Browndyke, Paskavitz, Sweet, Cohen, Tucker, Welsh-Bohmer and Schmechel2008; Larsen, Allen, Bigler, Goodrich-Hunsaker, & Hopkins, Reference Larsen, Allen, Bigler, Goodrich-Hunsaker and Hopkins2010). Goodrich-Hunsaker and Hopkins (Reference Goodrich-Hunsaker and Hopkins2009) and Wu, Allen, Goodrich-Hunsaker, Hopkins, & Bigler (Reference Wu, Allen, Goodrich-Hunsaker, Hopkins and Bigler2010) in case studies have shown that five patients with hippocampal damage (3 with anoxic injury, 2 with TBI) can pass SVTs. However, this woefully under samples the possible lesions and abnormalities that have the potential to directly disrupt SVT performance. While not a lesion study, Merten et al. (Reference Merten, Bossink and Schmand2007) have demonstrated SVT failure related to dementia severity and Gorissen, Sanz, and Schmand (Reference Gorissen, Sanz and Schmand2005) demonstrated high SVT failure rates in neuropsychiatric patients, particularly those with schizophrenia. If certain neuropathological conditions are more likely to affect SVT performance, such findings would be critical for SVT interpretation. Currently, there are no recommended adjustments in SVT cut-scores based on location, size or type of lesion that may be present.
Illness Behavior and Cognitive Performance
In non-litigating neurological and neuropsychiatric clinical populations, 15 to 30% or higher SVT failure rates have been reported (Williams, Reference Williams2011). Does this mean invalid neuropsychological test data occurs in up to one-third of all patients seen for neuropsychological assessment based on not passing an externally administered SVT? Are these patients malingering? What is the role of SVT performance in rendering differential diagnoses involving malingering, somatoform disorder and other neuropsychiatric conditions? Figure 4 from Metternich, Schmidtke, and Hull (Reference Metternich, Schmidtke and Hull2009) provides a model showing the overlap between functional and constitutional memory factors that may adversely affect cognition in the neuropsychiatric patient. In reviewing this model, the reader should note that there are numerous meta-cognitive as well as neural pathways that could legitimately disrupt SVT performance as a result of the disorder. Wager-Smith and Markou (Reference Wager-Smith and Markou2011) review the effects of stress, cytokine, and neuroinflammatory reactions that relate to “sickness behavior” and impaired cognition. In the context of the Metternich et al. (Reference Metternich, Schmidtke and Hull2009) model, sickness behaviors may interact with stress mediated biological factors that appear psychological yet disrupt cognition. Does any of this reflect differences in how SVT cut scores should be established if a known neuropsychiatric disorder is present?
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160704235124-76867-mediumThumb-S1355617712000252_fig4g.jpg?pub-status=live)
Fig. 4 Theoretical model postulating how stress may influence cognition. Primary influences may come from purely psychosocial variables, directly from physiologically activated stress variables or the combination of the two. Reprinted from Journal of Psychosomatic Research, Volume 66, Issue 05. “How are memory complaints in functional memory disorder related to measures of affect, metamemory and cognition?” by Birgitta Metternich, Kalaus Schmidtke and Michael Hull, pp. 435–444. Copyright © 2009 with permission from Elsevier.
Diagnosis Threat and SVT Performance
The “diagnosis threat” literature clearly demonstrates both experimentally and clinically that performance expectations influence actual cognitive test performance and the perceived influence of symptoms on cognitive test results (Ozen & Fernandes, Reference Ozen and Fernandes2011; Suhr & Gunstad, Reference Suhr and Gunstad2005). Likewise, placebo research on cognitive performance plainly demonstrates the influential role that expectations have on symptom generation and test performance (Pollo & Benedetti, Reference Pollo and Benedetti2009). Thus, psychological state and trait characteristics and the perception of well-being versus “illness” may influence cognitive performance (Pressman & Cohen, Reference Pressman and Cohen2005; Walkenhorst & Crowe, Reference Walkenhorst and Crowe2009). Is “near-pass” SVT performance an expected human dimension of diagnosis threat? Can these factors be disentangled from the degree and type of brain injury, medication status, litigation status, levels of psychological distress including premorbid conditions and other non-neurological factors by SVT performance (see discussion by Suhr, Tranel, Wefel, & Barrash, Reference Suhr, Tranel, Wefel and Barrash1997)?
Effort or Ability?
If SVT performance required minimal to no cognitive effort, then experimental paradigms using cognitive load as a distraction, should result in minimal to little change in SVT performance. Batt, Shores, and Chekaluk (Reference Batt, Shores and Chekaluk2008) examined non-litigating severe TBI patients on SVT measures during a task where distraction occurred during the SVT learning phase, demonstrating the influence of cognitive processing on SVT performance. Cognitive neuroscience uses simple cognitive tasks, as simple as SVT measures, to experimentally manipulate conditions to tease out neural and experimental effects on cognition (Graham, Barense, & Lee, Reference Graham, Barense and Lee2010). Unfortunately, other than the Batt et al. and a handful of other studies, the cognitive neuroscience of SVT performance has been ignored. For example, it is unknown whether the foil stimuli used in an SVT task are equivalent or not, or are uniquely influenced by certain types of structural damage, or neuropsychiatric or neurological condition.
Design Issues in SVT Research
The review to this point has raised several interpretative questions that occur in Near-Pass SVT subjects. The answer to these questions requires better designed SVT studies that address ambiguities of past SVT findings. Williams (Reference Williams2011) points out the necessity of some “messy” SVT research designs due to the impossibility of getting genuine malingerers to volunteer for standardization studies. Standardization studies have had to rely on simulator studies and clinical samples; mostly ones of convenience and mostly forensic samples.
Circular reasoning, tautology and SVT research
If one uses SVT performance as the only index of effort and then concludes that SVT failure is a sign of “poor effort,” yet there are no other independent measures of what may be test behavior compliance or willingness to engage, process, and perform, or even malinger, then is this not a tautological argument? In such studies the only classifier defining poor effort is the SVT performance itself. While such studies often classify subjects by secondary gain identifiers (litigation, disability determination, etc.), such identifiers are not direct measures of effort, only of secondary gain.
Tautology also involves the unnecessary repetitive use of different words with the same meaning. The list of terms being used interchangeably with SVT and effort include response bias, invalid or failed performance, symptom amplification, performance exaggeration, underperformance or distortion, symptom embellishment, disingenuous, sub-optimal, poor effort, non-credible, faked and malingered. It is not uncommon to see statements like – “Failed SVT performance was associated with invalid neuropsychological test performance that was deemed non-credible due to sub-optimal effort.” The tautological problems with such a statement should be obvious.
Rigor of SVT studies
Class I and II level research as endorsed by the National Institutes of Health (NIH) and all major medical societies involves independently conducted investigations that have some external review and monitoring where investigators are independent of the outcome at all levels of the investigation (Edlund, Gronseth, So, & Franklin, Reference Edlund, Gronseth, So and Franklin2004). The best Class I and II investigations are those where a priori consensus diagnostic standards are in place that are independent of any outcome measure, data collected prospectively and independently from those in charge of their analysis, where clinicians involved in diagnostic decision making are independent of those who analyze the data, and who are also blinded. In Class I or II investigations, clinicians and data managers cannot also be the statisticians. Explicitly different roles at all levels of data acquisition, tabulation, analysis, and report writing increases the likelihood of unbiased findings. Institutional based investigations require human subjects review, consent, and IRB approval.
Few current SVT studies meet Class I or II level research or are subject to IRB approval. SVT studies that come from clinical practitioners in private practice not affiliated with an institution do not fall under any external review process whatsoever. Important investigational research comes from clinicians in private practice but rarely does this research meet a Class I or II standard.
Is There a Neurobiology of Drive, Effort, Motivation, and Attention?
Much of the discussion up to this point has focused on cognitive elements of SVT performance but there is also a behavioral dimension that centers on the neurobiology of drive, effort, and motivation (Sarter, Gehring, & Kozak, Reference Sarter, Gehring and Kozak2006). Patients with frontotemporolimbic damage may be apathetic with problems sustaining drive and goal-directed behaviors (Lezak, Howieson & Loring, Reference Lezak, Howieson and Loring2004). Apathy is a common consequence of traumatic brain injury (TBI; Marin & Wilkosz, Reference Marin and Wilkosz2005). What happens during neuropsychological assessment of the patient with neurogenic drive and motivation problems? The following test scores were obtained in a patient approximately one year post-TBI who the family described as unmotivated; neuroimaging demonstrated extensive bi-frontal and right parietal encephalomalacia and generalized atrophy: TOMM: Trial 1 = 45/50; Trial 2 = 50/50; Rey 15-Item: 10/15; Word Memory Test (WMT): IR (immediate recognition) = 78%; DR (delayed recognition) = 85%; CNS (consistency response) = 78%). The 45/50 on TOMM Trial 1 represents a pass, but is right at the cut-score with a perfect 50/50 on Trial 2 representing a pass by all standards. The Rey 15-Item performance represents a pass (although a borderline score by some standards); however, a 78% correct on the WMT IR and CNS Scales represents a “failure” by WMT standards. How does brain damage to motivational and attentional networks affect SVT performance and should patients with obvious structural lesions be evaluated by different cut-score standards?
Which Test to Use?
There are now numerous SVT measures available for use in general neuropsychological practice (Grant & Adams, Reference Grant and Adams2009; Lezak, Howieson, & Loring, 2004; Mitrushina, Boone, Razani, & D'Elia, Reference Mitrushina, Boone, Razani and D'Elia2005; Strauss, Sherman, & Spreen, Reference Strauss, Sherman and Spreen2006; Tate, Reference Tate2010) and potentially even more in a forensic setting (Boone, Reference Boone2007; Larrabee, Reference Larrabee2007; Morgan & Sweet, Reference Morgan and Sweet2009). While SVT use is endorsed by professionals, SVT test selection relies solely on the judgment of the researcher and/or clinician. With such a broad array of SVT measures that can be used, which ones should be used and in what circumstances? Also, a reasonable argument has been made that multiple SVTs are needed, especially in any lengthy or forensic assessment (Boone, Reference Boone2009; Larrabee, Reference Larrabee2008), but again no agreed upon professional standards as to the correct number, in what order, and in what context. Administration of multiple SVT measures also raises other questions when failures on some but not others occur and whether there is an order effect in SVT test administration (Ryan, Glass, Hinds, & Brown, Reference Ryan, Glass, Hinds and Brown2010)?
The Problem With Cut Scores
Dwyer (Reference Dwyer1996) reviewed the methods of cut-score development concluding that cut-scores (a) always entail judgment; (b) inherently result in some misclassification, (c) impose artificial “pass/fail” dichotomies and (d) no “true” cut scores exist (p. 360). Given Dwyer's comments, should cut scores be adjusted or customized to specific clinical conditions? Kirkwood and Kirk (Reference Kirkwood and Kirk2010) in a university-based assessment clinic evaluated 193 consecutively referred children with mild TBI. This study had IRB approval and these investigators examined both the TOMM and Medical Symptom Validity Test (MSVT; Green, Reference Green2004). Using their terminology, Kirkwood and Kirk (Reference Kirkwood and Kirk2010) found 17% of the sample to exhibit “suboptimal effort.” Only one failure was thought to be influenced by litigation. They also attempted to identify other potential sources for SVT failure, including impulsivity, distractibility during testing, pre-injury neuropsychiatric diagnosis and potential effects of reading disability. This study unmistakably demonstrates the complexities and potential issues in a clinical sample that may lead to sub-optimal performance.
In the Merten et al. (Reference Merten, Bossink and Schmand2007) study mentioned earlier that examined bona fide neurological patients with and without clinically obvious symptoms only 15/48 (31%) passed all SVTs. Only 1/24 (4%) of those with clinically obvious cognitive symptoms was able to pass all SVTs. These authors conclude that the “…. Results clearly show that many of these bona fide patients fail on most SVTs. Had the recommended cutoffs been applied rigorously without taking the clinical picture into consideration, these patients would have been incorrectly classified as exerting insufficient effort.” (p. 314). Donders and Strong (Reference Donders and Strong2011) attempted to replicate a logistic regression method that had been developed by Wolfe et al. (Reference Wolfe, Millis, Hanks, Fichtenberg, Larrabee and Sweet2010) to identify embedded effort indicators on the California Verbal Learning Test (CVLT). They not only applied the logistic regression method of Wolfe et al. but also used an externally administered SVT. However, the limits of interrelationship and inter-test agreement between actual test performances, embedded indicators of effort and the external SVT led them to conclude that this method was not ready for clinical application.
These studies simply underscore the difficulties of what needs to be addressed and accounted for in the neuropsychological application of current SVT technology. Similarly, Powell et al. (Reference Powell, Locke, Smigielski and McCrea2011) show that supposed markers of suboptimal effort using the Trail Making Test, also have limited predictive ability; again demonstrating the problem of disentangling true neuropsychological performance from associated elements of effort, drive, motivation, and attention when attempting to use embedded methods that were not explicitly designed to simultaneously assess validity.
Cut-scores are a necessary part of contemporary neuropsychological testing as they provide a method for classification but cut-scores are best used in the context of guidelines, rather than a dichotomous defining point for presence or absence of a deficit (Strauss et al., Reference Strauss, Sherman and Spreen2006).
False Memory and Dissociative Reactions
Mental health issues as they relate to “false memory” have been the topic of considerable controversy in clinical, research, and legal settings (Loftus & Davis, Reference Loftus and Davis2006). Interestingly, even an animal model of false memory has been developed (McTighe, Cowell, Winters, Bussey, & Saksida, Reference McTighe, Cowell, Winters, Bussey and Saksida2010). Are some SVT failures by neurological or neuropsychiatric patients generated by “false memories”? For example, cognitive neuroscience often examines how confident a subject is in their response, when assessing false memory (Moritz & Woodward, Reference Moritz and Woodward2006). No SVT studies to date have tackled this dimension.
Does Failed SVT Always Equate With Invalid Performance for All Neuropsychological Measures?
In the Locke et al. (Reference Locke, Smigielski, Powell and Stevens2008) investigation not all neuropsychological test scores were significantly suppressed in the failed SVT group. The Category Test and two of the three Wisconsin Card Sorting measures did not differ between the group that “passed” the SVT and the group that “failed” nor did scores on the Beck Depression and Anxiety scales. Whitney, Shepard, Mariner, Mossbarger, and Herman (Reference Whitney, Shepard, Mariner, Mossbarger and Herman2010) found that Wechsler Test of Adult Reading (WTAR) scores were no different between those with passed or failed SVT performance suggesting that WTAR findings remain “robust even in the face of suboptimal effort” (p. 196). Does this mean valid neuropsychological test findings occur with some tests even in the presence of SVT “failure”?
SVT Deception
Is the SVT measure infallible? DenBoer and Hall (Reference DenBoer and Hall2007) have shown that simulators can be “taught” how to detect SVT tasks and pass them and go on to fail the more formal neuropsychological measures (see also Rüsseler, Brett, Klaue, Sailer, & Munte, Reference Rüsseler, Brett, Klaue, Sailer and Munte2008). If the SVT task can be faked how would the clinician and/or researcher know?
Conclusions
SVT findings may offer important information about neuropsychological test performance, but if an oversimplified view of SVT-test behavior dichotomizes neuropsychological performance as either valid (above cut-score) or invalid (below cut-score), clinically important Type I and II errors are unavoidable. As shown in this review, patients with legitimate neurological and/or neuropsychiatric conditions fail SVTs for likely neurogenic factors. There can be no debate that issues of poor effort and secondary gain may have such a profound effect as to completely invalidate neuropsychological test findings rendering them uninterruptable (see Stevens, Friedel, Mehren, & Merten, Reference Stevens, Friedel, Mehren and Merten2008). However, considerably more SVT research is needed to address the issues raised in this review.
Acknowledgments
Parts of this were presented as a debate at the 38th Annual International Neuropsychological Society meetings on “Admissibility and Appropriate Use of Symptom Validity Science in Forensic Consulting” in Acapulco, Mexico moderated by Paul M. Kaufmann, J.D., Ph.D. Dr. Bigler co-directs Brigham Young University's Neuropsychological Assessment and Research Clinic that provides a service to the community evaluating a broad spectrum of clinical referrals. For legal referrals the University is compensated for the evaluation. Dr. Bigler also performs forensic consultation for which he is directly compensated. His research is National Institutes of Health (NIH) funded, but no NIH grant funds directly supported the writing of this commentary. Dr. Bigler has no financial interest in any commercial symptom validity test (SVT) measure. The assistance of Jo Ann Petrie, Thomas J. Farrer, and Tracy J. Abildskov in the preparation of this manuscript is gratefully acknowledged.