The Albany Consistency Index for the Test of Memory Malingering

Similar documents
Performance profiles and cut-off scores on the Memory Assessment Scales

KEVIN J. BIANCHINI, PH.D., ABPN

Characterization of the Medical Symptom Validity Test in evaluation of clinically referred memory disorders clinic patients

Comparison of Performance of the Test of Memory Malingering and Word Memory Test in a Criminal Forensic Sample

Effects of severe depression on TOMM performance among disability-seeking outpatients

The Test of Memory Malingering (TOMM): normative data from cognitively intact, cognitively impaired, and elderly patients with dementia

The Repeatable Battery for the Assessment of Neuropsychological Status Effort Scale

WPE. WebPsychEmpiricist

Potential for interpretation disparities of Halstead Reitan neuropsychological battery performances in a litigating sample,

Commentary on Delis and Wetter, Cogniform disorder and cogniform condition: Proposed diagnoses for excessive cognitive symptoms

Differential diagnosis in the memory clinic: Exploring the value of improved neuropsychological examination Rienstra, A.

The unexamined lie is a lie worth fibbing Neuropsychological malingering and the Word Memory Test


NEUROCOGNITIVE VARIABLES UNDERLYING GROUP PERFORMANCE ON A MEASURE OF EFFORT: THE MEDICAL SYMPTOM VALIDITY TEST (MSVT) Julie Hart Covert, M.S.

Malingering detection in a Spanish population with a known-groups design

AN ESTABLISHMENT OF EMBEDDED SYMPTOM VALIDITY TESTING WITHIN THE DELIS-KAPLAN EXECUTIVE FUNCTIONING SYSTEM. A Dissertation by. Emanuel J.

Archives of Clinical Neuropsychology Advance Access published April 27, John H. Denning*

Improving the Methodology for Assessing Mild Cognitive Impairment Across the Lifespan

THE WORD READING TEST OF EFFORT IN ADULT LEARNING DISABILITY: A SIMULATION STUDY

Ecological Validity of the WMS-III Rarely Missed Index in Personal Injury Litigation. Rael T. Lange. Riverview Hospital.

Factors Influencing the Face Validity of Effort Tests: Timing of Warning and Feedback

Cutoff Scores for MMPI-2 and MMPI-2-RF Cognitive-Somatic Validity Scales for Psychometrically Defined Malingering Groups in a Military Sample

The effect of distraction on the Word Memory Test and Test of Memory Malingering performance in patients with a severe brain injury

Exaggerated MMPI-2 symptom report in personal injury litigants with malingered neurocognitive deficit

RBANS Embedded Measures of Suboptimal Effort in Dementia: Effort Scale Has a Lower Failure Rate than the Effort Index

Published online: 25 Aug 2014.

Chapter 23. Detection of Children s Malingering on Raven s Standard Progressive Matrices*

Criterion validity of the California Verbal Learning Test-Second Edition (CVLT-II) after traumatic brain injury

Simulated subaverage performance on the Block Span task of the Stanford-Binet Intelligence Scales- Fifth Edition

Detection and diagnosis of malingering in electrical injury

The vulnerability to coaching across measures of malingering

Improving Accuracy in the. through the use of Technology

A Multi-Method Assessment Approach to the Detection of Malingered Pain: Association with the MMPI-2 Restructured Form

International Journal of Forensic Psychology Copyright Volume 1, No. 3 SEPTEMBER 2006 pp. 1-21

Interpretive Report. Client Information

Background 6/24/2014. Validity Testing in Pediatric Populations. Michael Kirkwood, PhD, ABPP/CN. Conflict of Interest Statement

Experimental study examining observational and objective methods of assessing effort in an undergraduate sample

Determining causation of traumatic versus preexisting. conditions. David Fisher, Ph.D., ABPP, LP Chairman of the Board PsyBar, LLC

Medical Symptom Validity Test Performance Following Moderate-Severe Traumatic Brain Injury: Expectations Based on Orientation Log Classification

Elderly Norms for the Hopkins Verbal Learning Test-Revised*

Malingering Detection among Accommodation- Seeking University Students

Discriminant Function Analysis of Malingerers and Neurological Headache Patients Self- Reports of Neuropsychological Symptoms

Noncredible Explanations of Noncredible Performance on Symptom Validity Tests

Influence of poor effort on self-reported symptoms and neurocognitive test performance following mild traumatic brain injury

Effects of Coaching on Detecting Feigned Cognitive Impairment with the Category Test

The Importance of Symptom Validity Testing with Children: WMT and MSVT. Lloyd Flaro Ed.D., Paul Green Ph.D. & Nina Blaskewitz Dipl.Psych.

Published online: 12 Aug 2014.

Published online: 12 Dec 2014.

Head injury and the ability to feign neuropsychological deficits

Donald A. Davidoff, Ph.D., ABPDC Chief, Neuropsychology Department, McLean Hospital Assistant Professor of Psychology, Harvard Medical School

Assessment of Suboptimal Effort Using the CVLT- II Recognition Foils: A Known-Groups Comparison

Professional Practice Guidelines for the Use and Interpretation of SVTs: The Time Has Come. Jerry Sweet May 20, 2011 London, England

Financial Disclosure

Comparison of Oral and Computerized Versions of the Word Memory Test

Pain, Malingering, and Performance on the WAIS-III Processing Speed Index

Shane S. Bush & Robert L. Heilbronner & Ronald M. Ruff

A Dissertation. Submitted to the Faculty. Xavier University. in Partial Fulfillment of the. Requirements for the Degree of. Doctor of Psychology

Use of the Kaufman Brief Intelligence Test - Second Edition as an embedded measure of malingering in a college population

Enhancing neuropsychological performance in chronic cannabis users: The role of motivation

The Delis-Kaplan Executive Functions System Tower Test Resilience to Response Bias

Healthy Children Get Low Scores Too: Prevalence of Low Scores on the NEPSY-II in Preschoolers, Children, and Adolescents

Clinical Utility of Wechsler Memory Scale-Revised and Predicted IQ Discrepancies in Closed Head Injury

Normative Corrections for the Halstead Reitan Neuropsychological Battery

Malingering (AADEP Position Paper) The gross volitional exaggeration or fabrication of symptoms/dysfunction for the purpose of obtaining substantial m

Passing or Failing of Symptom Validity Tests in Academic Accessibility Populations: Neuropsychological Assessment of Near-Pass Patients

Using Neuropsychological Experts. Elizabeth L. Leonard, PhD

Minimizing Misdiagnosis: Psychometric Criteria for Possible or Probable Memory Impairment

MEGAN CIOTA, PH.D. Curriculum Vitae Address: 3939 Houma Boulevard, #223 Metairie, LA Phone Number: (504) Fax: (504)

Comparison of Predicted-difference, Simple-difference, and Premorbid-estimation methodologies for evaluating IQ and memory score discrepancies

International Journal of Forensic Psychology Copyright Volume 1, No. 3 SEPTEMBER 2006 pp

Validation of the Symptoms of Post- Concussion Syndrome Questionnaire as a Self-Report Symptom Validity Test: A Simulation Study. Victoria Jayne Reece

Malingering Detection Measure Utility and Concordance in a University Accommodation- Seeking Student Population

A Comparison Of Symptom Validity Tests For Academic Neuropsychological Evaluations: Medical Symptom Validity Test Versus The Word Reading Test

Archives of Clinical Neuropsychology 27 (2012)

Executive dysfunction in traumatic brain injury: The effects of injury severity and effort on the Wisconsin Card Sorting Test

Statistics as a Tool. A set of tools for collecting, organizing, presenting and analyzing numerical facts or observations.

Neuropsychological Testing (NPT)

Effortless Effort: Current Views on Assessing Malingering litigants in Neuropsychological Assessments

Cognitive-Behavioral Assessment of Depression: Clinical Validation of the Automatic Thoughts Questionnaire

Award Number: W81XWH

DETECTION OF MALINGERED MENTAL RETARDATION

Rapidly-administered short forms of the Wechsler Adult Intelligence Scale 3rd edition

APPLYING REACTION TIME (RT) AND EVENT-RELATED POTENTIAL (ERPS) MEASURES TO DETECT MALINGERED NEUROCOGNITIVE DEFICIT

What s Wrong With My Client: Understanding Psychological Testing in Order to Work Effectively With Your Expert

VALIDATION OF THE MILLER FORENSIC ASSESSMENT OF SYMPTOMS TEST (M- FAST) IN A CIVIL FORENSIC POPULATION

Attention and Memory Dysfunction in Pain Patients While Controlling for Effort on the California Verbal Learning Test-11

ISSN: (Print) (Online) Journal homepage:

Malingering: The Impact of Effort on Outcome. Gordon J. Horn, PhD Clinical Neuropsychologist National Deputy Director of Clinical Outcomes

Archives of Clinical Neuropsychology 31 (2016)

10/5/2015. Advances in Pediatric Neuropsychology Test Interpretation Part I: Importance of Considering Normal Variability. Financial Disclosures

THE EFFICIACY OF THE MMPI-2 LEES- HALEY FAKE BAD SCALE (FBS) FOR DIFFERENTIATING NEUROCOGNITIVE AND PSYCHIATRIC FEIGNERS

Neuropsychological Performance in Cannabis Users and Non-Users Following Motivation Manipulation

Chapter 31 Psychometric Foundations for the Interpretation of Neuropsychological Test Results *

MMPI-2 short form proposal: CAUTION

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

The significance of sensory motor functions as indicators of brain dysfunction in children

Utility of the MMPI 2-RF (Restructured Form) Validity Scales in Detecting Malingering in a Criminal Forensic Setting: A Known-Groups Design

THE VALIDITY OF THE LETTER MEMORY TEST AS A MEASURE OF MEMORY MALINGERING: ROBUSTNESS TO COACHING. A dissertation presented to.

Base Rates of Impaired Neuropsychological Test Performance Among Healthy Older Adults

Concurrent validity of WAIS-III short forms in a geriatric sample with suspected dementia: Verbal, performance and full scale IQ scores

Transcription:

Archives of Clinical Neuropsychology 27 (2012) 1 9 The Albany Consistency Index for the Test of Memory Malingering Jessica H. Gunner 1, *, Andrea S. Miele 1, Julie K. Lynch 2, Robert J. McCaffrey 1,2 1 Department of Psychology, University at Albany, State University of New York, Albany, NY, USA 2 Albany Neuropsychological Associates, Albany, NY, USA *Corresponding author at: Department of Psychology, University at Albany, State University of New York, Social Sciences 399, 1400 Washington Ave, Albany, NY, 12222, USA. E-mail address: JGunner@albany.edu (Jessica H. Gunner). Abstract Accepted 10 October 2011 The determination of examinee effort is an important component of a neuropsychological evaluation and relies heavily on the use of symptom validity tests (SVTs) such as the Test of Memory Malingering (TOMM) and the Word Memory Test (WMT). Diagnostic utility of SVTs varies. The sensitivity of traditional TOMM criteria to suboptimal effort is low. An index of response consistency across three trials of the TOMM was developed, denoted the Albany Consistency Index (ACI). This index identified a large proportion of examinees classified as optimal effort using traditional TOMM interpretive guidelines but suboptimal effort using the WMT profile analysis. In addition, previous research was extended, demonstrating a relationship between examinee performance on SVTs and neuropsychological tests. Effort classification using the ACI predicted the performance on the Global Memory Index from the Memory Assessment Scales. In conclusion, the ACI was a more sensitive indicator of suboptimal effort than traditional TOMM interpretive guidelines. Keywords: Symptom validity tests; Test of Memory Malingering; Word Memory Test; Albany Consistency Index; Memory Assessment Scales; Medico-legal; Forensic; Malingering Introduction In neuropsychological assessment, examinee performance may range from optimal effort with accurate symptom reporting to suboptimal effort (performance invalidity) with over-reporting of symptoms (Bush et al., 2005). Well-validated measures of effort are critical to the determination of an examinee s test performance. Failure to perform adequately on effort testing calls into question the validity and reliability of scores from the entire assessment, and conclusions regarding the presence of brain dysfunction cannot be made in the context of failed symptom validity testing (Heilbronner et al., 2009; Hom & Denny, 2002; Iverson & Binder, 2000; Rogers, 1997). Furthermore, Green, Rohling, Lees-Haley, and Allen (2001) found that effort accounted for as much as 53% of the variance in neuropsychological test scores, and that suboptimal effort examinees perform more poorly on neuropsychological tests than those with moderate or severe brain injuries. In fact, researchers have demonstrated that examinees putting forth suboptimal effort perform significantly worse on neuropsychological tests (Beetar & Williams, 1995; Constantinou, Bauer, Ashendorf, & McCaffrey, 2005; van Gorp et al., 1999). For example, examinees with suboptimal effort scored lower on sensory, motor, and attentional measures of the Halstead-Reitan Neuropsychological Battery (HRNB) than those with optimal effort (Mittenberg, Rotholc, Russell, & Heilbronner, 1996). More specifically, effort as determined by the Test of Memory Malingering (TOMM) predicted performance on various neuropsychological measures from the HRNB and Wechsler Adult Intelligence Scale-Revised (WAIS-R) (Constantinou et al., 2005). This paper is based upon the first author s Master thesis. Portions of the paper were presented at the 30th annual meeting of the National Academy of Neuropsychology, Vancouver, BC. Note: For information on the Albany Consistency Index for the Test of Memory Malingering scoring program please contact permission@mhs.com, do not contact the authors. # The Author 2011. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com. doi:10.1093/arclin/acr089 Advance Access publication on 21 November 2011

2 J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 In addition, suboptimal effort simulators were found to perform significantly lower and had other atypical features to their performance on the Memory Assessment Scales (MAS) than did optimal effort simulators (Beetar & Williams, 1995). Base rates of suboptimal effort are especially high in forensic and legal work compared with more traditional clinical settings; however, patients seen within a clinical context have also been found to demonstrate suboptimal effort (Larrabee, 2003; Mittenberg, Patton, Canyock, & Condit, 2002). For example, Mittenberg and colleagues (2002) found that 8% of medical cases exhibited suboptimal effort, whereas 29% of personal injury claimants, 30% of disability claimants, and 19% of criminal cases evidenced suboptimal effort. Binder (1993) found similar suboptimal effort rates for litigants, whereas Larrabee (2003) and Mittenberg and colleagues (2002) found that 40% of mild head injury litigants exhibited probable suboptimal effort. In light of these base rates, any neuropsychological evaluation should be considered incomplete without assessing the examinee s level of effort (Bush et al., 2005; Iverson & Binder, 2000). Although important, clinical judgment alone is not sufficient to detect suboptimal effort. Tests specifically designed to assess effort are necessary and should be utilized throughout a neuropsychological assessment (van Gorp et al., 1999). Both the National Academy of Neuropsychology (Bush et al., 2005) and the American Academy of Clinical Neuropsychology (Heilbronner et al., 2009) state that all neuropsychological evaluations should examine effort at various points during the assessment, preferably using multiple symptom validity measures from different cognitive domains. Numerous free-standing symptom validity tests (SVTs) have been developed which permit objective identification of suboptimal effort and are not sensitive to neurological deficits. In fact, individuals with brain damage, developmental disabilities, psychiatric problems, and dementia typically perform well on these measures (Heilbronner et al., 2009). The TOMM is one of the most frequently used SVTs (O Bryant & Lucas, 2006; Slick, Tan, Strauss, & Hultsch, 2004). A well-researched measure, TOMM cut-off scores have shown to distinguish between suboptimal and optimal effort regardless of age, education, psychological condition, laboratory induced pain, or cognitive impairment (Ashendorf, Constantinou, & McCaffrey, 2004; Etherton, Bianchini, Greve, & Ciota, 2005; O Bryant, Finlay, & O Jile, 2007; Rees, Tombaugh, & Boulay, 2001; Rees, Tombaugh, Gansler, & Moczynski, 1998; Tombaugh, 1996, 1997). Despite its favor among neuropsychologists, studies have shown that a subset of examinees classified as optimal effort by traditional TOMM criteria is classified as suboptimal effort by Word Memory Test (WMT) profile analysis during the same evaluation (Greve, Binder, & Binachini, 2009; Greve, Ord, Curtis, Bianchini, & Brennan, 2008). For example, Greve and colleagues (2008) found that TOMM cut-off scores had a sensitivity to suboptimal effort of 0.56, whereas WMT profile analysis had a sensitivity of 0.85 in a litigating, mild traumatic brain injury sample. Additionally, Gervais, Rohling, Green, and Ford (2004) reported that while 10% of claimants were identified as suboptimal effort by TOMM criteria, 30% were similarly identified by the WMT criteria. Given the TOMM s popularity among neuropsychologists, clinicians are likely to select this SVT as a measure of effort; however, the traditional TOMM criteria may not be as sensitive to suboptimal effort as criteria from other SVTs such as the WMT. The purpose of the present study was to examine a novel response consistency index to determine whether or not this index might improve the classification accuracy of the TOMM relative to the WMT. First, response consistency was examined across all possible permutations of TOMM trials, and a response consistency index, denoted as the Albany Consistency Index, was developed to identify examinees misclassified as optimal effort (false negatives) using the traditional TOMM criteria. Second, the relationship between effort and neuropsychological test performance was examined using the Global Memory Index of the MAS. This analysis involved a comparison of memory test performance by optimal and suboptimal effort examinees with effort level determined by the standard interpretive TOMM criteria, WMT criteria (profile analysis), and the Albany Consistency Index. Method Participants After obtaining IRB approval, an archival of analysis was completed on 48 cases from a private neuropsychological practice that were conducted for medico-legal reasons (i.e., compensation-seeking, litigation, or disability claims). This was a mixed neurological sample, although all but five examinees had a history of mild traumatic brain injury and met the mild traumatic brain injury criteria from the American Congress of Rehabilitation Medicine (ACRM) Mild Traumatic Brain Injury Committee of the Head Injury Interdisciplinary Special Interest Group (Committee on Mild Traumatic Brain Injury, 1993). Five examinees presented with other medical conditions, diagnosed by their treating neurologist. The diagnoses were fibromyalgia (n ¼ 2); transient ischemic attack (n ¼ 1); brain stem stroke and major depressive disorder (n ¼ 1); and multiple sclerosis (n ¼ 1).

J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 3 Materials The archival data consisted of scores from the HRNB (Reitan and Wolfson, 1993) and the MAS (Williams, 1991), a memory test that contains four summary index scores: Short-term Memory, Verbal Memory, Visual Memory, and Global Memory. Symptom validity measures consisted of the TOMM (Tombaugh, 1996) and WMT (Green, 2003). Procedure: Creating the Albany Consistency Index All examinees completed both the WMT and the TOMM. Examinee data were classified into eight groups for subsequent analyses. Groups 7 and 8 were classified using the Albany Consistency Index, which is described below. Group 1: TOMM Optimal Effort Group 2: TOMM Suboptimal Effort Group 3: WMT Optimal Effort Group 4: WMT Suboptimal Effort Group 5: TOMM Optimal Effort; WMT Optimal Effort Group 6: TOMM Optimal effort; WMT Suboptimal Effort Group 7: Albany Consistency Index Optimal Effort Group 8: Albany Consistency Index Suboptimal Effort TOMM individual item responses were used to create an Index of consistently incorrect, inconsistent, orconsistently correct responses in an effort to quantify the consistency of examinees responses across Trial 1, Trial 2, and the Retention Trial. Responses for items during each of these trials were coded as either one (1) for a correct response or zero (0) for an incorrect response. Coded responses were then evaluated across four different groupings of the three TOMM trials: Trial 1 and Trial 2 (Index A), Trial 1 and the Retention Trial (Index B), Trial 2 and the Retention Trial (Index C), and across all three trials (Index D). For each index, examinees responses were then identified as consistently incorrect, inconsistent, or consistently correct for each of the 50-items. Consistently incorrect responses were those answered incorrectly on both trials of Indices A, B, or C, or all three trials of Index D. Inconsistent responding was defined as one correct response and one incorrect response in any order on Indices A, B, and C. Inconsistent responding across three trials for Index D was defined as either responding to the item correctly on one trial and incorrectly on the other two trials, or correctly on two trials and incorrectly on one trial. Consistently correct responding was defined as responding correctly across both trials for Indices A, B, or C, or across all three trials of Index D. A description of each of the four indices is presented in Table 1. The response consistency patterns of examinees in Groups 5 and 6 were examined to assess for any variation between (i) examinees who exhibited optimal effort on both the TOMM and WMT criteria and (ii) examinees whose effort was optimal based on the TOMM criteria but whose effort was suboptimal based on WMT criteria. In this study, 16 examinees exhibited optimal effort based on TOMM criteria but suboptimal effort based on WMT profile analysis. These optimal effort performances identified using the TOMM were considered to represent false negatives; that is, the TOMM criteria failed to detect suboptimal effort that was identified by the WMT criteria. Twenty-three examinees exhibited optimal effort based on both the TOMM and WMT criteria. The TOMM performances in this group were considered true negatives. The frequency of examinees in Groups 5 and 6 responding consistently incorrect, inconsistent, and consistently correct was then compared for each index. After examining the frequency graphs for false positives and false negatives for all groups across all indices, Index D was selected for development of a cutoff score, and from this point forward will be referred to as the Albany Consistency Index. This index was selected because none of the examinees exhibiting optimal effort based Table 1. Trial comparisons and types of consistency Index A B C D Trial comparison Trials 1 and 2 Trial 1 and Retention Trial Trial 2 and Retention Trial Trials 1, 2, and Retention Trial Consistently incorrect Incorrect on both trials Incorrect on both trials Incorrect on both trials Incorrect on all three trials Inconsistent Incorrect on one trial Incorrect on one trial Incorrect on one trial Incorrect on one trial OR two trials Consistently correct Correct on both trials Correct on both trials Correct on both trials Correct on all three trials

4 J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 on the TOMM and WMT criteria responded inconsistently on ten or more items of the Albany Consistency Index. As such, the cutoff score for the identification of suboptimal effort using the Albany Consistency Index was set at 10 inconsistent responses. Examinees who responded inconsistently on 10 items on the Albany Consistency Index were classified as putting forth suboptimal effort (Group 8), while those who responded inconsistently on,10 items on the Albany Consistency Index were classified as putting forth optimal effort (Group 7). Procedure: Remaining Analyses Statistical tests included three logistic regressions and univariate ANOVAs. Logistic regression was used to examine the accuracy with which the standard TOMM criteria and the Albany Consistency Index classified optimal and suboptimal effort relative to WMT criteria. This was used to determine the predictive validity of the Albany Consistency Index over and above the standard TOMM criteria in determining suboptimal effort as established by the WMT criteria. Univariate tests were utilized to examine the differences on the Global Memory Index of the MAS between optimal and suboptimal effort groups. Three separate univariate tests were conducted to compare the MAS test performances between the effort groups as determined by the standard TOMM criteria, Albany Consistency Index, and WMT criteria. Results Demographics and Classification Accuracy Detailed demographic characteristics for all groups can be found in Table 2. Examinees average age and years of education were 45.3 (9.8) and 14.2 (2.9), respectively. Of these subjects, 23 were male and 45 were right handed. The average scores on the WAIS-R Verbal IQ (VIQ), Performance IQ (PIQ), and Full-scale IQ (FSIQ) scores were 97 (12), 94 (11), and 95 (11), respectively. Chi-square analyses showed no significant differences in age, education, gender, handedness, or WAIS-R scores between optimal and suboptimal effort examinees based on standard TOMM criteria, Albany Consistency Index, or WMT criteria. One examinee was classified as suboptimal effort on the TOMM but classified as optimal effort on the WMT (false positive) and 16 examinees were classified as optimal effort on the TOMM but suboptimal effort on the WMT (false negatives). This indicated that, based on the WMT criteria, the standard TOMM criteria had a sensitivity of 0.33 and a specificity of 0.96 (see Fig. 1). The Albany Consistency Index correctly identified suboptimal effort in nine of the 16 examinees (56%) classified as Table 2. Means, Standard Deviations, and ranges for the eight participant groups Group 1 2 3 4 5 6 7 8 SVT TOMM (n ¼ 48) WMT (n ¼ 48) TOMM Pass (n ¼ 39) TOMM Fail (n ¼ 9) WMT Pass (n ¼ 24) WMT Fail (n ¼ 24) TOMM Pass and WMT Pass (n ¼ 23) TOMM Pass and WMT Fail (n ¼ 16) Pass Consistency Index (n ¼ 30) Fail Consistency Index (n ¼ 18) Age M (SD) 45 (10) 46 (10) 41 (10) 45 (10) 44 (10) 46 (9) 44 (10) 49 (7) 45 (10) 46 (10) Range 19 67 21 67 19 49 19 67 21 67 19 58 21 67 35 58 21 67 19 58 Gender (M/F) 23/25 21/18 2/7 23/25 14/10 9/15 13/10 8/8 14/16 5/12 Education M (SD) 14.2 (2.9) 14.3 (2.8) 14.0 (3.3) 14.2 (2.9) 14.5 (3.0) 14.0 (2.8) 14.5 (3.1) 13.9 (2.5) 14.6 (2.9) 13.6 (2.8) Range 9 21 9 21 9 18 9 21 9 21 9 18 9 21 10 18 9 21 9 18 Handedness (R/L) 45/3 36/3 9/0 45/3 23/1 22/2 22/1 14/2 28/2 17/1 WAIS-R VIQ M (SD) 97 (12) 98 (12) 91 (12) 97 (12) 101 (11) 93 (12) 101 (11) 94 (12) 100 (12) 93 (10) Range 62 122 62 122 67 106 62 122 82 122 62 113 82 122 62 113 62 122 67 107 PIQ M (SD) 94 (11) 96 (11) 86 (10) 94 (11) 98 (12) 90 (10) 99 (12) 91 (9) 98 (11) 88 (8) Range 74 126 75 126 74 105 74 126 75 126 74 111 75 126 77 111 75 126 74 105 FSIQ M (SD) 95 (11) 97 (11) 88 (11) 95 (11) 99 (11) 91 (10) 100 (10) 93 (10) 98 (11) 90 (9) Range 67 121 67 121 69 102 67 121 85 121 67 107 85 121 67 107 67 121 69 102

J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 5 Fig. 1. Calculating classification accuracy for the TOMM based on WMT classification. Fig. 2. Calculating classification accuracy for the Albany Consistency Index based on WMT classification. optimal effort using the TOMM criteria but as suboptimal effort based on WMT criteria. In addition, while identifying 56% of the false negatives, the Albany Consistency Index did not result in any false positives. In fact, this index correctly classified the one false positive from the standard TOMM criteria (i.e., the examinee was correctly classified as optimal effort). Thus, evaluating effort using the Albany Consistency Index increased the sensitivity of the TOMM cut-off scores from 0.31 to 0.71 and also increased specificity to 1.00 (see Fig. 2). Validation of the Albany Consistency Index Logistic regression was used to evaluate the incremental validity of the Albany Consistency Index. Three direct logistic regressions were executed. The first (Model 1) utilized only the standard TOMM criteria as an independent variable in order to classify examinees into optimal and suboptimal effort groups as defined by the WMT criteria. Table 3 shows the regression coefficients, Wald statistics, odds ratios, and 95% confidence intervals for the odds ratios for the TOMM criteria as a sole predictor variable. This model had a x 2 -value of 59.08. According to the Wald criterion, the TOMM criteria reliably predicted effort as determined by the WMT criteria, x 2 (1, N ¼ 48) ¼ 4.85, p ¼.028. The variance in effort accounted for by the standard TOMM criteria was small (McFadden s D ¼ 0.11; 95% confidence interval of the effect ¼ 2.44). The overall accuracy rate was 64.6%, in which 95.8% of the optimal effort examinees and only 33.3% of the suboptimal effort examinees were correctly classified. This model produced an odds ratio of 11.50, which indicates that a failing performance on the WMT is 11.5 times more likely if an examinee has failed the TOMM.

6 J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 Table 3. Logistic regression analysis of effort determined by the WMT as a function of suboptimal and optimal effort based on the standard TOMM criteria Variable B Wald chi-square Odds ratio 95% Confidence interval for odds ratio Lower Upper TOMM 2.442 4.846 11.500 1.307 101.182 Table 4. Logistic regression analysis of effort determined by the WMT as a function of suboptimal and optimal effort based on the standard TOMM criteria and the Albany Consistency Index Variable b Wald chi-square Odds ratio 95% Confidence interval for odds ratio Lower Upper TOMM 0.000 0.000 1.000 0.053 18.915 Albany Consistency Index 3.091 7.375 22.000 2.364 204.759 The second direct logistic regression (Model 2) examined the classification accuracy of the standard TOMM criteria and the Albany Consistency Index together to determine whether the Albany Consistency Index increased the correct classification of suboptimal and optimal effort based on the WMT criteria compared with the standard TOMM criteria alone (Model 1). Table 4 shows the results from the logistic regression for both the standard TOMM criteria and the Albany Consistency Index. According to the Wald criterion, the Albany Consistency Index reliably predicted effort as determined by the WMT, x 2 (1, n ¼ 48) ¼ 7.38, p ¼.007, while the standard TOMM criteria did not significantly predict effort, x 2 (1, n ¼ 48) ¼ 0.00, p ¼ 1.000. The variance in effort accounted for by the standard TOMM criteria and Albany Consistency Index in this second model was larger than that obtained in the previous logistic regression (McFadden s D ¼ 0.29; 95% confidence interval for the effect ¼ 3.09). This analysis correctly classified 91.7% of the optimal effort examinees and 66.7% of the suboptimal effort examinees, with an overall accuracy rate of 79.2% compared with 64.6% in the first model. The odds ratio of 1.00 associated with the TOMM criteria indicates that suboptimal effort as determined by these criteria did not increase the likelihood of also having suboptimal effort as determined by the WMT criteria. However, according to the second odds ratio, identifying suboptimal effort on the WMT is 22.00 times more likely if an examinee exhibited suboptimal effort on the Albany Consistency Index. The second model had an overall x 2 -value of 47.35, which is significantly smaller than the first model with a x 2 -value of 59.08 (p,.05). Statistically, the use of both the standard TOMM criteria and Albany Consistency Index as predictors of WMT performance is more accurate in classifying effort than the TOMM alone. The final logistic regression (Model 3) analyzed the utility of the Albany Consistency Index alone in predicting effort as determined by the WMT criteria. See Table 5 for results from the logistic regression including only the Albany Consistency Index. According to the Wald criterion, the Albany Consistency Index reliably predicted effort as determined by the WMT criteria, x 2 (1, N ¼ 48) ¼ 13.04, p,.005. The variance in effort accounted for by the Albany Consistency Index was identical to the second model that utilized both the standard TOMM and the Albany Consistency Index (McFadden s D ¼ 0.29; 95% confidence interval for the effect ¼ 3.09). Classification results were also identical to the second model with 91.7% of the optimal effort examinees and 66.7% of the suboptimal effort examinees correctly classified, for an overall accuracy rate of 79.2%. Similar to the second model, the odds ratio of 22.00 indicates that suboptimal effort based on the WMT criteria is 22 times more likely if an examinee s score on the Albany Consistency Index indicates suboptimal effort. Additionally, this third model had the exact same x 2 value (x 2 ¼ 47.35) as the second model. Therefore, the standard TOMM criteria do not add any incremental validity to the Albany Consistency Index in classifying suboptimal and optimal effort as determined by the WMT criteria. Relationship of Effort to Memory Performance Three univariate ANOVAs were conducted to compare the performance on the Global Memory Index of the MAS between examinees with optimal versus suboptimal performance on the TOMM, Albany Consistency Index, and WMT. The Short-Term Memory Index, Verbal Memory Index, and Visual Memory Index were all significantly correlated with the Global Memory Index (r ¼.46, p,.0005, r ¼.88, p,.0005, and r ¼.86, p,.0005, respectively); thus, the Short-term Memory, Verbal Memory, and Visual Memory Indices were regressed on the Global Memory Index. This resulted in a significant adjusted R 2 value of.995 (F (3, 48) ¼ 3,695.3, p,.0005), indicating that these three indices accounted for 99.5% of the

J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 7 Table 5. Logistic regression analysis of effort determined by the WMT as a function of suboptimal and optimal effort based on the Albany Consistency Index Variable b Wald chi-square Odds ratio 95% Confidence interval for odds ratio Lower Upper Albany Consistency Index 3.091 13.036 22.000 4.109 117.804 Table 6. Means, standard deviations, and ranges for the Global Memory Index of the Memory Assessment Scale Global Memory Index TOMM optimal effort TOMM suboptimal effort Albany Consistency Index optimal effort Albany Consistency Index suboptimal effort WMT optimal effort WMT suboptimal effort Mean 87.95 75.00 90.09 77.82 93.79 77.65 Standard 14.74 15.78 15.13 13.19 14.30 13.02 deviation Range 53 114 56 102 53 114 56 102 63 114 53 102 variance in the Global Memory Index scores. This was expected, as the Global Memory Index is comprised of scores from the other three indices. Based on this analysis, only the Global Memory Index of the MAS was used for group comparisons to avoid redundancy and multicollinearity. Group means, ranges, and standard deviations on the Global Memory Index can be found in Table 6. ANOVAs were used instead of t-tests to obtain the eta-squared values (a measure of effect size) for each comparison. Eta-squared values provide the variance in the dependent variable (Global Memory Index) uniquely attributable to the predictor variable (TOMM, Albany Consistency Index, or WMT). Following a Bonferroni correction, the alpha level was adjusted to p,.0166. The results indicated that there were no significant differences in scores on the Global Memory Index between examinees with optimal versus suboptimal effort as determined by the standard TOMM criteria, F (1, 47) ¼ 5.06, p ¼.029. Effort based on the standard TOMM criteria accounted for only 9.7% of the variance in the Global Memory Index (eta-squared obtained ¼.097). In contrast, there were significant differences on Global Memory Index scores between optimal and suboptimal effort examinees as determined by the Albany Consistency Index, F (1, 47) ¼ 7.95, p ¼.007. Examinees exhibiting optimal effort on the Albany Consistency Index performed significantly better on the Global Memory Index compared with those putting forth suboptimal effort. In this analysis, effort as determined by the Albany Consistency Index accounted for 14.5% of the variance in scores on the Global Memory Index (eta-squared obtained ¼.145). There were also significant differences in scores on the Global Memory Index between optimal versus suboptimal effort examinees as determined by WMT criteria, F (41, 5) ¼ 16.33, p,.005, with optimal effort examinees performing significantly better on the Global Memory Index. Effort as indexed by WMT criteria accounted for 26.6% of the variance in scores on the Global Memory Index (eta-squared obtained ¼.266). Discussion This study examined whether a response consistency index would improve the TOMM s sensitivity to suboptimal effort based on WMT classifications. The Albany Consistency Index identified 56% of the examinees who exhibited optimal effort on the TOMM but suboptimal effort on the WMT, while not identifying any optimal effort examinees as putting forth suboptimal effort. The Albany Consistency Index increased the sensitivity of the traditional TOMM cut-off scores from 0.31 to 0.71 while maintaining high specificity (0.96 1.00). These results indicate that when effort was determined by the WMT criteria, the Albany Consistency Index provided more accurate classification of suboptimal effort and perfect classification of optimal effort compared with the traditional TOMM criteria. Three logistic regressions were conducted to determine the ability of the standard TOMM criteria alone (Model 1), standard TOMM criteria and Albany Consistency Index together (Model 2), and the Albany Consistency Index alone (Model 3) to predict effort as classified by the WMT criteria. Model 1 indicated that the standard TOMM criteria are much better at identifying optimal rather than suboptimal effort. Model 2 showed an increased correct classification of effort groups and also indicated that the standard TOMM criteria did not significantly increase the predictive accuracy of the Albany Consistency Index. Thus, Model 3, which used only the Albany Consistency Index, was the simplest and most accurate model for classifying effort as determined by the WMT criteria. This model resulted in the exact same x 2 value, classification results, and variance accounted for in the WMT as obtained in Model 2, indicating that the standard TOMM criteria did not increase the incremental

8 J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 validity of the Albany Consistency Index. Additionally, those examinees identified as suboptimal effort by the Albany Consistency Index were almost two times more likely than those identified as suboptimal effort by the standard TOMM criteria to have also been identified as suboptimal effort by the WMT criteria. Thus, together, logistic regression and classification accuracy statistics demonstrate the utility of the Albany Consistency Index in identifying suboptimal effort. A second goal of this study was to replicate and expand prior research, showing that suboptimal effort examinees perform lower than the optimal effort examinees on neuropsychological tests (Beetar & Williams, 1995; Constantinou et al., 2005). The results of the current study demonstrated that suboptimal effort examinees performed significantly lower on the Global Memory Index of the MAS when using either the WMT criteria or the Albany Consistency Index to classify effort. However, when classifying effort using the standard TOMM criteria, suboptimal and optimal effort groups obtained similar Global Memory Index scores. It is likely that effort as determined by these criteria were not predictive of scores on the Global Memory Index due to the TOMM criteria s limited ability to accurately classify suboptimal effort. Examinees identified as optimal rather than suboptimal effort by the WMT criteria or the Albany Consistency Index were more likely to have higher scores on the Global Memory Index. Therefore, effort classification by both WMT criteria and the Albany Consistency Index was predictive of scores on the Global Memory Index. Effort as determined by the WMT criteria, the Albany Consistency Index, and standard TOMM criteria accounted for 26.6%, 14.5%, and only 9.7% of the variance on the Global Memory Index, respectively. Limitations and Future Directions Results from the logistic regressions and univariate ANOVAs may only generalize to a similar litigating sample. In addition, when utilizing logistic regression, it is important to consider both correlations between the predictor variables as well as between the predictor and outcome variables. Although within acceptable limits for a logistic regression, the variables used in the three models (the TOMM, Albany Consistency Index, and WMT) were moderately correlated and, therefore, results from the inferential statistics may overestimate their classification ability. For this reason, causal relationships should not be inferred from the above results. Additionally, due to conducting multiple comparisons, a Bonferroni correction was utilized to protect against Type I error; however, these corrected alpha values may be overly stringent and mask true differences between the groups. Finally, although computing the consistency scores from individual TOMM data may require additional time, a computer spreadsheet application can provide assistance. According to Bush and colleagues (2005), increased time allocated to the accurate assessment of effort is both necessary and justified. Future research may benefit from utilizing a larger sample that would allow for more powerful statistical tests. In addition, the Albany Consistency Index should be cross-validated using both litigating and clinical samples. Conclusions The major findings of this study validate the use of the Albany Consistency Index for the TOMM and support previous research, showing that suboptimal effort examinees perform more poorly on neuropsychological tests than do optimal effort examinees (Beetar & Williams, 1995; Constantinou et al., 2005). Examinees performing better on the Global Memory Index were significantly more likely to have optimal effort scores on both the Albany Consistency Index and WMT compared with those who performed poorly on the Global Memory Index. More importantly, the Albany Consistency Index identified a large proportion of examinees who exhibited optimal effort based on the traditional TOMM criteria but suboptimal effort based on the WMT criteria. The Albany Consistency Index more than doubled the sensitivity of the traditional TOMM criteria for identifying suboptimal effort, thereby demonstrating that it is a more sensitive indicator of effort than the traditional TOMM criteria. Conflict of Interest The first author would like to acknowledge the generous travel grant support from the Neuropsychology Division of the New York State Psychological Association. Acknowledgements The authors are grateful to Cecil R. Reynolds, former editor of Archives of Clinical Neuropsychology and current editor of Psychological Assessment, who served as the guest action editor. This manuscript was submitted to the blind peer review

process. The authors also wish to acknowledge the assistance provided by Derek Stenclik with the scoring program for the Albany Consistency Index. References J.H. Gunner et al. / Archives of Clinical Neuropsychology 27 (2012) 1 9 9 Ashendorf, L., Constantinou, M., & McCaffrey, R. J. (2004). The effect of depression and anxiety on the TOMM in community-dwelling older adults. Archives of Clinical Neuropsychology, 18, 125 130. Beetar, J. T., & Williams, J. M. (1995). Malingering response styles on the Memory Assessment Scales and Symptom Validity Tests. Archives of Clinical Neuropsychology, 10, 57 72. Binder, L. M. (1993). Assessment of malingering after mild head trauma with the Portland Digit Recognition Test. Journal of Clinical and Experimental Neuropsychology, 15, 170 182. Bush, S. S., Ruff, R. M., Troster, A. I., Barth, J. T., Koffler, S. P., Pliskin, N. H., et al. (2005). Symptom validity assessment: Practice issues and medical necessity NAN policy & planning committee. Archives of Clinical Neuropsychology, 20, 419 426. Committee on Mild Traumatic Brain Injury, American Congress of Rehabilitation Medicine (ACRM) (1993). Definition of mild traumatic brain injury. Journal of Head Trauma Rehabilitation, 8 (3), 86 87. Constantinou, M., Bauer, L., Ashendorf, L., & McCaffrey, R. J. (2005). Is poor performance on recognition memory effort measures indicative of generalized poor performance on neuropsychological tests? Archives of Clinical Neuropsychology, 20, 191 198. Etherton, J. L., Bianchini, K. J., Greve, K. W., & Ciota, M. A. (2005). Test of Memory Malingering: Performance is unaffected by laboratory-induced pain: Implications for clinical use. Archives of Clinical Neuropsychology, 20, 375 384. Gervais, R. O., Rohling, M. L., Green, O., & Ford, W. (2004). A comparison of WMT, CARB, and TOMM failure rates in non-head injury disability claimants. Archives of Clinical Neuropsychology, 19, 475 487. Green, P. (2003). Green s Word Memory Test. Edmonton, CA: Green s Publishing Inc. Green, P., Rohling, M. L., Lees-Haley, O. R., & Allen, L. M. (2001). Effort has a greater effect on test scores than sever brain injury in compensation claimants. Brain Injury, 15, 1045 1060. Greve, K. W., Binder, L. M., & Bianchini, K. J. (2009). Rates of below-chance performance in forced-choice symptom validity tests. The Clinical Neuropsychologist, 23, 534 544. Greve, K. W., Ord, J., Curtis, K. L., Bianchini, K. J., & Brennan, A. (2008). Detecting malingering in traumatic brain injury and chronic pain: A comparison of three forced-choice symptom validity tests. The Clinical Neuropsychologist, 22, 896 918. Heilbronner, R. L., Sweet, J. J., Morgan, J. E., Larrabee, G. J., & Millis, S. R., & Conference Participants. (2009). American academy of clinical neuropsychology consensus conference statement on the neuropsychological assessment of effort, response bias, and malingering. The Clinical Neuropsychologist, 23, 1093 1129. J. Hom, & R. Denny (Eds.). (2002). Detection of response bias in forensic neuropsychology. Binghamptom, NY: The Haworth Medical Press. Iverson, G. L., & Binder, L. M. (2000). Detecting exaggeration and malingering in neuropsychological assessment. Journal of Head Trauma and Rehabilitation, 15, 829 858. Larrabee, G. J. (2003). Detection of malingering using atypical performance patterns on standard neuropsychological tests. The Clinical Neuropsychologist, 17, 410 425. Mittenberg, W., Patton, C., Canyock, E. M., & Condit, D. C. (2002). Base rates of malingering and symptom exaggeration. Journal of Clinical and Experimental Neuropsychology, 24, 1094 1102. Mittenberg, W., Rotholc, A., Russell, E., & Heilbronner, R. (1996). Identification of malingered head injury on the Halstead-Reitan Battery. Archives of Clinical Neuropsychology, 11, 271 281. O Bryant, S. E., Finlay, C. G., & O Jile, J. R. (2007). TOMM performances and self-reported symptoms of depression and anxiety. Journal of Psychopathology and Behavior Assessment, 29, 111 114. O Bryant, S. E., & Lucas, J. A. (2006). Estimating the predictive value of the test of memory malingering: An illustrative example for clinicians. The Clinical Neuropsychologist, 20, 533 540. Rees, L. M., Tombaugh, T. N., & Boulay, L. (2001) Depression and the Test of Memory Malingering. Archives of Clinical Neuropsychology, 16, 501 506. Rees, L. M., Tombaugh, T. N., Gansler, D. A., & Moczynski, N. P. (1998). Five validation experiments of the Test of Memory Malingering (TOMM). Psychological Assessment, 10, 10 20. Reitan, R. M., & Wolfson, D. (1993). The Halstead-Reitan neuropsychological test battery. Theory and clinical interpretation (2nd ed.). Tucson, AZ: Neuropsychology Press. R. Rogers (Eds.). (1997). Researching dissimulation. New York: Guilford Press. Slick, D. J., Tan, J.E., Strauss, E. H., & Hultsch, D. F. (2004). Detecting malingering: A survey of experts practices. Archives of Clinical Neuropsychology, 19, 465 473. Tombaugh, T. N. (1996). Test of Memory Malingering (TOMM). New York: Multi Health Systems. Tombaugh, T. N. (1997). The Test of Memory Malingering (TOMM): Normative data from cognitively intact and cognitively impaired individuals. Psychological Assessment, 9, 260 268. van Gorp, W. G., Humphrey, L. A., Kalechstein, A., Brumm, V. L., McMullen, W. J., Stoddard, M., et al. (1999). How well do standard neuropsychological tests identify malingering? A preliminary analysis. Journal of Clinical and Experimental Neuropsychology, 21, 245 250. Williams, J. M. (1991). Memory Assessment Scales. Odessa, FL: Psychological Assessment Resources.