Introduction

The field of psychological injury and law lacks a malingering detection system dedicated to detecting malingered posttraumatic stress disorder (PTSD) in forensic disability and related evaluations. In this regard, I constructed a new system that builds on prior models, but ones on other psychological injuries and not on PTSD. That is, I examined the malingering detection systems for neurocognitive and pain-related complaints that had been developed by Slick et al. (1999) for Malingered Neurocognitive Dysfunction (MND) and by Bianchini et al. (2005) for Malingered Pain-Related Disability (MPRD) (see the Appendix). As well, I considered the work of Rogers, Bender, and Johnson (2011a, b) and Boone (2011) on their recommendations on malingering detection, and the test protocol developed by Rubenzer (2009) to detect malingered PTSD. The model that I developed for PTSD was constructed so that it could be general and easily transformed by using slight alterations (mostly just by using some differing examples), and, thus, create new systems for detecting malingered pain and neurocognitive function. That is, in constructing the system for PTSD to begin with, I was quite aware that I wanted later to generalize it to ones for pain and neurocognition. These additional malingering detection systems are presented in Young (2014a).

It is beyond the scope of the present article to explain the system in depth, and aside from the brief text that follows in the article, the appendix that presents it includes an introductory explanation in this regard. The reader can consult Chapter 5 in Young (2014a) for a detailed description of the system. Also, it is beyond the scope of this article to summarize the nature of PTSD, how to assess and diagnose it, and the controversies related to it. The reader can consult Young (2014a) for further information, as well as Young (2014b) and Young, Lareau, and Pierre (2014). For further description of malingering, as well as the field of psychological injury and law, in general, Young and Drogin (2014) and Young (2014c) offer pertinent background. Young (2015) provides a review of the recent literature on malingering since the Young (2014a) book. There are many other books related to malingering (e.g., Carone and Bush 2013; Larrabee, 2012a in Larrabee, 2012b; Rogers, 2008), psychological injury law, and assessment (e.g., Kane & Dvoskin, 2011), and PTSD (e.g., Friedman, Keane, & Resick, 2014; Vasterling, Bryant, & Keane, 2012) that offer valuable information (and that are integrated into Young, 2014a). As for critical articles by others on the topic, among many others, informative articles have been written by Demakis and Elhai (2011) on PTSD and malingering, Sleep, Petty, and Wygant (2015) and Odland, Lammy, Martin, Grote, and Mittenberg (2015) on testing in forensic disability and related cases, and Zoellner, Bedard-Gilligan, Jun, Marks, and Garcia (2013) and Biehn, Elhai, Seligman, Tamburrino, and Forbes (2013) on PTSD in the DSM-5.

The MND and MPRD systems include different degrees of possible malingering and related qualifications, such as definite and probable. Similarly, the present system is a dimensional one that organizes categories on a continuum from significant to minimal likelihood of malingering. Specifically, I propose a model with a range of seven different response biases, including ones ranging from, at one end, frank malingering or related attributions to, at the other end, an absence of malingering, with an indeterminate “gray zone” category, too.

Note that the attribution of malingering can only take place in the context of a full-scale assessment that includes interviews, collateral information, and so on, and not just testing. Malingering might be apparent in cases having incontrovertible evidence, such as videographic evidence, but usually this type of information is not available. Therefore, generally in cases in which the appropriate conclusion is that malingering is present, the decision is based on test results and not on other evidence of this latter nature. In this regard, when the evaluator refers to definite malingering for a case, generally, it is because there are test results that indicate definite negative response bias; for example, the case includes statistically significant below-chance failure on two forced-choice tests. In such circumstances, the attribution of malingering is given when the file as a whole dictates that this conclusion is appropriate.

In the following, first, I review the prior malingering detection systems and procedures that informed the development of my own system. Then, I describe the approach that I took in creating my system.

Prior Models

Slick et al. (1999) developed a system of “specific, clearly articulated” criteria for rating different degrees of malingering. In particular, they focused on definite, probable, and possible malingering. Aside from this critical aspect to their system, it is notable for articulating many types of possible inconsistencies in data gathered in an assessment, e.g., between test data and observed behavior. There are four categories of criteria in their system, numbered in capitalized letters: (A) presence of substantial external incentive; (B) evidence from neuropsychological testing; (C) evidence from self-report; and (D) exclusion criteria even if the evidence for (B) and (C) are present.

For their model of MPRD, Bianchini et al. (2005) maintained the Slick et al. (1999) distinction of definite, probable, and possible malingering-related disability. As for the criteria, they are grouped into the same four categories in Slick et al. (1999), and another one was added on evidence from physical evaluation, which consists of four criteria. The criteria in the critical categories were altered relative to the MND model. When I had to decide which model could serve me better in constructing my own, I found the language in the MPRD a good starting point. However, I consulted other material, as well.

Rubenzer (2009) did not create a diagnostic system related to malingering, but he developed a weighting scale for tests that can be used in the detection of malingering and other biases evident in PTSD assessment. For each measure, he assigned a weight of either one or two, except for the case of failure of any SVT at a level that is below-chance (statistically significant), for which he assigned the weight of five. This decision is consistent with the importance given in the literature to statistically significant below-chance level failure on forced-choice tests. Rubenzer’s system informed the present one because he included a weighting protocol, which led me to develop the 60 rules concerning tests in my system. Also, he listed multiple tests that could be useful, and I did the same in my approach.

It is informative to view that Rubenzer listed tests from all the major categories in the literature—personality tests, stand-alone tests, and cognitive tests as might be used in neuropsychological evaluation. It is also informative that he allowed several indicators/scales/measures to be taken for weighting from one instrument, such as was the case for each of the personality tests listed, as well as for the interview schedule for malingering detection. In the diagnostic system that I developed, I followed the same procedures of integrating the different types of feigning-detection instruments and allowing more than one index from critical tests, such as from the personality inventories.

In the MND and MPRD systems, inconsistencies and discrepancies are considered revealing with respect to malingering detection. In Young (2014a), I noted the following about their treatment in the two prior systems: (a) The types of inconsistencies/discrepancies in the two systems need more clarity in definition/examples. (b) The two systems do not list all types of inconsistencies/discrepancies possible. (c) They do not list all of the combinations possible of types of inconsistencies/discrepancies. (d) They do not include uniformly within- and not only across-category inconsistencies/discrepancies. (e) The prior systems include multiple types that overlap, but they should be separated according to: (i) standard test data; (ii) self-report; (iii) observations; (iv) known patterns of brain functioning; (v) known patterns of physiological functioning; (vi) collateral information; and (vii) documented information. (f) Information in these inconsistency/discrepancy categories could be about pre-event, event, or post-event factors. It might refer to either pre-event history, such as prior police or criminal record, or event/post-event symptoms, impairments, dysfunctions, and disabilities, if any. (g) The inconsistencies/discrepancies could be compelling/marked/substantial or otherwise, but, in the two systems, no clear guidelines are offered to differentiate the more severe compelling type. (h) In the two systems, test data could derive from measures of exaggeration, fabrication, and suspected malingering, such as in symptom validity tests (SVTs), but also tests from tests like the MMPIs. In Young (2014a), I noted that better ways of combining the different types of tests data in detecting malingering need to be created.

Note that Slick and Sherman (2012, 2013) modified extensively their 1999 MND model but, in Young (2014a), I showed that the revision has clear difficulties that compromise its value. Also, Greve, Curtis, and Bianchini (2013) recommended that the MPRD system could be used in PTSD malingering detection, but I had noted this suggestion is not feasible, given the multiple pain factors embedded in the system.

System for Detecting Malingered PTSD

The malingered PTSD detection system that I created built on the influences just described, and they constitute advances in the field that cannot be minimized. Nevertheless, by considering them together, as well as additional sources, such as the criticisms of the MND in Rogers et al. (2011a, b) and Boone (2011), the new system that I created not only took the best of the prior work but added extensively to it.

The major innovations that I incorporated into the system to help in detecting malingered PTSD and related negative response biases concern developing elaborate protocols involving: (a) rules for testing that integrate the different types of tests that can be used in the detecting of feigning/malingering—such as forced-choice tests, structured interview tests, embedded cognitive/neuropsychological tests/measures/scales, and validity indicators on personality tests—and (b) inconsistencies/discrepancies that might be evident in evaluations—such as compelling ones in and between testing, self-report, observations, collateral information, documents, and known brain function/psychology.

These considerations led to much complexity and length in the system that I created, but these factors make it precise, usable, and applicable across a broad range of assessments. Moreover, it can account for multiple contingencies in an assessment across its major parts—those related to testing and also from other information gathered, such as from the interview of the examinee, records, and collateral information. Therefore, given these advantages, what some might consider weaknesses (complexity, length) of the system for detection of malingered PTSD in the appended table, to the contrary, add to its potential value and validity, and justify that it took over 20 printed pages to publish in Young (2014a). The MND and MPRD systems take much less space to describe in depth compared to my own, which attests further to the comprehensive nature of the system that I developed.

That said, to repeat, my system is based on theirs, and could not have been constructed without them. In this regard, in the table in the appendix that presents my system, it is noted that the sections not in italics are taken from their prior work. This accomplishes two things. On the one hand, it acknowledges their important contributions to the field. On the other hand, it indicates that most of the system that I created greatly elaborated the prior systems.

As for the contents of my system, the crux lies in how I defined definite malingering and related negative response biases. Specifically, aside from cases with extremely compelling evidence, such as frank admission or indisputable videographic evidence, in the present detection system, definite malingering can be attributed in cases in which: (a) two or more forced-choice measures are failed at the statistically-significantly-below-chance level; or (b) there are five or more test failures on other valid psychometric measures; or (c) there are three or more compelling inconsistencies; or (d) any combinations of these types of evidence are found, or (e) other evidence replaces the weighting of some of these types of evidence, such as extreme scores related to negative response bias on valid psychometric tests or the presence of an overall judgment of the file adding weight to the determination that malingering might be taking place. Therefore, by examining other factors beyond those listed above for definite malingering, the parameters of the system allow that three test failures could be sufficient to attribute malingering, everything else being equal, which makes it comparable to other systems to a degree.

In this regard, the reader will note that Larrabee (2012b) emphasized three if not two failures on relevant tests as very strong evidence of malingering. All things considered, the present system arrives at a protocol that gives a comparable weighting to such test failures. However, in the present case, there are safeguards built into the system such that the number of tests used for it is limited, so that the risk of Type I error is minimized.

As for concluding the presence of definite response bias, the criteria above apply, except that they involve: one forced-choice test, not two; four other tests, not five or more; and two compelling inconsistencies, not three or more, with none of the extreme nature involved. In terms of probable response bias, the criteria exclude forced-choice test failure, but consider three other test failures, not four, and one compelling inconsistency, not two. About other levels in the system, for example, for the indeterminate gray zone or for an absence of any negative response bias, the reader should consult the appendix.

The 60 rules that I created on test usage in the present system constitute its major advance. These rules comprise a rigorous packet that, if followed judiciously, will facilitate reliable use of the system. Moreover, this would be true no matter what type of examinee, the referral source, and so on, permitting generalizability and assuring validity of the system. The 60 rules were constructed to apply equally to the other two systems developed (for pain and neurocognition), and so do not apply just to the one for PTSD.

The 60 rules of the present system were constructed according to 10 pertinent principles and parameters, as specified in the following. In describing them, I elaborate with material from other parts of the book (Young, 2014a) that were not included at the point at which I presented them, but the additional points are consistent with them. (a) There are two tracks in the system, Regular (for PTSD, pain) and Neuropsychological/Cognitive (Neurocognition). (b) There are multiple test types, including forced-choice and personality ones. All types can be used in the system, and the specific ones chosen should be scientifically supported for the question at hand. (c–e) Some test types are more critical than others (e.g., two option forced-choice ones). Some criteria are more critical than others, e.g., statistically significant below-chance performance. And some tests more reliable and valid than others for the purposes at hand, e.g., the MMPI-2-RF [Minnesota Multiphasic Personality Inventory, Second Edition, Restructured Form; Ben-Porath & Tellegen, 2008/2011; see Young (2014a) on the research demonstrating the value of the MMPI-2-RF in the forensic disability and related context]. (f) Any one test can provide one to several validity indicators, depending on the research findings in the area (e.g., the MMPI-2-RF has a family of F and related scales). (g) For use in the present system, the tests should include 10–15 primary measures specified before undertaking an assessment. Moreover, these 10–15 measures should give 5–8 positive findings (and, at most, 3–4 of them from any one instrument) in order to conclude that the examinee has manifested feigning or related response bias, including of malingering. (h) Tests that are correlated in the literature can be used within specified limits (i.e., moderate, at most), but the correlations need to be acknowledged. The present system does not call for use of tests or scales that are independent, unlike systems that are more statistically oriented. (i) Malingering can be concluded only when there is introconvertible evidence after examination of the full reliable data set gathered. That being said, evidence of problematic presentations and performances that do not reach the level of outright malingering can still be qualified in ways that cast sufficient doubt on the credibility of the examinee. (j) In general, test selection and score interpretations must be undertaken scientifically, impartially, and comprehensively, while considering the limits of the evaluees (cognitive deficits secondary to TBI, culture issues, etc.).

Conclusions

To conclude, the new malingering detection system that I have created for case of PTSD needs research on its reliability and validity before it can be used confidently in practice and court. However, if applied prudently, as presently constituted, it might be useful as a guide in assessment. Moreover, the present paper seeks comments, critiques, suggested changes, additions, deletions, and so on, toward improving the system. Just as I built on prior work in building the present system, so can the reader build on mine to improve it. Once responses are received, I will integrate them toward changing the present system, or defending it, or both, as required, and present. [Note that the journal’s procedure for reviewing suggested revisions to the proposed malingering detection system, as well as my response to those that are accepted for publication, will be the same as that for any similar submission to the journal, that is, by independent review.]

The present system has been modified to create two other very similar systems, related to detection of malingered neurocognition and pain. Changes applicable to the system on PTSD will necessarily impact the other two.

Finally, given the extensive portion in the present malingering detection system that concerns inconsistencies, it can be used by mental health professionals who focus on inconsistencies only because they do not have competence or formal training in administering, scoring, and interpreting psychological tests. Other professionals in the field are involved in malingering detection, and the systems created need to be general enough for their use, and not just psychologists. That being said, mental health professionals who do not use tests bear a greater burden in establishing the presence of malingering and related negative response biases when using the current system because they will only be using half of it.

Once all the systems in the Young (2014a) are considered justified one way or another (with or without changes), it could be integrated with other ones that are more statistical in nature. The protocol that I created is rationally derived, based on other rational ones. It includes algorithms that account for the number of tests, scales, and scores that can be used in it, but it does not include statistical procedures, such as Bayesian ones that might be informative, as well, toward malingering determination (e.g., Odland et al., 2015; possible variations of Larrabee, 2012b; Schutte, Millis, Axelrod, & VanDyke, 2011). However, there might be ways to use advanced statistical procedures in the present approach.

The most important future work on the present system concerns reliability and validity, though. As far as I know, the prior ones have not been studied for inter-evaluator agreement. The present approach needs to prove its mettle in this regard. Moreover, it needs to demonstrate efficacy in differentiating known malingering-related groups or malingering simulator groups from relevant control groups. This is no mean feat, because it would take more than the use of small convenience samples or university samples, given its complexity and length and that assessments for the research would rely on a lot more data than in typical study to date. The MND and MPRD systems have given the basis for creating the present one, and if agreement can be reached that a better one than the prior ones has been created (whether as presently constituted or after changes through commentaries or subsequent research) that applies not only to PTSD but, with slight modification, also to MND and MPRD cases, then the impetus for determining its reliability and validity might be set in motion.

In addition, the new malingering detection system that I have created should increase the likelihood that evaluation data analyzed with it will meet court and related requirements. Admissibility challenges in court based on Daubert (1993) and related decisions are meant to ensure that evidence proffered to court meet the criteria of good science. By using a protocol that is comprehensive, scientifically-informed, and balanced, as the present one seeks to be, the chances of meeting successfully any admissibility challenges related to malingering determinations are increased. Finally, the work described here might facilitate research on the prevalence or base rate of malingering, as well as other negative response styles in psychological injury populations. The question of base rate of malingering is critical not only for individual cases but also for statistical calculation of relevant psychometric properties of malingering-related tests, e.g., positive predictive power, sensitivity, and specificity. These are crucial in determining test and scale cut scores, for example, which can be quite contentious in the field.