Abstract:Self-reports may be affected by two primary sources of distortion, i.e., content-related (CRD) and content-unrelated (CUD) distortions. CRD and CUD, however, may co-vary, and similar detection strategies have been used to capture both. Thus, we hypothesized that a scale developed to detect random responding-arguably, one the most evident examples of CUDwould likely be sensitive to both CUD and, albeit to a lesser extent, CRD. Study 1 (N = 1,901) empirically tested this hypothesis by developing a random respond… Show more
“…A comparison with TOMM based on 100 adult volunteers showed excellent values for sensitivity and specificity, but even higher values were obtained by combining the two measures (Giromini, Barbosa, et al, 2019). The IoP-29 has proven its usefulness in separating honest from random responders (Giromini et al, 2020a(Giromini et al, , 2020bWinters et al, 2021), and honest responders from those simulating depression, PTSD, and/or schizophrenia (Giromini et al, 2020a(Giromini et al, , 2020bPignolo et al, 2021;Šömen et al, 2021). A study of 154 possibly traumatized firefighters showed that IoP-29 FDS scores based on honest responding were almost consistently below cutoff, despite high variation in symptom presentation (Carvalho et al, 2021).…”
Objective: We report on the validation of the Norwegian version of the Inventory of Problems-29 (IoP-29) using an experimental simulation setup with honest responders and responders feigning depression. Method: The sample consisted of 275 participants recruited by convenience by spreading invitational links to participate in the study on social media. They were randomly assigned to either an experimental simulation or an honest condition with the IoP-29. Participants in both conditions were asked to confirm whether they had followed the instructions or not. The honest respondents (n = 138) were asked to respond to the IoP-29 as honestly as they could, whereas the experimental simulators (n = 137) were told to simulate depression based on a case vignette and descriptions of depressive symptoms. Results: We found that participants in the simulation group scored significantly higher than the control group (d = 2.39, p < .000). We found no significant effects from age, gender, or education on the resulting scores. Area under the curve (AUC) for IoP-29 was .94 (SE = .01), meaning a good ability to separate the two conditions. The cutoff for the IoP-False Disorder Score ≥.50 yielded sensitivity = .77, specificity = .93, positive predictive power = .91, negative predictive power = .81, and an overall correct classification = .85. Conclusions: The Norwegian version of IoP-29 demonstrated good validity in discriminating between experimental simulators of depression and a nonclinical control group.
Public Significance StatementOur study shows that the Norwegian translation of Inventory of Problems-29 can effectively discriminate between honest nonclinical responders and feigners of depression. This study marks the first step of the validation of the Norwegian IoP-29 version for clinical use.
“…A comparison with TOMM based on 100 adult volunteers showed excellent values for sensitivity and specificity, but even higher values were obtained by combining the two measures (Giromini, Barbosa, et al, 2019). The IoP-29 has proven its usefulness in separating honest from random responders (Giromini et al, 2020a(Giromini et al, , 2020bWinters et al, 2021), and honest responders from those simulating depression, PTSD, and/or schizophrenia (Giromini et al, 2020a(Giromini et al, , 2020bPignolo et al, 2021;Šömen et al, 2021). A study of 154 possibly traumatized firefighters showed that IoP-29 FDS scores based on honest responding were almost consistently below cutoff, despite high variation in symptom presentation (Carvalho et al, 2021).…”
Objective: We report on the validation of the Norwegian version of the Inventory of Problems-29 (IoP-29) using an experimental simulation setup with honest responders and responders feigning depression. Method: The sample consisted of 275 participants recruited by convenience by spreading invitational links to participate in the study on social media. They were randomly assigned to either an experimental simulation or an honest condition with the IoP-29. Participants in both conditions were asked to confirm whether they had followed the instructions or not. The honest respondents (n = 138) were asked to respond to the IoP-29 as honestly as they could, whereas the experimental simulators (n = 137) were told to simulate depression based on a case vignette and descriptions of depressive symptoms. Results: We found that participants in the simulation group scored significantly higher than the control group (d = 2.39, p < .000). We found no significant effects from age, gender, or education on the resulting scores. Area under the curve (AUC) for IoP-29 was .94 (SE = .01), meaning a good ability to separate the two conditions. The cutoff for the IoP-False Disorder Score ≥.50 yielded sensitivity = .77, specificity = .93, positive predictive power = .91, negative predictive power = .81, and an overall correct classification = .85. Conclusions: The Norwegian version of IoP-29 demonstrated good validity in discriminating between experimental simulators of depression and a nonclinical control group.
Public Significance StatementOur study shows that the Norwegian translation of Inventory of Problems-29 can effectively discriminate between honest nonclinical responders and feigners of depression. This study marks the first step of the validation of the Norwegian IoP-29 version for clinical use.
“… Specifically computed to convert traditional scores into a binomial experiment; 1–3, Acquisition trials; Animals, Category fluency (Curtis et al., 2008; Hurtubise et al., 2020; Sugarman & Axelrod, 2015); BC, Below chance level (below the 95% confidence interval around the mean); BIN, Cutoff based on the binomial distribution; BNT‐15, Boston Naming Test—Short Form (Abeare et al., 2022; Deloria et al., 2021; Erdodi, Dunn, et al., 2018); C, At chance level (within the 95% confidence interval around the mean); CD WAIS‐IV , Coding (Ashendorf et al., 2017; Erdodi, Abeare, et al., 2017); CIM, Complex Ideational Material (Erdodi, 2019; Erdodi et al., 2016; Erdodi & Roth, 2017); COL, Color Naming; CPT‐3, Conners' Continuous Performance Test—Third Edition ( T = 90 is the highest score possible; Ord et al., 2020; Robinson et al., 2022); COM, Combination score (FR + true positives—false positives); DCT, Dot Counting Test (Boone et al., 2002; Hansen et al., 2022); Dem ADJ , Demographically adjusted score; DH, Dominant hand; D‐KEFS, Delis Kaplan Executive System (Cutler et al., 2022; Eglit et al., 2020; Erdodi, Sagar, et al., 2018); DR, Delayed recall; DS WAIS‐IV , Digit Span subtest of the Wechsler Adult Intelligence Scale—Fourth Edition (Shura et al., 2020; Whitney et al., 2009); EMP, Empirically derived cutoffs; EWFT, Emotion Word Fluency Test (Abeare, Hurtubise, et al., 2021); FAS, Letter fluency (Curtis et al., 2008; Deloria et al., 2021; Hurtubise et al., 2020); FCR, Forced choice recognition; FR, Free recall; FTT, Finger Tapping Test (Arnold et al., 2005; Axelrod et al., 2014; Erdodi, Taylor, et al., 2019); GPB, Grooved Pegboard Test (Erdodi, Kirsch, et al., 2018; Erdodi, Seke, et al., 2017; Link et al., 2021); HVLT‐R, Hopkins Verbal Learning Test—Revised (Cutler et al., 2021; Sawyer et al., 2017); IOP‐M, Inventory of Problems—29 memory module (Giromini et al., 2020a, 2020b; Holcomb, Pyne, et al., 2022); IR, DR & CNS, Immediate, Delayed & Consistency of Recognition trials (% correct); LM Recognition, Yes/No recognition trial of the Logical Memory subtest of the Wechsler Memory Scale—Fourth Edition (Bortnik et al., 2010; Dunn et al., 2021); LNS WAIS‐IV , Letter‐Number Sequencing (Erdodi &...…”
This study was designed to empirically evaluate the classification accuracy of various definitions of invalid performance in two forced-choice recognition performance validity tests (PVTs; FCR CVLT-II and Test of Memory Malingering [TOMM-2]). The proportion of at and below chance level responding defined by the binomial theory and making any errors was computed across two mixed clinical samples from the United States and Canada (N = 470) and two sets of criterion PVTs. There was virtually no overlap between the binomial and empirical distributions. Over 95% of patients who passed all PVTs obtained a perfect score. At chance level responding was limited to patients who failed ≥2 PVTs (91% of them failed 3 PVTs). No one scored below chance level on FCR CVLT-II or TOMM-2. All 40 patients with dementia scored above chance. Although at or below chance level performance provides very strong evidence of non-credible responding, scores above chance level have no negative predictive value. Even at chance level scores on PVTs provide compelling evidence for non-credible presentation.A single error on the FCR CVLT-II or TOMM-2 is highly specific (0.95) to psychometrically defined invalid performance.Defining non-credible responding as below chance level
“…Thus, one might question whether our findings generalize to reallife evaluation contexts which use these precautions. On the other hand, inattentive item review or inconsistent effort would likely obscure discrimination of feigners from honest responders (Giromini et al, 2020c). Indeed, we had no control over the participants, so feigners might have searched the internet for the answers or asked family members and friends how to respond to the items, etc.…”
While the psychometric equivalence of computerized versus paper-and-pencil administration formats has been documented for some tests, so far very few studies have focused on the comparability and validity of test scores obtained via in-person versus remote administrations, and none of them have researched a symptom validity test (SVT). To contribute to fill this gap in the literature, we investigated the scores of the Inventory of Problems-29 (IOP-29) generated by various administration formats. More specifically, Study 1 evaluated the equivalence of scores from nonclinical individuals administered the IOP-29 remotely (n = 146) versus in-person via computer (n = 140) versus in-person via paper-and-pencil format (n = 140). Study 2 reviewed published IOP-29 studies conducted using remote/online versus in-person, paper-and-pencil test administrations to determine if remote testing could adversely influence the validity of IOP-29 test results. Taken together, our findings suggest that the effectiveness of the IOP-29 is preserved when alternating between face-to-face and online/remote formats.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.