2017
DOI: 10.1111/jedm.12155
|View full text |Cite
|
Sign up to set email alerts
|

How to Compare Parametric and Nonparametric Person‐Fit Statistics Using Real Data

Abstract: Person‐fit assessment (PFA) is concerned with uncovering atypical test performance as reflected in the pattern of scores on individual items on a test. Existing person‐fit statistics (PFSs) include both parametric and nonparametric statistics. Comparison of PFSs has been a popular research topic in PFA, but almost all comparisons have employed simulated data. This article suggests an approach for comparing the performance of parametric and nonparametric PFSs using real data. This article then shows that there … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 40 publications
0
9
0
Order By: Relevance
“…We finally considered two indicators that were based on the similarity of response patterns from different examinees (Maynes, 2017 ; Zopluoglu, 2017 ). The first indicator (PT) was formed on basis of the number of identical responses and the second indicator (PI) on basis of the number of identical incorrect responses; for earlier studies on the performance of some of these indicators in the field of educational testing, see Karabatsos ( 2003 ), Tendeiro and Meijer ( 2014 ), Kim et al ( 2017 ), Sinharay ( 2017 ), and Man et al ( 2018 ). In the following, we will briefly review the employed indicators.…”
Section: Indicators Of Cheatingmentioning
confidence: 99%
“…We finally considered two indicators that were based on the similarity of response patterns from different examinees (Maynes, 2017 ; Zopluoglu, 2017 ). The first indicator (PT) was formed on basis of the number of identical responses and the second indicator (PI) on basis of the number of identical incorrect responses; for earlier studies on the performance of some of these indicators in the field of educational testing, see Karabatsos ( 2003 ), Tendeiro and Meijer ( 2014 ), Kim et al ( 2017 ), Sinharay ( 2017 ), and Man et al ( 2018 ). In the following, we will briefly review the employed indicators.…”
Section: Indicators Of Cheatingmentioning
confidence: 99%
“…In the context of person‐fit analysis using item scores, researchers such as Meijer and Tendeiro () and Sinharay () emphasized the importance of first assessing the overall fit of the IRT model before performing any person‐fit analysis. A similar strategy should be used for person‐fit analysis using response times.…”
Section: Conclusion and Recommendationsmentioning
confidence: 99%
“…Based on G p and G n p , response patterns can be classified as severely inconsistent, suggesting that the resulting total score is probably invalid or “‘suspect.” We computed cutoff values using a bootstrap procedure based on the graded response model (i.e., an item response theory for polytomous data) with 500 replications and a one-sided α-level of .05 (e.g., Seo & Weiss, 2013; Sinharay, 2017). The latent trait values and item parameters for generating the bootstrap replications were estimated using the “mirt” R package (Chalmers, 2012), using semiparametric Davidian curves to account for a skewed distribution of the latent trait values (Woods & Lin, 2009).…”
Section: Methodsmentioning
confidence: 99%