2014
DOI: 10.1007/s11336-014-9408-y
|View full text |Cite
|
Sign up to set email alerts
|

A Statistical Test for Differential Item Pair Functioning

Abstract: This paper presents an IRT-based statistical test for differential item functioning (DIF). The test is developed for items conforming to the Rasch (Probabilistic models for some intelligence and attainment tests, The Danish Institute of Educational Research, Copenhagen, 1960) model but we will outline its extension to more complex IRT models. Its difference from the existing procedures is that DIF is defined in terms of the relative difficulties of pairs of items and not in terms of the difficulties of individ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

2
52
1

Year Published

2016
2016
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 52 publications
(61 citation statements)
references
References 53 publications
2
52
1
Order By: Relevance
“…In contrast to our study, Crins et al did not find DIF for any of the items flagged for DIF in our study that were also included in the PROMIS Physical Function v1.2 item bank [49]. It has been suggested that such differences can occur because most available DIF methods can detect whether there is DIF but cannot identify the exact DIF items due to parameter identification issues [56]. Our study and the study of Crins et al, found minimal impact of language DIF on T-scores, which suggests that the original US item parameters can be used for calculating the T-scores of the DF-PROMIS-UE v2.0 bank.…”
Section: Discussioncontrasting
confidence: 99%
“…In contrast to our study, Crins et al did not find DIF for any of the items flagged for DIF in our study that were also included in the PROMIS Physical Function v1.2 item bank [49]. It has been suggested that such differences can occur because most available DIF methods can detect whether there is DIF but cannot identify the exact DIF items due to parameter identification issues [56]. Our study and the study of Crins et al, found minimal impact of language DIF on T-scores, which suggests that the original US item parameters can be used for calculating the T-scores of the DF-PROMIS-UE v2.0 bank.…”
Section: Discussioncontrasting
confidence: 99%
“…Although the concept of DIF seems straightforward, some problems have been highlighted in among others a recent study by Bechger and Maris (2014) and are mostly related to comparing parameters that are not identified from the observations. Bechger and Maris (2014) proposed using a differential item pair functioning DIF test, which focuses on comparing item pairs instead of seeing DIF as an item property.…”
Section: Differential Item Functioningmentioning
confidence: 99%
“…The procedure starts with a separate calibration of the data within each group. There exists an overall test for DIF, which under the null hypothesis that there is no DIF follows a Chi-square distribution with the number of items minus one degrees of freedom (Bechger and Maris 2014). If an item pair in the calibration of one group has a different relative difficulty when compared to the relative difficulty in the calibration of the second group, that item pair is subject to DIF.…”
Section: Differential Item Functioningmentioning
confidence: 99%
See 1 more Smart Citation
“…For practical analysis, therefore, further assumptions need to be made. One that is often used in practice is to expect the majority of items to be DIF‐free (e.g., Angoff, ; Bechger & Maris, ; Koretz & McCaffrey, ; Pohl, Stets, & Carstensen, ). This approach follows the logic that the majority of test items work as intended and only a few items show DIF.…”
mentioning
confidence: 99%