2001
DOI: 10.3102/10769986026003283
|View full text |Cite
|
Sign up to set email alerts
|

The Rater Bundle Model

Abstract: In this article an item response model is introduced for repeated ratings of student work, which we have called the Rater Bundle Model (RBM). Development of this model was motivated by the observation that when repeated ratings occur, the assumption of conditional independence is violated, and hence current state-ofthe-art item response models, such as the rater facets model, that ignore this violation, underestimate measurement error, and overestimate reliability. In the rater bundle model these dependencies … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

2
57
0

Year Published

2008
2008
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 51 publications
(59 citation statements)
references
References 20 publications
2
57
0
Order By: Relevance
“…Rater severity effects have been included in some measurement models and studied by DeCarlo, Kim, and Johnson (2011) ;Donoghue, McClellan, and Gladkova (2006); Engelhard (2002); Longford (1995) ;Patz, Junker, Johnson, and Mariano (2002); Wilson and Hoskens (2001); and Wolfe and Myford (1997). The results from their studies indicate that the bias or systematic error may be caused by varying degrees of rater leniency or strictness.…”
mentioning
confidence: 99%
See 2 more Smart Citations
“…Rater severity effects have been included in some measurement models and studied by DeCarlo, Kim, and Johnson (2011) ;Donoghue, McClellan, and Gladkova (2006); Engelhard (2002); Longford (1995) ;Patz, Junker, Johnson, and Mariano (2002); Wilson and Hoskens (2001); and Wolfe and Myford (1997). The results from their studies indicate that the bias or systematic error may be caused by varying degrees of rater leniency or strictness.…”
mentioning
confidence: 99%
“…There are a few research studies incorporating rater severity effects into the item response theory (IRT) models (Donoghue et al, 2006;Engelhard, 1996;Patz, 1997;Patz & Junker, 1999;Wilson & Hoskens, 2001). The FACETS model (Linacre, 1991) is the IRT model that allows for the estimation of differences in severity between raters, and thus eliminates rater bias from the estimates of the items and examinees' ability.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…Viewed more generally, methods developed in this article extend existing statistical methodology for the analysis of multirater ordinal data (4-7) and item response data (8)(9)(10)(11)(12) to provide a framework for the analysis of panel rating data collected by using the Delphi method and related interactive rating schemes (13).…”
mentioning
confidence: 99%
“…For example, the test specifications might call for a task of moderate difficulty at site S, with the constraint that the task was not previously exposure at site S. The parameters from the across-site, task-only model might be interpreted as preliminary estimates, to be used for initial task selection. These initial parameter estimates would be then updated using the within-site, taskrater model after the task was used for a period of time at site S. Parameter estimates for task-raters can be periodically evaluated to examine rater leniency and discrimination parameters over time, for example, to delineate rater drift (Wilson & Hoskens, 2001;Harik et al, 2009), to evaluate the accuracy of equating processes and to identify gaps in the task pool. Availability of rater generic and rater specific task and test statistics would help monitor rater performances and spot problematic raters.…”
Section: Discussionmentioning
confidence: 99%