2013
DOI: 10.1002/j.2333-8504.2013.tb02330.x
|View full text |Cite
|
Sign up to set email alerts
|

The Effects of Rater Severity and Rater Distribution on Examinees' Ability Estimation for Constructed‐response Items

Abstract: Since its 1947 founding, ETS has conducted and disseminated scientific research to support its products and services, and to advance the measurement and education fields. In keeping with these goals, ETS is committed to making its research freely available to the professional community and to the general public. Published accounts of ETS research, including papers in the ETS Research Report series, undergo a formal peer-review process by ETS staff to ensure that they meet established scientific and professiona… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
10
1
2

Year Published

2014
2014
2020
2020

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 10 publications
(13 citation statements)
references
References 11 publications
0
10
1
2
Order By: Relevance
“…Unfortunately, the literature on effective quality control procedures using quality control tools on automated scores or long‐term monitoring on both human and automated scores is sparse. However, many studies have been conducted on human scoring and rater effects (DeCarlo, ; Donoghue, McClellan, & Gladkova, ; Engelhard, , ; Longford, ; Myford & Wolfe, ; Patz, Junker, Johnson, & Mariano, ; Wang & Yao, ; Wilson & Hoskens, ; Wolfe & Myford, ). The results from these studies indicate that biases of examinee ability estimates or systematic error may be caused by varying degrees of rater leniency or central tendency.…”
mentioning
confidence: 99%
See 3 more Smart Citations
“…Unfortunately, the literature on effective quality control procedures using quality control tools on automated scores or long‐term monitoring on both human and automated scores is sparse. However, many studies have been conducted on human scoring and rater effects (DeCarlo, ; Donoghue, McClellan, & Gladkova, ; Engelhard, , ; Longford, ; Myford & Wolfe, ; Patz, Junker, Johnson, & Mariano, ; Wang & Yao, ; Wilson & Hoskens, ; Wolfe & Myford, ). The results from these studies indicate that biases of examinee ability estimates or systematic error may be caused by varying degrees of rater leniency or central tendency.…”
mentioning
confidence: 99%
“…The results from these studies indicate that biases of examinee ability estimates or systematic error may be caused by varying degrees of rater leniency or central tendency. Additionally, rater effects can increase these bias estimates and lower test reliability (Donoghue et al, ; Wang & Yao, ).…”
mentioning
confidence: 99%
See 2 more Smart Citations
“…A number of models exist that take into account rater effects. Examples include the many-faceted Rasch model (Linacre, 1989); FACETS model (Lunz, Wright , & Linacre, 1990); an IRT model for multiple raters (Verhelst & Verstralen, 2001); the rater bundle model ; the hierarchical rater model (Patz, Junker, Johnson, & Mariano, 2002) and its signal detection theory version (DeCarlo, 2010;DeCarlo, Kim, and Johnson, 2011); and Yao's rater model (Wang & Yao, 2013). These models are most useful when all the CR items of an assessment have been scored and merged with the multiple choice items (Sgammato & Donoghue, 2018).…”
Section: Subgroup/feature Biasesmentioning
confidence: 99%