2016
DOI: 10.1016/j.asw.2015.06.002
|View full text |Cite
|
Sign up to set email alerts
|

Features of difficult-to-score essays

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
27
0
2

Year Published

2016
2016
2023
2023

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(34 citation statements)
references
References 27 publications
5
27
0
2
Order By: Relevance
“…This finding supports the findings of previous research such as those of Raczynski, Cohen, Engelhard, and Lu (2015), which demonstrated that some essays are significantly more difficult to be scored accurately than other essays for professional raters like professors and reviewers. This finding is also in line with what Wolfe, Song, and Jiao (2016) found as features of difficult-to-score essays for professional raters.…”
Section: A Comparison Of Teachers' and Students' Perceptions Of L2 Wrsupporting
confidence: 91%
“…This finding supports the findings of previous research such as those of Raczynski, Cohen, Engelhard, and Lu (2015), which demonstrated that some essays are significantly more difficult to be scored accurately than other essays for professional raters like professors and reviewers. This finding is also in line with what Wolfe, Song, and Jiao (2016) found as features of difficult-to-score essays for professional raters.…”
Section: A Comparison Of Teachers' and Students' Perceptions Of L2 Wrsupporting
confidence: 91%
“…Furthermore, results from these analyses do not provide substantive information regarding the implications of DRF, potential causes of DRF, or specific issues that individuals who train raters need to address during rater training in order to improve the fairness of an assessment. In order to understand better the substantive implications of DRF and identify potential issues to address in rater training, practitioners might consider using qualitative approaches such as cognitive interviews (Wang, Engelhard, Raczynski, Song, & Wolfe, 2017;Wolfe, Kao, & Ranney, 1998) and textual analyses of student compositions (Wind, Stager, & Patil, 2017;Wolfe, Song, & Jiao, 2016) in conjunction with the quantitative results.…”
Section: Discussionmentioning
confidence: 99%
“…To model cumulative response processes, the many‐facet Rasch model (Linacre, ) is proposed by Engelhard () to examine rater accuracy. This model is also named the rater accuracy model (RAM; Wolfe, Song, & Jiao, ), because it provides rater accuracy estimates on a latent continuum. The RAM can be specified as follows: lnπij,kπij,k1=δiλjτk,where πij,k = probability of receiving an accuracy rating k on essay i for rater j ; πij,false(k1false) = probability of receiving an accuracy rating k – 1 on essay i for rater j ; δi = difficulty of essay i to be scored accurately; λj = accuracy of rater j ; τk = difficulty of reaching category k relative to category k – 1 of accuracy ratings. …”
Section: Evaluating Rater Accuracymentioning
confidence: 99%
“…To model cumulative response processes, the many-facet Rasch model (Linacre, 1989) is proposed by Engelhard (1996) to examine rater accuracy. This model is also named the rater accuracy model (RAM; Wolfe, Song, & Jiao, 2016), because it provides rater accuracy estimates on a latent continuum. The RAM can be specified as follows:…”
Section: Evaluating Rater Accuracymentioning
confidence: 99%