“…Over the last two decades, automated scoring has been widely developed and used in a variety of content domains, such as mathematics (Bennett & Sebrechts, ; Sandene, Horkay, Bennett, Braswell, & Oranje, ), science (Linn et al, ; Liu et al, ; Nehm, Ha, & Mayfield, ), and language testing (Bernstein, Van Moere, & Cheng, ; Higgins, Zechner, Xi, & Williamson, ), to name a few. Furthermore, in assessing written responses across content domains, automated scoring has been used to evaluate rubric dimensions, such as content (Attali & Powers, ; Dzikovska et al, ; Leacock & Chodorow, ; Mitchell, Russell, Broomhead, & Aldridge, ; Nielsen, Ward, & Martin, ; Sukkarieh & Bolge, ) and quality (Burstein & Marcu, ; Foltz, Laham, & Landauer, ). The accuracy of automated scores depends on a number of factors, including the content domain, the complexity of the tasks, the levels of the scoring rubrics, and the number of responses available to build the automated scoring models.…”