2012
DOI: 10.1111/j.1745-3992.2011.00223.x
|View full text |Cite
|
Sign up to set email alerts
|

A Framework for Evaluation and Use of Automated Scoring

Abstract: A framework for evaluation and use of automated scoring of constructed‐response tasks is provided that entails both evaluation of automated scoring as well as guidelines for implementation and maintenance in the context of constantly evolving technologies. Consideration of validity issues and challenges associated with automated scoring are discussed within the framework. The fit between the scoring capability and the assessment purpose, the agreement between human and automated scores, the consideration of as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
259
0
18

Year Published

2014
2014
2018
2018

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 261 publications
(279 citation statements)
references
References 31 publications
2
259
0
18
Order By: Relevance
“…Assessment experts characterize this as a "low-stakes" application of AWE (Chapelle & Chung, 2010;Williamson, Xi, & Breyer, 2012;Weigle, 2013a), but for students and instructors, time, effort, and funding are limited resources, and decisions about where to invest them are not inconsequential. More work is needed not only to address concerns about the pedagogical value of AWE feedback but to ensure quality experiences for end users, particularly those working in a second or foreign language.…”
Section: Resultsmentioning
confidence: 99%
“…Assessment experts characterize this as a "low-stakes" application of AWE (Chapelle & Chung, 2010;Williamson, Xi, & Breyer, 2012;Weigle, 2013a), but for students and instructors, time, effort, and funding are limited resources, and decisions about where to invest them are not inconsequential. More work is needed not only to address concerns about the pedagogical value of AWE feedback but to ensure quality experiences for end users, particularly those working in a second or foreign language.…”
Section: Resultsmentioning
confidence: 99%
“…Construct validity and its association with instructional activities have also fallen within the scope of AES system-centric research (Attali & Burstein, 2006;Page, Keith, & Lavoie, 1995). In the past decade, an important trend examining AES within the larger context of the argument-based approach to test validation (Kane, 1992(Kane, , 2006 has focused on applying, refining, and expanding conceptual validation frameworks to particular applications of automated scoring (e.g., Bennett & Bejar, 1998;Xi, 2008;Williamson, Xi, & Breyer, 2012).…”
Section: Aes For Assessment and Awe For Instructionmentioning
confidence: 99%
“…There is a large body of literature with regards to ATS systems of text produced by nonnative English-language learners (Page, 1968;Attali and Burstein, 2006;Rudner and Liang, 2002;Elliot, 2003;Landauer et al, 2003;Briscoe et al, 2010;Yannakoudakis et al, 2011;Sakaguchi et al, 2015, among others), overviews of which can be found in various studies (Williamson, 2009;Dikli, 2006;Shermis and Hammer, 2012). Implicitly or explicitly, previous work has primarily treated text scoring as a supervised text classification task, and has utilized a large selection of techniques, ranging from the use of syntactic parsers, via vectorial semantics combined with dimensionality reduction, to generative and discriminative machine learning.…”
Section: Introductionmentioning
confidence: 99%