2018
DOI: 10.1177/0731948718803296
|View full text |Cite
|
Sign up to set email alerts
|

The Potential for Automated Text Evaluation to Improve the Technical Adequacy of Written Expression Curriculum-Based Measurement

Abstract: Written-expression curriculum-based measurement (WE-CBM) is used for screening and progress monitoring students with or at risk of learning disabilities (LD) for academic supports; however, WE-CBM has limitations in technical adequacy, construct representation, and scoring feasibility as grade-level increases. The purpose of this study was to examine the structural and external validity of automated text evaluation with Coh-Metrix versus traditional WE-CBM scoring for narrative writing samples (7-min duration)… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(16 citation statements)
references
References 45 publications
0
16
0
Order By: Relevance
“…First, the findings of greater validity and diagnostic accuracy for all scoring approaches when based on three versus one screening sample per student build on prior generalizability theory studies demonstrating that multiple writing samples are needed for adequate reliability (Keller-Margulis et al, 2016;Kim et al, 2017). Second, the findings of comparable validity and diagnostic accuracy across complex (Wilson et al, 2016) and (b) comparable relations to holistic quality for 7 min samples scored with free automated text evaluation and complex WE-CBM metrics (Mercer et al, 2019). These findings illustrate that it may be difficult to improve validity of automated text evaluation beyond complex WE-CBM scoring; instead, investigating the optimal number of writing samples and writing duration may yield more benefits.…”
Section: Discussionmentioning
confidence: 54%
See 3 more Smart Citations
“…First, the findings of greater validity and diagnostic accuracy for all scoring approaches when based on three versus one screening sample per student build on prior generalizability theory studies demonstrating that multiple writing samples are needed for adequate reliability (Keller-Margulis et al, 2016;Kim et al, 2017). Second, the findings of comparable validity and diagnostic accuracy across complex (Wilson et al, 2016) and (b) comparable relations to holistic quality for 7 min samples scored with free automated text evaluation and complex WE-CBM metrics (Mercer et al, 2019). These findings illustrate that it may be difficult to improve validity of automated text evaluation beyond complex WE-CBM scoring; instead, investigating the optimal number of writing samples and writing duration may yield more benefits.…”
Section: Discussionmentioning
confidence: 54%
“…Consistent with recent calls for adopting open science practices in special education (Cook et al, 2018), developing and evaluating open-source tools can facilitate critical review and refinement of scoring models, replication efforts by other research teams, and greater usage of tools by removing cost barriers. For example, Mercer et al (2019) examined the validity of scoring models based on Coh-Metrix (Graesser et al, 2014), a free, but proprietary, tool for originally designed to predict reading comprehension difficulty of texts, relative to WE-CBM scores for 7-minute narrative writing samples from students in second through fifth grade. Results indicated that composite scores based on Coh-Metrix could predict raters' holistic quality ratings on the screening samples, both for the samples used to generate the Coh-Metrix scores and on similar samples collected 3 months apart, and that correlations with holistic quality were similar for composites based on Coh-Metrix (r = .73 -.81) and WE-CBM scores (r = .74 -.77).…”
Section: Applying Automated Text Evaluation To We-cbmmentioning
confidence: 99%
See 2 more Smart Citations
“…As the theory developed further, computational tools such as Coh-Metrix were presented to produces indices of the linguistic and discourse representations of pieces of text. Mercer et al [5] studied the validity of automated text evaluation with this tool -using a number of algorithms to train different models-and found it generally valid for written-expression curriculum-based measurement.…”
Section: Related Workmentioning
confidence: 99%