1986
DOI: 10.1111/j.1745-3984.1986.tb00250.x
|View full text |Cite
|
Sign up to set email alerts
|

The Impact of Item Deletion on Equating Conversions and Reported Score Distributions

Abstract: A formal analysis of the effects of item deletion on equating/scaling functions and reported score distributions is presented. There are two components of the present analysis: analytical and empirical. The analytical decomposition demonstrates how the effects of item characteristics, test properties, individual examinee responses, and rounding rules combine to produce the item deletion effect on the equating/scaling function and candidate scores, In addition to demonstrating how the deleted item's psychometri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2005
2005
2019
2019

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 3 publications
0
6
0
Order By: Relevance
“…Lord and Wild (1985) compared the contribution of the four verbal item types to measurement accuracy of the GRE General Test, finding that the reading comprehension item type measures something slightly different from what is measured by sentence completion, analogy, or antonym item types. Dorans (1986) used IRT to study the effects of item deletion on equating functions and the score distribution on the SAT, concluding that reequating should be done when an item is dropped. Kingston and Holland (1986) compared equating errors using IRT and several other equating methods, and several equating designs, for equating the GRE General Test, with varying results depending on the specific design and method.…”
Section: Lord's Book Applications Of Item Response Theory To Practicmentioning
confidence: 99%
“…Lord and Wild (1985) compared the contribution of the four verbal item types to measurement accuracy of the GRE General Test, finding that the reading comprehension item type measures something slightly different from what is measured by sentence completion, analogy, or antonym item types. Dorans (1986) used IRT to study the effects of item deletion on equating functions and the score distribution on the SAT, concluding that reequating should be done when an item is dropped. Kingston and Holland (1986) compared equating errors using IRT and several other equating methods, and several equating designs, for equating the GRE General Test, with varying results depending on the specific design and method.…”
Section: Lord's Book Applications Of Item Response Theory To Practicmentioning
confidence: 99%
“…Lord and Cheryl Wild (1985) compared the contribution of the four verbal item types to measurement accuracy of the GRE General Test, finding that the reading comprehension item type measures something slightly different from what is measured by sentence completion, analogy, or antonym item types. Dorans (1986) used IRT to study the effects of item deletion on equating functions and the score distribution on the SAT, concluding that reequating should be done when an item is dropped. Kingston and Holland (1986) compared equating errors using IRT and several other equating methods, and several equating designs, for equating the GRE General Test, with varying results depending on the specific design and method.…”
Section: Lord's Book Applications Of Item Response Theory To Practicmentioning
confidence: 99%
“…The comparison tells us what the relationship is between the score scale of the new test and the score scale used previously. We use this information to identify the corresponding scores on the old and new tests (Dorans, 1986;Haertel, 2004 Table 2 shows that the tests were long and assessed a variety of subject areas and grades. They had been rigorously developed and subjected to extensive internal and external reviews.…”
Section: Definition Of Some Termsmentioning
confidence: 99%
“…The comparison tells us what the relationship is between the score scale of the new test and the score scale used previously. We use this information to identify the corresponding scores on the old and new tests (Dorans, 1986; Haertel, 2004). Once this is done, we can compare old and new test scores, apply previously established cut scores to the new test, and the like.…”
Section: Definition Of Some Termsmentioning
confidence: 99%