2001
DOI: 10.1037/0021-9010.86.2.215
|View full text |Cite
|
Sign up to set email alerts
|

Are performance appraisal ratings from different rating sources comparable?

Abstract: The purpose of this study was to test whether a multisource performance appraisal instrument exhibited measurement invariance across different groups of raters. Multiple-groups confirmatory factor analysis as well as item response theory (IRT) techniques were used to test for invariance of the rating instrument across self, peer, supervisor, and subordinate raters. The results of the confirmatory factor analysis indicated that the rating instrument was invariant across these rater groups. The IRT analysis yiel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

17
139
1
1

Year Published

2002
2002
2017
2017

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 138 publications
(158 citation statements)
references
References 44 publications
17
139
1
1
Order By: Relevance
“…Unfortunately, there is consistent evidence that raters do not agree in their evaluations of ratees. Disagreement is probably more substantial when raters differ in terms of their relationship with the persons to be rated (e.g., supervisors vs. peers), but even when raters are at the same level in an organization, they often do not agree in their evaluations (Facteau & Craig, 2001;Harris & Schaubroeck, 1988;Heneman, 1974;Murphy, Cleveland, & Mohler, 2001;Viswesvaran, Ones, & Schmidt, 1996). There have been intense disagreements about precisely how disagreements among raters should be interpreted and about the more general implications of these disagreements for the psychometric quality of ratings (e.g., Murphy & DeShon, 2000;Ones, Viswesvaran, & Schmidt, 2008), but there is little controversy about the fact that raters do not show the level of agreement one might expect from, for example, two different forms of the same paper-and-pencil test.…”
Section: Disagreement Among Ratersmentioning
confidence: 99%
“…Unfortunately, there is consistent evidence that raters do not agree in their evaluations of ratees. Disagreement is probably more substantial when raters differ in terms of their relationship with the persons to be rated (e.g., supervisors vs. peers), but even when raters are at the same level in an organization, they often do not agree in their evaluations (Facteau & Craig, 2001;Harris & Schaubroeck, 1988;Heneman, 1974;Murphy, Cleveland, & Mohler, 2001;Viswesvaran, Ones, & Schmidt, 1996). There have been intense disagreements about precisely how disagreements among raters should be interpreted and about the more general implications of these disagreements for the psychometric quality of ratings (e.g., Murphy & DeShon, 2000;Ones, Viswesvaran, & Schmidt, 2008), but there is little controversy about the fact that raters do not show the level of agreement one might expect from, for example, two different forms of the same paper-and-pencil test.…”
Section: Disagreement Among Ratersmentioning
confidence: 99%
“…The Black and White IRT-based IRFs are quite similar in Figure 9, whereas the CFAbased IRFs in Figure 4 are substantially different for the Black and White groups. Facteau and Craig (2001), Maurer, Raju, andCollins (1998), andLaffitte, Raju, Scott, andFasolo (1998) also reported some measurement nonequivalence results that were not consistent across the two perspectives.…”
Section: Irt Perspectivementioning
confidence: 99%
“…Facteau and Craig (2001), using confirmatory factor analysis (CFA) and item response theory (IRT) with a large dataset of 360°feedback ratings, found construct equivalence between peers, supervisors, and subordinates. Maurer, Raju, and Collins (1998), employing both CFA and IRT, found evidence of construct-level agreements between peer and subordinate ratings of performance.…”
Section: Construct-level Disagreements and Rater Reliability In Supermentioning
confidence: 99%