2005
DOI: 10.1002/j.2333-8504.2005.tb01982.x
|View full text |Cite
|
Sign up to set email alerts
|

An Examination of Rater Orientations and Test-Taker Performance on English-for-Academic-Purposes Speaking Tasks

Abstract: )test development efforts. As part of the foundation for the development of the next generation TOEFL test, papers and research reports were commissioned from experts within the fields of measurement, language teaching, and testing through the TOEFL 2000 project. The resulting critical reviews, expert opinions, and research results have helped to inform TOEFL program development efforts with respect to test construct, test user needs, and test delivery. Opinions expressed in these papers are those of the autho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

16
178
1
4

Year Published

2008
2008
2021
2021

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 122 publications
(199 citation statements)
references
References 43 publications
16
178
1
4
Order By: Relevance
“…Based on Brown, Iwashita, & McNamara (2005), the rubrics for the TOEFL iBT Speaking test were reflective of what teachers of English as a second language and applied linguists thought were important in evaluating candidates' speaking performance in an academic environment. However, the features used in the automated scoring model were only a subset of the criteria used by the human raters, reducing the model's power in explaining candidates' performance on realworld speaking tasks.…”
Section: Rebuttals Counterevidencementioning
confidence: 99%
“…Based on Brown, Iwashita, & McNamara (2005), the rubrics for the TOEFL iBT Speaking test were reflective of what teachers of English as a second language and applied linguists thought were important in evaluating candidates' speaking performance in an academic environment. However, the features used in the automated scoring model were only a subset of the criteria used by the human raters, reducing the model's power in explaining candidates' performance on realworld speaking tasks.…”
Section: Rebuttals Counterevidencementioning
confidence: 99%
“…In particular, descrepancy between multiple raters' judgement has always been an area of intest since 1920s. Some interesting findings relevant to vocabulary assessment in oral examinations have been identified in the past studies (Brown, Iwashita, & McNamara, 2005;Lorenzo-Dus & Meara, 2005;Malvern & Richards, 2002;Read, 2000Read, , 2005. To date, however, few studies have focused solely on rater performance in assessing vocabulary.…”
Section: Introductionmentioning
confidence: 86%
“…Some findings include: (1) Raters' judgements on vocabulary do not correlate with lexical diversity (D) in oral proficiency interviews (Lorenzo-Dus & Meara, 2005;Malvern & Richards, 2002); (2) Raters exhibit idiosyncratic approaches regarding saliency of lexical features in assessing vocabulary in oral interviews. They typically make more negative than positive comments on vocabulary (Brown, 2006); (3) Raters' judgements are sensitive to word types, tokens, and difficult words in OPIs (Brown et al, 2005;Lorenzo-Dus & Meara, 2005); (4) Raters have conflicting views on assessing linguistic aspects vis-à-vis pragmatic aspects of vocabulary in oral examinations (Brown et al, 2005); (5) It is difficult for raters to assess vocabulary at adjacent IELTS band levels (Read, 2005); (6) High correlations have been found between subcategories in oral examinations, such as vocabulary, grammar, fluency, etc. (Brown and Taylor, 2006;Malvern and Richards, 2002;Taylor and Jones, 2001); and (7) Vocabulary and grammar were prone to be rated more harshly than other constructs in oral examinations (Galaczi, 2005).…”
Section: Literature Reviewmentioning
confidence: 99%
“…has assumed that different ratings can be controlled by rater training with explicit assessment criteria and samples of performance at different levels [8].…”
Section: Brownmentioning
confidence: 99%
“…English language oral proficiency is usually evaluated by human raters, mostly native speakers [8]. Raters play a major role in the assessment process and influence the quality and meaning of scores obtained.…”
mentioning
confidence: 99%