2011
DOI: 10.1609/aaai.v25i1.7847
|View full text |Cite
|
Sign up to set email alerts
|

How to Calibrate the Scores of Biased Reviewers by Quadratic Programming

Abstract: Peer reviewing is the key ingredient of evaluating the quality of scientific work. Based on the review scores assigned by the individual reviewers to the submissions, program committees of conferences and journal editors decide which papers to accept for publication and which to reject. However, some reviewers may be more rigorous than others, they may be biased one way or the other, and they often have highly subjective preferences over the papers they review. Moreover, each reviewer usually has only a very l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…Finally, there are various other types of strategic or dishonest behavior in peer assessment Hvistendahl 2013;Ferguson, Marcus, and Oransky 2014;Fanelli 2009;Resnik, Gutierrez-Ford, and Peddada 2008;Vijaykumar 2020;Littman 2021;Jecmen et al 2020;Wu et al 2021) and the design of computational methods to mitigate such behavior is vital. More generally, peer assessment is an important application with a broad set of challenges including subjectivity (Lee 2015;Noothigattu, Shah, and Procaccia 2021), miscalibration (Roos, Rothe, and Scheuermann 2011;Wang and Shah 2019), biases (Tomkins, Zhang, and Heavlin 2017;Stelmakh, Shah, and Singh 2019a;Manzoor and Shah 2021), and others (Meir et al 2020;Fiez, Shah, and Ratliff 2020;Wang et al 2021;Shah 2021).…”
Section: Discussionmentioning
confidence: 99%
“…Finally, there are various other types of strategic or dishonest behavior in peer assessment Hvistendahl 2013;Ferguson, Marcus, and Oransky 2014;Fanelli 2009;Resnik, Gutierrez-Ford, and Peddada 2008;Vijaykumar 2020;Littman 2021;Jecmen et al 2020;Wu et al 2021) and the design of computational methods to mitigate such behavior is vital. More generally, peer assessment is an important application with a broad set of challenges including subjectivity (Lee 2015;Noothigattu, Shah, and Procaccia 2021), miscalibration (Roos, Rothe, and Scheuermann 2011;Wang and Shah 2019), biases (Tomkins, Zhang, and Heavlin 2017;Stelmakh, Shah, and Singh 2019a;Manzoor and Shah 2021), and others (Meir et al 2020;Fiez, Shah, and Ratliff 2020;Wang et al 2021;Shah 2021).…”
Section: Discussionmentioning
confidence: 99%
“…We show the results in Fig 5 . Observe that interestingly, among the pairs where both authors gave a strict ranking, they disagreed 32% This high amount of disagreement between co-authors about the scientific contribution of their jointly authored papers has some implications for research on peer review. Many models of peer review [21,38,[46][47][48][49][50] assume existence of some "true quality" of each paper. This result raises questions about such an assumption-if there were such a true quality, then it is perhaps the authors who would know them well at least in a relative sense, but as we saw above, authors do not seem to agree.…”
Section: Agreements Between Co-authors and Between Authors And Peer-r...mentioning
confidence: 99%
“…Mrowinski et al [17] exploit evolutionary computation to improve editorial strategies in peer review. Roos et al [18] propose a method for calibrating the ratings of potentially biased reviewers via a maximum likelihood estimation (MLE) approach. Stelmakh et al [19] discuss biases due to demographics in single-blind peer review and study associated hypothesis testing problems.…”
Section: Related Workmentioning
confidence: 99%