Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1149
|View full text |Cite
|
Sign up to set email alerts
|

A Dataset of Peer Reviews (PeerRead): Collection, Insights and NLP Applications

Abstract: Peer reviewing is a central component in the scientific publishing process. We present the first public dataset of scientific peer reviews available for research purposes (PeerRead v1), 1 providing an opportunity to study this important artifact. The dataset consists of 14.7K paper drafts and the corresponding accept/reject decisions in top-tier venues including ACL, NIPS and ICLR. The dataset also includes 10.7K textual peer reviews written by experts for a subset of the papers. We describe the data collectio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
164
3

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 131 publications
(168 citation statements)
references
References 13 publications
1
164
3
Order By: Relevance
“…a paper is, the less confident the reviewers will be. Note that our correlation results are different from those reported by Kang et al (2018), who report that the OVAL has low Pearson correlation with SND (0.01) and ORG (0.08). While the differences might be caused by a variation in aspect definitions, we believe that our estimate is more reliable as the dataset analyzed in Kang et al (2018) is substantially smaller than ours.…”
Section: Reviewscontrasting
confidence: 99%
“…a paper is, the less confident the reviewers will be. Note that our correlation results are different from those reported by Kang et al (2018), who report that the OVAL has low Pearson correlation with SND (0.01) and ORG (0.08). While the differences might be caused by a variation in aspect definitions, we believe that our estimate is more reliable as the dataset analyzed in Kang et al (2018) is substantially smaller than ours.…”
Section: Reviewscontrasting
confidence: 99%
“…For Task 2, we observe that the handcrafted feature-based system by Kang et al (2018) performs inferior compared to the baselines. This is because the features were very naive and did not 4 https://github.com/aritzzz/DeepSentiPeer address the complexity involved in such a task.…”
Section: Results and Analysismentioning
confidence: 97%
“…For Task 1, we can see that our review sentiment augmented approach outperforms the baselines and the comparing systems by a wide margin (∼ 29% reduction in error) on the ICLR 2017 dataset. With only using review+sentiment information, we are still able to outperform Kang et al (2018) by a margin of 11% in terms of RMSE. A further relative error reduction of 19% with the addition of paper features strongly suggests that only review is not sufficient for the final recommendation.…”
Section: Methodsmentioning
confidence: 94%
See 1 more Smart Citation
“…We collect review data from three sources: (1) openreview.net-an online peer reviewing platform for ICLR 2017, ICLR 2018, and UAI 2018 2 ; (2) reviews released for accepted papers at NeurIPS from 2013 to 2017; and (3) opted-in reviews for ACL 2017 from Kang et al (2018).…”
Section: Ampere Datasetmentioning
confidence: 99%