Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.705
|View full text |Cite
|
Sign up to set email alerts
|

Benchmarking Commonsense Knowledge Base Population with an Effective Evaluation Dataset

Abstract: Reasoning over commonsense knowledge bases (CSKBs) whose elements are in the form of free-text is an important yet hard task in NLP. While CSKB completion only fills the missing links within the domain of the CSKB, CSKB population is alternatively proposed with the goal of reasoning unseen assertions from external resources. In this task, CSKBs are grounded to a large-scale eventuality (activity, state, and event) graph to discriminate whether novel triples from the eventuality graph are plausible or not. Howe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 46 publications
0
2
0
Order By: Relevance
“…The overall inter-annotator agreement (IAA) stands at 81% in terms of pairwise agreement, and the Fleiss kappa [20] is 0.56. These statistics are generally comparable to or slightly higher than those of other high-quality dataset construction works [54,18,19,29], which indicates that the annotators are close to achieving a strong internal agreement.…”
Section: Inference 3: ${Event3_id}mentioning
confidence: 54%
“…The overall inter-annotator agreement (IAA) stands at 81% in terms of pairwise agreement, and the Fleiss kappa [20] is 0.56. These statistics are generally comparable to or slightly higher than those of other high-quality dataset construction works [54,18,19,29], which indicates that the annotators are close to achieving a strong internal agreement.…”
Section: Inference 3: ${Event3_id}mentioning
confidence: 54%
“…While existing methods bank on expensive and timeconsuming crowdsourcing to collect commonsense knowledge (Sap et al, 2019a;Mostafazadeh et al, 2020), it remains infeasible to obtain CSKGs that are large enough to cover numerous entities and situations in the world (He et al, 2022;Tandon et al, 2014). To overcome this limitation, various automatic CSKG construction methods have been proposed to acquire commonsense knowledge at scale (Bosselut et al, 2019), including prompting Large Language Model (LLM) (West et al, 2022;Yu et al, 2022), rule mining from massive corpora (Tandon et al, 2017;Zhang et al, 2022a), and knowledge graph population (Fang et al, 2021a(Fang et al, ,b, 2023. Although those methods are effective, they still suffer from noises introduced by construction bias and the lack of human supervision.…”
Section: Introductionmentioning
confidence: 99%