Proceedings of the Workshop on Generalization in the Age of Deep Learning 2018
DOI: 10.18653/v1/w18-1002
|View full text |Cite
|
Sign up to set email alerts
|

Commonsense mining as knowledge base completion? A study on the impact of novelty

Abstract: Commonsense knowledge bases such as Con-ceptNet represent knowledge in the form of relational triples. Inspired by the recent work by Li et al. (2016), we analyse if knowledge base completion models can be used to mine commonsense knowledge from raw text. We propose novelty of predicted triples with respect to the training set as an important factor in interpreting results. We critically analyse the difficulty of mining novel commonsense knowledge, and show that a simple baseline method outperforms the previou… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
22
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 19 publications
(24 citation statements)
references
References 8 publications
2
22
0
Order By: Relevance
“…While still in use today (Davis, 2017;Gordon and Hobbs, 2017), computational advances have allowed for more data-driven knowledge collection and representation (e.g., automatic extraction; Etzioni et al, 2008;Elazar et al, 2019). We will cover recent approaches that use natural language to represent commonsense (Speer et al, 2017;Sap et al, 2019a), and while noting the challenges that come with using datadriven methods (Gordon and Van Durme, 2013;Jastrzebski et al, 2018).…”
Section: Descriptionmentioning
confidence: 99%
“…While still in use today (Davis, 2017;Gordon and Hobbs, 2017), computational advances have allowed for more data-driven knowledge collection and representation (e.g., automatic extraction; Etzioni et al, 2008;Elazar et al, 2019). We will cover recent approaches that use natural language to represent commonsense (Speer et al, 2017;Sap et al, 2019a), and while noting the challenges that come with using datadriven methods (Gordon and Van Durme, 2013;Jastrzebski et al, 2018).…”
Section: Descriptionmentioning
confidence: 99%
“…Numerous research efforts aim at extracting knowledge from text corpora but research on the exact purpose of commonsense knowledge (commonsense knowledge presents facts or individual features about the concept, such as 'Lemons are sour') which is machine-readable, is comparatively rare [33]. Automatically inferring missing facts from existing ones has thus become an increasingly important task.…”
Section: Related Workmentioning
confidence: 99%
“…Several approaches have been proposed for training models for commonsense knowledge base completion (Li et al, 2016;Jastrzebski et al, 2018). Each of these approaches uses some sort of supervised training on a particular knowledge base, evaluating the model's performance on a held-out test set from the same database.…”
Section: Introductionmentioning
confidence: 99%
“…These works use relations from ConceptNet, a crowd-sourced database of structured commonsense knowledge, to train and validate their models (Liu and Singh, 2004). However, it has been shown that these methods generalize poorly to novel data (Li et al, 2016;Jastrzebski et al, 2018). Jastrzebski et al (2018) demonstrated that much of the data in the ConceptNet test set were simply rephrased relations from the training set, and that this train-test set leakage led to artificially inflated test performance metrics.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation