Proceedings of the Fifth International Conference on Knowledge Capture 2009
DOI: 10.1145/1597735.1597763
|View full text |Cite
|
Sign up to set email alerts
|

Large-scale extraction and use of knowledge from text

Abstract: A large amount of empirically derived world knowledge is essential for many languageprocessing tasks, to create expectations that can help assess plausibility and guide disambiguation. Following Schubert (2002), we present our work on creating a large database of "tuples" from the output of a parser, thus implicitly capturing simple world knowledge expectations, and then utilizing it for two tasks, namely improving parsing and improving the plausibility assessment of paraphrase rules used in textual entailment… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0

Year Published

2009
2009
2016
2016

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 32 publications
(20 citation statements)
references
References 15 publications
0
20
0
Order By: Relevance
“…Peter Clark and Phil Harrison [26] worked on knowledge extraction by making database of "tuples" and thus capturing the simple word knowledge. And then using it in improving parsing and the plausibility assessment of paraphrase rules used in textual entailment.…”
Section: A Text Miningmentioning
confidence: 99%
“…Peter Clark and Phil Harrison [26] worked on knowledge extraction by making database of "tuples" and thus capturing the simple word knowledge. And then using it in improving parsing and the plausibility assessment of paraphrase rules used in textual entailment.…”
Section: A Text Miningmentioning
confidence: 99%
“…Many other authors have pro-426 cessed text with a morphosyntactic parser and then selected one 427 or more surface patterns to extract a set of candidate semantic 428 classes, that in turn are refined [18,19]. 429 In a different direction, [20] argues that text contains general 430 knowledge in form of assertions and they may be exploited after 431 aggregating big amounts of data, like in KNEXT [20], TextRunner 432 [21] or DART [22]. 433 We use both ideas going one step beyond by combining induced 434 classes with named entity types.…”
Section: Tablementioning
confidence: 99%
“…theory NELL [3] 24 × 7 learning fixed dynamic ML techniques DART [7] world knowledge × × semi-automated RTE [2], and [13] entailment × × ATP NLU [20] commonsense rules × × semi-supervised Text2Onto [6] ontology learning √ √ semi-supervised LexO [24] complex classes √ × semi-supervised FCA [5] taxonomy √ × FCA OP [4], and [23] ontology population available available semi-/supervised a corpus many be large, it might not contain all the necessary evidence of an event of interest. A corpus contains ambiguous statements about an event that leads to a non-determinism of the state of the event.…”
Section: Workmentioning
confidence: 99%