Proceedings of the 35th Annual Meeting on Association for Computational Linguistics - 1997
DOI: 10.3115/976909.979638
|View full text |Cite
|
Sign up to set email alerts
|

A DOP model for semantic interpretation

Abstract: In data-oriented language processing, an annotated language corpus is used as a stochastic grammar. The most probable analysis of a new sentence is constructed by combining fragments from the corpus in the most probable way. This approach has been successfully used for syntactic analysis, using corpora with syntactic annotations such as the Penn Tree-bank. If a corpus with semantically annotated sentences is used, the same approach can also generate the most probable semantic interpretation of an input sentenc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
33
0

Year Published

1999
1999
2012
2012

Publication Types

Select...
4
3
2

Relationship

3
6

Authors

Journals

citations
Cited by 30 publications
(34 citation statements)
references
References 8 publications
1
33
0
Order By: Relevance
“…Note that the subtree probabilities in Likelihood-DOP are directly estimated from their relative frequencies in the treebank-trees. While the relative-frequency estimator obtains competitive results on several domains (Bonnema et al 1997;Bod, 2001a;De Pauw, 2000), it does not maximize the likelihood of the training data (Johnson, 2002). This is because there may be hidden derivations which the relative-frequency estimator cannot deal with.…”
Section: Likelihood-dopmentioning
confidence: 95%
“…Note that the subtree probabilities in Likelihood-DOP are directly estimated from their relative frequencies in the treebank-trees. While the relative-frequency estimator obtains competitive results on several domains (Bonnema et al 1997;Bod, 2001a;De Pauw, 2000), it does not maximize the likelihood of the training data (Johnson, 2002). This is because there may be hidden derivations which the relative-frequency estimator cannot deal with.…”
Section: Likelihood-dopmentioning
confidence: 95%
“…Accuracies on the Homecentre Tables 3 and 4 show that there is a consistent increase in parse accuracy for all metrics if larger fragments are included, but that the increase itself decreases. This phenomenon is also known as the DOP hypothesis (Bod 1998), and has been confirmed for Tree-DOP on the ATIS, OVIS and Wall Street Journal treebanks (see Bod 1993Bod , 1998Bod , 1999Bod , 2000aSima'an 1999;Bonnema et al 1997;Hoogweg 2000). The current result thus extends the validity of the DOP hypothesis to LFG annotations.…”
Section: Comparing Different Fragment Sizesmentioning
confidence: 96%
“…But the model has also been applied to several other grammatical frameworks, e.g. Tree-Insertion Grammar (Hoogweg 2000), Tree-Adjoining Grammar (Neumann 1998), Lexical-Functional Grammar (Bod & Kaplan 1998;Cormons 1999), Head-driven Phrase Structure Grammar (Neumann & Flickinger 1999), and Montague Grammar (Bonnema et al 1997;Bod 1999). Most probability models for DOP use the relative frequency estimator to estimate fragment probabilities, although Bod (2000b) trains fragment probabilities by a maximum likelihood reestimation procedure belonging to the class of expectation-maximization algorithms.…”
Section: Introductionmentioning
confidence: 99%
“…Another approach within treebank parsing is Data-Oriented Parsing (DOP) (Bod 1992(Bod , 1993Bonnema, Bod and Scha 1997), which collects statistics on the occurrence frequency of all tree fragments within a corpus, derives any sentence by assembling such fragments, and scores any parse in terms of the sum of the probabilities of all 380 A. Krotov and others of its derivations (and hence this model presents significant problems in terms of computational overhead 3 ).…”
Section: Treebank Grammars and Parsingmentioning
confidence: 99%