2001
DOI: 10.1006/csla.2000.0159
|View full text |Cite
|
Sign up to set email alerts
|

Whole-sentence exponential language models: a vehicle for linguistic-statistical integration

Abstract: We introduce an exponential language model which models a whole sentence or utterance as a single unit. By avoiding the chain rule, the model treats each sentence as a "bag of features", where features are arbitrary computable properties of the sentence. The new model is computationally more efficient, and more naturally suited to modeling global sentential phenomena, than the conditional exponential (e.g. maximum entropy) models proposed to date. Using the model is straightforward. Training the model requires… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
56
0
1

Year Published

2001
2001
2016
2016

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 68 publications
(58 citation statements)
references
References 19 publications
(14 reference statements)
1
56
0
1
Order By: Relevance
“…Future experiments are needed to evaluate system performance on isolated words, and the performance on isolated words and complete sentences must be compared in order to determine whether the additional context information available in complete sentences improves system performance. Since previous studies indicate that semantic context is important to human listeners in the perceptual restoration task, future experimenters may wish to evaluate top-down completion approaches with a whole-sentence language model (Rosenfeld et al, 2001).…”
Section: Discussionmentioning
confidence: 99%
“…Future experiments are needed to evaluate system performance on isolated words, and the performance on isolated words and complete sentences must be compared in order to determine whether the additional context information available in complete sentences improves system performance. Since previous studies indicate that semantic context is important to human listeners in the perceptual restoration task, future experimenters may wish to evaluate top-down completion approaches with a whole-sentence language model (Rosenfeld et al, 2001).…”
Section: Discussionmentioning
confidence: 99%
“…Another type of maximum entropy LMs is the whole sentence model given in [111]. This model predicts the probability of the whole sentence by using feature functions over the entire sentence.…”
Section: Maximum Entropy Lmsmentioning
confidence: 99%
“…Different form generative models requiring the joint probability of the observations and the labels, it calculates the conditional probability of labels over the observations directly. Rosenfeld proposed a maximum entropy based language model that employed shallow syntactic features [8]. Magdin used Maximum Mutual Information Estimation (MMIE) method to optimize an objective function that involved a metric between correct transcriptions and their competing hypotheses, which are encoded as word graphs generated by Viterbi decoding process [18].…”
Section: Enhanced Language Modelsmentioning
confidence: 99%
“…N-gram language model can be improved by directly integrated with higher level models, such as dependency relations and probabilistic top-down parsing [6,7]. Discriminative methods such as maximum entropy models can also be used to utilize global features from word sequences and syntactic structures [8]. Discriminative re-scoring methods…”
Section: Introductionmentioning
confidence: 99%