2005
DOI: 10.1007/s10994-005-0916-y
|View full text |Cite
|
Sign up to set email alerts
|

A Neural Syntactic Language Model

Abstract: Abstract. This paper presents a study of using neural probabilistic models in a syntactic based language model. The neural probabilistic model makes use of a distributed representation of the items in the conditioning history, and is powerful in capturing long dependencies. Employing neural network based models in the syntactic based language model enables it to use efficiently the large amount of information available in a syntactic parse in estimating the next word in a string. Several scenarios of integrati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
0

Year Published

2007
2007
2016
2016

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 49 publications
(36 citation statements)
references
References 31 publications
0
36
0
Order By: Relevance
“…For example, we could mention exponential models (such as "model M" (Chen, 2009)), syntactic models (Emami and Jelinek, 2005), Bayesian models (Teh, 2006), etc. For the sake of brevity, we omit further discussion of such techniques because they are largely orthogonal to deep learning.…”
Section: Deep Learning For Asr Language Modellingmentioning
confidence: 99%
“…For example, we could mention exponential models (such as "model M" (Chen, 2009)), syntactic models (Emami and Jelinek, 2005), Bayesian models (Teh, 2006), etc. For the sake of brevity, we omit further discussion of such techniques because they are largely orthogonal to deep learning.…”
Section: Deep Learning For Asr Language Modellingmentioning
confidence: 99%
“…A well established efficiency trick assigns each possible output to a unique class and then uses a two-step process to find the probability of an MTU, instead of computing the probability of all possible outputs (Goodman, 2001;Emami and Jelinek, 2005;Mikolov et al, 2011b). Under this scheme we compute the probability of an MTU by multiplying the probability of its class c i t with the probability of the This factorization reduces the complexity of computing the output probabilities from O(|V |) to O(|C| + max i |c i |) where |C| is the number of classes and |c i | is the number of minimal units in class c i .…”
Section: Atomic Mtu Rnn Modelmentioning
confidence: 99%
“…Many existing formulations of (symbolic and connectionist) language processing models belong to this framework (e.g., [20] and the review therein), in the sense that they are classifiers for input sequences.…”
Section: B Recursive Temporal Abstraction For Sensory Inputsmentioning
confidence: 99%