Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Confere 2015
DOI: 10.3115/v1/p15-2084
|View full text |Cite
|
Sign up to set email alerts
|

Dependency Recurrent Neural Language Models for Sentence Completion

Abstract: Recent work on language modelling has shifted focus from count-based models to neural models. In these works, the words in each sentence are always considered in a left-to-right order. In this paper we show how we can improve the performance of the recurrent neural network (RNN) language model by incorporating the syntactic dependencies of a sentence, which have the effect of bringing relevant contexts closer to the word being predicted. We evaluate our approach on the Microsoft Research Sentence Completion Ch… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 36 publications
(28 citation statements)
references
References 20 publications
(25 reference statements)
0
28
0
Order By: Relevance
“…They tested a variety of language modeling approaches using their task, and report that well-trained generative n-gram models achieve correct predictions ≈ 30% of the time. State-of-the-art performance on the their word prediction task using recurrent neural network langage models, 6 report highest scores are in the mid-50% range (Mirowski and Vlachos, 2015;Mnih and Kavukcuoglu, 2013).…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…They tested a variety of language modeling approaches using their task, and report that well-trained generative n-gram models achieve correct predictions ≈ 30% of the time. State-of-the-art performance on the their word prediction task using recurrent neural network langage models, 6 report highest scores are in the mid-50% range (Mirowski and Vlachos, 2015;Mnih and Kavukcuoglu, 2013).…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…Neural network language models [22,24,43] have attracted a lot of attention recently given their dense and learnable representation form and generalization property, as a contrast to the traditional bag-of-words representations. Word2vec skip-gram [23] (cf.…”
Section: Related Workmentioning
confidence: 99%
“…The best performing LSTM is worse than a LDTREEL-STM (d = 300). The input and output embeddings (W e and W ho ) dominate the number of parameters in all neural models except for RNNME, depRNN+3gram and ldepRNN+4gram, which include a ME model that contains 1 billion sparse n-gram features (Mikolov, 2012;Mirowski and Vlachos, 2015). The number of parameters in TREELSTM and LDTREELSTM is not much larger compared to LSTM due to the tied W e and W ho matrices.…”
Section: Microsoft Sentence Completion Challengementioning
confidence: 99%
“…Emami et al (2003) and Sennrich (2015) estimate the parameters of a structured language model using feed-forward neural networks (Bengio et al, 2003). Mirowski and Vlachos (2015) re-implement the model of Gubbins and Vlachos (2013) with RNNs. They view sentences as sequences of words over a tree.…”
Section: Introductionmentioning
confidence: 99%