Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2016
DOI: 10.18653/v1/n16-1035
|View full text |Cite
|
Sign up to set email alerts
|

Top-down Tree Long Short-Term Memory Networks

Abstract: Long Short-Term Memory (LSTM) networks, a type of recurrent neural network with a more complex computational unit, have been successfully applied to a variety of sequence modeling tasks. In this paper we develop Tree Long Short-Term Memory (TREELSTM), a neural network model based on LSTM, which is designed to predict a tree rather than a linear sequence. TREELSTM defines the probability of a sentence by estimating the generation probability of its dependency tree. At each time step, a node is generated based o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
39
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 56 publications
(39 citation statements)
references
References 28 publications
0
39
0
Order By: Relevance
“…The architecture of tree LSTMs varies depending on the task. Some options include summarizing over the children, adding a separate forget gate for each child (Tai et al, 2015), recurrent propagation among siblings (Zhang et al, 2016), or use of stack LSTMs (Dyer et al, 2015). Our work differs from these studies in two respects: the tree structure here characterizes a discussion rather than a single sentence; 7 3 7 7 that is terrifying.…”
Section: Related Workmentioning
confidence: 99%
“…The architecture of tree LSTMs varies depending on the task. Some options include summarizing over the children, adding a separate forget gate for each child (Tai et al, 2015), recurrent propagation among siblings (Zhang et al, 2016), or use of stack LSTMs (Dyer et al, 2015). Our work differs from these studies in two respects: the tree structure here characterizes a discussion rather than a single sentence; 7 3 7 7 that is terrifying.…”
Section: Related Workmentioning
confidence: 99%
“…Another drawback is that they needed special tokens for predicting branches, which are not necessary for MWPs because all operators are binary operators. The similar framework is also used in code generation (Zhang et al, 2016;Yin and Neubig, 2017). Alvarez-Melis and Jaakkola (2017) presented doubly recurrent neural networks to predict tree topology explicitly.…”
Section: Seq2tree Architecturesmentioning
confidence: 99%
“…Kementchedjhieva and Lopez (2018) found a single unit in their English CNLM that seems, qualitatively, to be tracking morpheme/word boundaries. Since they trained the model with whitespace, the main function of this unit could simply (Mikolov, 2012), Word RNN (Zweig et al, 2012), Word LSTM and LdTreeLSTM (Zhang et al, 2016). We further report models incorporating distributional encodings of semantics (right): Skipgram(+RNNs) from Mikolov et al (2013a), the PMI-based model of Woods (2016), and the Context-Embedding-based approach of Melamud et al (2016). be to predict the very frequent whitespace character.…”
Section: Boundary Tracking In Cnlmsmentioning
confidence: 99%