Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1130
|View full text |Cite
|
Sign up to set email alerts
|

Stack-Pointer Networks for Dependency Parsing

Abstract: We introduce a novel architecture for dependency parsing: stack-pointer networks (STACKPTR). Combining pointer networks (Vinyals et al., 2015) with an internal stack, the proposed model first reads and encodes the whole sentence, then builds the dependency tree top-down (from root-to-leaf) in a depth-first fashion. The stack tracks the status of the depthfirst search and the pointer networks select one child for the word at the top of the stack at each step. The STACKPTR parser benefits from the information of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
213
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 134 publications
(215 citation statements)
references
References 39 publications
2
213
0
Order By: Relevance
“…3.1 Pointer Networks for Parsing. Ma et al (2018) and Lin et al (2019) both use a pointer network as the backbone of their parsing models and achieve state-of-the-art performance in dependency and discourse parsing tasks, respectively. As shown in Figures 2(a) and 3, in both cases, the parsing algorithm is implemented in a top-down depth-first order.…”
Section: Hierarchical Pointer Networkmentioning
confidence: 99%
See 2 more Smart Citations
“…3.1 Pointer Networks for Parsing. Ma et al (2018) and Lin et al (2019) both use a pointer network as the backbone of their parsing models and achieve state-of-the-art performance in dependency and discourse parsing tasks, respectively. As shown in Figures 2(a) and 3, in both cases, the parsing algorithm is implemented in a top-down depth-first order.…”
Section: Hierarchical Pointer Networkmentioning
confidence: 99%
“…Given a sentence of length n, the number of decoding steps to build a parse tree is linear. The attention mechanism at each decoding step computes an attention vector of length n. The overall decoding complexity is O(n 2 ), which is same as the StackPointer Parser (Ma et al, 2018).…”
Section: Hierarchical Decodermentioning
confidence: 99%
See 1 more Smart Citation
“…The encoder and the decoder jointly form the parsing model and we consider two alternatives 2 from (Ahmad et al, 2019): "SelfAtt-Graph" and "RNN-Stack". The "SelfAtt-Graph" parser consists of a modified self-attentional encoder (Shaw et al, 2018) and a graph-based deep bi-affine decoder (Dozat and Manning, 2017), while the "RNN-Stack" parser is composed of a Recurrent Neural Network (RNN) based encoder and a stack-pointer decoder (Ma et al, 2018).…”
Section: Architecturementioning
confidence: 99%
“…To train the parser, we adopt both cross-entropy objectives for these two types of parsers as in (Dozat and Manning, 2017;Ma et al, 2018). The encoder and the decoder are jointly trained to optimize the probability of the dependency trees (y) given sentences (x):…”
Section: Parsingmentioning
confidence: 99%