Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1073
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Neural Dependency Parsing

Abstract: Unsupervised dependency parsing aims to learn a dependency grammar from text annotated with only POS tags. Various features and inductive biases are often used to incorporate prior knowledge into learning. One useful type of prior information is that there exist correlations between the parameters of grammar rules involving different POS tags. Previous work employed manually designed features or special prior distributions to encode such information. In this paper, we propose a novel approach to unsupervised d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
74
0
1

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
3

Relationship

3
4

Authors

Journals

citations
Cited by 54 publications
(75 citation statements)
references
References 8 publications
0
74
0
1
Order By: Relevance
“…WSJ10 WSJ Unlexicalized Approaches, with WSJ10 EVG (Headden III et al, 2009) 65.0 -TSG-DMV (Blunsom and Cohn, 2010) 65.9 53.1 PR-S (Gillenwater et al, 2010) 64.3 53.3 HDP-DEP (Naseem et al, 2010) 73.8 -UR-A E-DMV (Tu and Honavar, 2012) 71.4 57.0 Neural E-DMV (Jiang et al, 2016) 72.5 57.6 Systems Using Lexical Information and/or More Data LexTSG-DMV (Blunsom and Cohn, 2010) 67.7 55.7 L-EVG (Headden III et al, 2009) 68.8 -CS (Spitkovsky et al, 2013) 72.0 64.4 MaxEnc (Le and Zuidema, 2015) 73. Again, we find that good initialization leads to better performance than KM initialization, and both good initialization and KM initialization are significantly better than random and uniform initialization.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…WSJ10 WSJ Unlexicalized Approaches, with WSJ10 EVG (Headden III et al, 2009) 65.0 -TSG-DMV (Blunsom and Cohn, 2010) 65.9 53.1 PR-S (Gillenwater et al, 2010) 64.3 53.3 HDP-DEP (Naseem et al, 2010) 73.8 -UR-A E-DMV (Tu and Honavar, 2012) 71.4 57.0 Neural E-DMV (Jiang et al, 2016) 72.5 57.6 Systems Using Lexical Information and/or More Data LexTSG-DMV (Blunsom and Cohn, 2010) 67.7 55.7 L-EVG (Headden III et al, 2009) 68.8 -CS (Spitkovsky et al, 2013) 72.0 64.4 MaxEnc (Le and Zuidema, 2015) 73. Again, we find that good initialization leads to better performance than KM initialization, and both good initialization and KM initialization are significantly better than random and uniform initialization.…”
Section: Methodsmentioning
confidence: 99%
“…Here we employ a neural approach to smoothing. Specifically, we propose a lexicalized extension of neural DMV (Jiang et al, 2016) and we call the resulting approach L-NDMV.…”
Section: Lexicalized Ndmvmentioning
confidence: 99%
See 2 more Smart Citations
“…where D(x) is the set of all possible dependency arcs of sentence x, 1[·] is the indicator function, and µ(x, i, j) is the expected count defined as follows, (Jiang et al, 2016), and Convex-MST (Grave and Elhadad, 2015) Methods WSJ10 WSJ Basic Setup Feature DMV (Berg-Kirkpatrick et al, 2010) 63.0 -UR-A E-DMV (Tu and Honavar, 2012) 71.4 57.0 Neural E-DMV (Jiang et al, 2016) 69.7 52.5 Neural E-DMV (Good Init) (Jiang et al, 2016) 72.5 57.6 Basic Setup + Universal Linguistic Prior Convex-MST (Grave and Elhadad, 2015) 60.8 48.6 HDP-DEP (Naseem et al, 2010) 71.9 -CRFAE 71.7 55.7 Systems Using Extra Info LexTSG-DMV (Blunsom and Cohn, 2010) 67.7 55.7 CS (Spitkovsky et al, 2013) 72.0 64.4 MaxEnc (Le and Zuidema, 2015) 73.2 65.8 Table 3: Comparison of recent unsupervised dependency parsing systems on English. Basic setup is the same as our setup except that linguistic prior is not used.…”
Section: Algorithmmentioning
confidence: 99%