2017
DOI: 10.1613/jair.5259
|View full text |Cite
|
Sign up to set email alerts
|

A Neural Probabilistic Structured-Prediction Method for Transition-Based Natural Language Processing

Abstract: We propose a neural probabilistic structured-prediction method for transition-based natural language processing, which integrates beam search and contrastive learning. The method uses a global optimization model, which can leverage arbitrary features over nonlocal context. Beam search is used for efficient heuristic decoding, and contrastive learning is performed for adjusting the model according to search errors. When evaluated on both chunking and dependency parsing tasks, the proposed method achieves signif… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 51 publications
0
6
0
Order By: Relevance
“…The end pointer of the stack changed position as the stack of tree nodes could be pushed and popped. Zhou et al [145] integrated beam search and contrastive learning for better optimization.…”
Section: B Parsingmentioning
confidence: 99%
“…The end pointer of the stack changed position as the stack of tree nodes could be pushed and popped. Zhou et al [145] integrated beam search and contrastive learning for better optimization.…”
Section: B Parsingmentioning
confidence: 99%
“…Thus all features in the neural segmentation model are dense features, alleviating the feature sparsity problem naturally, and meanwhile it is free of feature engineering. For better comparison between discrete and neural features, the overall segmentation framework of the baseline is kept, which includes the incremental segmentation process, the beam-search decoder and the training process integrated with beam-search (Zhang & Clark, 2011;Zhou et al, 2017). In addition, the neural network scorer takes the similar feature sources as the baseline, which includes both the input character sequence and the partially constructed output word sequence.…”
Section: Transition-based Neural Modelmentioning
confidence: 99%
“…Since it is challenging to integrate word features to the CRF inference framework of the existing character-based methods, we take inspiration from word-based discrete segmentation model instead. In particular, we follow Zhang and Clark (2007), using a transition-based framework (Zhang & Clark, 2011;Zhou, Zhang, Cheng, Huang, Dai, & Chen, 2017) to segment a sentence incrementally from left to right, scoring partially segmented results by using both character-level and word-level features. We replace the discrete word and character features of Zhang and Clark (2007) with neural word and character representations, respectively, and change their linear model into a deep neural network.…”
Section: Introductionmentioning
confidence: 99%
“…This is especially true with the advent of deep learning-based NLP. Currently, some of the best algorithms for part-of-speech tagging (Huang, Xu, & Yu, 2015), parsing (D. Chen & Manning, 2014;Dyer, Ballesteros, Ling, Matthews, & Smith, 2015;Zhou et al, 2017;Zhu, Zhang, Chen, Zhang, & Zhu, 2013), named entity recognition (Chiu & Nichols, 2016;Passos, Kumar, & McCallum, 2014), sentiment classification (Mikolov, Sutskever, Chen, Corrado, & Dean, 2013), machine translation (Koehn, Och, & Marcu, 2003), and contextual embeddings (Q. Chen, Zhu, Ling, Wei, & Jiang, 2016;Liu, Shen, Duh, & Gao, 2017) are done with deep learning-based NLP.…”
Section: Introductionmentioning
confidence: 99%