2018
DOI: 10.1609/aaai.v32i1.11928
|View full text |Cite
|
Sign up to set email alerts
|

Deep Semantic Role Labeling With Self-Attention

Abstract: Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied. Recent years, end-to-end SRL with recurrent neural networks (RNN) has gained increasing attention. However, it remains a major challenge for RNNs to handle structural information and long range dependencies. In this paper, we present a simple and effective architecture for SRL which aims to address these problems. Our model is based on self-attention which can directly capture the r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
33
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 227 publications
(33 citation statements)
references
References 24 publications
0
33
0
Order By: Relevance
“…Semantic Extractor: We adopt the bidirectional long shortterm memory (Bi-LSTM) (Zhou et al 2016) with selfattention mechanism (Tan et al 2018) to learn the document representation. With the input word w q in a document x, the hidden states of Bi-LSTM are updated as,…”
Section: Learning Head Label Classifiersmentioning
confidence: 99%
“…Semantic Extractor: We adopt the bidirectional long shortterm memory (Bi-LSTM) (Zhou et al 2016) with selfattention mechanism (Tan et al 2018) to learn the document representation. With the input word w q in a document x, the hidden states of Bi-LSTM are updated as,…”
Section: Learning Head Label Classifiersmentioning
confidence: 99%
“…Cheng et al integrated the long short-term memory architecture with self-attention mechanism to render sequence-level networks better at handling structured input [20]. Tan et al presented a simple and effective architecture for semantic role labeling based on self-attention to handle structural information and long range dependencies [21]. The attention mechanism has enjoyed great popularity in the machine translation as well as NLP communities.…”
Section: B Self-attention Mechanismmentioning
confidence: 99%
“…Two main classes of approaches to span-based SRL are sequence labeling approaches (Yang and Mitchell 2017;He et al 2017;Tan et al 2018;Strubell et al 2018) and graphbased approaches (He et al 2018;Li et al 2019;Ouchi, Shindo, and Matsumoto 2018;Marcheggiani and Titov 2020). Sequence labeling approaches often employ the BIO tagging scheme and predict a BIO tag for each word in the input sentence.…”
Section: Introductionmentioning
confidence: 99%