2020
DOI: 10.1609/aaai.v34i05.6511
|View full text |Cite
|
Sign up to set email alerts
|

SG-Net: Syntax-Guided Machine Reading Comprehension

Abstract: For machine reading comprehension, the capacity of effectively modeling the linguistic knowledge from the detail-riddled and lengthy passages and getting ride of the noises is essential to improve its performance. Traditional attentive models attend to all words without explicit constraint, which results in inaccurate concentration on some dispensable words. In this work, we propose using syntax to guide the text modeling by incorporating explicit syntactic constraints into attention mechanism for better lingu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
102
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 157 publications
(103 citation statements)
references
References 23 publications
1
102
0
Order By: Relevance
“…It directly proves the necessity of incorporating structural information and reasoning mechanism in document-level RE, and many kinds of basic encoder could be well integrated into our model due to the loosely-coupled architecture. We consider the performance gain on BERT-DRE is mainly because that DHG makes full use of the semantic information of BERT, and the graph structure complements the weakness of BERT in capturing long-range syntactic structure, which is consistent with the conclusion of some recent studies (Clark et al, 2019;Zhang et al, 2020b). Meanwhile, another phenomenon is that the DHG-based models outperform all baseline models in both intra-and inter-sentence scenarios, especially the inter-sentence setting, which demonstrates that the majority of DHG mainly comes from inter-sentence relational facts.…”
Section: Resultssupporting
confidence: 84%
“…It directly proves the necessity of incorporating structural information and reasoning mechanism in document-level RE, and many kinds of basic encoder could be well integrated into our model due to the loosely-coupled architecture. We consider the performance gain on BERT-DRE is mainly because that DHG makes full use of the semantic information of BERT, and the graph structure complements the weakness of BERT in capturing long-range syntactic structure, which is consistent with the conclusion of some recent studies (Clark et al, 2019;Zhang et al, 2020b). Meanwhile, another phenomenon is that the DHG-based models outperform all baseline models in both intra-and inter-sentence scenarios, especially the inter-sentence setting, which demonstrates that the majority of DHG mainly comes from inter-sentence relational facts.…”
Section: Resultssupporting
confidence: 84%
“…From a practical point of view, finetuning a pre-trained model is resource-saving. Besides, a variety of studies show that fine-tuning a pre-trained model has a great impact on MCRC performance [8]- [10].…”
Section: Introductionmentioning
confidence: 99%
“…To do so, we implement the QA model [16]. In fact, recently published QA models including BERT and XLNet [17] are shown to achieve human-like performance when tested on the Stanford Question Answering data set [18]. Once the detailed traffic information is obtained, we implement an automated system that converts collected posts from social media into real-time traffic information that are used to update navigation maps and warn drivers about any traffic events.…”
Section: Introductionmentioning
confidence: 99%