2019
DOI: 10.1609/aaai.v33i01.33016586
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Sentence Matching with Densely-Connected Recurrent and Co-Attentive Information

Abstract: Sentence matching is widely used in various natural language tasks such as natural language inference, paraphrase identification, and question answering. For these tasks, understanding logical and semantic relationship between two sentences is required but it is yet challenging. Although attention mechanism is useful to capture the semantic relationship and to properly align the elements of two sentences, previous methods of attention mechanism simply use a summation operation which does not retain original fe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
124
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 172 publications
(124 citation statements)
references
References 9 publications
0
124
0
Order By: Relevance
“…We compare single models and ensemble models. For a fair comparison, we only compare with results obtained without external contextualized embed- (Chen et al, 2017) 4.3M 88.0 DIIN (Gong et al, 2018) 4.4M 88.0 MwAN (Tan et al, 2018) 14M 88.3 CAFE (Tay et al, 2018b) 4.7M 88.5 HIM (Chen et al, 2017) 7.7M 88.6 SAN (Liu et al, 2018) 3.5M 88.6 CSRAN (Tay et al, 2018a) 13.9M 88.7 DRCN (Kim et al, 2018) 6.7M 88.9 RE2 ( Model Acc(%) ESIM (Chen et al, 2017) 70.6 DecompAtt (Parikh et al, 2016) 72.3 DGEM (Khot et al, 2018) 77.3 HCRN (Tay et al, 2018c) 80.0 CAFE (Tay et al, 2018b) 83.3 CSRAN (Tay et al, 2018a) 86.7 RE2 (ours) 86.0±0.6 dings. In the ensemble experiment, we train 8 models with different random seeds and ensemble the results by a voting strategy.…”
Section: Results On Natural Language Inferencementioning
confidence: 99%
See 3 more Smart Citations
“…We compare single models and ensemble models. For a fair comparison, we only compare with results obtained without external contextualized embed- (Chen et al, 2017) 4.3M 88.0 DIIN (Gong et al, 2018) 4.4M 88.0 MwAN (Tan et al, 2018) 14M 88.3 CAFE (Tay et al, 2018b) 4.7M 88.5 HIM (Chen et al, 2017) 7.7M 88.6 SAN (Liu et al, 2018) 3.5M 88.6 CSRAN (Tay et al, 2018a) 13.9M 88.7 DRCN (Kim et al, 2018) 6.7M 88.9 RE2 ( Model Acc(%) ESIM (Chen et al, 2017) 70.6 DecompAtt (Parikh et al, 2016) 72.3 DGEM (Khot et al, 2018) 77.3 HCRN (Tay et al, 2018c) 80.0 CAFE (Tay et al, 2018b) 83.3 CSRAN (Tay et al, 2018a) 86.7 RE2 (ours) 86.0±0.6 dings. In the ensemble experiment, we train 8 models with different random seeds and ensemble the results by a voting strategy.…”
Section: Results On Natural Language Inferencementioning
confidence: 99%
“…POS tags are found in many previous works including Tay et al (2018b) and Gong et al (2018). The exact match of lemmatized tokens is reported as a powerful binary feature in Gong et al (2018) and Kim et al (2018). The second way is adding complexity to the alignment computation.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…They reach an accuracy of 91.1%. Kim et al (2019) exploit denselyconnected co-attentive recurrent neural network, and reach 90% accuracy. In our scenario, we generate pseudo premises and hypotheses, then apply the standard transformer encoder (Ashish et al, 2017;Devlin et al, 2019) to train two NLI models.…”
Section: Related Workmentioning
confidence: 99%