Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence 2017
DOI: 10.24963/ijcai.2017/579
|View full text |Cite
|
Sign up to set email alerts
|

Bilateral Multi-Perspective Matching for Natural Language Sentences

Abstract: Natural language sentence matching is a fundamental technology for a variety of tasks. Previous approaches either match sentences from a single direction or only apply single granular (wordby-word or sentence-by-sentence) matching. In this work, we propose a bilateral multi-perspective matching (BiMPM) model. Given two sentences P and Q, our model first encodes them with a BiL-STM encoder. Next, we match the two encoded sentences in two directions P against Q and Q against P . In each matching direction, each … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
494
0
2

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 581 publications
(498 citation statements)
references
References 0 publications
2
494
0
2
Order By: Relevance
“…Lei et al (2016) consider a related task leveraging the AskUbuntu corpus (dos Santos et al, 2015), but it contains two orders of magnitude less annotations, thus limiting the quality of any model. Most relevant to this work is that of Wang et al (2017), who present the best results on the Quora dataset prior to this work. The bilateral multi-perspective matching model (BIMPM) of Wang et al uses a character-based LSTM (Hochreiter and Schmidhuber, 1997) at its input representation layer, a layer of bi-LSTMs for computing context information, four different types of multi-perspective matching layers, an additional bi-LSTM aggregation layer, followed by a two-layer feedforward network for prediction.…”
Section: Related Workmentioning
confidence: 99%
“…Lei et al (2016) consider a related task leveraging the AskUbuntu corpus (dos Santos et al, 2015), but it contains two orders of magnitude less annotations, thus limiting the quality of any model. Most relevant to this work is that of Wang et al (2017), who present the best results on the Quora dataset prior to this work. The bilateral multi-perspective matching model (BIMPM) of Wang et al uses a character-based LSTM (Hochreiter and Schmidhuber, 1997) at its input representation layer, a layer of bi-LSTMs for computing context information, four different types of multi-perspective matching layers, an additional bi-LSTM aggregation layer, followed by a two-layer feedforward network for prediction.…”
Section: Related Workmentioning
confidence: 99%
“…• (Wang et al, 2016a) 0.734 0.742 BiMPM (Wang et al, 2017) 0 thogonal decomposition (OD) strategy has a superior performance to direct (DI) strategy on all datasets. The comparison results are posted in Table 1.…”
Section: Resultsmentioning
confidence: 99%
“…Model r ρ MSE Meaning Factory (Jiménez et al, 2014) 0.8268 0.7721 0.3224 ECNU (Zhao et al, 2014) 0.8414 --BiLSTM (Tai et al, 2015) 0.8567 0.7966 0.2736 Tree-LSTM (Tai et al, 2015) 0.8676 0.8083 0.2532 MPCNN (He et al, 2015) 0 Wang and Ittycheriah (2015) 0.746 0.820 QA-LSTM (Tan et al, 2015) 0.728 0.832 Att-pooling (dos Santos et al, 2016) 0.753 0.851 LDC (Wang et al, 2016b) 0.771 0.845 MPCNN (He et al, 2015) 0.777 0.836 PWIM 0.738 0.827 NCE-CNN (Rao et al, 2016) 0.801 0.877 BiMPM (Wang et al, 2017) 0.802 0.875 IWAN-att (Proposed) 0.822 0.889 IWAN-skip (Proposed) 0.801 0.861 Table 3: Test results on Clean version TrecQA.…”
Section: Training Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…Baselines. We compare the performance of our models with that of the state-of-the-art models on the clean version of the TREC-QA dataset (Shen et al, 2017;Bian et al, 2017;Wang et al, 2017;Rao et al, 2016;Tay et al, 2017). We do not have access to the original implementation of IWAN, hence we use our implementation of the IWAN model as the basis for our models.…”
Section: Contextual Language Modelmentioning
confidence: 99%