Proceedings of the First Workshop on Subword and Character Level Models in NLP 2017
DOI: 10.18653/v1/w17-4121
|View full text |Cite
|
Sign up to set email alerts
|

Neural Paraphrase Identification of Questions with Noisy Pretraining

Abstract: We present a solution to the problem of paraphrase identification of questions. We focus on a recent dataset of question pairs annotated with binary paraphrase labels and show that a variant of the decomposable attention model (Parikh et al., 2016) results in accurate performance on this task, while being far simpler than many competing neural architectures. Furthermore, when the model is pretrained on a noisy dataset of automatically collected question paraphrases, it obtains the best reported performance on … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
25
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 40 publications
(26 citation statements)
references
References 19 publications
1
25
0
Order By: Relevance
“…They find that Shortcut-Stacked Sentence Encoder Model (SSE) [39] performs the best, giving testing accuracy of 87.8%. The previous best accuracy is 88.4% [13]. Our multi-cascaded model beats this result without any data augmentation with an accuracy of 89.4%.…”
Section: Quora Datasetsupporting
confidence: 52%
See 4 more Smart Citations
“…They find that Shortcut-Stacked Sentence Encoder Model (SSE) [39] performs the best, giving testing accuracy of 87.8%. The previous best accuracy is 88.4% [13]. Our multi-cascaded model beats this result without any data augmentation with an accuracy of 89.4%.…”
Section: Quora Datasetsupporting
confidence: 52%
“…In contrast to [7], we avoid using ensemble model approach (which is computationally costly). Similarly, contrary to results in [13], our results are based on word features only and do not use computationally expensive character-based features.…”
Section: Quora Datasetmentioning
confidence: 88%
See 3 more Smart Citations