Proceedings of the 55th Annual Meeting of the Association For Computational Linguistics (Volume 1: Long Papers) 2017
DOI: 10.18653/v1/p17-1152
|View full text |Cite
|
Sign up to set email alerts
|

Enhanced LSTM for Natural Language Inference

Abstract: Reasoning and inference are central to human and artificial intelligence. Modeling inference in human language is very challenging. With the availability of large annotated data (Bowman et al., 2015), it has recently become feasible to train neural network based inference models, which have shown to be very effective. In this paper, we present a new state-of-the-art result, achieving the accuracy of 88.6% on the Stanford Natural Language Inference Dataset. Unlike the previous top models that use very complicat… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

9
825
0
2

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 931 publications
(858 citation statements)
references
References 28 publications
9
825
0
2
Order By: Relevance
“…Enhanced Sequential Information Model ESIM (Chen et al, 2017) performs inference in three stages. First, Input Encoding uses BiLSTMs to produce representations of each word in its context within premise or hypothesis.…”
Section: Modelsmentioning
confidence: 99%
“…Enhanced Sequential Information Model ESIM (Chen et al, 2017) performs inference in three stages. First, Input Encoding uses BiLSTMs to produce representations of each word in its context within premise or hypothesis.…”
Section: Modelsmentioning
confidence: 99%
“…As the only large human-annotated corpus for NLI currently available, the Stanford NLI Corpus (SNLI; Bowman et al, 2015) has enabled a good deal of progress on NLU, serving as a major benchmark for machine learning work on sentence understanding and spurring work on core representation learning techniques for NLU, such as attention (Wang and Jiang, 2016;Parikh et al, 2016), memory (Munkhdalai and Yu, 2017), and the use of parse structure (Mou et al, 2016b;Bowman et al, 2016;Chen et al, 2017). However, SNLI falls short of providing a sufficient testing ground for machine learning models in two ways.…”
Section: Introductionmentioning
confidence: 99%
“…For example, Part-Of-Speech (POS) tags are used for syntactic parsers. The parsers are used to improve higher-level tasks, such as natural language inference (Chen et al, 2016) and machine translation (Eriguchi et al, 2016). These systems are often pipelines and not trained end-to-end.…”
Section: Introductionmentioning
confidence: 99%