Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1224
|View full text |Cite
|
Sign up to set email alerts
|

Neural Natural Language Inference Models Enhanced with External Knowledge

Abstract: Modeling natural language inference is a very challenging task. With the availability of large annotated data, it has recently become feasible to train complex models such as neural-network-based inference models, which have shown to achieve the state-of-the-art performance. Although there exist relatively large annotated data, can machines learn all knowledge needed to perform natural language inference (NLI) from these data? If not, how can neural-network-based NLI models benefit from external knowledge and … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
166
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 222 publications
(166 citation statements)
references
References 48 publications
0
166
0
Order By: Relevance
“…Task-specific KB architectures Other work has focused on integrating KBs into neural architectures for specific downstream tasks (Yang and Mitchell, 2017;Sun et al, 2018;Chen et al, 2018;Bauer et al, 2018;Mihaylov and Frank, 2018;Wang and Jiang, 2019;Yang et al, 2019). Our approach instead uses KBs to learn more generally transferable representations that can be used to improve a variety of downstream tasks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Task-specific KB architectures Other work has focused on integrating KBs into neural architectures for specific downstream tasks (Yang and Mitchell, 2017;Sun et al, 2018;Chen et al, 2018;Bauer et al, 2018;Mihaylov and Frank, 2018;Wang and Jiang, 2019;Yang et al, 2019). Our approach instead uses KBs to learn more generally transferable representations that can be used to improve a variety of downstream tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Then, the model recontextualizes the entity-span representations with word-toentity attention to allow long range interactions between contextual word representations and all entity spans in the context. The entire KAR is inserted between two layers in the middle of a pretrained model such as BERT. In contrast to previous approaches that integrate external knowledge into task-specific models with task supervision (e.g., Yang and Mitchell, 2017;Chen et al, 2018), our approach learns the entity linkers with self-supervision on unlabeled data. This results in general purpose knowledge enhanced representations that can be applied to a wide range of downstream tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Future work includes the following directions: (1) We plan to explore approaches for effectively representing and incorporating external knowledge (Chen et al, 2018b) in the ESIM model and the BERT model, such as knowledge graph and user profile. It is important to advance the understanding of how to effectively represent the interactions between the context and the external knowledge for the response selection task.…”
Section: Resultsmentioning
confidence: 99%
“…These two attempts show a direction towards solving medical NLI problem where the pretrained embeddings are fine-tuned on medical corpus and are used in the state-of-the-art NLI architecture. Chen et al (2018) proposed the use of external knowledge to help enrich neural-network based NLI models by applying Knowledge-enriched coattention, Local inference collection with Exter-nal Knowledge, and Knowledge-enchanced inference composition components. Another line of solution tries to bring in the extra domain knowledge from sources like Unified Medical Language System (UMLS) (Bodenreider, 2004).…”
Section: Introductionmentioning
confidence: 99%
“…Another line of solution tries to bring in the extra domain knowledge from sources like Unified Medical Language System (UMLS) (Bodenreider, 2004). Romanov and Shivade (2018) used the knowledge-directed attention based methods in (Chen et al, 2018) for Medical NLI. Another such attempt is made by , where they incorporate domain knowledge in terms of the definitions of medical concepts from UMLS with the state-of-the-art NLI model ESIM (Chen et al, 2017) and vanilla word embeddings of Glove (Pennington et al, 2014) and fastText (Bojanowski et al, 2017).…”
Section: Introductionmentioning
confidence: 99%