2020
DOI: 10.1016/j.eswa.2020.113214
|View full text |Cite
|
Sign up to set email alerts
|

Multiple premises entailment recognition based on attention and gate mechanism

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 4 publications
0
2
0
Order By: Relevance
“…At the same time, in the process of model training, facing the pain points such as limited data volume of cold chain logistics prediction, most of the existing researches use Attention mechanism to train the prediction model. Although Attention mechanism is widely used in AI and other scenarios as the mainstream, it usually needs a large amount of data to learn reasonable weight parameters, and cannot learn the sequence relationship in the sequence [17]. Therefore, the Attention mechanism may not be effective in predicting continuous values with limited data.…”
Section: Innovationmentioning
confidence: 99%
“…At the same time, in the process of model training, facing the pain points such as limited data volume of cold chain logistics prediction, most of the existing researches use Attention mechanism to train the prediction model. Although Attention mechanism is widely used in AI and other scenarios as the mainstream, it usually needs a large amount of data to learn reasonable weight parameters, and cannot learn the sequence relationship in the sequence [17]. Therefore, the Attention mechanism may not be effective in predicting continuous values with limited data.…”
Section: Innovationmentioning
confidence: 99%
“…In addition, the authors in [21] propose a model for the task of multiplepremise inference that uses a gate mechanism and use multiple LSTMs. Another model for multiplepremise inference is proposed in [30] by applying local matching-integration and concatenate-matching. In local matching, each premise sentence is individually matched with the hypothesis sentence to obtain multiple local classi cation results, which are then merged to get the nal inference classi cation.…”
Section: Related Workmentioning
confidence: 99%
“…In practical applications, RNN has been able to process some simple correlation information while its memory capacity is not strong. When the sequence is too long, error back propagation will cause larger gradient dispersion and gradient explosion problems, which can be effectively alleviated by introducing a "gate" mechanism [47,48] and memory unit [25,49] in the LSTM network.…”
Section: Feature Extraction Based On Lstm Networkmentioning
confidence: 99%