Interspeech 2016 2016
DOI: 10.21437/interspeech.2016-354
|View full text |Cite
|
Sign up to set email alerts
|

Attention-Based Convolutional Neural Networks for Sentence Classification

Abstract: Sentence classification is one of the foundational tasks in spoken language understanding (SLU) and natural language processing (NLP). In this paper we propose a novel convolutional neural network (CNN) with attention mechanism to improve the performance of sentence classification. In traditional C-NN, it is not easy to encode long term contextual information and correlation between non-consecutive words effectively. In contrast, our attention-based CNN is able to capture these kinds of information for each wo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
44
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 88 publications
(44 citation statements)
references
References 13 publications
0
44
0
Order By: Relevance
“…• ATT-CNN [43]: It uses attention mechanism to capture long-term dependence information and correlation between nonconsecutive words automatically and then sents them to CNN. The parameters are the same with [10] in CNN.…”
Section: B Baseline Methodsmentioning
confidence: 99%
“…• ATT-CNN [43]: It uses attention mechanism to capture long-term dependence information and correlation between nonconsecutive words automatically and then sents them to CNN. The parameters are the same with [10] in CNN.…”
Section: B Baseline Methodsmentioning
confidence: 99%
“…The probability module calculates the activation probability of the capsule according to the semantic feature , combined with formula (18).…”
Section: Figure 6 Capsule Structure Diagrammentioning
confidence: 99%
“…The attention mechanism can achieve selective focus on important information. Zhao et al [18] proposed an ATT-CNN model combining attention mechanism and CNN to effectively identify the importance of words in a sentence. Vaswani et al [19] proposed multi-head attention mechanism adopted in the transformer translation model allows the model to obtain more levels of information in sentences from different spaces and improve the feature expression ability of the model.…”
Section: Introductionmentioning
confidence: 99%
“…The parameter setting in convolution neural network for all above models refer to Kim's paper, the convolution filter widths H are set to [2,3,4]. Each width has a set of 100 convolution filters.We conduct grid search on MR datasets and find that Model III can get a good performance when the hidden state dimension k in Bi-LSTM is set to 100, the dimension of Bi-LSTM and the global feature selection vector is set to 100 during training.…”
Section: 2baselines and Parametersmentioning
confidence: 99%