2019
DOI: 10.3233/ida-184311
|View full text |Cite
|
Sign up to set email alerts
|

An attention-gated convolutional neural network for sentence classification

Abstract: In this paper, we investigate the effect of different hyperparameters as well as different combinations of hyperparameters settings on the performance of the Attention-Gated Convolutional Neural Networks (AGCNNs), e.g., the kernel window size, the number of feature maps, the keep rate of the dropout layer, and the activation function. We draw practical advice from a wide range of empirical results. Through the sensitivity analysis, we further improve the hyperparameters settings of AGCNNs. Experiments show tha… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(15 citation statements)
references
References 36 publications
0
13
0
Order By: Relevance
“…Since the weights and the abstract features are convolved at the same level, the significant features identified by the gating weights is very monotonous. To better discover contextual information in text classification, Yang Liu et al introduced a new CNN model (AGCNN) for sentence classification, which generated the gating weights by a variety of specialized convolution kernels to integrate the contextual information of a particular context window into the control weights [31]. And to achieve better performance with aspect-based sentiment analysis, Wei Xue et al proposed a model based on gated convolutional neural networks, which can selectively output the sentiment features according to the given aspect or entity [32].…”
Section: Related Workmentioning
confidence: 99%
“…Since the weights and the abstract features are convolved at the same level, the significant features identified by the gating weights is very monotonous. To better discover contextual information in text classification, Yang Liu et al introduced a new CNN model (AGCNN) for sentence classification, which generated the gating weights by a variety of specialized convolution kernels to integrate the contextual information of a particular context window into the control weights [31]. And to achieve better performance with aspect-based sentiment analysis, Wei Xue et al proposed a model based on gated convolutional neural networks, which can selectively output the sentiment features according to the given aspect or entity [32].…”
Section: Related Workmentioning
confidence: 99%
“…Attention models have also been used with CNN and time series data. The work in [43] proposed an attention gated CNN for sentence classification, and the work in [44] introduced a temporal causal discovery framework (TCDF) for learning causal relationships in time series data. However, many innovative CNN architectures mentioned above have not yet been explored for renewable energy applications.…”
Section: Of 29mentioning
confidence: 99%
“…The works in [65][66][67] are a few examples where attention blocks were proposed and used with LSTM architectures for time series forecasting. Attention blocks have also been used with CNN architectures [40,43,68,69] for image classification and time series data. One of the noteworthy contributions of the attention mechanism for time series forecasting can be found in [70].…”
Section: The Proposed Modelmentioning
confidence: 99%
“…Specifically, they have used CNN as the feature extractor of short texts and SVM as the classifier, and SVMCNN shows a better performance than each of CNN and SVM. Liu et al have proposed an attention-gated CNN for the sentence classification by generating attention weights from the feature's context windows before the pooling layer [36], which shows a better performance than standard CNN models.…”
Section: Cnnmentioning
confidence: 99%