2020
DOI: 10.3233/jifs-191171
|View full text |Cite
|
Sign up to set email alerts
|

Attention-based LSTM, GRU and CNN for short text classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 40 publications
(21 citation statements)
references
References 3 publications
0
21
0
Order By: Relevance
“…In the subsequent experiments, we use HingeLoss, which has the best overall effect of efficiency and accuracy, as the loss function for labeled data, and RampLoss as the loss function for unlabeled data, as shown in Figure 6(b). Then, in the following chapters [17,18], our optimization objective (12), as shown in Figure 6(d):…”
Section: Application Of Nlp In Text Classificationmentioning
confidence: 99%
“…In the subsequent experiments, we use HingeLoss, which has the best overall effect of efficiency and accuracy, as the loss function for labeled data, and RampLoss as the loss function for unlabeled data, as shown in Figure 6(b). Then, in the following chapters [17,18], our optimization objective (12), as shown in Figure 6(d):…”
Section: Application Of Nlp In Text Classificationmentioning
confidence: 99%
“…The introduction of transformers‐based embeddings has shown a significant increase in the quality of contextual information extraction 10,87,89 . However, the use of transformers is affected by some limitations, viz., out of vocabulary (OOV) limitation, increased complexity, and importantly the method overfitting in small networks 99,100 . Due to the highlighted drawbacks of transformers, an ensemble of attention and neuro‐fuzzy networks 101 would help reduce the limiting effects, thus increasing classification performance.…”
Section: Open Issues and Future Research Directionsmentioning
confidence: 99%
“…e GRU model [16], a variant of the recurrent neural network (RNN) gated recurrent unit [17], is particularly suitable for time-series data and also addresses the longtime dependence problem that is important for predictive regression results, coupled with the convolutional neural network (CNN) model that can effectively extract potentially important features between data. e combination of GRU and CNN models [18] is therefore ideally suited to our prediction and regression of corporate financial reports.…”
Section: Building Deep Learning Prediction Models Based On Gru and Cnnmentioning
confidence: 99%