2023
DOI: 10.3390/s23031481
|View full text |Cite
|
Sign up to set email alerts
|

Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU

Abstract: In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the proposed Sliced Bidirectional Gated Recurrent Unit (Sliced Bi-GRU), Multi-head Self-Attention mechanism, and Bidirectional Encoder Representations from Transformers embedding. First, the word vector representation obta… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 31 publications
(10 citation statements)
references
References 33 publications
0
4
0
1
Order By: Relevance
“…This approach enables a deeper contextual understanding of the data, crucial for handling the complex nature of protein sequences. Additionally, the model incorporates a multi-head attention mechanism, allowing it to capture diverse contextual representations from multiple perspectives. This mechanism divides the input into multiple heads, with each head focusing on different parts of the sequence, thus enabling a more nuanced and comprehensive understanding of the protein sequences .…”
Section: Methodsmentioning
confidence: 99%
“…This approach enables a deeper contextual understanding of the data, crucial for handling the complex nature of protein sequences. Additionally, the model incorporates a multi-head attention mechanism, allowing it to capture diverse contextual representations from multiple perspectives. This mechanism divides the input into multiple heads, with each head focusing on different parts of the sequence, thus enabling a more nuanced and comprehensive understanding of the protein sequences .…”
Section: Methodsmentioning
confidence: 99%
“…Other metrics such as SPICE [2] account for such variety by comparing parse trees, however, they still can not account for different word choice. Inspired by advances in language models as well as approaches like BERTScore [110] and CLIPScore [40], we use a pretrained language model to compute caption similarity.…”
Section: Sampling Image Pairs Using Languagementioning
confidence: 99%
“…Overall, the proposed LeBERT model was 88.2% accurate when used for binary classification. Zhang et al [21] presented a Sliced BI-GRU (bidirectional-gated recurrent unit) architecture that employs BERT embedding in conjunction with the multi-head self-attention mechanism. The BERT models' word vector representation first, which plays a role in the neural network's embedding layer, and then they divide the input sequence into equal-length chunks.…”
Section: Literature Reviewmentioning
confidence: 99%