2020
DOI: 10.3233/faia200602
|View full text |Cite
|
Sign up to set email alerts
|

Pretraining and Fine-Tuning Strategies for Sentiment Analysis of Latvian Tweets

Abstract: In this paper, we present various pre-training strategies that aid in improving the accuracy of the sentiment classification task. At first, we pre-train language representation models using these strategies and then fine-tune them on the downstream task. Experimental results on a time-balanced tweet evaluation set show the improvement over the previous technique. We achieve 76% accuracy for sentiment analysis on Latvian tweets, which is a substantial improvement over previous work.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…We adapted the sentiment analysis tool 1 used for general Latvian tweets [11] and used it along with a Latvian-specific language model [13], which we fine-tuned onto the entire LTEC. This allowed us to train a sentiment classifier to distinguish negative, positive and neutral tweets in the sentiment training data subset of LTEC other Latvian tweet data sets 2 [7,8,12].…”
Section: Methodsmentioning
confidence: 99%
“…We adapted the sentiment analysis tool 1 used for general Latvian tweets [11] and used it along with a Latvian-specific language model [13], which we fine-tuned onto the entire LTEC. This allowed us to train a sentiment classifier to distinguish negative, positive and neutral tweets in the sentiment training data subset of LTEC other Latvian tweet data sets 2 [7,8,12].…”
Section: Methodsmentioning
confidence: 99%
“…The most recent works proposed language models specifically pre-trained on tweet corpora: Thakkar and Pinnis [ 16 ] achieved encouraging performance leveraging a time-balanced evaluation set for sentiment analysis on Latvian tweets, comparing several BERT-based architectures, and Nguyen et al [ 12 ] presented BERTweet, the first public large-scale pre-trained language model for English tweets; Ángel González et al [ 15 ] proposed TWiLBERT, a specialization of the BERT architecture both for the Spanish language and the Twitter domain. For languages other than English, such as Persian [ 53 ] and Arabic [ 54 ], recent studies have also focused on deep neural networks such as CNN and LSTM.…”
Section: Background and Related Workmentioning
confidence: 99%
“…In the field of sentiment analysis of tweets, most of the scientific literature has obtained state-of-the-art results adopting the approach of training language models directly from scratch starting from corpora made up exclusively of tweets, so that the models could better handle the specific tweet jargon, characterized by a particular syntax and grammar not containing punctuation, with contracted or elongated words, keywords, hashtags, emoticons, emojis and so on. These approaches, working not only in English [ 11 , 12 ], but also in other languages such as Italian [ 13 ], Spanish [ 14 , 15 ], and Latvian [ 16 ], necessarily impose two constraints: the first requires the building of large corpora of tweets to be used for training the language models in the specific language considered, and the second is the need for substantial resources, of both hardware and time, to train the models from scratch starting from these corpora.…”
Section: Introductionmentioning
confidence: 99%
“…They used the Italian language as a case study but generally, their approach is applicable to different languages. Several papers have focused on pre-training on specific-language tweet corpora: Thakkar et al [16] presented different pre-training strategies for the sentiment classification task in Latvian; Jose Angel et al [17] proposed TWilBert, a BERT-based architecture that outperforms multilingual BERT on classification tasks; and Tuan Anh et al [18] used CNN and LSTM architecture for sentiment analysis of informal Indonesian tweets. As a summary of the limitations and difficulties of sentiment analysis, a multi-class classification approach for Twitter is proposed in [19].…”
Section: Introductionmentioning
confidence: 99%