2021
DOI: 10.1002/asi.24493
|View full text |Cite
|
Sign up to set email alerts
|

Lexical data augmentation for sentiment analysis

Abstract: Machine learning methods, especially deep learning models, have achieved impressive performance in various natural language processing tasks including sentiment analysis. However, deep learning models are more demanding for training data. Data augmentation techniques are widely used to generate new instances based on modifications to existing data or relying on external knowledge bases to address annotated data scarcity, which hinders the full potential of machine learning techniques. This paper presents our w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0
2

Year Published

2022
2022
2024
2024

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 25 publications
(18 citation statements)
references
References 34 publications
0
16
0
2
Order By: Relevance
“…It has undergone instruction fine-tuning on the basis of LoRA (Hu et al, 2021), and the constructed prompt's instruction is "Please determine whether the following content expresses a positive sentiment, and output 0 or 1 -->," with specific content included as input, corresponding labels as output. In contrast, the comparative methods encompass the commonly used techniques such as DICT (Zhang et al, 2015), EDA (Wei & Zou, 2019), and PLSDA (Xiang et al, 2021). The metrics is the accuracy (%).…”
Section: Experiments Settingmentioning
confidence: 99%
See 2 more Smart Citations
“…It has undergone instruction fine-tuning on the basis of LoRA (Hu et al, 2021), and the constructed prompt's instruction is "Please determine whether the following content expresses a positive sentiment, and output 0 or 1 -->," with specific content included as input, corresponding labels as output. In contrast, the comparative methods encompass the commonly used techniques such as DICT (Zhang et al, 2015), EDA (Wei & Zou, 2019), and PLSDA (Xiang et al, 2021). The metrics is the accuracy (%).…”
Section: Experiments Settingmentioning
confidence: 99%
“…The manual aggregation of domain‐specific training datasets can be a labor‐intensive and costly endeavor. This challenge underscores the critical need for an exploration into data augmentation strategies, which has secured a broad footprint in the realm of image processing, typically implemented through transformations (Alqudah et al, 2023; Niu et al, 2023; Shorten & Khoshgoftaar, 2019; Xiang et al, 2021).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The UDA approach offers a new way of effectively detecting and utilizing noise in unlabeled data (Xie et al, 2020). It unifies the learning process between labeled and unlabeled data through a specific form of contrastive learning and corresponding data augmentation (Xiang et al, 2021), and achieves superior results in image and regular text classification (Xie et al, 2020). The potential of UDA for short BSCs is worth investigating.…”
Section: Semi-supervised Learningmentioning
confidence: 99%
“…Thus many methods have been tried out in research so far. Among them are methods for swapping [ 59 ], deleting [ 16 , 38 ], inducing spelling mistakes [ 6 , 10 ], paraphrasing [ 28 ], and replacing of synonyms [ 25 , 61 , 66 ], close embeddings [ 2 , 58 ] and words predicted by a language model [ 11 , 18 , 24 ] on word-level. On a broader level, methods which change the dependency tree [ 45 , 62 ], perform round-trip-translation [ 27 , 47 ], or interpolate the input instances [ 9 , 65 ] are used.…”
Section: Related Workmentioning
confidence: 99%