Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1056
|View full text |Cite
|
Sign up to set email alerts
|

Refining Word Embeddings for Sentiment Analysis

Abstract: Word embeddings that can capture semantic and syntactic information from contexts have been extensively used for various natural language processing tasks. However, existing methods for learning contextbased word embeddings typically fail to capture sufficient sentiment information. This may result in words with similar vector representations having an opposite sentiment polarity (e.g., good and bad), thus degrading sentiment analysis performance. Therefore, this study proposes a word vector refinement model t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
69
0
1

Year Published

2017
2017
2021
2021

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 148 publications
(72 citation statements)
references
References 18 publications
2
69
0
1
Order By: Relevance
“…Past studies have tried to incorporate sentiment during the training process of the embedding, 17,18 concatenation of pretrained embedding with additional linguistic features, 19 and refinement of the pretrained embedding. 20 Here we incorporated a polarity one-dimensional vector (Fig 1B). We built the dictionary on the basis of a previous lexicon with known sentiments 21 and manually added the words “plus” and “minus.” These added words do not exist in our medical data set and were later used to validate our out-of-vocabulary predictions.…”
Section: Methodsmentioning
confidence: 99%
“…Past studies have tried to incorporate sentiment during the training process of the embedding, 17,18 concatenation of pretrained embedding with additional linguistic features, 19 and refinement of the pretrained embedding. 20 Here we incorporated a polarity one-dimensional vector (Fig 1B). We built the dictionary on the basis of a previous lexicon with known sentiments 21 and manually added the words “plus” and “minus.” These added words do not exist in our medical data set and were later used to validate our out-of-vocabulary predictions.…”
Section: Methodsmentioning
confidence: 99%
“…Compared with GloVe+DeepMoji, GloVe+Emo2Vec achieves same or better results on 11/14 datasets, which on average gives 1.0% improvement. GloVe+Emo2Vec achieves better performances on SOTA results on three datasets (SE0714, stress and tube tablet) and comparable result to SOTA on dataset Previous SOTA results GloVe GloVe+DeepMoji GloVe+Emo2Vec SS-Twitter bi-LSTM (Felbo et al, 2017) 0.88 0.78 0.81 0.81 SS-Youtube bi-LSTM (Felbo et al, 2017) 0.93 0.84 0.86 0.87 SS-binary bi-LSTM (Yu et al, 2017) another four datasets (tube auto, SemEval, SCv1-GEN and SCv2-GEN). We believe the reason why we achieve a much better performance than SOTA on the SE0714 is that headlines are usually short and emotional words exist more commonly in headlines.…”
Section: Resultsmentioning
confidence: 80%
“…There are recent studies that aims to project not only semantic and syntactic but also sentiment content of text before creating a model [15], [16]. [17] emphasizes the same problem with an approach distinctly using existing word embedding model.…”
Section: Training Word2vec Model Resultsmentioning
confidence: 99%