Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1049
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Learning of Sentence Embeddings Using Compositional n-Gram Features

Abstract: The recent tremendous success of unsupervised word embeddings in a multitude of applications raises the obvious question if similar methods could be derived to improve embeddings (i.e. semantic representations) of word sequences as well. We present a simple but efficient unsupervised objective to train distributed representations of sentences. Our method outperforms the state-of-the-art unsupervised models on most benchmark tasks, highlighting the robustness of the produced general-purpose sentence embeddings.

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
403
0
2

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 494 publications
(407 citation statements)
references
References 30 publications
2
403
0
2
Order By: Relevance
“…However, word embeddings came to the foreground by Mikolov, Chen, Corrado, and Dean (), who presented the popular Continuous Bag‐of‐Words model (CBOW) and the Continuous Skip‐gram model. Additionally, sentence embeddings (Doc2Vec (Lau & Baldwin, ) or Sent2vec (Pagliardini, Gupta, & Jaggi, )) as well as the popular GloVe (Global Vectors) (Pennington, Socher, & Manning, ) method are utilized by keyphrase extraction methods.…”
Section: Unsupervised Methodsmentioning
confidence: 99%
“…However, word embeddings came to the foreground by Mikolov, Chen, Corrado, and Dean (), who presented the popular Continuous Bag‐of‐Words model (CBOW) and the Continuous Skip‐gram model. Additionally, sentence embeddings (Doc2Vec (Lau & Baldwin, ) or Sent2vec (Pagliardini, Gupta, & Jaggi, )) as well as the popular GloVe (Global Vectors) (Pennington, Socher, & Manning, ) method are utilized by keyphrase extraction methods.…”
Section: Unsupervised Methodsmentioning
confidence: 99%
“…7 Yin et al (2016) [23] 78.9 84. 8 Pagliardini et al (2018) [43] 76.4 83. 4 Ferreira et al (2018) [44] 74.08 83.…”
Section: Msrp Datasetmentioning
confidence: 98%
“…The classifiers tried to predict the respective labels from the text of the tweet alone. In the process, we analyzed the performance of four different classifier models: Bag of Words, Sent2Vec sentence embeddings [38] coupled with Support Vector Machines (SVMs) [39], FastText [40], and BERT [34]. The tokenization and word character encoding process was different for each model class.…”
Section: Trainingmentioning
confidence: 99%