Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1098
|View full text |Cite
|
Sign up to set email alerts
|

Better Word Embeddings by Disentangling Contextual n-Gram Information

Abstract: Pre-trained word vectors are ubiquitous in Natural Language Processing applications. In this paper, we show how training word embeddings jointly with bigram and even trigram embeddings, results in improved unigram embeddings. We claim that training word embeddings along with higher n-gram embeddings helps in the removal of the contextual information from the unigrams, resulting in better stand-alone word embeddings. We empirically show the validity of our hypothesis by outperforming other competing word repres… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
26
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 28 publications
(26 citation statements)
references
References 11 publications
0
26
0
Order By: Relevance
“…From the processed data, uni-, bi-, and trigrams are jointly trained in deriving our custom word embeddings. Previous work suggests that multigram can improve on the quality of the obtained word embeddings (Gupta et al, 2019). In addition to word embeddings, we also derive document embeddings based on the cleaned text using pre-trained contextual multilingual universal sentence encoders.…”
Section: Text Mining Pipelinementioning
confidence: 99%
“…From the processed data, uni-, bi-, and trigrams are jointly trained in deriving our custom word embeddings. Previous work suggests that multigram can improve on the quality of the obtained word embeddings (Gupta et al, 2019). In addition to word embeddings, we also derive document embeddings based on the cleaned text using pre-trained contextual multilingual universal sentence encoders.…”
Section: Text Mining Pipelinementioning
confidence: 99%
“…Research works have been carried out to model the order of words when learning the distributed sentence representation (Le and Mikolov, 2014;Kiros et al, 2015;Conneau et al, 2017;Pagliardini et al, 2018;Gupta et al, 2019;Shen et al, 2019). Le and Mikolov propose Doc2vec (Le and Mikolov, 2014) to add a paragraph vector to represent the missing information from the current context.…”
Section: Sentence Embeddingmentioning
confidence: 99%
“…Sent2Vec (Pagliardini et al, 2018) aims to strike a balance between matrix factorization and deep learning. Gupta et al (2019) propose two modifications of Word2vec by considering higher-order word n-grams along with uni-gram during training. Shen et al (2019) use InferSent (Conneau et al, 2017) for sentence embeddings based on word vectors learned by Glove (Pennington et al, 2014) or FastText (Joulin et al, 2017).…”
Section: Sentence Embeddingmentioning
confidence: 99%
See 2 more Smart Citations