2021
DOI: 10.1609/aaai.v35i8.16857
|View full text |Cite
|
Sign up to set email alerts
|

Extending Multi-Sense Word Embedding to Phrases and Sentences for Unsupervised Semantic Applications

Abstract: Most unsupervised NLP models represent each word with a single point or single region in semantic space, while the existing multi-sense word embeddings cannot represent longer word sequences like phrases or sentences. We propose a novel embedding method for a text sequence (a phrase or a sentence) where each sequence is represented by a distinct set of multi-mode codebook embeddings to capture different semantic facets of its meaning. The codebook embeddings can be viewed as the cluster centers which summarize… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 44 publications
0
1
0
Order By: Relevance
“…This particular framework provides context representations without ignoring word order information or long-term dependencies. Chang et al 9 outlined a novel embedding method for a text sequence (i.e., a phrase or a sentence), in which the sequence is represented by a set of multi-mode codebook embeddings intended to capture different semantic facets. Manchanda et al 10 extended the skip-gram model by clustering the occurrences of the multi-sense words and accounting for their diversity in the contexts of Word2Vec to obtain accurate and efficient vector representations for each sense.…”
Section: Related Workmentioning
confidence: 99%
“…This particular framework provides context representations without ignoring word order information or long-term dependencies. Chang et al 9 outlined a novel embedding method for a text sequence (i.e., a phrase or a sentence), in which the sequence is represented by a set of multi-mode codebook embeddings intended to capture different semantic facets. Manchanda et al 10 extended the skip-gram model by clustering the occurrences of the multi-sense words and accounting for their diversity in the contexts of Word2Vec to obtain accurate and efficient vector representations for each sense.…”
Section: Related Workmentioning
confidence: 99%