2020
DOI: 10.1109/access.2020.2997973
|View full text |Cite
|
Sign up to set email alerts
|

Collaboratively Modeling and Embedding of Latent Topics for Short Texts

Abstract: Deriving a successful document representation is the critical challenge in many downstream tasks in NLP, especially when documents are very short. It is challenging to handle the sparsity and the noise problems confronting short texts. Some approaches employ latent topic models, based on global word co-occurrence, to obtain topic distribution as the representation. Others leverage word embeddings, which consider local conditional dependencies, to map a document as a summation vector of them. Unlike the existin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 33 publications
0
9
0
Order By: Relevance
“…These models seem to outperform both the DMM and individual PDMM methods but also involve high computation costs. A new model called a Collaboratively Modeling and Embedding DMM (CME-DMM) was proposed by Liu [67] for capturing coherent hidden topics from STs. All these models were suggested for topic modeling over short text.…”
Section: Short Text Topic Modellingmentioning
confidence: 99%
“…These models seem to outperform both the DMM and individual PDMM methods but also involve high computation costs. A new model called a Collaboratively Modeling and Embedding DMM (CME-DMM) was proposed by Liu [67] for capturing coherent hidden topics from STs. All these models were suggested for topic modeling over short text.…”
Section: Short Text Topic Modellingmentioning
confidence: 99%
“…Liu et al [22] proposed Collaboratively Modeling and Embedding -Dirichlet Multinomial Mixtures (CME-DMM) to apply topic modeling on short texts that were collected from a social media platform. Liu identified major challenges in modeling latent topics in short texts, including the inadequacy of word co-occurrence instances, the need to capture local contextual dependencies in noisy short texts due to their limited length, and maintaining consistency between the latent topic models based on word co-occcurrences and their representation based on local contextual dependencies.…”
Section: Related Workmentioning
confidence: 99%
“…Topic modeling is well known for text mining researchers in Natural Language Processing (NLP), which aims to extract and interpret knowledge from data [13]. A popular method, such as Latent Dirichlet Allocation (LDA) has been proven beneficial in finding useful words that are associated with topics [13].…”
Section: B Topic Modelingmentioning
confidence: 99%