2019
DOI: 10.1016/j.knosys.2018.08.011
|View full text |Cite
|
Sign up to set email alerts
|

Experimental explorations on short text topic mining between LDA and NMF based Schemes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
122
0
5

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 174 publications
(128 citation statements)
references
References 23 publications
1
122
0
5
Order By: Relevance
“…It can be used to automatically extract valuable hidden elements. Conventional topic modeling techniques such as probabilistic latent semantic analysis (PLSA) and latent Dirichlet allocation (LDA) are widely used to infer latent topical structures from documents [9], [14], [20], [23].…”
Section: B Topic Modeling Based On Contentsmentioning
confidence: 99%
“…It can be used to automatically extract valuable hidden elements. Conventional topic modeling techniques such as probabilistic latent semantic analysis (PLSA) and latent Dirichlet allocation (LDA) are widely used to infer latent topical structures from documents [9], [14], [20], [23].…”
Section: B Topic Modeling Based On Contentsmentioning
confidence: 99%
“…Vector representation is transforming meaningful words with context terms represented as number, where each word is distributed with weights [26]. Various forms of extracting words are: Frequency based Embedding [27], Prediction based Embedding [28], Non-Negative Matrix Factorization [29]. TF-IDF (Term-Frequency -Inverse document frequency) [30], Count Vector, Co-occurrence vector and document frequency.…”
Section: Feature Extraction and Selectionmentioning
confidence: 99%
“…Google News - [22], [66], [80] Snippets 8 [8], [29], [36], [44], [45], [49][50][51][52], [68], [74], [75] NIPS - [38] DBLP 6 [31], [37], [62] Yahoo! Answers 11 [38], [45], [77], [82] Online News 7 [26][27][28][29], [31], [32], [37], [47], [62], [68], [81] Baidu Q & A 35 [49], [52], [37], [81][35] [15], [39], [75] Satck Overflow Q & A - [29], [44], [47] Tweets {August to Oct 2008}…”
Section: Short Textmentioning
confidence: 99%
“…As we shall see, we propose a new topic extraction strategy to mitigate these problems. In [32] the authors proposed a novel model called ''knowledge-guided nonnegative matrix factorization for better short text topic mining'' (abbreviated as KGNMF), KGNMF integrated the word-word semantic graph regularization which can be learned from external knowledge base e.g., Wikipedia into the basic NMF for improvement in short text topic learning. In [33] the authors developed a general non-probabilistic topic modeling framework.…”
Section: Latent Topic Decompositionmentioning
confidence: 99%
“…Due to the semantic text representation approach specially WSD in addition the semantic similarity, the conventional topic models are not effectively to capture semantic topics, which leads to the poor performance in [33] X √ Schneider et al [22] X X Zhao et al [23] √ X Chen et al [32] X √ Hong et al [34] X X Peng et al [47] X X Xu et al [26] X X Li et al [24] √ X Allahyari et al [25] √ X Izquierdo et al [48] √ X Viegas et al [35] √ (word embedding) √ semantic topic learning. To tackle this problem, we introduce semantic topic modelling SNNMF.…”
Section: Problem Statementmentioning
confidence: 99%