2018
DOI: 10.1093/comjnl/bxy037
|View full text |Cite
|
Sign up to set email alerts
|

Relational Biterm Topic Model: Short-Text Topic Modeling using Word Embeddings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 34 publications
(18 citation statements)
references
References 5 publications
0
18
0
Order By: Relevance
“…The application of topic models builds on the identification of coherently recurring word patterns (topics) within documents (Blei, 2012); short documents, however, often feature an insufficient frequency of word co-occurrences so that the topic modeling algorithm is unable to detect coherent patterns. This so-called "sparsity problem" leads to low quality topic models (Hong & Davison, 2010;Li et al, 2019). While a growing branch of research is concerned with developing new specialized models for collections of short texts (see e.g., Li et al, 2019), 9 a simple and established approach to address the sparsity-problem is to concatenate several short documents to a longer one (Guo et al, 2016).…”
Section: Topic Modelingmentioning
confidence: 99%
See 1 more Smart Citation
“…The application of topic models builds on the identification of coherently recurring word patterns (topics) within documents (Blei, 2012); short documents, however, often feature an insufficient frequency of word co-occurrences so that the topic modeling algorithm is unable to detect coherent patterns. This so-called "sparsity problem" leads to low quality topic models (Hong & Davison, 2010;Li et al, 2019). While a growing branch of research is concerned with developing new specialized models for collections of short texts (see e.g., Li et al, 2019), 9 a simple and established approach to address the sparsity-problem is to concatenate several short documents to a longer one (Guo et al, 2016).…”
Section: Topic Modelingmentioning
confidence: 99%
“…This so-called "sparsity problem" leads to low quality topic models (Hong & Davison, 2010;Li et al, 2019). While a growing branch of research is concerned with developing new specialized models for collections of short texts (see e.g., Li et al, 2019), 9 a simple and established approach to address the sparsity-problem is to concatenate several short documents to a longer one (Guo et al, 2016). While this artificial document concatenation distorts the structure of the original data, the method leads to a significant increase of topic coherence in the resulting models (Steinskog et al, 2017).…”
Section: Topic Modelingmentioning
confidence: 99%
“…This model tries to infer the topics in a corpus and the joint probability distribution for LDA can be factored as P( w,z ) = P( w|z )× P( z|r ). The success of LDA has motivated numerous models to extend it in order to analyse opinions (Blei, Ng, & Jordan, ; Koller & Friedman, ; Duan, Li, Li, Lu, & Wen, ; Chen, Jose, Yu, & Yuan, ; Li et al, ; Yang, Ma, Silva, Liu, & Hua, 2015; Yang et al, ; Li, Ouyang, & Zhou, ; Chien, Lee, & Tan, ).…”
Section: Related Workmentioning
confidence: 99%
“…In addition to the aforementioned aggregation or pooling based models, another popular research direction for short text topic modelling is using word correlations or embedding to enhance topic models. For example, Biterm Topic Model (BTM) (Yan et al, 2013) and Relational BTM (Li et al, 2018b) enrich each document with word pairs (i.e., biterms). Instead of using all the word pairs, Yang et al (2015a) considers pre-trained phrases.…”
Section: Related Workmentioning
confidence: 99%