2009
DOI: 10.1016/j.neucom.2008.06.011
|View full text |Cite
|
Sign up to set email alerts
|

A density-based method for adaptive LDA model selection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
409
1
7

Year Published

2018
2018
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 624 publications
(422 citation statements)
references
References 8 publications
5
409
1
7
Order By: Relevance
“…In the resulting scree plot, the log-likelihood will initially increase as a function of T, flattens at optimal models, and may decrease after that, indicating a large number of topics (Griffiths & Steyvers, 2004;Kosinski et al, 2016). Cao et al (2009) developed a density-based method for adaptive LDA model selection. They started from the observation that when K is too small (i.e., with only a few topics), the discrimination between the topics is low, once there are words that overlap across topics.…”
Section: Selecting the Number Of Topics: Cross-validation Analysesmentioning
confidence: 99%
See 2 more Smart Citations
“…In the resulting scree plot, the log-likelihood will initially increase as a function of T, flattens at optimal models, and may decrease after that, indicating a large number of topics (Griffiths & Steyvers, 2004;Kosinski et al, 2016). Cao et al (2009) developed a density-based method for adaptive LDA model selection. They started from the observation that when K is too small (i.e., with only a few topics), the discrimination between the topics is low, once there are words that overlap across topics.…”
Section: Selecting the Number Of Topics: Cross-validation Analysesmentioning
confidence: 99%
“…The main feature of the approach of Cao et al (2009) is that it integrates a clustering process based on density, considering a topic as equivalent to a semantic cluster. The best model will have the largest possible intra-cluster similarity, which means that the cluster (i.e., the topic) includes coherent semantic content.…”
Section: Selecting the Number Of Topics: Cross-validation Analysesmentioning
confidence: 99%
See 1 more Smart Citation
“…The LDA uses the parameters shown in Table 2. In addition, several methods for selecting the optimal number of topics have already been proposed [23,24,[31][32][33]. We used the method Cao et al proposed, which is considered to be the most appropriate method for this study [31].…”
Section: Experiments Resultsmentioning
confidence: 99%
“…There are various opinions on how to determine the optimal number of topics [23,24,[30][31][32]. For our purposes, the topic is determined by the topic probability θ of each document.…”
Section: Componentmentioning
confidence: 99%