2019
DOI: 10.1016/j.ins.2019.07.048
|View full text |Cite
|
Sign up to set email alerts
|

A nonparametric model for online topic discovery with word embeddings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
15
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 54 publications
(15 citation statements)
references
References 10 publications
0
15
0
Order By: Relevance
“…Thereafter, MStreamF (Yin et al, 2018) was thus proposed by incorporating a forgetting mechanism to cope with cluster evolution, and allows processing each batch only one time. The NPMM model (Chen et al, 2019) was recently introduced by using the word-embeddings to eliminate a cluster generating parameter of the model.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Thereafter, MStreamF (Yin et al, 2018) was thus proposed by incorporating a forgetting mechanism to cope with cluster evolution, and allows processing each batch only one time. The NPMM model (Chen et al, 2019) was recently introduced by using the word-embeddings to eliminate a cluster generating parameter of the model.…”
Section: Related Workmentioning
confidence: 99%
“…To evaluate the performance of the proposed algorithm, we conduct experiments on three real and two synthetic datasets. These datasets were also used in (Yin and Wang, 2016a;Qiang et al, 2018;Yin et al, 2018;Jia et al, 2018;Chen et al, 2019) to evaluate short text clustering models. In the preprocessing step, we removed stop words, converted all text into lowercase, and stemming.…”
Section: Datasets and Evaluation Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…Besides some other detection methods-including probability graph methods, deep learning methods, representation learning methods, and so on-are applied to perform event detection. For example, Chen et al 17 proposed a nonparametric model for online topic discovery with word embeddings, where they mainly exploited auxiliary word embeddings to infer the topic number and employed a ''spike and slab'' function to alleviate the sparsity problem of topic-word distributions in online short text analyses. Hu et al 18 presented a transformation-gated long short-term memory (LSTM) to enhance the ability of capturing short-term mutation information, where the function of transformation gate, input gate information, and the value range of the partial derivative corresponding to the transformation gate were studied.…”
Section: Related Workmentioning
confidence: 99%
“…Subsequent works are mainly dedicated to designing elaborate features to improve detection performance. For example, textual features like psychological and linguistic clues [3], syntactic stylometry [4], review topic [5], [6], and behavioral features like rating deviation [7], [8] are explored in many works. However, designing effective features is usually time-consuming [9] and heavily rely on expert knowledge in particular areas.…”
Section: Introductionmentioning
confidence: 99%