2019
DOI: 10.1016/j.ipm.2019.102098
|View full text |Cite
|
Sign up to set email alerts
|

ATM: Adversarial-neural Topic Model

Abstract: Topic models are widely used for thematic structure discovery in text. But traditional topic models often require dedicated inference procedures for specific tasks at hand. Also, they are not designed to generate word-level semantic representations. To address the limitations, we propose a neural topic modeling approach based on the Generative Adversarial Nets (GANs), called Adversarial-neural Topic Model (ATM) in this paper. To our best knowledge, this work is the first attempt to use adversarial training for… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
2

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 72 publications
(38 citation statements)
references
References 23 publications
0
36
2
Order By: Relevance
“…To be best of our knowledge, W-LDA is the first topic model based on the WAE framework. Recently, Adversarial Topic model (ATM) (Wang et al, 2018) proposes using GAN with Dirichlet prior to learn topics. The generator takes in samples from Dirichlet distribution and maps to a document-word distribution layer to form the fake samples.…”
Section: Related Workmentioning
confidence: 99%
“…To be best of our knowledge, W-LDA is the first topic model based on the WAE framework. Recently, Adversarial Topic model (ATM) (Wang et al, 2018) proposes using GAN with Dirichlet prior to learn topics. The generator takes in samples from Dirichlet distribution and maps to a document-word distribution layer to form the fake samples.…”
Section: Related Workmentioning
confidence: 99%
“…We evaluate the performance of proposed models on four datasets: NYTimes 2 (NYT), Grolier 3 (GRL), DBpedia ontology classification dataset (DBP) (Zhang et al, 2015) and 20 Newsgroups 4 (20NG). For NYTimes and Grolier datasets, we use the processed version of (Wang et al, 2019a). For the DBpedia dataset, we first sample 100, 000 documents from the whole training set, and then perform preprocessing including tokenization, lemmatization, removal of stopwords, and low-frequency words.…”
Section: Methodsmentioning
confidence: 99%
“…Apart from VAE-based approaches, Adversarialneural Topic Model (ATM) (Wang et al, 2019a)) was proposed to model topics with GANs. The generator of ATM projects randomly sampled topic distributions to word distributions, and is adversarially trained with a discriminator that tries to distinguish real and generated word distributions.…”
Section: Neural Topic Modelingmentioning
confidence: 99%
See 2 more Smart Citations