Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1099
|View full text |Cite
|
Sign up to set email alerts
|

Lessons from the Bible on Modern Topics: Low-Resource Multilingual Topic Model Evaluation

Abstract: Multilingual topic models enable document analysis across languages through coherent multilingual summaries of the data. However, there is no standard and effective metric to evaluate the quality of multilingual topics. We introduce a new intrinsic evaluation of multilingual topic models that correlates well with human judgments of multilingual topic coherence as well as performance in downstream applications. Importantly, we also study evaluation for lowresource languages. Because standard metrics fail to acc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
8
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 32 publications
0
8
0
Order By: Relevance
“…We choose topic coherence (Hao, Boyd-Graber, and Paul 2018) and crosslingual document classification (Smet, Tang, and Moens 2011) as intrinsic and extrinsic evaluation tasks, respectively. The reason for choosing these two tasks is that they examine the models from different angles: Topic coherence looks at topic-word distributions, whereas classification focuses on document-topic distributions.…”
Section: Discussionmentioning
confidence: 99%
“…We choose topic coherence (Hao, Boyd-Graber, and Paul 2018) and crosslingual document classification (Smet, Tang, and Moens 2011) as intrinsic and extrinsic evaluation tasks, respectively. The reason for choosing these two tasks is that they examine the models from different angles: Topic coherence looks at topic-word distributions, whereas classification focuses on document-topic distributions.…”
Section: Discussionmentioning
confidence: 99%
“…Additionally, we propose a supervised topic quality estimator that by combining multiple metrics delivers even better results. For future work, we intend to work with larger datasets to investigate neural solutions to combine features from different metrics, as well as to apply our findings to other variants of LDA models trained on low-resource languages, where high-quality external corpora are usually not available (Hao et al, 2018).…”
Section: Discussionmentioning
confidence: 99%
“…We relied on the python scripts and instructions provided by Lau et al (2014) and chose a topic cardinality (i.e., number of top words) of 10 for the calculation. Following explanations by Hao et al (2018), NPMI scores are then first calculated individually for each language and averaged subsequently per model. The higher the NPMI value, the better.…”
Section: Coherence Metricsmentioning
confidence: 99%