Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing 2022
DOI: 10.18653/v1/2022.emnlp-main.176
|View full text |Cite
|
Sign up to set email alerts
|

Mitigating Data Sparsity for Short Text Topic Modeling by Topic-Semantic Contrastive Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…This could be achieved by training a decoder for summarization, generating a summarization for each topic, and subsequently concatenating them. This framework can also be extended to hierarchical topic modeling (Chen et al, 2023;Shahid et al, 2023;Eshima and Mochihashi, 2023), mitigate data sparsity for short text topic modeling (Wu et al, 2022), generate topic-relevant and coherent long texts (Yang et al, 2022), and construct a network of topics together with meaningful relationships between them (Byrne et al, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…This could be achieved by training a decoder for summarization, generating a summarization for each topic, and subsequently concatenating them. This framework can also be extended to hierarchical topic modeling (Chen et al, 2023;Shahid et al, 2023;Eshima and Mochihashi, 2023), mitigate data sparsity for short text topic modeling (Wu et al, 2022), generate topic-relevant and coherent long texts (Yang et al, 2022), and construct a network of topics together with meaningful relationships between them (Byrne et al, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…To enhance text representations through self-supervision in an unsupervised environment, some researchers applied contrastive learning to pre-trained language models like BERT (Kim et al, 2021;Yan et al, 2021). Wu et al (2022) proposed a topic-semantic contrastive topic model (TSCTM) to solve the issue of data sparsity.…”
Section: Contrastive Learningmentioning
confidence: 99%
“…propose the contrastive learning on doc-topic distributions where they build positive and negative pairs by sampling salient words of documents. Differently, Wu et al (2022) directly sample positive and negative pairs based on the topic semantics of documents to capture relations among samples. Specifically, they quantize doc-topic distributions following Wu et al (2020b) and then sample documents with the same quantization indices as positive pairs and different indices as negative pairs.…”
Section: Ntms With Contrastive Learningmentioning
confidence: 99%
“…They learn entity vector representations from manually edited knowledge graphs. Based on Wu et al (2020b), Wu et al (2022) further propose a contrastive learning method according to the topic semantics of short texts, which better captures the similarity relations among them. This refines the representations of short texts and thus their doctopic distributions.…”
Section: Short Text Ntmsmentioning
confidence: 99%