Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.630
|View full text |Cite
|
Sign up to set email alerts
|

tBERT: Topic Models and BERT Joining Forces for Semantic Similarity Detection

Abstract: Semantic similarity detection is a fundamental task in natural language understanding. Adding topic information has been useful for previous feature-engineered semantic similarity models as well as neural models for other tasks. There is currently no standard way of combining topics with pretrained contextual representations such as BERT. We propose a novel topic-informed BERT-based architecture for pairwise semantic similarity detection and show that our model improves performance over strong neural baselines… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
45
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 119 publications
(46 citation statements)
references
References 28 publications
1
45
0
Order By: Relevance
“…In this study, we characterized the responses via feature extraction by employing a combination of the latent Dirichlet allocation model (LDA) and semantic representation transformers called Bidirectional Encoder Representations from Transformers (BERT), then grouped similar responses using k-means clustering. Based on studies demonstrating that this combination improved performance on numerous language analysis tasks (Rangrej et al, 2011;Peinelt et al, 2020;Xie et al, 2020), this approach allowed us to better reveal the most common types of silver linings in participants' responses.…”
Section: (B) Sentiment Analysismentioning
confidence: 99%
“…In this study, we characterized the responses via feature extraction by employing a combination of the latent Dirichlet allocation model (LDA) and semantic representation transformers called Bidirectional Encoder Representations from Transformers (BERT), then grouped similar responses using k-means clustering. Based on studies demonstrating that this combination improved performance on numerous language analysis tasks (Rangrej et al, 2011;Peinelt et al, 2020;Xie et al, 2020), this approach allowed us to better reveal the most common types of silver linings in participants' responses.…”
Section: (B) Sentiment Analysismentioning
confidence: 99%
“…Thereafter, we refer to this enhanced modeling method of PBERT as TPBERT, whose model structure is schematically depicted in Figure 2. Our TPBERT bears some resemblance to a recently proposed BERT-based model for semantic similarity detection [29].…”
Section: Incorporation Of Task-specific Topic Information Into Pbertmentioning
confidence: 80%
“…Due to the large number of parameters required by the BERT model and its advantages in local semantic representation, this model cannot represent the macro-domain information of the input document other than its own semantics. Inspired by reference [8] and reference [9], we propose a document semantic acquisition method based on tALBERT. We obtain the topic information at the word level and document level through an LDA topic model, obtain a semantic representation at the wordlevel and document-level through ALBERT, and fuse the above information through a concatenation mechanism to represent the document.…”
Section: Model Frameworkmentioning
confidence: 99%
“…In 2020, Lan et al [8] proposed "A Lite BERT" (ALBERT) model, which greatly simplifies the number of required parameters. Peinelt et al [9] combined a topic model with a BERT model for the task of semantic similarity detection. We have reason to use ALBERT and topic models to extract important information of different granularities form documents to further improve the effect of multi-label text classification.…”
Section: Introductionmentioning
confidence: 99%