2021 International Conference on Advanced Computer Science and Information Systems (ICACSIS) 2021
DOI: 10.1109/icacsis53237.2021.9631322
|View full text |Cite
|
Sign up to set email alerts
|

Topic Modeling for Customer Service Chats

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 9 publications
0
5
0
Order By: Relevance
“…Hence, Top2Vec can scale a large number of topics and vast quantities of data. Such strength is especially required when multiple languages emerge within a corpus (Hendry et al, 2021). The main disadvantage of Top2Vec, however, is that it is unqualified to work with a small amount of data (Abuzayed and Al-Khalifa, 2021; e.g., <1,000 documents).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Hence, Top2Vec can scale a large number of topics and vast quantities of data. Such strength is especially required when multiple languages emerge within a corpus (Hendry et al, 2021). The main disadvantage of Top2Vec, however, is that it is unqualified to work with a small amount of data (Abuzayed and Al-Khalifa, 2021; e.g., <1,000 documents).…”
Section: Discussionmentioning
confidence: 99%
“…In this study, a pretrained embedding models, the Universal Sentence Encoder, was used to create word and document embeddings. Since word vectors that emerge closest to the document vectors seem to best describe the topic of the document, the number of documents that can be grouped together represents the number of topics (Hendry et al, 2021).…”
Section: Model 3: Top2vecmentioning
confidence: 99%
“…Dirichlet Allocation, LDA) 기법이 위의 선행연구를 포함해 주로 적용되어왔으나, 최근 제안된 BERTopic 기법이 다른 기 법들에 비해 성능이 월등하게 우수한 것이 알려지면서 많은 관심을 받고 있다 (Abuzayed and Al-Khalifa, 2021;Egger and Yu, 2022;Hendry et al, 2021;Grootendorst, 2022). 1).…”
Section: 토픽 모델링의 기법으로는 잠재 디리클레 할당(Latentunclassified
“…In this research, pre-trained embedding models, namely, All-MiniLM-L6-v2, Doc2Vec, and Universal Sentence Encoder, were compared based on the coherence score, and word and document embeddings were generated. Since word vectors that are closely aligned with document vectors tend to provide the most accurate representation of a document's underlying topic, the number of documents that can be grouped together indicates the number of distinct topics (Hendry et al, 2021 ).…”
Section: Methodsmentioning
confidence: 99%