Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.181
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Conversation Disentanglement through Co-Training

Abstract: Conversation disentanglement aims to separate intermingled messages into detached sessions, which is a fundamental task in understanding multi-party conversations. Existing work on conversation disentanglement relies heavily upon human-annotated datasets, which are expensive to obtain in practice. In this work, we explore to train a conversation disentanglement model without referencing any human annotations. Our method is built upon a deep co-training algorithm, which consists of two neural networks: a messag… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(14 citation statements)
references
References 28 publications
0
14
0
Order By: Relevance
“…Disentanglement is worthy of study. Decoupling messages or clustering conversation threads help with screening concerned parts among contexts, therefore it may be naturally required by passage comprehension, and related downstream dialogue tasks (Elsner and Charniak, 2010;Jia et al, 2020;Liu et al, 2021a), such as response selection, question-answering, etc.…”
Section: … …mentioning
confidence: 99%
See 2 more Smart Citations
“…Disentanglement is worthy of study. Decoupling messages or clustering conversation threads help with screening concerned parts among contexts, therefore it may be naturally required by passage comprehension, and related downstream dialogue tasks (Elsner and Charniak, 2010;Jia et al, 2020;Liu et al, 2021a), such as response selection, question-answering, etc.…”
Section: … …mentioning
confidence: 99%
“…Earlier works mainly depend on feature engineering Elsner and Charniak, 2010;Yu and Joty, 2020), and use well-constructed handcrafted features to train a naive classifier (Elsner and Charniak, 2010) or linear feed-forward network . Recent works are mostly based on two strategies: 1) two-step (Mehri and Carenini, 2017;Yu and Joty, 2020;Li et al, 2020b;Liu et al, 2021a) and 2) end-to-end (Tan et al, 2019;. In terms of the two-step method, the disentanglement task is divided into matching and clustering.…”
Section: … …mentioning
confidence: 99%
See 1 more Smart Citation
“…Due to the aforementioned features of dialogue development, there are always multiple ongoing conversation threads developing in a dialogue history simultaneously, which cause troubles for both human and machine to understand dialogue context or further deal with various reading comprehension tasks (Kummerfeld et al, 2019;Elsner and Charniak, 2010;Joty et al, 2019;Shen et al, 2006;Jiang et al, 2018Jiang et al, , 2021. Therefore, to some extent, disentangling context or clustering conversation threads can make an effective prerequisite for downstream tasks on dialogues (Elsner and Charniak, 2010;Liu et al, 2021a;Jia et al, 2020), as which contributes to screening concerned parts for further machine reading comprehension (MRC) applications.…”
Section: Introductionmentioning
confidence: 99%
“…Earlier works mainly depend on feature engineering (Kummerfeld et al, 2019;Elsner and Charniak, 2010;Yu and Joty, 2020), and use well-constructed handcrafted features to train a naive classifier (Elsner and Charniak, 2010) or linear feed-forward network (Kummerfeld et al, 2019). Recent works mainly based on two strategies: 1) two-step (Mehri and Carenini, 2017;Yu and Joty, 2020;Li et al, 2020c;Liu et al, 2021a) and 2) end-to-end (Tan et al, 2019;. In the two-step method, disentanglement task is divided into matching and clustering, which means firstly matching utterance pairs to detect reply-to relations and then dividing utterances into clusters according to the matching score.…”
Section: Introductionmentioning
confidence: 99%