2022
DOI: 10.1609/aaai.v36i10.21432
|View full text |Cite
|
Sign up to set email alerts
|

DialogLM: Pre-trained Model for Long Dialogue Understanding and Summarization

Abstract: Dialogue is an essential part of human communication and cooperation. Existing research mainly focuses on short dialogue scenarios in a one-on-one fashion. However, multi-person interactions in the real world, such as meetings or interviews, are frequently over a few thousand words. There is still a lack of corresponding research and powerful tools to understand and process such long dialogues. Therefore, in this work, we present a pre-training framework for long dialogue understanding and summarization. Consi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 53 publications
(31 citation statements)
references
References 43 publications
0
31
0
Order By: Relevance
“…Since the introduction of BERT (Devlin et al, 2019), the research community has witnessed remarkable progress in the field of language model pre-training on a large amount of free text. Such advancements have led to significant progresses in a wide range of natural language understanding (NLU) tasks Yang et al, 2019;Clark et al, 2020;Lan et al, 2021) and text generation tasks (Radford et al, 2019;Lewis et al, 2020;Raffel et al, 2020;Su et al, 2021a,e,g,d,f,c;Zhong et al, 2021) Contrastive Learning. Generally, contrastive learning methods distinguish observed data points from fictitious negative samples.…”
Section: B Related Workmentioning
confidence: 99%
“…Since the introduction of BERT (Devlin et al, 2019), the research community has witnessed remarkable progress in the field of language model pre-training on a large amount of free text. Such advancements have led to significant progresses in a wide range of natural language understanding (NLU) tasks Yang et al, 2019;Clark et al, 2020;Lan et al, 2021) and text generation tasks (Radford et al, 2019;Lewis et al, 2020;Raffel et al, 2020;Su et al, 2021a,e,g,d,f,c;Zhong et al, 2021) Contrastive Learning. Generally, contrastive learning methods distinguish observed data points from fictitious negative samples.…”
Section: B Related Workmentioning
confidence: 99%
“…Recent work on QMSum has introduced taskspecific denoising objectives for meeting summarization (Zhong et al, 2021a), generated final fine-grained summaries based on multiple coarsegrained steps (Zhang et al, 2021a), and treated the extractive text of an extractive-abstractive model as a latent variable (Mao et al, 2021). Zhang et al (2021b) analyze the challenges of long dialogue summarization such as the input length, the role of queries, and domain adaptation.…”
Section: Query-focused Summarizationmentioning
confidence: 99%
“…For instance, BART proposes a full-text denoising pre-training objective for seq2seq models, whose cost is prohibitive for long text, and not tailored to dialogues. Zhong et al (2022) present the DialogLM model, along with a dialoguededicated, window-based denoising pretraining approach. Windows of consecutive utterances are first selected from the conversation, which is then disrupted with arbitrary dialogue-related noises, e.g., speaker mask, turn splitting, turn merging, text infilling and turn permutation.…”
Section: Domain Adaptationmentioning
confidence: 99%
“…As mentioned in Subsection 4.3, pretrained on free-form text makes these models difficult to process specifically structured dialogue transcriptions. Future research could follow the work of, notably DialoGPT (Zhang et al, 2020) and Di-alogLM (Zhong et al, 2022), to develop better languge models dedicated to spoken language, which will potentially offer large gains in task performance of abstractive meeting summarization. Commonsense incorporation.…”
Section: Future Directionsmentioning
confidence: 99%