Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1210
|View full text |Cite
|
Sign up to set email alerts
|

Keep Meeting Summaries on Topic: Abstractive Multi-Modal Meeting Summarization

Abstract: Transcripts of natural, multi-person meetings differ significantly from documents like news articles, which can make Natural Language Generation models generate unfocused summaries. We develop an abstractive meeting summarizer from both videos and audios of meeting recordings. Specifically, we propose a multi-modal hierarchical attention mechanism across three levels: topic segment, utterance and word. To narrow down the focus into topically-relevant segments, we jointly model topic segmentation and summarizat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
41
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 92 publications
(42 citation statements)
references
References 17 publications
1
41
0
Order By: Relevance
“…Additional gains can be obtained with transfer learning (TL) where the model was initialized using the CNN/DailyMail data. The hierarchical models achieve higher ROUGE scores, consistent with [12]. Table 2: ROUGE F1 on the AMI test set -Baseline Models.…”
Section: Resultssupporting
confidence: 64%
See 4 more Smart Citations
“…Additional gains can be obtained with transfer learning (TL) where the model was initialized using the CNN/DailyMail data. The hierarchical models achieve higher ROUGE scores, consistent with [12]. Table 2: ROUGE F1 on the AMI test set -Baseline Models.…”
Section: Resultssupporting
confidence: 64%
“…This relatively small drop is likely because even at 30% WER, the sentence/utterance embedding similarity between a manual source and an ASR source is about 0.70-0.85% [33,34]. This system achieves higher all ROUGE measures than extractive method CoreRank [6], and when compared to abstractive method Top-icSeg (without visual signals) [12] our system achieves higher ROUGE-2 and ROUGE-L although lower ROUGE-1. Furthermore, when using transcripts with lower WER (A2), ROUGE scores are closer to those obtained from the manual transcripts, and yield higher ROUGE-2 and ROUGE-L scores than the state-of-the-art multi-modal TopicSeg+VFOA.…”
Section: Modelmentioning
confidence: 76%
See 3 more Smart Citations