Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1098
|View full text |Cite
|
Sign up to set email alerts
|

Improving the Similarity Measure of Determinantal Point Processes for Extractive Multi-Document Summarization

Abstract: The most important obstacles facing multidocument summarization include excessive redundancy in source descriptions and the looming shortage of training data. These obstacles prevent encoder-decoder models from being used directly, but optimization-based methods such as determinantal point processes (DPPs) are known to handle them well. In this paper we seek to strengthen a DPP-based method for extractive multi-document summarization by presenting a novel similarity measure inspired by capsule networks. The ap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
66
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 59 publications
(66 citation statements)
references
References 50 publications
0
66
0
Order By: Relevance
“…We are particularly interested in leveraging BERT for better sentence quality and diversity estimates. This paper extends on previous work (Cho et al, 2019) by incorporating deep contextualized representations into DPP, with an emphasis on better sentence selection for extractive multi-document summarization. The major research contributions of this work include the following: (i) we make a first attempt to combine DPP with BERT representations to measure sentence quality and diversity and report encouraging results on benchmark summarization datasets; (ii) our findings suggest that it is best to model sentence quality, i.e., how important a sentence is to the summary, by combining semantic representations and surface indicators of the sentence, whereas pairwise sentence dissimilarity can be determined by semantic repre-sentations only; (iii) our analysis reveals that combining contextualized representations with surface features (e.g., sentence length, position, centrality, etc) remains necessary, as deep representations, albeit powerful, may not capture domain-specific semantics/knowledge such as word frequency.…”
Section: Introductionmentioning
confidence: 77%
See 4 more Smart Citations
“…We are particularly interested in leveraging BERT for better sentence quality and diversity estimates. This paper extends on previous work (Cho et al, 2019) by incorporating deep contextualized representations into DPP, with an emphasis on better sentence selection for extractive multi-document summarization. The major research contributions of this work include the following: (i) we make a first attempt to combine DPP with BERT representations to measure sentence quality and diversity and report encouraging results on benchmark summarization datasets; (ii) our findings suggest that it is best to model sentence quality, i.e., how important a sentence is to the summary, by combining semantic representations and surface indicators of the sentence, whereas pairwise sentence dissimilarity can be determined by semantic repre-sentations only; (iii) our analysis reveals that combining contextualized representations with surface features (e.g., sentence length, position, centrality, etc) remains necessary, as deep representations, albeit powerful, may not capture domain-specific semantics/knowledge such as word frequency.…”
Section: Introductionmentioning
confidence: 77%
“…Opinosis (Ganesan et al, 2010) 25.15 5.12 8.12 Extract+Rewrite (Song et al, 2018) 29.07 6.11 9.20 Pointer-Gen (See et al, 2017) 31.44 6.40 10.20 SumBasic (Vanderwende et al, 2007) 31.58 6.06 10.06 KLSumm (Haghighi et al, 2009) 31.23 7.07 10.56 LexRank (Erkan and Radev, 2004) 33.10 7.50 11.13 DPP (Kulesza and Taskar, 2012) † 36.95 9.83 13.57 DPP-Caps (Cho et al, 2019) 36.61 9.30 13.09 DPP-Caps-Comb (Cho et al, 2019) observe that DPP-BERT-Combined yields the best performance, achieving 10.23% and 11.06% Fscores respectively on DUC-04 and TAC-11. This finding suggests that sentence similarity scores and importance features from the DPP-BERT system and TF-IDF based features can complement each other to boost system performance.…”
Section: Summarization Resultsmentioning
confidence: 99%
See 3 more Smart Citations