Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.493
|View full text |Cite
|
Sign up to set email alerts
|

Fact-level Extractive Summarization with Hierarchical Graph Mask on BERT

Abstract: Most current extractive summarization models generate summaries by selecting salient sentences. However, one of the problems with sentence-level extractive summarization is that there exists a gap between the human-written gold summary and the oracle sentence labels. In this paper, we propose to extract fact-level semantic units for better extractive summarization. We also introduce a hierarchical structure, which incorporates the multi-level of granularities of the textual information into the model. In addit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…To mitigate the redundancy between extracted sentences, Zhou et al [32] extract sub-sentential units based on the constituency parsing tree to construct a summary. Wu et al [31] apply dependency analysis to sentences and propose an empirical method to split sentences into smaller units called "fact". However, these small units [28] decrease the readability of summary which is unacceptable in our task.…”
Section: Text Summarizationmentioning
confidence: 99%
“…To mitigate the redundancy between extracted sentences, Zhou et al [32] extract sub-sentential units based on the constituency parsing tree to construct a summary. Wu et al [31] apply dependency analysis to sentences and propose an empirical method to split sentences into smaller units called "fact". However, these small units [28] decrease the readability of summary which is unacceptable in our task.…”
Section: Text Summarizationmentioning
confidence: 99%
“…Over the past few years, summarization task has witnessed a huge deal of progress in extractive (Nallapati et al, 2017;Liu and Lapata, 2019;Yuan et al, 2020;Cui et al, 2020;Jia et al, 2020;Feng et al, 2018) and abstractive (See et al, 2017;Gehrmann et al, 2018;Zhang et al, 2019;Tian et al, 2019;Zou et al, 2020) [Introductory] Neural machine translation (@xcite), directly applying a single neural network to transform the source sentence into the target sentence, has now reached impressive performance (@xcite […] Motivated by recent success in unsupervised cross-lingual embeddings (@xcite), the models proposed for unsupervised NMT often assume that a pair of sentences from two different languages can be mapped to a same latent representation in a shared-latent space (@xcite) […] Although the shared encoder is vital for mapping sentences from different languages into the shared-latent space, it is weak in keeping the uniqueness and internal characteristics of each language, such as the style, terminology and sentence structure. […] For each language, the encoder and its corresponding decoder perform an AE, where the encoder generates the latent representations from the perturbed input sentences and the decoder reconstructs the sentences from the latent representations.…”
Section: Introductionmentioning
confidence: 99%
“…Over the past few years, summarization task has witnessed a huge deal of progress in extractive (Nallapati et al, 2017;Liu and Lapata, 2019;Yuan et al, 2020;Cui et al, 2020;Jia et al, 2020;Feng et al, 2018) and abstractive (See et al, 2017;Cohan et al, 2018;Gehrmann et al, 2018;Zhang et al, 2019;Tian et al, 2019;Zou et al, 2020) [Introductory] Neural machine translation (@xcite), directly applying a single neural network to transform the source sentence into the target sentence, has now reached impressive performance (@xcite […] Motivated by recent success in unsupervised cross-lingual embeddings (@xcite), the models proposed for unsupervised NMT often assume that a pair of sentences from two different languages can be mapped to a same latent representation in a shared-latent space (@xcite) […] Although the shared encoder is vital for mapping sentences from different languages into the shared-latent space, it is weak in keeping the uniqueness and internal characteristics of each language, such as the style, terminology and sentence structure. […] For each language, the encoder and its corresponding decoder perform an AE, where the encoder generates the latent representations from the perturbed input sentences and the decoder reconstructs the sentences from the latent representations.…”
Section: Introductionmentioning
confidence: 99%