Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop 2019
DOI: 10.18653/v1/p19-2034
|View full text |Cite
|
Sign up to set email alerts
|

STRASS: A Light and Effective Method for Extractive Summarization Based on Sentence Embeddings

Abstract: This paper introduces STRASS: Summarization by TRAnsformation Selection and Scoring. It is an extractive text summarization method which leverages the semantic information in existing sentence embedding spaces. Our method creates an extractive summary by selecting the sentences with the closest embeddings to the document embedding. The model learns a transformation of the document embedding to minimize the similarity between the extractive summary and the ground truth summary. As the transformation is only com… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
7
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(7 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…M7 -STRASS (Bouscarrat et al, 2019) extracts a summary by selecting the sentences with the closest embeddings to the document embedding, learning a transformation to maximize the similarity between the summary and the ground truth reference. See et al (2017) propose a variation of encoder-decoder models, the Pointer Generator Network, where the decoder can choose to generate a word from the vocabulary or copy a word from the input.…”
Section: M1mentioning
confidence: 99%
“…M7 -STRASS (Bouscarrat et al, 2019) extracts a summary by selecting the sentences with the closest embeddings to the document embedding, learning a transformation to maximize the similarity between the summary and the ground truth reference. See et al (2017) propose a variation of encoder-decoder models, the Pointer Generator Network, where the decoder can choose to generate a word from the vocabulary or copy a word from the input.…”
Section: M1mentioning
confidence: 99%
“…vance labels of sentences in a document as binarylatent variables M4 -REFRESH(Narayan et al, 2018) propose using REINFORCE (Williams, 1992) to extract summaries, approximating the search space during training by limiting to combinations of individually high-scoring sentences. M5 -RNES (Wu and Hu, M7 -STRASS(Bouscarrat et al, 2019) extracts a summary by selecting the sentences with the closest embeddings to the document embedding, learning a transformation to maximize the similarity between the summary and the ground truth reference. M11 -Improve-abs(Kryściński et al, 2018) extend the model ofPaulus et al (2017) by augmenting the decoder with an external LSTM language model and add a novelty RL-based objective during training.…”
mentioning
confidence: 99%
“…Paulus et al [4] propose a network with a novel intra-attention for the abstractive summarization of long documents. Bouscarrat et al [6] propose a light model based on Transformer for extractive summarization. However, most of these deep learning mechanisms are supervised models and they need a large number of training data.…”
Section: Related Workmentioning
confidence: 99%
“…Many sequence-to-sequence models [1][2][3] and attention-based models [4] are proposed for this task. BERT [5] is also used in existing deep learning models to achieve a better performance [6,7]. These studies have achieved promising results in document summarization.…”
Section: Introductionmentioning
confidence: 99%