Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.495
|View full text |Cite
|
Sign up to set email alerts
|

WSL-DS: Weakly Supervised Learning with Distant Supervision for Query Focused Multi-Document Abstractive Summarization

Abstract: In the Query Focused Multi-Document Summarization (QF-MDS) task, a set of documents and a query are given where the goal is to generate a summary from these documents based on the given query. However, one major challenge for this task is the lack of availability of labeled training datasets. To overcome this issue, in this paper, we propose a novel weakly supervised learning approach via utilizing distant supervision. In particular, we use datasets similar to the target dataset as the training data where we l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 19 publications
(6 citation statements)
references
References 34 publications
0
6
0
Order By: Relevance
“…In the future, we would like to extend ChartSumm to a multilingual dataset to address the scarcity of well-formatted datasets in other low-resource languages. We will also study how to incorporate query relevance [33][34][35], question-answering [36][37][38][39], and entity recognition [40][41][42] capabilities in this task.…”
Section: Discussionmentioning
confidence: 99%
“…In the future, we would like to extend ChartSumm to a multilingual dataset to address the scarcity of well-formatted datasets in other low-resource languages. We will also study how to incorporate query relevance [33][34][35], question-answering [36][37][38][39], and entity recognition [40][41][42] capabilities in this task.…”
Section: Discussionmentioning
confidence: 99%
“…In the future, we will extend our proposed approach to a more challenging task of validating a claim using extracted evidence sentences from a retrieval system. We will also explore how domain adaptation [27] or transfer learning [28] from answer selection models [29][30][31] impacts the overall performance. Meanwhile, target word probing, where the performance of prompt-based language models is affected by the words' linguistic similarity, is also one of the potential future directions of our work.…”
Section: Discussionmentioning
confidence: 99%
“…Portions of this work have been published as short papers at the COLING 2020 (Laskar, Hoque, and Huang 2020c) and the Canadian AI 2020 (Laskar, Hoque, and Huang 2020a) conference proceedings. However, this work substantially extends the published papers in several ways, most notably: (i) we investigate the performance of different attentions in query-focused summary generation (Section 3.1); (ii) we propose a novel sequential fine-tuning approach to utilize all the available multi-document gold reference summaries for supervised training (Section 3.2); (iii) for the query-focused abstractive summarization task in single-document scenarios, we conduct several ablation tests to investigate the effectiveness of different components used in our model (Section 5.1.3) as well as case studies to analyze the effectiveness of our model in the zero-shot learning setup (Section 5.1.4), as well as summarize the key limitations of the Debatepedia dataset (Section 5.1.5); (v) for the multi-document scenario, we study how incorporating recent transformer-based pre-trained summarizers in our proposed model impact performance (Section 5.2.3); and finally, (vi) in addition to extensive experiments on benchmark datasets, we also conduct human evaluation to qualitatively compare among different models proposed for query-focused multi-document summarization (Section 5.2.4).…”
Section: Bibliographic Notementioning
confidence: 99%