Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.330
|View full text |Cite
|
Sign up to set email alerts
|

MATINF: A Jointly Labeled Large-Scale Dataset for Classification, Question Answering and Summarization

Abstract: Recently, large-scale datasets have vastly facilitated the development in nearly all domains of Natural Language Processing. However, there is currently no cross-task dataset in NLP, which hinders the development of multi-task learning. We propose MATINF, the first jointly labeled large-scale dataset for classification, question answering and summarization. MAT-INF contains 1.07 million question-answer pairs with human-labeled categories and usergenerated question descriptions. Based on such rich information, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 44 publications
0
12
0
Order By: Relevance
“…Yadav et al (2021a) proposed question-aware transformer models for question summarization. Xu et al (2020) automatically created a Chinese dataset (MATINF) for medical question answering, summarization, and classification tasks focusing on maternity and infant categories. Some of the other prominent works in the abstractive summarization of long and short documents include Cohan et al (2018); Zhang et al (2019a); MacAvaney et al (2019); Sotudeh et al (2020).…”
Section: Related Workmentioning
confidence: 99%
“…Yadav et al (2021a) proposed question-aware transformer models for question summarization. Xu et al (2020) automatically created a Chinese dataset (MATINF) for medical question answering, summarization, and classification tasks focusing on maternity and infant categories. Some of the other prominent works in the abstractive summarization of long and short documents include Cohan et al (2018); Zhang et al (2019a); MacAvaney et al (2019); Sotudeh et al (2020).…”
Section: Related Workmentioning
confidence: 99%
“…For ProphetNet-En, we reports the results on summarization tasks CNN/DM (Hermann et al, 2015), Gigaword (Rush et al, 2015), and MSNews (Liu et al, 2020a); question generation tasks SQuAD 1.1 (Rajpurkar et al, 2016) and MSQG (Liu et al, 2020a). (Vinyals and Le, 2015) 0.336 0.238 0.03 0.128 0.183 0.448 0.353 0.004 0.016 0.205 iVAE MI (Fang et al, 2019) 0 Table 6: Results of ProphetNet-Dialog-En on DailyDialog and PersonaChat.…”
Section: Finetuning Benchmarksmentioning
confidence: 99%
“…We provide six pre-trained models with downstream task finetuning scripts, including ProphetNet-En pre-trained with 160GB English raw text, ProphetNet-Zh pre-trained with 160GB Chinese raw text, ProphetNet-Multi with 101GB Wiki-100 corpus and 1.5TB Common Crawl 2 data, ProphetNet-Dialog-En with 60 million sessions Reddit open-domain dialog corpus, ProphetNet-Dialog-Zh with collected Chinese dialog corpus over 30 million sessions, and ProphetNet-Code pre-trained with 10 million codes and documents. ProphetNet-X achieves new state-of-the-art results on 10 benchmarks, including Chinese summarization (MATINF-SUMM (Xu et al, 2020a) and LC-STS (Hu et al, 2015)), Chinese question answering (MATINF-QA (Xu et al, 2020a)), cross-lingual generation (XGLUE NTG (Liang et al, 2020) and Figure 1: A diagram of ProphetNet-X framework. ProphetNet-X models share the same model structure and cover various languages and domains.…”
Section: Introductionmentioning
confidence: 99%
“…Previous studies (Dergousoff and Mandryk, 2015;van Berkel et al, 2017) reveal that gamification can often improve the quality of collected data. Instead of collecting data from the wild like most datasets (Zhang et al, 2014(Zhang et al, , 2015Xu et al, 2020b), we collect the data from historical game records of Decrypto Online, a well-designed online board game. The screenshot of the user interface is shown in Figure 2.…”
Section: Data Collectionmentioning
confidence: 99%