Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1262
|View full text |Cite
|
Sign up to set email alerts
|

Data-efficient Neural Text Compression with Interactive Learning

Abstract: Neural sequence-to-sequence models have been successfully applied to text compression. However, these models were trained on huge automatically induced parallel corpora, which are only available for a few domains and tasks. In this paper, we propose a novel interactive setup to neural text compression that enables transferring a model to new domains and compression tasks with minimal human supervision. This is achieved by employing active learning, which intelligently samples from a large pool of unlabeled dat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 28 publications
0
1
0
Order By: Relevance
“…Our methods are also reminiscent of active learning methods (Settles, 2009;Peris and Casacuberta, 2018;P.V.S and Meyer, 2019), such as uncertainty sampling (Lewis and Gale, 1994) which selects (unlabeled) data points, which a model trained on a small labeled subset, has least confidence in, or predicts as farthest (in vector space, based on cosine similarity) (Sener and Savarese, 2018;Wolf, 2011). Our approach uses labeled data for selection, similar to core-set selection approaches (Wei et al, 2013).…”
Section: Related Workmentioning
confidence: 99%
“…Our methods are also reminiscent of active learning methods (Settles, 2009;Peris and Casacuberta, 2018;P.V.S and Meyer, 2019), such as uncertainty sampling (Lewis and Gale, 1994) which selects (unlabeled) data points, which a model trained on a small labeled subset, has least confidence in, or predicts as farthest (in vector space, based on cosine similarity) (Sener and Savarese, 2018;Wolf, 2011). Our approach uses labeled data for selection, similar to core-set selection approaches (Wei et al, 2013).…”
Section: Related Workmentioning
confidence: 99%