Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume 2021
DOI: 10.18653/v1/2021.eacl-main.73
|View full text |Cite
|
Sign up to set email alerts
|

DRAG: Director-Generator Language Modelling Framework for Non-Parallel Author Stylized Rewriting

Abstract: Author stylized rewriting is the task of rewriting an input text in a particular author's style. Recent works in this area have leveraged Transformer-based language models in a denoising autoencoder setup to generate author stylized text without relying on a parallel corpus of data. However, these approaches are limited by the lack of explicit control of target attributes and being entirely data-driven. In this paper, we propose a Director-Generator framework to rewrite content in the target author's style, sp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 28 publications
0
5
0
Order By: Relevance
“…Recently, we see more datasets that are generated by a machine [106,124,199,230], which has the advantage of being relatively cheap and fast to generate at scale compared to human-generated datasets. Finally, there are other types of creators such as non-experts, unspecified individuals, or a broad set of creators (e.g., in the case of web crawled data) [13,228,268,277].…”
Section: Dimensions and Codesmentioning
confidence: 99%
See 2 more Smart Citations
“…Recently, we see more datasets that are generated by a machine [106,124,199,230], which has the advantage of being relatively cheap and fast to generate at scale compared to human-generated datasets. Finally, there are other types of creators such as non-experts, unspecified individuals, or a broad set of creators (e.g., in the case of web crawled data) [13,228,268,277].…”
Section: Dimensions and Codesmentioning
confidence: 99%
“…For larger datasets (around tens of 1000s of examples) we denote this as large [56,207,257]. For models that undergo extensive large-scale pre-training, we categorized data used in this process as extremely large to indicate a dataset of millions of examples [43,223,228,276] or more. We also included an unknown if the paper did not explicitly mention the dataset used for training [180,192,238].…”
Section: Dimensions and Codesmentioning
confidence: 99%
See 1 more Smart Citation
“…a Note that we did not mention transfer works that shift the styles from one author to the other (e.g.,Syed et al 2020;Singh et al 2021). As opposed to the Shakespeare-Joyce example given above, which is moving texts along some diachronical dimensions and with respect to their poem or poetry nature, these works take style as a persistent characteristics of specific individuals.…”
mentioning
confidence: 99%
“…alNote that we did not mention transfer works that shift style from one author to the other by including multiple authors (e.g., Syed et al 2020; Singh et al 2021). As opposed to the Shakespeare-Joyce example given above, which paraphrases texts conditioned on a diachronical dimensions and with respect to their poem or poetry nature, these works take style as persistent characteristics of specific individuals.…”
mentioning
confidence: 99%