Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1041
|View full text |Cite
|
Sign up to set email alerts
|

Disentangled Representation Learning for Non-Parallel Text Style Transfer

Abstract: This paper tackles the problem of disentangling the latent variables of style and content in language models. We propose a simple yet effective approach, which incorporates auxiliary multi-task and adversarial objectives, for label prediction and bag-of-words prediction, respectively. We show, both qualitatively and quantitatively, that the style and content are indeed disentangled in the latent space. This disentangled latent representation learning method is applied to style transfer on non-parallel corpora.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
254
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 236 publications
(276 citation statements)
references
References 27 publications
2
254
0
Order By: Relevance
“…Hu et al (2017) propose a new neural generative model which combines variational auto-encoders and holistic attribute discriminators for the effective imposition of semantic structures. Following their work, many methods (Fu et al, 2018;John et al, 2018;Zhang et al, 2018a,b) has been proposed based on standard encoder-decoder architecture.…”
Section: Related Workmentioning
confidence: 99%
“…Hu et al (2017) propose a new neural generative model which combines variational auto-encoders and holistic attribute discriminators for the effective imposition of semantic structures. Following their work, many methods (Fu et al, 2018;John et al, 2018;Zhang et al, 2018a,b) has been proposed based on standard encoder-decoder architecture.…”
Section: Related Workmentioning
confidence: 99%
“…where p(w|z syn ) is the predicted distribution. BoW has been explored by previous work (Weng et al, 2017;John et al, 2018), showing good ability of preserving semantics. For the syntactic space, the multi-task loss trains a model to predict syntax on z syn .…”
Section: Disentangling Syntax and Semantics Into Different Latent Spacesmentioning
confidence: 99%
“…In particular, we introduce two continuous latent variables to capture semantics and syntax, respectively. To separate the semantic and syntactic information from each other, we borrow the adversarial approaches from the text style-transfer research (Hu et al, 2017;Fu et al, 2018;John et al, 2018), but adapt it into our scenario of syntactic modeling. We also observe that syntax and semantics are highly interwoven, and therefore further propose an adversarial reconstruction loss to regularize the syntactic and se-mantic spaces.…”
Section: Introductionmentioning
confidence: 99%
“…There are other perspectives on what constitutes a disentangled representation not addressed in this paper [1], [16], including definitions not statistical in nature, instead taking into account the manifold structure and symmetry transformations in data [1], [12], [20]. Other deep learning approaches to disentangling include the adversarial setting [21]- [23]. Disentangled representations have also been studied in supervised and semi-supervised contexts [24].…”
Section: Discussionmentioning
confidence: 99%
“…We can calculate the model posterior distribution p θ (z|x) at the network optimum, eqs. (22) and (23). Using Bayes' rule we find (see Appendix B-D)…”
Section: B Optimal β Values In An Analytically Tractable Modelmentioning
confidence: 91%