2022
DOI: 10.1109/tip.2022.3142527
|View full text |Cite
|
Sign up to set email alerts
|

Towards Disentangling Latent Space for Unsupervised Semantic Face Editing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(5 citation statements)
references
References 26 publications
0
5
0
Order By: Relevance
“…This metric describes how drastically the image changes when the intermediate latent space is interpolated along a certain direction, and its small value represents a relatively smooth latent space and low entanglement. Referring to STIA-WO [ 22 ] the PPL value is calculated for the intermediate latent space with a certain range of a sampling point along its orthogonal semantic attribute direction, instead of randomly sampling two latent spaces for calculating the PPL value as …”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This metric describes how drastically the image changes when the intermediate latent space is interpolated along a certain direction, and its small value represents a relatively smooth latent space and low entanglement. Referring to STIA-WO [ 22 ] the PPL value is calculated for the intermediate latent space with a certain range of a sampling point along its orthogonal semantic attribute direction, instead of randomly sampling two latent spaces for calculating the PPL value as …”
Section: Resultsmentioning
confidence: 99%
“…As shown in Figure 2 , when editing along with the gender semantic attribute direction, the faces in red boxes 1 and 2 have added glasses, and the hair color has changed. STGAN-WO [ 22 ] controls image generation at two scales separately in the latent space and achieves disentanglement of image structure from texture semantic attribute using weight orthogonal regularization, but its regularization method limits the weight space in the network; consequently, reduces the quality of the generated images. Chen et al [ 23 , 24 ].…”
Section: Introductionmentioning
confidence: 99%
“…Much like transferring selection 'styles', this example provides the obvious benefit of not requiring pre-defined population parameters for generating all aspects of the desired training data. Finally, a particularly intriguing use would involve combining work on creating an interpretable latent space, where population parameters can be translated into different parts of a multi-dimensional Gaussian (Liu et al 2022), with the ability to project generated examples onto this latent space (Karras et al 2021). While the latent space interpretation would allow for a simulator that is instantaneous once trained, the combination with latent space projection would make the network capable of inferring the demographic parameters of real data and testing evolutionary hypotheses using the network directly.…”
Section: Discussionmentioning
confidence: 99%
“…Much like transferring selection ‘styles’, this example provides the obvious benefit of not requiring pre-defined population parameters for generating all aspects of the desired training data. Finally, a particularly intriguing use would involve combining work on creating an interpretable latent space, where population parameters can be translated into different parts of a multi-dimensional Gaussian (Liu et al . 2022), with the ability to project generated examples onto this latent space (Karras et al .…”
Section: Discussionmentioning
confidence: 99%
“…A perfectly disentangled latent representation where each dimension represents a human-understandable concept would naturally be interpretable. However, achieving a fully disentangled representation is not feasible in the general case, as approaches require specific training data [241], supervision [182], or predefined concepts [145]. Nevertheless, there are approaches using SCMs (Section 2.3) to identify confounders and then disentangle representations to produce more interpretable models.…”
Section: Disentangled Representationsmentioning
confidence: 99%