2024
DOI: 10.1109/tnnls.2022.3212620
|View full text |Cite
|
Sign up to set email alerts
|

Disentanglement by Cyclic Reconstruction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 25 publications
0
1
0
Order By: Relevance
“…Therefore, it is always advantageous in obtaining representations that carry certain attributes or extract discriminative features. Reconstruction based training (Gonzalez-Garcia et al, 2018;Zhang et al, 2019;Bertoin and Rachelson, 2022) is widely adopted in disentanglement learning and used to obtain disentangled representations. The application of representation disentanglement is extensive, including speech (Chan and Ghosh, 2022;), computer vision (Gonzalez-Garcia et al, 2018Lee et al, 2021) and natural language precessing (Bao et al, 2019;Cheng et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Therefore, it is always advantageous in obtaining representations that carry certain attributes or extract discriminative features. Reconstruction based training (Gonzalez-Garcia et al, 2018;Zhang et al, 2019;Bertoin and Rachelson, 2022) is widely adopted in disentanglement learning and used to obtain disentangled representations. The application of representation disentanglement is extensive, including speech (Chan and Ghosh, 2022;), computer vision (Gonzalez-Garcia et al, 2018Lee et al, 2021) and natural language precessing (Bao et al, 2019;Cheng et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Cyclic Reconstruction To effectively disentangle speech representations, we employ the cyclic reconstruction method proposed by Bertoin and Rachelson (2022) to reduce the mutual information between the content and non-content representations in an unsupervised learning setting. Specifically, after the extraction of content and non-content representations from the content and non-content encoder respectively, we stack two sub-networks: the content feature predictor ϕ content and non-content feature predictor ϕ non−content to cyclically reconstruct the content and non-content representations.…”
Section: Speech Representation Disentanglement Learningmentioning
confidence: 99%
“…However, recent studies [54,25] show that certain data augmentation techniques may lead to a decrease in sample efficiency and even cause divergence. Other recent works improve the generalization performance by leveraging pre-trained image encoder [56] or segmenting important pixels from the test environment [2], etc. Unfortunately, most of them rely on knowledge or data from outer sources, e.g., ImageNet [9].…”
Section: Introductionmentioning
confidence: 99%