2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00787
|View full text |Cite
|
Sign up to set email alerts
|

Towards Photo-Realistic Virtual Try-On by Adaptively Generating↔Preserving Image Content

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
267
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 235 publications
(267 citation statements)
references
References 32 publications
0
267
0
Order By: Relevance
“…We perform visual comparison of our proposed method with CP-VTON [8] and ACGPN [10]. DP-VTON combines pixel transformation with feature transformation to generate accurate and clear warped clothing.…”
Section: Qualitative Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…We perform visual comparison of our proposed method with CP-VTON [8] and ACGPN [10]. DP-VTON combines pixel transformation with feature transformation to generate accurate and clear warped clothing.…”
Section: Qualitative Resultsmentioning
confidence: 99%
“…On the other hand, we adopt Inception Score (IS) [17], Fréchet Inception Distance (FID) [18] and Peak Signal to Noise Ratio (PSNR) to evaluate the performance of our models on the quality of the generated try-on images. Table . 1 lists image quality (IS, FID and PSNR) and pair-wise structural similarity (SSIM) scores by CP-VTON [8], ACGPN [10] and DP-VTON. The quantitative metrics demonstrate the superiority of DP-VTON over other methods.…”
Section: Quantitative Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Our method outperforms the state of the art [Men et al 2020;Wang et al 2018a;Yang et al 2020] with respect to three components: body shape, photorealism, and skin color preservation on real images, as well as generalizing to other datasets, while only using unpaired data. While our method outperforms SOTA as it is, it can be further improved on real images to allow for complex texture patterns such as plaid, and specularities.…”
Section: Introductionmentioning
confidence: 82%
“…On a similar line, Jandial et al [16] presented a two-stage training pipeline consisting of a coarse-to-fine warping network and a texture transfer network conditioned on a learned segmentation mask and trained with a triplet loss strategy to further improve the quality of try-on results. More recently, Yang et al [17] proposed to generate the semantic layout of the target person and predict whether the corresponding image content needs to be generated or preserved, thus leading to more photo-realistic try-on results.…”
Section: Related Workmentioning
confidence: 99%