2020 IEEE International Conference on Multimedia and Expo (ICME) 2020
DOI: 10.1109/icme46284.2020.9102761
|View full text |Cite
|
Sign up to set email alerts
|

Attentive Generative Adversarial Network To Bridge Multi-Domain Gap For Image Synthesis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 10 publications
0
8
0
Order By: Relevance
“…AGAN-CL [114] consists of a network which is trained to produce masks, thereby providing fine-grained information such as the number of objects, location, size and shape. The authors employed a multi-scale loss between real and generated masks, and an additional perceptual loss for global coherence.…”
Section: Semantic Masksmentioning
confidence: 99%
See 2 more Smart Citations
“…AGAN-CL [114] consists of a network which is trained to produce masks, thereby providing fine-grained information such as the number of objects, location, size and shape. The authors employed a multi-scale loss between real and generated masks, and an additional perceptual loss for global coherence.…”
Section: Semantic Masksmentioning
confidence: 99%
“…Input Method caption [16], [33], [40], [42], [41], [48], [35], [54], [43], [55], [58], [61], [65], [67], [68], [69], [70], [75], [80], [86], [34], [87], [128] caption + dialogue [93], [95], [99] caption + layout [104], [97], [108], [103] caption + semantic masks [109], [110], [113], [114], [115], [38] scene graphs [116], [121], [122], [124],…”
Section: Evaluation Of T2i Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Fréchet Inception Distance (FID) With visual features extracted by a pre-trained Inception-v3 [205] model, FID ---GAWWN [182] 3.62 67.22 -StackGAN [16] 3.70 51.89 -StackGAN++ [17] 4.04 15.30 -CVAEGAN [183] 4.97 --HDGAN [108] 4.15 --FusedGAN [184] 3.92 --PPAN [109] 4.38 --HfGAN [185] 4.48 --LeicaGAN [186] 4.62 --AttnGAN [14] 4.36 -67.82 MirrorGAN [18] 4.56 -57.67 SEGAN [114] 4.67 18.17 -ControlGAN [116] 4.58 -69.33 DM-GAN [187] 4 [188] 4.67 --textStyleGAN [120] 4.78 -74.72 AGAN-CL [189] 4.97 -63.87 TVBi-GAN [125] 5.03 11.83 -Souza et al [124] 4.23 11.17 -RiFeGAN [190] 5.23 --Wang et al [191] 5.06 12.34 86.50 Bridge-GAN [122] 4.74 -- [206] measures the distance between the real image distribution and generated image distribution. Compared with IS, FID is a more consistent evaluation metric as it captures various kinds of disturbances [206].…”
Section: Image Quality Metricsmentioning
confidence: 99%
“…On the other hand, FID share the same problem with IS such as struggling to detect overfitting results. Except for above common image quality metrics, some evaluation metrics are specially designed for certain gen- 3.52 -C4Synth [192] 3.52 -HfGAN [185] 3.57 -LeicaGAN [186] 3.92 -Text-SeGAN [193] 4.03 -RiFeGAN [190] 4.53 -AGAN-CL [189] 4.72 -Souza et al [124] 3.71 16.47 eration tasks. For image synthesis conditioned on semantic map, the image quality can be assessed by leveraging pretrained segmentation model to compute the mean average precision (mAP) and pixel accuracy (Acc).…”
Section: Image Quality Metricsmentioning
confidence: 99%