2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.00663
|View full text |Cite
|
Sign up to set email alerts
|

When do GANs replicate? On the choice of dataset size

Abstract: Do GANs replicate training images? Previous studies have shown that GANs do not seem to replicate training data without significant change in the training procedure. This leads to a series of research on the exact condition needed for GANs to overfit to the training data. Although a number of factors has been theoretically or empirically identified, the effect of dataset size and complexity on GANs replication is still unknown. With empirical evidence from BigGAN and StyleGAN2, on datasets CelebA, Flower and L… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
3
3
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 30 publications
(22 citation statements)
references
References 24 publications
(35 reference statements)
0
22
0
Order By: Relevance
“…For instance, numerous metrics exist to measure similarity with the training data [32,3], the extent of mode collapse [61,15], and the impact of individual training samples [4,75]. Moreover, other work provides insights into when and why GANs may replicate training examples [50,26], as well as how to mitigate such effects [50]. Our work extends these lines of inquiry to conditional diffusion models, where we measure novelty by computing how frequently models regenerate training instances when provided with textual prompts.…”
Section: Related Workmentioning
confidence: 99%
“…For instance, numerous metrics exist to measure similarity with the training data [32,3], the extent of mode collapse [61,15], and the impact of individual training samples [4,75]. Moreover, other work provides insights into when and why GANs may replicate training examples [50,26], as well as how to mitigate such effects [50]. Our work extends these lines of inquiry to conditional diffusion models, where we measure novelty by computing how frequently models regenerate training instances when provided with textual prompts.…”
Section: Related Workmentioning
confidence: 99%
“…Learningbased models can also be vulnerable to inference attacks on models aiming to leak sensitive information about training data [22]. Differential privacy between training and synthetic samples in adversarial models substantially improves for large and diverse training datasets as encountered in FL settings [67]. Furthermore, FedGIMP uses a shared generator without direct access to data and unshared discriminators that are not communicated [68].…”
Section: Discussionmentioning
confidence: 99%
“…Learning-based models can also be vulnerable to inference attacks on models aiming to leak sensitive information about training data [21]. Differential privacy between training and synthetic samples in adversarial models substantially improves for large and diverse training datasets as encountered in FL settings [64]. Furthermore, FedGIMP uses a shared generator without direct access to data and unshared discriminators that are not communicated [65].…”
Section: Discussionmentioning
confidence: 99%