2021
DOI: 10.1140/epjc/s10052-021-09338-8
|View full text |Cite
|
Sign up to set email alerts
|

Compressing PDF sets using generative adversarial networks

Abstract: We present a compression algorithm for parton densities using synthetic replicas generated from the training of a generative adversarial network (GAN). The generated replicas are used to further enhance the statistics of a given Monte Carlo PDF set prior to compression. This results in a compression methodology that is able to provide a compressed set with smaller number of replicas and a more adequate representation of the original probability distribution. We also address the question of whether the GAN coul… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 18 publications
(21 citation statements)
references
References 28 publications
0
21
0
Order By: Relevance
“…Finally, there are many simulation-related questions in fundamental physics, where AImethods allow us to make significant progress. Examples going beyond immediate applications to event generation include symbolic regression [130], sample and data compression [57,131], detection of symmetries [132][133][134][135], and many other fascinating new ideas and concepts.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, there are many simulation-related questions in fundamental physics, where AImethods allow us to make significant progress. Examples going beyond immediate applications to event generation include symbolic regression [130], sample and data compression [57,131], detection of symmetries [132][133][134][135], and many other fascinating new ideas and concepts.…”
Section: Discussionmentioning
confidence: 99%
“…The successful implementation of this idea has led to the NNPDF family of proton PDF determinations [46,[48][49][50] as well as to variants in the context of polarised PDF [51] and nuclear PDF [52,53] global analyses. The current implementation frontier, which has led to the recent NNPDF4.0 determination, involves a suite of contemporary machine learning methods and tools, specifically cross-validation to avoid overtraining, hyperoptimization [54] combined with K-folding for the automatic selection of the methodology, feature scaling of the input for the optimization of the neural networks used as basic underlying model [55], and GAN-enhanced compression for final efficient delivery [56,57].…”
Section: Parton Distribution Functionsmentioning
confidence: 99%
“…However, it usually means encoding a given data point into a more storage-efficient representation. This is different from our proposal which aims to encode the full underlying distribution in the network parameters [28] and to represent all aspects of the LHC training data in the generative network. Modulo limits of expressivity, the online-trained network output can, in principle, replace the training data completely.…”
Section: Introductionmentioning
confidence: 87%
“…More details about the Monte Carlo compression strategy adopted in this work can be found in App. A.1, with technical information about the algorithmic settings following [167].…”
Section: Monte Carlo Compressionmentioning
confidence: 99%