2018
DOI: 10.48550/arxiv.1806.00322
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Direct optimisation of the discovery significance when training neural networks to search for new physics in particle colliders

Adam Elwood,
Dirk Krücker

Abstract: We introduce two new loss functions designed to directly optimise the statistical significance of the expected number of signal events when training neural networks to classify events as signal or background in the scenario of a search for new physics at a particle collider. The loss functions are designed to directly maximise commonly used estimates of the statistical significance, s/ √ s + b, and the Asimov estimate, Z A . We consider their use in a toy SUSY search with 30 fb −1 of 14 TeV data collected at t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(14 citation statements)
references
References 19 publications
0
14
0
Order By: Relevance
“…Finally, it is important to consider the task of background estimation in the broader context of analysis optimization. A variety of methods have been proposed to directly optimize analysis sensitivity including uncertainty [107][108][109][110]. Background estimation is a key part of analysis design and could be integrated into the ABCD method in order to further optimize the overall discovery potential.…”
Section: Discussionmentioning
confidence: 99%
“…Finally, it is important to consider the task of background estimation in the broader context of analysis optimization. A variety of methods have been proposed to directly optimize analysis sensitivity including uncertainty [107][108][109][110]. Background estimation is a key part of analysis design and could be integrated into the ABCD method in order to further optimize the overall discovery potential.…”
Section: Discussionmentioning
confidence: 99%
“…In [12,13], an evolutionary method for training neural networks called NeuroEvolution of Augmenting Topologies (NEAT) [14] was used in the direct optimization of event selectors. Another approach to directly optimizing neural networks for discovery significance was introduced in [15], using cost functions defined over batches of training events. Our prescription, on the other hand, relies on simply using the output of typical ML-based classifiers in an optimal manner.…”
Section: The Landscape Of Related Methodsmentioning
confidence: 99%
“…Henceforth, p(e i ) and p i will both refer to this estimate of p(e i ). 15 Out of the six statistical distance measures, only DPearχ 2 maximization leads to a categorizer that is independent of the value of the overall signal to background ratio S/B under H b+s . Training using the other five measures is sensitive to S/B; if training is to be done using these measures, w i and p i both need to be appropriately scaled before proceeding further.…”
Section: Preprocessing: Learning P(e)mentioning
confidence: 99%
See 1 more Smart Citation
“…The scale factors can also include correction factors such as trigger efficiencies and should account for the sample size when only a subset of the generated data is used to compute the loss. A similar approach of building a differentiable metric based on a FOM was taken by Elwood and Krücker [7], with a loss function based on the discovery significance.…”
Section: Punzi-lossmentioning
confidence: 99%