2020
DOI: 10.48550/arxiv.2007.14400
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ABCDisCo: Automating the ABCD Method with Machine Learning

Gregor Kasieczka,
Benjamin Nachman,
Matthew D. Schwartz
et al.

Abstract: The ABCD method is one of the most widely used data-driven background estimation techniques in high energy physics. Cuts on two statistically-independent classifiers separate signal and background into four regions, so that background in the signal region can be estimated simply using the other three control regions. Typically, the independent classifiers are chosen "by hand" to be intuitive and physically motivated variables. Here, we explore the possibility of automating the design of one or both of these cl… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 52 publications
(101 reference statements)
0
7
0
Order By: Relevance
“…In both approaches, the data are used directly to transport predictions from a control region to the signal sensitive region instead of relying on simulation for this extrapolation. In the ABCD method, two classifiers f and g and two working points a and b are constructed and then four regions called A, B, C and D are defined by f ≶ a and g ≶ b (for a machine learning version of ABCD, see [21][22][23]). If f and g are independent, then one can relate the background prediction in the region f > a and g > b to the other three regions.…”
Section: Model Dependence In Hep Data Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…In both approaches, the data are used directly to transport predictions from a control region to the signal sensitive region instead of relying on simulation for this extrapolation. In the ABCD method, two classifiers f and g and two working points a and b are constructed and then four regions called A, B, C and D are defined by f ≶ a and g ≶ b (for a machine learning version of ABCD, see [21][22][23]). If f and g are independent, then one can relate the background prediction in the region f > a and g > b to the other three regions.…”
Section: Model Dependence In Hep Data Analysismentioning
confidence: 99%
“…One can simply remove mass-sensitive features from the training, but powerful classifiers can learn the mass indirectly through subtle correlations with other useful features. A variety of decorrelation techniques exist to solve this problem [21,[26][27][28][29][30][31][32][33][34][35][36][37][38][39]. In the context of neural networks, one can add terms to the loss function to achieve automatic decorrelation:…”
Section: Model Dependence In Hep Data Analysismentioning
confidence: 99%
“…A key challenge facing such methods is that the machine learning classifiers must be relatively independent from the resonant feature, for otherwise artificial bumps can be formed. Many automated decorrelation methods have been proposed to ensure that classifiers are decorrelated from particular features by construction [39][40][41][42][43][44][45][46][47][48][49][50], but they may not apply in all cases. In particular, weakly supervised approaches that learn directly on the signal region cannot be simply combined with a decorrelation scheme because such an approach could degrade the performance in the presence of a signal.…”
Section: Introductionmentioning
confidence: 99%
“…In particular, the advent of unsupervised and weaklysupervised Machine Learning (ML) techniques has allowed for the development of broad model independent NP search and characterisation strategies [1]. Simultaneously, there have been important efforts to reduce reliance of LHC measurements on Monte Carlo (MC) simulations of hadronic processes [2][3][4][5][6].…”
mentioning
confidence: 99%
“…So the model would over-parameterize the data making the inclusion of mixtures redundant. 2 Therefore the key insight is to instead write down a mixture model in terms of p(j n |z n ) and p(b n |z n ), such that the correlations between N j and N b in the dataset are parameterized by the class label alone. The number of parameters in this model is 2 × (d j + d b − 2) + 1.…”
mentioning
confidence: 99%