2024
DOI: 10.1007/jhep02(2024)220
|View full text |Cite
|
Sign up to set email alerts
|

Anomaly detection in the presence of irrelevant features

Marat Freytsis,
Maxim Perelstein,
Yik Chuen San

Abstract: Experiments at particle colliders are the primary source of insight into physics at microscopic scales. Searches at these facilities often rely on optimization of analyses targeting specific models of new physics. Increasingly, however, data-driven model-agnostic approaches based on machine learning are also being explored. A major challenge is that such methods can be highly sensitive to the presence of many irrelevant features in the data. This paper presents Boosted Decision Tree (BDT)-based techniques to i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 31 publications
0
2
0
Order By: Relevance
“…As the classifiers themselves are not the focus on this work, we use the same hyperparameters and architecture used in refs. [17,18] without further optimisation for the high level features, despite promising performance improvements observed with decision trees [96,132] or for mitigating sculpting after applying cuts [20]. This enables easier direct comparison with CurtainsF4F.…”
Section: Weakly-supervised Classifiersmentioning
confidence: 99%
See 1 more Smart Citation
“…As the classifiers themselves are not the focus on this work, we use the same hyperparameters and architecture used in refs. [17,18] without further optimisation for the high level features, despite promising performance improvements observed with decision trees [96,132] or for mitigating sculpting after applying cuts [20]. This enables easier direct comparison with CurtainsF4F.…”
Section: Weakly-supervised Classifiersmentioning
confidence: 99%
“…This drop in performance, however, can be expected and is reproducible in the idealised setting. CWoLa classifiers based on neural networks have been observed to lose sensitivity as the dimensionality of the input increases, especially with inputs which are not sensitive to signal processes [20,96,132]. This effect is amplified as the number of signal events in the training data decreases.…”
Section: Jhep04(2024)109mentioning
confidence: 99%