2021
DOI: 10.1088/2632-2153/abec21
|View full text |Cite
|
Sign up to set email alerts
|

GPU coprocessors as a service for deep learning inference in high energy physics

Abstract: In the next decade, the demands for computing in large scientific experiments are expected to grow tremendously. During the same time period, CPU performance increases will be limited. At the CERN Large Hadron Collider (LHC), these two issues will confront one another as the collider is upgraded for high luminosity running. Alternative processors such as graphics processing units (GPUs) can resolve this confrontation provided that algorithms can be sufficiently accelerated. In many cases, algorithmic speedups … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2
1

Relationship

4
5

Authors

Journals

citations
Cited by 24 publications
(18 citation statements)
references
References 66 publications
0
18
0
Order By: Relevance
“…This includes GPUs and potentially even field-programmable gate arrays (FPGAs) or ML-specific processors such as the GraphCore intelligence processing units (IPUs) [67] through specialized ML compilers [68][69][70]. These coprocessing accelerators can be integrated into existing CPU-based experimental software frameworks as a scalable service that grows to meet the transient demand [71][72][73].…”
Section: Resultsmentioning
confidence: 99%
“…This includes GPUs and potentially even field-programmable gate arrays (FPGAs) or ML-specific processors such as the GraphCore intelligence processing units (IPUs) [67] through specialized ML compilers [68][69][70]. These coprocessing accelerators can be integrated into existing CPU-based experimental software frameworks as a scalable service that grows to meet the transient demand [71][72][73].…”
Section: Resultsmentioning
confidence: 99%
“…This works shows how a tracking pipeline based on geometric deep learning can achieve state-of-the-art computing performance that scales linearly with the number of spacepoints, showing great promise for the next generation of HEP experiments. The inference pipeline has been optimized on GPU systems, on the assumption that the next generation of HEP experiments will have widespread access to accelerators either locally in heterogeneous systems [27,53] or remotely [54,55].…”
Section: Discussionmentioning
confidence: 99%
“…Results are given in Fig 9, separately for CPUs and GPUs. Having in mind an offline application, one could maximally benefit if the network throughput by running the network at once across batches of events, e.g., implementing the inference-as-a-service concept discussed in [106].…”
Section: Latency and Power Measurementsmentioning
confidence: 99%