2018 IEEE Real-Time Systems Symposium (RTSS) 2018
DOI: 10.1109/rtss.2018.00021
|View full text |Cite
|
Sign up to set email alerts
|

Deadline-Based Scheduling for GPU with Preemption Support

Abstract: Modern automotive-grade embedded computing platforms feature high-performance Graphics Processing Units (GPUs) to support the massively parallel processing power needed for next-generation autonomous driving applications (e.g., Deep Neural Network (DNN) inference, sensor fusion, path planning, etc). As these workload-intensive activities are pushed to higher criticality levels, there is a stronger need for more predictable scheduling algorithms that are able to guarantee predictability without overly sacrifici… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
26
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 63 publications
(27 citation statements)
references
References 27 publications
1
26
0
Order By: Relevance
“…A case study based on a real implementation demonstrated improvements on both average-case (up to 24%) and longest-observed (up to 30%) response times of DNNs, with respect to the adoption of standard approaches. Possible future research directions target the design of predictable mechanisms for scheduling DNN on heterogeneous platforms, such as GPUs 29 and reprogrammable FPGAs, 53 and the extension of other protocols (eg, proxy execution 34,35 and BROE 38,39 ) to handle budget exhaustion inside critical sections to parallel workloads scheduled by reservation servers.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…A case study based on a real implementation demonstrated improvements on both average-case (up to 24%) and longest-observed (up to 30%) response times of DNNs, with respect to the adoption of standard approaches. Possible future research directions target the design of predictable mechanisms for scheduling DNN on heterogeneous platforms, such as GPUs 29 and reprogrammable FPGAs, 53 and the extension of other protocols (eg, proxy execution 34,35 and BROE 38,39 ) to handle budget exhaustion inside critical sections to parallel workloads scheduled by reservation servers.…”
Section: Resultsmentioning
confidence: 99%
“…Similar mechanisms have been proposed by Capodieci et al 27 and Ali and Yun. 28 Capodieci et al 29 implemented the constant bandwidth server (also used in this paper to provide timing isolation of Tensorflow threads) for scheduling CUDA kernels on Nvidia GPUs. Unfortunately, their work is not publicly available.…”
Section: Scheduling Of Dnnmentioning
confidence: 99%
“…However, this is a parasite effect given by the TensorRT API implementation, which submits input and output network conversion layers to the iGPU. In this way, the measured latencies for the GPU are also perturbed by the GPU context scheduler [28].…”
Section: Interference On the Igpumentioning
confidence: 99%
“…This allows the system designer to easily account for such periods and CPU/GPU processing time for scheduling purposes. We refer the reader to the following paper for an in-depth understanding of GPU scheduling on NVIDIA platforms [6].…”
Section: Iso Regulation Compliancementioning
confidence: 99%