2019 IEEE Real-Time and Embedded Technology and Applications Symposium (RTAS) 2019
DOI: 10.1109/rtas.2019.00011
|View full text |Cite
|
Sign up to set email alerts
|

Fractional GPUs: Software-Based Compute and Memory Bandwidth Reservation for GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
29
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 52 publications
(29 citation statements)
references
References 17 publications
0
29
0
Order By: Relevance
“…Instead, those variations are caused by the initial core state (e.g. branch predictor state), or changes in instruction cache behavior due to changes in the memory alignment of the binaries with and No [12], [28], [31], [32], [4], [25], [33], [34] GPU Partially [2] No [7], [19], [38], [39] without thread redundancy. In the case of FAC benchmark, since it is a small benchmark (around 700 instructions only), these tiny effects have a visible impact in relative terms (e.g.…”
Section: B Execution Time Overheadmentioning
confidence: 99%
“…Instead, those variations are caused by the initial core state (e.g. branch predictor state), or changes in instruction cache behavior due to changes in the memory alignment of the binaries with and No [12], [28], [31], [32], [4], [25], [33], [34] GPU Partially [2] No [7], [19], [38], [39] without thread redundancy. In the case of FAC benchmark, since it is a small benchmark (around 700 instructions only), these tiny effects have a visible impact in relative terms (e.g.…”
Section: B Execution Time Overheadmentioning
confidence: 99%
“…No [18], [19], [20], [21], [22], [23], [24], [25] GPU Partially [3] No [26], [27], [28], [29] [30] implement DCLS, whereas some Arm Cortex-R5 designs implement Triple-Core Lockstep [5], but fail to provide enough performance for AD systems [31]. Some improvements shorten time-to-detection for errors [32] or enhance recovery processes [33], but do not improve performance.…”
Section: Our Approachmentioning
confidence: 99%
“…Analogous solutions for accelerators (e.g. GPUs or the Kalray MPPA family) have been proposed, either with hardware support [27], [15], [16], [17] or with software-only support [26], [27], [28], but none of them guarantees diversity to protect against CCFs. Some preliminary solutions guarantee diversity to some extent for GPUs either with [13] or without hardware support [3].…”
Section: Our Approachmentioning
confidence: 99%
“…In [39], the authors show how to partition GPU memory resources, including cache and main memory, to enforce strong isolation between concurrent kernels. However, the approach is highly platform-specific, requiring a great deal of reverse engineering, it is focused on discreet GPUs rather than integrated CPU-GPU SoCs, and does not protect the GPU from CPU interference.…”
Section: Memory-aware Framework On Gpumentioning
confidence: 99%
“…We further assume that only one GPU kernel is executed at a time. While recent work has shown that co-scheduling multiple kernels can improve GPU resource utilization [76,39], it also complicates the issue of timing analysis. For this reason, we reserve such an extension to future work.…”
Section: System Model and Assumptionsmentioning
confidence: 99%