2017 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2017
DOI: 10.1109/hpca.2017.14
|View full text |Cite
|
Sign up to set email alerts
|

Controlled Kernel Launch for Dynamic Parallelism in GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 36 publications
(18 citation statements)
references
References 33 publications
0
18
0
Order By: Relevance
“…In this paper, it is not our intention to discuss, compare and/or quantify dynamic parallelism overhead with the counterpart approaches. Some of the works on the comparison and discussion of dynamic parallelism overhead are [59,[62][63][64].…”
Section: Dynamic Parallelismmentioning
confidence: 99%
“…In this paper, it is not our intention to discuss, compare and/or quantify dynamic parallelism overhead with the counterpart approaches. Some of the works on the comparison and discussion of dynamic parallelism overhead are [59,[62][63][64].…”
Section: Dynamic Parallelismmentioning
confidence: 99%
“…They show that there is potential for speedup in several problems with inhomogeneous workload but that the greater overhead of launching kernels on the device can negate the benefits. Tang et al [12] discuss a dynamic platform which seeks to launch device side kernels only when the potential computation time outweighs the launch overhead. They show good speedup for several benchmark problems.…”
Section: Dynamic Parallelismmentioning
confidence: 99%
“…Maestro [36] dynamically selects SMK versus spatial multitasking. A number of papers target dynamic parallelism (DP), in which a kernel launches child kernels to increase resource utilization, and reduce the launch overhead, exploit data locality and improve load balancing [9], [20], [41], [46], [47]. All of these prior works focus on resource partitioning and optimization within a conventional GPU; none of these prior works explore the opportunity for exploiting TLP-resource diversity.…”
Section: Related Workmentioning
confidence: 99%