2012 IEEE 26th International Parallel and Distributed Processing Symposium 2012
DOI: 10.1109/ipdps.2012.101
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Large Scale Image Analyses on Parallel, CPU-GPU Equipped Systems

Abstract: The past decade has witnessed a major paradigm shift in high performance computing with the introduction of accelerators as general purpose processors. These computing devices make available very high parallel computing power at low cost and power consumption, transforming current high performance platforms into heterogeneous CPU-GPU equipped systems. Although the theoretical performance achieved by these hybrid systems is impressive, taking practical advantage of this computing power remains a very challengin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0
2

Year Published

2013
2013
2019
2019

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 35 publications
(34 citation statements)
references
References 22 publications
0
32
0
2
Order By: Relevance
“…The specification of a parallel task consists of a set of input and output parameters with a corresponding data type (lines [16][17][32][33]) and a list of implementation variants. Each implementation variant defines the data distributions for the input and output parameters (lines [19][20][35][36] and either defines a runtime prediction (basic parallel tasks that are not further decomposed) or includes an entire task graph (complex parallel tasks). In the case of a basic parallel task, a set of symbolic runtime formulas is given (lines 22-25).…”
Section: ) Application Specificationmentioning
confidence: 99%
See 1 more Smart Citation
“…The specification of a parallel task consists of a set of input and output parameters with a corresponding data type (lines [16][17][32][33]) and a list of implementation variants. Each implementation variant defines the data distributions for the input and output parameters (lines [19][20][35][36] and either defines a runtime prediction (basic parallel tasks that are not further decomposed) or includes an entire task graph (complex parallel tasks). In the case of a basic parallel task, a set of symbolic runtime formulas is given (lines 22-25).…”
Section: ) Application Specificationmentioning
confidence: 99%
“…8 shows an example schedule for a hybrid platform. A schedule consists of communication operations (lines 2-12) and the execution of parallel tasks (lines [13][14][15][16][17][18][19] where each operation has an associated start time and an associated finish time. A communication operation additionally contains the identifiers of the source and target nodes in the underlying task graph and the definition of the source and target execution units.…”
Section: ) Schedule Specificationmentioning
confidence: 99%
“…Most of the research on power management for the CPU-GPU subsystems has been geared towards optimizing for general purpose GPU (GPGPU) computing [4,15,3,8,17,11]. Suda and Ren [13] proposed a method to estimate the power consumption of CUDA kernels in discrete GPU subsystems for the purpose of energy optimization.…”
Section: Related Workmentioning
confidence: 99%
“…Unlike the performance controller output described in (14) and (15), this algorithm does not take into account any performance jitter since it does not use the reference constraint on the number of packets in the system queues. In addition, it considers the GPU and the CPU in isolation and does not guarantee selection of the frequency for the producer (CPU) that would enable sufficient performance to the consumer (GPU).…”
Section: Adding Energy Optimization To the Controlmentioning
confidence: 99%
“…The feature computations on objects are generally more regular and compute intensive than the operations in the segmentation stage. This characteristics of the feature computation operations lead to better GPU acceleration [13]. …”
Section: Application Parallelization For High Throughput Executionmentioning
confidence: 99%