Proceedings of the 7th International Conference on Principles and Practice of Programming in Java 2009
DOI: 10.1145/1596655.1596670
|View full text |Cite
|
Sign up to set email alerts
|

Automatic parallelization for graphics processing units

Abstract: Accelerated graphics cards, or Graphics Processing Units (GPUs), have become ubiquitous in recent years. On the right kinds of problems, GPUs greatly surpass CPUs in terms of raw performance. However, because they are difficult to program, GPUs are used only for a narrow class of special-purpose applications; the raw processing power made available by GPUs is unused most of the time.This paper presents an extension to a Java JIT compiler that executes suitable code on the GPU instead of the CPU. Both static an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2009
2009
2018
2018

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 27 publications
(25 citation statements)
references
References 29 publications
0
25
0
Order By: Relevance
“…It will be even harder to extract the fine-grained parallelism necessary for efficient use of many-core systems like GPUs with thousands of threads. Therefore, several automatic static parallelization techniques for GPUs have been proposed to exploit more parallelism [Han and Abdelrahman 2010;Baskaran et al 2010;Wolfe 2010;Leung et al 2009;.…”
Section: Motivationmentioning
confidence: 99%
See 2 more Smart Citations
“…It will be even harder to extract the fine-grained parallelism necessary for efficient use of many-core systems like GPUs with thousands of threads. Therefore, several automatic static parallelization techniques for GPUs have been proposed to exploit more parallelism [Han and Abdelrahman 2010;Baskaran et al 2010;Wolfe 2010;Leung et al 2009;.…”
Section: Motivationmentioning
confidence: 99%
“…There are previous works that have focused on generating CUDA code from sequential input [Han and Abdelrahman 2010;Baskaran et al 2010;Wolfe 2010;Leung et al 2009;. HiCUDA [Han and Abdelrahman 2010] is a high-level directive-based compiler framework for CUDA programming where programmers need to insert directives into sequential C code to define the boundaries of kernel functions.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Some of the more recent such research aims to effectively employ powerful dedicated and specialized co-processors like graphic cards. Notable VM designs in this direction are the CellVM [15], [30] for the Cell Broadband Engine, a VM with a distributed Java heap on a homogeneous TILE-64 system [28]; an extension of the JikesVM to detect and offload loops on CUDA devices [12]; and VMs for Intel's Larrabee GPGPU architecture [22]. Most of these designs are specific to VMs for Java-like languages which are based on a shared-memory concurrent programming model.…”
Section: Vm Support For High-level Concurrencymentioning
confidence: 99%
“…However, in 2006 NVidia introduced a programming environment called CUDA [25], which allows the GPU to be programmed through more traditional means. At this moment a dedicated programming effort is still required to develop algorithms that perform efficiently on GPU hardware, but efforts are underway for automatic transformation of CPU programs into GPU counterparts [24].…”
Section: General Purpose Computation On Graphics Hardwarementioning
confidence: 99%