2019
DOI: 10.1007/978-3-030-20656-7_12
|View full text |Cite
|
Sign up to set email alerts
|

GPUMixer: Performance-Driven Floating-Point Tuning for GPU Scientific Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 30 publications
(7 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…Although GPU-CUDA math operations are double-precision capable (Whitehead and Fit-Florea 2011), increased peak performance (e.g. higher speed-up) is found when single-precision operations are used in their place which may differ in precision compared to CPU math operations (Laguna et al 2019). Our findings from table 1 and table 2 show that larger L-M initialization difference at the start of MCMC sampling results in larger S-scores in significantly different distributions.…”
Section: Discussionmentioning
confidence: 99%
“…Although GPU-CUDA math operations are double-precision capable (Whitehead and Fit-Florea 2011), increased peak performance (e.g. higher speed-up) is found when single-precision operations are used in their place which may differ in precision compared to CPU math operations (Laguna et al 2019). Our findings from table 1 and table 2 show that larger L-M initialization difference at the start of MCMC sampling results in larger S-scores in significantly different distributions.…”
Section: Discussionmentioning
confidence: 99%
“…Decreasing the reliance on these massive datacenters and using the edge servers (with the surge of decentralized "datacenters") will allow pre-processing and selective forwarding of processed data sets to the cloud, not only making computing more efficient but also automatically improving data privacy because of the proximity of these servers to the client sites. Other techniques for energy optimization include optimization of DRAM refresh rates on the hardware side [89] and optimizing the mixing of low-and high-precision floating point operations for mixed precision settings using techniques as described recently in the GPUmixer [90].…”
Section: Energy-aware Computingmentioning
confidence: 99%
“…If it is not possible to avoid temporary data storage or data usage, precision reduction becomes popular [9,23,24,27,29]. Machine learning pushes the introduction of precision reduction [17], but it is natural to exploit new native hardware formats with reduced memory footprint in scientific computations, too.…”
Section: -Terminologymentioning
confidence: 99%
“…data delivery. While the cores with their vector registers can yield an impressive number of computations per second and while there are many cores, we struggle to feed them with data [10,16,17,24,25,27].…”
mentioning
confidence: 99%