Proceedings of the 8th Workshop on Latest Advances in Scalable Algorithms for Large-Scale Systems 2017
DOI: 10.1145/3148226.3148237
|View full text |Cite
|
Sign up to set email alerts
|

Investigating half precision arithmetic to accelerate dense linear system solvers

Abstract: The use of low-precision arithmetic in mixed-precision computing methods has been a powerful tool to accelerate numerous scientific computing applications. Artificial intelligence (AI) in particular has pushed this to current extremes, making use of half-precision floating-point arithmetic (FP16) in approaches based on neural networks. The appeal of FP16 is in the high performance that can be achieved using it on today's powerful manycore GPU accelerators, e.g., like the NVIDIA V100, that can provide 120 TeraF… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
31
0
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 52 publications
(32 citation statements)
references
References 17 publications
0
31
0
1
Order By: Relevance
“…A first step in this direction is the recent performance study of Haidar et al [16], which shows promising results.…”
Section: Discussionmentioning
confidence: 99%
“…A first step in this direction is the recent performance study of Haidar et al [16], which shows promising results.…”
Section: Discussionmentioning
confidence: 99%
“…An investigation of similar iterative refinement methods on earlier generations of GPUs can be found in [11]. With the announcement of NVIDIA's V100 Tensor Cores.…”
Section: Related Workmentioning
confidence: 99%
“…Compared to our previous work in [11], the primary contribution of this paper is to propose and implement a high-performance framework for the mixed-precision iterative refinement solvers that makes use for the first time of GPU Tensor Core-accelerated FP16-TC. To this end, we will:…”
Section: Contributionsmentioning
confidence: 99%
“…While in many scientific applications the use of double-precision floating-point is most common, this precision is not always required. For example, iterative methods can exhibit resilience against low precision arithmetic as has been shown for the computation of inverse matrix roots [Lass et al 2018a] and for solving systems of linear equations [Angerer et al 2016;Haidar et al 2018Haidar et al , 2017Klavík et al 2014]. Mainly driven by the growing popularity of artificial neural networks [Gupta et al 2015], we can observe growing support of low-precision data types in hardware accelerators.…”
Section: Approximate Computingmentioning
confidence: 99%