1990
DOI: 10.1145/77626.79170
|View full text |Cite
|
Sign up to set email alerts
|

A set of level 3 basic linear algebra subprograms

Abstract: This paper describes an extension to the set of Basic Linear Algebra Subprograms. The extensions are targeted at matrix-vector operations that should provide for efficient and portable implementations of algorithms for high-performance computers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
778
0
10

Year Published

1996
1996
2019
2019

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 1,442 publications
(788 citation statements)
references
References 17 publications
0
778
0
10
Order By: Relevance
“…All computations were performed on a Sun Fire X4600 M2 with 16 AMD 2.8GHz cores and 32GB of RAM. The original MATLAB R NTF codes were rewritten in C++ and compiled with several libraries including LAPACK [1], ScaLAPACK [2], BLACS [4], BLAS [5] and MPICH [8].…”
Section: Data and Experimental Resultsmentioning
confidence: 99%
“…All computations were performed on a Sun Fire X4600 M2 with 16 AMD 2.8GHz cores and 32GB of RAM. The original MATLAB R NTF codes were rewritten in C++ and compiled with several libraries including LAPACK [1], ScaLAPACK [2], BLACS [4], BLAS [5] and MPICH [8].…”
Section: Data and Experimental Resultsmentioning
confidence: 99%
“…The first strategy is implemented using calls to ScaLAPACK; the second strategy is implemented with calls to LAPACK and BLAS [12]. They compare the strategies using Cholesky factorization on a network of workstations.…”
Section: Literature Surveymentioning
confidence: 99%
“…Moreover C++ interfaces well with other programming languages, and we can easily encapsulate functions and libraries implemented in other languages. In fact, in future releases we plan to use BLAS [10] and LAPACK [2] libraries for speeding-up the computation: BLAS offers the most efficient routines (implemented in FORTRAN) for vector and matrix multiplications, operations that are fundamental for speeding-up execution both in the forward and in the backward step of the backpropagation algorithm. LAPACK, that uses BLAS for low-level subroutines, offers a set of very efficient implementation of linear algebra functions, such matrix QR decomposition or pseudoinverse matrix calculation.…”
Section: Efficiency Of C++ Librariesmentioning
confidence: 99%