2016
DOI: 10.14778/2994509.2994515
|View full text |Cite
|
Sign up to set email alerts
|

Compressed linear algebra for large-scale machine learning

Abstract: Large-scale machine learning (ML) algorithms are often iterative, using repeated read-only data access and I/Obound matrix-vector multiplications to converge to an optimal model. It is crucial for performance to fit the data into single-node or distributed main memory. General-purpose, heavy-and lightweight compression techniques struggle to achieve both good compression ratios and fast decompression speed to enable block-wise uncompressed operations. Hence, we initiate work on compressed linear algebra (CLA),… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 55 publications
(27 citation statements)
references
References 43 publications
0
27
0
Order By: Relevance
“…Unlike existing work [9,22,48], we made the conscious design decision not to generate the data access into the fused operators. Instead, the handcoded skeleton implements the data access-depending on its sparse-safeness over cells or non-zero values-of dense, sparse, or compressed [28] matrices and calls an abstract (virtual) genexec method for each value. Generated operators inherit this skeleton and only override the specific genexec, which yields very lean yet efficient operators.…”
Section: Code Generation Plansmentioning
confidence: 99%
See 1 more Smart Citation
“…Unlike existing work [9,22,48], we made the conscious design decision not to generate the data access into the fused operators. Instead, the handcoded skeleton implements the data access-depending on its sparse-safeness over cells or non-zero values-of dense, sparse, or compressed [28] matrices and calls an abstract (virtual) genexec method for each value. Generated operators inherit this skeleton and only override the specific genexec, which yields very lean yet efficient operators.…”
Section: Code Generation Plansmentioning
confidence: 99%
“…Compressed Linear Algebra (CLA): All templates support operations over compressed matrices (column-wise compression, heterogeneous encoding formats, and column co-coding) [28]. Figure 9 shows the runtime of Base, Fused, and Gen for computing the sparse-safe expression sum(X 2 ) over Airline78 and Mnist8m.…”
Section: Operations Performancementioning
confidence: 99%
“…ShinyLearner is limited to datasets that fit into computer memory. For larger datasets, frameworks such as Apache SystemML support distributed algorithm execution[79]; however, the number of algorithms implemented in these frameworks is still relatively small.…”
Section: Discussionmentioning
confidence: 99%
“…These ideas have been applied to speed up ML workloads. SystemML [24,40,48] ScalOps [98], Pig latin [79], and KeystoneML [92] propose high-level ML languages for automatic parallelization and materialization, as well as easier programming. Hamlet [64] and others [63,86] avoid expensive denormalizations.…”
Section: Dbms-inspired Optimizationmentioning
confidence: 99%