Proceedings of the 2nd ACM SIGPLAN International Workshop on Machine Learning and Programming Languages 2018
DOI: 10.1145/3211346.3211354
|View full text |Cite
|
Sign up to set email alerts
|

Diesel: DSL for linear algebra and neural net computations on GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
27
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 43 publications
(27 citation statements)
references
References 3 publications
0
27
0
Order By: Relevance
“…Matrix multiplication is the most tuned computation kernel in history: The missing optimizations are all well known and may be found in use cases and open-source implementations such like CUTLASS [36]. Alternatively, polyhedral compilation has been shown to match or outperform cuBLAS, provided sufficient target-and operator-specific information has been captured in the optimization heuristic and code generator [20]. While our scientific focus was on covering a wide range of layers with TC, a production release would need to embed such operator-specific strategies as well.…”
Section: Performance Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Matrix multiplication is the most tuned computation kernel in history: The missing optimizations are all well known and may be found in use cases and open-source implementations such like CUTLASS [36]. Alternatively, polyhedral compilation has been shown to match or outperform cuBLAS, provided sufficient target-and operator-specific information has been captured in the optimization heuristic and code generator [20]. While our scientific focus was on covering a wide range of layers with TC, a production release would need to embed such operator-specific strategies as well.…”
Section: Performance Resultsmentioning
confidence: 99%
“…The polyhedral framework of compilation emerged as a natural candidate to design a versatile optimization flow satisfying the needs of the domain and target hardware. It has demonstrated strong results in domain-specific optimization [5,9,20,46], expert-driven meta-programming [6,15,26], embedding of third-party library code [40], and automatic generation of efficient code for heterogeneous targets [5,7,43,51,70,77]. We attempt to take the best of both worlds, defining a domain-specific language rich enough to capture full sub-graphs of modern Machine Learning (ML) models while enabling aggressive compilation competitive to native libraries.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…(Vasilache et al 2018). Originally, Halide is designed for image processing pipeline on NVIDIA GPU, similar to Diesel (Elango et al 2018), and Tensor Comprehension is designed for machine learning applications. DSLs will abstract complex control logic comparing to traditional language like C/C++/Java.…”
Section: Discussionmentioning
confidence: 99%
“…Diesel [9], NOVA [8], and PPCG [31] make heavy use of the polyhedral model for optimization. Fireiron generates nested affine loops and might therefore profit using polyhedral techniques too.…”
Section: Related Workmentioning
confidence: 99%