2019
DOI: 10.48550/arxiv.1903.05895
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Fast Algorithms for Linear Transforms Using Butterfly Factorizations

Abstract: Fast linear transforms are ubiquitous in machine learning, including the discrete Fourier transform, discrete cosine transform, and other structured transformations such as convolutions. All of these transforms can be represented by dense matrix-vector multiplication, yet each has a specialized and highly efficient (subquadratic) algorithm. We ask to what extent hand-crafting these algorithms and implementations is necessary, what structural priors they encode, and how much knowledge is required to automatical… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 27 publications
0
2
0
Order By: Relevance
“…The procedure explained in algorithm 1 can be represented by a butterfly graph similar to the FFT's graph. The butterfly network structure has been used for function representation [26] and fast factorization for approximating linear transformation [6]. We adopt this graph as an architecture design for the layers of a neural network.…”
Section: Butterfly Neural Networkmentioning
confidence: 99%
“…The procedure explained in algorithm 1 can be represented by a butterfly graph similar to the FFT's graph. The butterfly network structure has been used for function representation [26] and fast factorization for approximating linear transformation [6]. We adopt this graph as an architecture design for the layers of a neural network.…”
Section: Butterfly Neural Networkmentioning
confidence: 99%
“…Another way to reduce the model (and optimizer) memory required for storing and training a neural network is to replace weight matrices with special structured matrices, such as low-rank matrices [Sainath et al, 2013], Toeplitz-like matrices [Sindhwani et al, 2015], block-circulant matrices [Cheng et al, 2015, Ding et al, 2017, Fastfood transforms [Yang et al, 2015], low displacement rank matrices [Thomas et al, 2018], and butterfly matrices [Dao et al, 2019].…”
Section: Other Techniquesmentioning
confidence: 99%