2022
DOI: 10.1038/s41586-022-05172-4
|View full text |Cite
|
Sign up to set email alerts
|

Discovering faster matrix multiplication algorithms with reinforcement learning

Abstract: Improving the efficiency of algorithms for fundamental computations can have a widespread impact, as it can affect the overall speed of a large amount of computations. Matrix multiplication is one such primitive task, occurring in many systems—from neural networks to scientific computing routines. The automatic discovery of algorithms using machine learning offers the prospect of reaching beyond human intuition and outperforming the current best human-designed algorithms. However, automating the algorithm disc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
140
0
2

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 275 publications
(142 citation statements)
references
References 39 publications
0
140
0
2
Order By: Relevance
“…The development of mathematical proofs and arguments seems to be one of the most difficult challenges. Nevertheless, some barriers have already fallen with the discovery of new multiplication algorithms [5].…”
Section: Discussionmentioning
confidence: 99%
“…The development of mathematical proofs and arguments seems to be one of the most difficult challenges. Nevertheless, some barriers have already fallen with the discovery of new multiplication algorithms [5].…”
Section: Discussionmentioning
confidence: 99%
“…Finally, we have to note that our scheme is based on computer-aided efficient search procedures for local relation enumerations. However, we are also actively looking for systematic redundancy generation schemes for recursive matrix multiplication methods as well as deep learning techniques recently fueled by DeepMind initiatives for faster matrix multiplications [18].…”
Section: B Proposed Methodsmentioning
confidence: 99%
“…In the past decade, machine learning has drawn great attention from almost all natural science and engineering communities, such as mathematics [1][2][3], physics [4][5][6][7][8][9][10], biology [11][12][13], and materials sciences [14][15][16], and has been widely used in various aspects of modern society, e.g., automatic driving systems, face recognition, fraud detection, expert recommendation system, speech enhancement, and natural language processing, etc. Especially, the deep learning techniques based on the artificial neural networks [17,18] have become the most popular and dominant machine learning approaches progressively, and their interactions with many-body physics have been intensively explored in recent years.…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, the 2nd scale transformation is composed of two isometries U (2) s, and each U (2) maps {σ (1) } to variables {σ (2) } sitting on the red lines. U (3) constitutes the last scale transformation, and maps {σ (2) } to variables {σ (3) } sitting on the black lines. Eventually, the operator is represented in terms of {σ (3) }, and this completes the full RG process.…”
Section: Introductionmentioning
confidence: 99%