2022
DOI: 10.1088/2632-2153/aca6cd
|View full text |Cite
|
Sign up to set email alerts
|

Training neural networks using Metropolis Monte Carlo and an adaptive variant

Abstract: We examine the zero-temperature Metropolis Monte Carlo algorithm as a tool for training a neural network by minimizing a loss function. We find that, as expected on theoretical grounds and shown empirically by other authors, Metropolis Monte Carlo can train a neural net with an accuracy comparable to that of gradient descent, if not necessarily as quickly. The Metropolis algorithm does not fail automatically when the number of parameters of a neural network is large. It can fail when a neural network’s structur… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 44 publications
0
5
0
Order By: Relevance
“…However, compared to physically motivated ansatzes such as ADAPT that suffer significantly less from the barren plateau problem, QCMPS shows even better optimization performance for some systems (Table S5). Several techniques, such as DMRG-sweep-like method, identity block initialization, layer-wise optimization, and Monte-Carlo-based algorithm, may be able to alleviate the optimization problem. However, a detailed investigation in this direction is out of the scope of this study.…”
Section: Numerical Resultsmentioning
confidence: 99%
“…However, compared to physically motivated ansatzes such as ADAPT that suffer significantly less from the barren plateau problem, QCMPS shows even better optimization performance for some systems (Table S5). Several techniques, such as DMRG-sweep-like method, identity block initialization, layer-wise optimization, and Monte-Carlo-based algorithm, may be able to alleviate the optimization problem. However, a detailed investigation in this direction is out of the scope of this study.…”
Section: Numerical Resultsmentioning
confidence: 99%
“…In the field of statistical physics, we usually rely on Monte Carlo algorithms for optimization, like simulated annealing [31]. Some studies have been reported on the line [32,33]. However, it is not so easy and requires some modifications [33].…”
Section: Discussionmentioning
confidence: 99%
“…Some studies have been reported on the line [32,33]. However, it is not so easy and requires some modifications [33]. Specifically, it fails when the network is highly heterogeneous.…”
Section: Discussionmentioning
confidence: 99%
“…In the field of statistical physics, we usually rely on Monte Carlo algorithms for optimization, like simulated annealing [30]. Some studies have been reported on the line [31,32]. However, it is not so easy and requires some modifications [32].…”
Section: Discussionmentioning
confidence: 99%
“…Some studies have been reported on the line [31,32]. However, it is not so easy and requires some modifications [32]. Specifically, it fails when the network is highly heterogeneous.…”
Section: Discussionmentioning
confidence: 99%