2018
DOI: 10.1007/978-3-030-02465-9_43
|View full text |Cite
|
Sign up to set email alerts
|

Taking Gradients Through Experiments: LSTMs and Memory Proximal Policy Optimization for Black-Box Quantum Control

Abstract: In this work we introduce the application of black-box quantum control as an interesting reinforcement learning problem to the machine learning community. We analyze the structure of the reinforcement learning problems arising in quantum physics and argue that agents parameterized by long short-term memory (LSTM) networks trained via stochastic policy gradients yield a general method to solving them. In this context we introduce a variant of the proximal policy optimization (PPO) algorithm called the memory pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
40
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
10

Relationship

0
10

Authors

Journals

citations
Cited by 38 publications
(40 citation statements)
references
References 26 publications
0
40
0
Order By: Relevance
“…In this paper, we adopt a radically different approach to this problem based on machine learning (ML) [40][41][42][43][44][45][46]. ML has recently been applied successfully to several problems in equilibrium condensed matter physics [47,48], turbulent dynamics [49,50] and experimental design [51,52], and here we demonstrate that Reinforcement Learning (RL) provides deep insights into nonequilibrium quantum dynamics [53][54][55][56][57][58]. Specifically, we use a modified version of the Watkins Q-Learning algorithm [40] to teach a computer agent to find driving protocols which prepare a quantum system in a target state |ψ * starting from an initial state |ψ i by controlling a time-dependent field.…”
Section: Introductionmentioning
confidence: 99%
“…In this paper, we adopt a radically different approach to this problem based on machine learning (ML) [40][41][42][43][44][45][46]. ML has recently been applied successfully to several problems in equilibrium condensed matter physics [47,48], turbulent dynamics [49,50] and experimental design [51,52], and here we demonstrate that Reinforcement Learning (RL) provides deep insights into nonequilibrium quantum dynamics [53][54][55][56][57][58]. Specifically, we use a modified version of the Watkins Q-Learning algorithm [40] to teach a computer agent to find driving protocols which prepare a quantum system in a target state |ψ * starting from an initial state |ψ i by controlling a time-dependent field.…”
Section: Introductionmentioning
confidence: 99%
“…In the mean time, numerically, the state preparation paradigm has been formulated as an optimisation problem [44][45][46][47][48][49][50][51][52]. Recently, stochastic descent, gradientbased GRAPE [53] and CRAB [54], and model-free Machine Learning [48,[55][56][57][58][59][60][61][62][63][64][65][66] have proven useful algorithms to find approximate fast-forward Hamiltonians in singleparticle and many-body systems.…”
Section: Introductionmentioning
confidence: 99%
“…Unfortunately, these theories have limited application in non-integrable manybody systems, for which no exact closed-form expressions can be obtained. This has motivated the development of efficient numerical algorithms, such as GRAPE [16,17], CRAB [18], and Machine learning based approaches [19][20][21][22][23][24][25][26][27][28][29][30][31]. State preparation can be formulated as an optimal control problem for which the objective is to find the set of controls that extremize a cost function, i.e.…”
mentioning
confidence: 99%