2022
DOI: 10.48550/arxiv.2204.04785
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Driving black-box quantum thermal machines with optimal power/efficiency trade-offs using reinforcement learning

Abstract: The optimal control of non-equilibrium open quantum systems is a challenging task but has a key role in improving existing quantum information processing technologies. We introduce a general model-free framework based on Reinforcement Learning to identify out-of-equilibrium thermodynamic cycles that are Pareto optimal trade-offs between power and efficiency for quantum heat engines and refrigerators. The method does not require any knowledge of the quantum thermal machine, nor of the system model, nor of the q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 108 publications
(186 reference statements)
0
3
0
Order By: Relevance
“…QOCT for heat engines has been almost exclusively applied to the discrete Carnot and Otto four stroke cycles. Typical optimization targets are maximum efficiency, maximum power, or minimal fluctuations [217,582,641]. A trade-off has been identified between these tasks.…”
Section: Quantum Thermodynamicsmentioning
confidence: 99%
“…QOCT for heat engines has been almost exclusively applied to the discrete Carnot and Otto four stroke cycles. Typical optimization targets are maximum efficiency, maximum power, or minimal fluctuations [217,582,641]. A trade-off has been identified between these tasks.…”
Section: Quantum Thermodynamicsmentioning
confidence: 99%
“…While MPS-based algorithms have been used in the context of optimal many-body control to find high-fidelity protocols [17][18][19][20] , the advantages of deep RL for quantum control 21 have so far been investigated using exact simulations of only a small number of interacting quantum degrees of freedom. Nevertheless, policy-gradient and value-function RL algorithms have recently been established as useful tools in the study of quantum state preparation [22][23][24][25][26][27][28][29][30][31][32][33][34][35][36][37][38][39] , quantum error correction and mitigation [40][41][42][43] , quantum circuit design [44][45][46][47] , quantum metrology 48,49 , and quantum heat engines 50,51 ; quantum reinforcement learning algorithms have been proposed as well [52][53][54][55][56] . Thus, in times of rapidly developing quantum simulators which exceed the computational capabilities of classical computers 57 , the natural question arises regarding scaling up the size of quantum systems in RL control studies beyond exact diagonalization methods.…”
Section: State-informed Many-body Controlmentioning
confidence: 99%
“…QOCT for heat engines has been almost exclusively applied to the discrete Carnot and Otto four stroke cycles. Typical optimization targets are maximum efficiency, maximum power, or minimal fluctuations [214,577,635]. A trade-off has been identified between these tasks.…”
Section: Quantum Thermodynamicsmentioning
confidence: 99%