2020
DOI: 10.1038/s41586-020-2939-8
|View full text |Cite
|
Sign up to set email alerts
|

Autonomous navigation of stratospheric balloons using reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
109
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 201 publications
(109 citation statements)
references
References 37 publications
0
109
0
Order By: Relevance
“…controllers) in the physical world is more limited in comparison, due to difficulties in scaling parallel data collection, and higher variability in real-world data. Bellemare et al (2020) present a case study on autonomous balloon navigation, adopting a Q-learning approach, rather than model-based like ours. Akkaya et al (2019) use domain randomization in simulation to close the sim2real gap for a difficult dexterous manipulation task.…”
Section: Related Workmentioning
confidence: 99%
“…controllers) in the physical world is more limited in comparison, due to difficulties in scaling parallel data collection, and higher variability in real-world data. Bellemare et al (2020) present a case study on autonomous balloon navigation, adopting a Q-learning approach, rather than model-based like ours. Akkaya et al (2019) use domain randomization in simulation to close the sim2real gap for a difficult dexterous manipulation task.…”
Section: Related Workmentioning
confidence: 99%
“…RL has been making large inroads in solving ever more complex problems. For instance, newer systems have managed to learn and optimize over systems for which they are not given explicit rules [4], [5], instrument controls [6], and building controls based on edge computing [7]. This would fundamentally change how we are able to design experimental field campaigns, speeding up scientific discovery by improving data quality and usefulness, and facilitating more effective usage of research funding.…”
Section: Narrativementioning
confidence: 99%
“…Bellemare et al applied the distributional QR-DQN algorithm to realize stagnation point maintenance of high-altitude stratospheric balloons by determining the wind field use strategy [22]. The RL framework was applied to create a highperforming flight controller.…”
Section: Introductionmentioning
confidence: 99%
“…First, the basic Q-learning algorithm and the modified algorithms have very strong robustness and can be applied to a variety of tasks, such as video games [33] and quadrotor control [34]. Second, a deep Qnetwork algorithm has been used to guide the path of air vehicles and has achieved very good results [22] [35]. To our knowledge, this is the first time that the reinforcement learning method has been completely applied to HALE solarpowered aircraft flight trajectory planning.…”
Section: Introductionmentioning
confidence: 99%