2019
DOI: 10.1016/j.actaastro.2018.12.033
|View full text |Cite|
|
Sign up to set email alerts
|

Integrated guidance for Mars entry and powered descent using reinforcement learning and pseudospectral method

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 35 publications
(12 citation statements)
references
References 18 publications
0
11
0
1
Order By: Relevance
“…Then, combined with the methods of fuzzy reinforcement learning (FRL) (Jouffe, 1998) and proximal policy optimization (PPO) (Schulman et al, 2017), a repeated learning process can be implemented which will converge to the output network under the consideration of possible interferences and uncertainties in different scenarios. During a landing phase, the algorithm outputs control commands in real time according to the current state (Furfaro and Linares, 2017;Gaudet et al, 2018;Jiang et al, 2018).…”
Section: Learning-based Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Then, combined with the methods of fuzzy reinforcement learning (FRL) (Jouffe, 1998) and proximal policy optimization (PPO) (Schulman et al, 2017), a repeated learning process can be implemented which will converge to the output network under the consideration of possible interferences and uncertainties in different scenarios. During a landing phase, the algorithm outputs control commands in real time according to the current state (Furfaro and Linares, 2017;Gaudet et al, 2018;Jiang et al, 2018).…”
Section: Learning-based Methodsmentioning
confidence: 99%
“…However, the ZEM/ZEV method can derive only the guidance command when omitting the aerodynamic forces. Recently, the reinforcement learning based guidance algorithm has also been adopted to enhance the robustness of the ZEM/ZEV algorithms (Furfaro and Linares, 2017;Gaudet et al, 2018;Jiang et al, 2018), but its effectiveness is affected by the training samples, and it is not adaptable to model uncertainties.…”
Section: Introductionmentioning
confidence: 99%
“…Instead, control methods using AI will most often give statistical guarantees on performance when subject to specified uncertainties. One example of RL applied to powered descent is shown in Jiang et al (2019), which also incorporates pseudospectral optimal control methods. In this case, the RL controller is used to determine the 'handover' point at which the controller switches from entry-phase to powered descent-phase.…”
Section: Related Workmentioning
confidence: 99%
“…Наконец, отметим еще один крупный класс работ, использующих методы обучения с подкреплением, это работы американских специалистов Р. Фурфаро (Roberto Furfaro) и Б. Гаудета (Brian Gaudet) [20,21,22,23,24]. Все эти работы посвящены задаче адаптивного управления аппаратом во время спуска на поверхность Марса или другого небесного тела.…”
Section: Introductionunclassified