2020 IEEE International Conference on Robotics and Automation (ICRA) 2020
DOI: 10.1109/icra40945.2020.9197443
|View full text |Cite
|
Sign up to set email alerts
|

Aggressive Online Control of a Quadrotor via Deep Network Representations of Optimality Principles

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 16 publications
(11 citation statements)
references
References 15 publications
0
11
0
Order By: Relevance
“…Please note though that we and others are quickly developing deep learning approaches that can cross the reality gap for performing visual odometry Sanket et al (2020), tracking of predetermined optimal trajectories Kaufmann et al (2020) or even for full optimal control Li et al (2020b). AIRR has already been a driving force to develop AI methods that will successfully bridge the reality gap, even for robots that are difficult to model in detail upfront.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…Please note though that we and others are quickly developing deep learning approaches that can cross the reality gap for performing visual odometry Sanket et al (2020), tracking of predetermined optimal trajectories Kaufmann et al (2020) or even for full optimal control Li et al (2020b). AIRR has already been a driving force to develop AI methods that will successfully bridge the reality gap, even for robots that are difficult to model in detail upfront.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…One significant advantage of direct methods is their ability to easily handle more complicated OCPs, such as those with path constraints. In the context of supervised learning, [33,20] use Hermite-Simpson direct collocation to generate data for finite-horizon OCPs. Alternatively, Radau pseudospectral methods [30,9] are ideal for solving infinite-horizon open loop OCPs, though they have not yet been used for supervised learning.…”
Section: Data Generationmentioning
confidence: 99%
“…This line of work has been futher developed using nonlinear regression with NNs [12,11,24,25,26] and sparse polynomials [3], significantly increasing the maximum feasible problem dimension. A variation of the method in [25] is proposed by [7], in which the value gradient is directly approximated without learning the value function itself, while [31,33,20,12,11] use NNs to directly approximate the control policy without solving the HJB equation.…”
Section: Introductionmentioning
confidence: 99%
“…This line of work has been futher developed using nonlinear regression with NNs [14,13,29,30,32] and sparse polynomials [2], significantly increasing the maximum feasible problem dimension. Alternatively, one can directly approximate the value gradient [8,31] or control policy [37,40,25,14,13,31].…”
Section: Introductionmentioning
confidence: 99%