2020
DOI: 10.1109/mra.2020.2980547
|View full text |Cite
|
Sign up to set email alerts
|

Decoding Motor Skills of Artificial Intelligence and Human Policies: A Study on Humanoid and Human Balance Control

Abstract: AI-guided controller development time per iteration System definition transferred from AI policy Tuning and evaluation Development Time per iteration Development time Performance Max. performance AI policy development time per iteration System definition for DRL framework Tuning and evaluation Reformulation Development time per iteration AI policy development time per iteration System definition for DRL framework Tuning and evaluation Reformulation Development time per iteration Classical controller developmen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…Lastly, Yan et al [3] delve into the stability of humanoid robots constructed using cutting-edge machine learning and AI techniques. A deep reinforcement learning-based control framework, meticulously crafted to empower humanoid robots with the dexterity needed for push recovery, takes center stage.…”
Section: Human-inspired Approachesmentioning
confidence: 99%
See 1 more Smart Citation
“…Lastly, Yan et al [3] delve into the stability of humanoid robots constructed using cutting-edge machine learning and AI techniques. A deep reinforcement learning-based control framework, meticulously crafted to empower humanoid robots with the dexterity needed for push recovery, takes center stage.…”
Section: Human-inspired Approachesmentioning
confidence: 99%
“…Cloning approaches have been discussed in Sect. 3. Inverted pendulum-type schemes are discussed in Sect.…”
Section: Introductionmentioning
confidence: 99%
“…Upright standing offers a large variety of recovery strategies that can be leveraged in case of emergency to avoid falling down, among them: ankle, hip, stepping, height modulation and foot-tilting for any legged robot, plus angular momentum modulation for humanoid robots [13]. For small perturbations, in-place recovery strategies controlling the Center of Pressure (CoP) [14], the centroidal angular momentum [15], or using foot-tilting [16], [17] are sufficient.…”
Section: A Classical Non-linear Controlmentioning
confidence: 99%
“…We analysed the diversity of skills of MELA and MoE by a t-distributed Stochastic Neighbor Embedding (t-SNE) analysis ( Maaten and Hinton, 2008 ). T-SNE projects high-dimensionals NN activation on a 2D plane by clustering similar NN activations together but keeping dissimilar data points distant, which can be used to analyse robotic behaviours ( Yuan et al, 2020 ).…”
Section: Comparison Of Multi-expert Learning Architecture and Mixture...mentioning
confidence: 99%