2020
DOI: 10.1007/978-3-030-43089-4_45
|View full text |Cite
|
Sign up to set email alerts
|

Bandit-Based Model Selection for Deformable Object Manipulation

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 13 publications
0
10
0
Order By: Relevance
“…This approach was recently extended by (McConachie and Berenson, 2016) by choosing among a set of models for this Jacobian using a multi-arm bandit formulation. Under this formulation, the "arms" were defined as models based on their diminishing rigidity and an adaptive Jacobian (as proposed by Navarro-Alarcón et al; see Section 6.3.2) with various sets of parameter.…”
Section: Multi-point Control Fanson and Patriciumentioning
confidence: 99%
“…This approach was recently extended by (McConachie and Berenson, 2016) by choosing among a set of models for this Jacobian using a multi-arm bandit formulation. Under this formulation, the "arms" were defined as models based on their diminishing rigidity and an adaptive Jacobian (as proposed by Navarro-Alarcón et al; see Section 6.3.2) with various sets of parameter.…”
Section: Multi-point Control Fanson and Patriciumentioning
confidence: 99%
“…Yet, since it is a local model, it should be continuously updated during task execution. Model updating methods include: Broyden rules [9], receding horizon adaption [24], local gradient descent [25], QP-based optimization [26], and Multi-armed Banditbased methods [27]. Furthermore, one can design a simple controller by inverting the Jacobian model.…”
Section: Modelingmentioning
confidence: 99%
“…To accelerate learning, we exploit the correlation between arms, calculated as the cosine similarity between the cost weights. Our formulation fits the problem from [27] and we implement their framework for nonstationary correlated multi-arm bandits. Finally, we return to exploit mode after a fixed number of steps N R , if we returned to nominal dynamics, or if we stopped moving after leaving the initial trap state.…”
Section: B Online: Trap-aware Mpcmentioning
confidence: 99%