2021
DOI: 10.48550/arxiv.2109.05143
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Bundled Gradients through Contact via Randomized Smoothing

Abstract: The empirical success of derivative-free methods in reinforcement learning for planning through contact seems at odds with the perceived fragility of classical gradient-based optimization methods in these domains. What is causing this gap, and how might we use the answer to improve gradient-based methods? We believe a stochastic formulation of dynamics is one crucial ingredient. We use tools from randomized smoothing to analyze sampling-based approximations of the gradient, and formalize such approximations th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 38 publications
0
2
0
Order By: Relevance
“…General-purpose LCP solvers that are typically used rely on a pivoting method like Lemke's algorithm [6]. Randomized smoothing has been proposed as a method for returning gradients through contact [32] with this model. An alternative soft-contact model is also available for patch contacts [9], but it is more computationally expensive, requiring sophisticated higher-order implicit integrators, and does not natively provide gradients.…”
Section: Existing State-of-the-artmentioning
confidence: 99%
See 1 more Smart Citation
“…General-purpose LCP solvers that are typically used rely on a pivoting method like Lemke's algorithm [6]. Randomized smoothing has been proposed as a method for returning gradients through contact [32] with this model. An alternative soft-contact model is also available for patch contacts [9], but it is more computationally expensive, requiring sophisticated higher-order implicit integrators, and does not natively provide gradients.…”
Section: Existing State-of-the-artmentioning
confidence: 99%
“…MuJoCo overcomes such difficulties and returns continuous gradients by employing a soft-contact model and a finite difference scheme. For LCP-based simulators that return subgradients, randomized smoothing via gradient bundles has been proposed [32]. We compare Dojo's smooth gradients with the latter approach.…”
Section: A Simulationmentioning
confidence: 99%