2022
DOI: 10.1109/lra.2022.3146931
|View full text |Cite
|
Sign up to set email alerts
|

Bundled Gradients Through Contact Via Randomized Smoothing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 24 publications
(15 citation statements)
references
References 45 publications
0
14
0
Order By: Relevance
“…To enforce the convergence towards an optimal (local) solution, it remains crucial to reduce the noise injected via the randomized smoothing across the iterations. A first possible strategy [13] consists in relying on Robbins-Monro rule [42] by decreasing the variance in a way such that k 2 k < ∞ to guarantee convergence towards a local minima. In this work, we propose to decrease k in a way that adapts to the problem and avoids the smoothing being reduced too quickly, which would lead to performance similar to classical DDP, or too slowly, which would induce an unnecessary large number of iterations.…”
Section: B Adaptive Smoothingmentioning
confidence: 99%
See 1 more Smart Citation
“…To enforce the convergence towards an optimal (local) solution, it remains crucial to reduce the noise injected via the randomized smoothing across the iterations. A first possible strategy [13] consists in relying on Robbins-Monro rule [42] by decreasing the variance in a way such that k 2 k < ∞ to guarantee convergence towards a local minima. In this work, we propose to decrease k in a way that adapts to the problem and avoids the smoothing being reduced too quickly, which would lead to performance similar to classical DDP, or too slowly, which would induce an unnecessary large number of iterations.…”
Section: B Adaptive Smoothingmentioning
confidence: 99%
“…Concurrently to our work, [13] introduced the notion of randomized smoothing in order to get gradients through contacts. We take a different point of view by establishing links between RL and OC through the lens of RS, which is a first step towards a stronger interplay between these two fields.…”
Section: Introductionmentioning
confidence: 99%
“…Although it is possible to solve nonlinear optimization problems without access to the gradients of the objective or constraint functions, either by estimating gradients [16] or using zero-order methods [17], it is often much faster to use exact gradient information when it is available. However, exact gradients can be difficult to derive symbolically for complex optimization problems.…”
Section: A Differentiable Simulation and Temporal Logicmentioning
confidence: 99%
“…On the other hand, AD necessarily incurs some overhead at runtime, making each AD function call more expensive than those used in an FD scheme. Additionally, some arguments [17] suggest that exact gradients may be less useful than finitedifference or stochastic approximations when the objective is stiff or discontinuous. We compare AD with a 3-point finitedifference method by re-solving problem (2) for both case studies, keeping all parameters constant (N = 512, λ = 0.1, same random seed) and substituting the gradients obtained using AD for those computed using finite differences.…”
Section: Design Optimization Ablation Studymentioning
confidence: 99%
“…In recent years, the robotics community has also developed special-purpose differentiable simulators for robotic systems, particularly those involving rigid body contact dynamics [14,15,16,17]. These simulators have been used to solve system identification and controller design tasks, but they do not represent a general-purpose framework, as gradients are often derived by hand and the simulators are not expressive enough to model full-stack robotic systems (e.g.…”
Section: Introductionmentioning
confidence: 99%