2022
DOI: 10.1109/lra.2022.3187843
|View full text |Cite
|
Sign up to set email alerts
|

Learning Deformable Object Manipulation From Expert Demonstrations

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(9 citation statements)
references
References 19 publications
0
9
0
Order By: Relevance
“…However, these model-based approaches require time-consuming planning like Cross Entropy Method or Model Predictive Control to achieve the goal cloth state. As for model-free approaches, some approaches learn task-specific cloth manipulation policies from expert demonstrations [7], [8], [9]. However, these task-specific policies fail to reuse information for different tasks efficiently.…”
Section: A Sequential Multi-step Cloth Manipulationmentioning
confidence: 99%
See 1 more Smart Citation
“…However, these model-based approaches require time-consuming planning like Cross Entropy Method or Model Predictive Control to achieve the goal cloth state. As for model-free approaches, some approaches learn task-specific cloth manipulation policies from expert demonstrations [7], [8], [9]. However, these task-specific policies fail to reuse information for different tasks efficiently.…”
Section: A Sequential Multi-step Cloth Manipulationmentioning
confidence: 99%
“…Some of these approaches learn task-specific sequential multi-step cloth manipulation policies from expert demonstrations. Although these approaches are demonstrated to be effective, different models should be trained to perform different tasks [7], [8], [9]. Others learn a goal-conditioned policy from random data and can achieve arbitrary cloth goal states.…”
mentioning
confidence: 99%
“…Prior works [40,41,42,43,44] propose methods to detect or model physics properties of granular media or liquids and test on scooping and pouring. In contrast, we propose a general method of predicting 3D tool flow which does not require modeling properties of deformables and is not specialized to scooping or pouring tasks, and which uses point clouds as inputs instead of RGB images [45].…”
Section: Related Workmentioning
confidence: 99%
“…Overall, these results may provide evidence for the benefits of imitation learning in these environments over reinforcement learning. An interesting future direction to explore for these tasks would be to combine imitation learning with reinforcement learning [45,69,70].…”
Section: Figures S2 and S3mentioning
confidence: 99%
“…In contrast, others attempt to reconstruct the corresponding mesh to guide the manipulation [10,155,233]. Policy noise [130], demonstration data [244,311], specially engineered data [156,386], HER [9], MaxEnt-RL objective [392,146,234] and advantage-weighted loss exploration term [311] are leverage to overcome the exploration problem for DRL applications.…”
Section: Data-driven Control In Cloth-shapingmentioning
confidence: 99%