2019
DOI: 10.1007/978-3-030-28619-4_30
|View full text |Cite
|
Sign up to set email alerts
|

A Multiview Approach to Learning Articulated Motion Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
13
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(13 citation statements)
references
References 46 publications
0
13
0
Order By: Relevance
“…Consequently, a plenty of research has focused on developing representations that can faciliate planning and reasoning for highly specific situated tasks. These representations vary significantly depending on the application, from two-dimensional costmaps (Elfes, 1987), volumetric 3D voxel representations (Hornung et al, 2013(Hornung et al, , 2010, primitive shape based object approximations (Miller et al, 2003;Huebner and Kragic, 2008) to more rich representations that model high level semantic properties (Galindo et al, 2005;Pronobis and Jensfelt, 2012), 6 DOF pose of the objects of interest (Hudson et al, 2012) or affordances between objects (Daniele et al, 2017). Since inferring exhaustively detailed world models is impractical, one solution is to design perception pipelines that infer task relevant world models (Eppner et al, 2016;Fallon et al, 2014).…”
Section: Introductionmentioning
confidence: 99%
“…Consequently, a plenty of research has focused on developing representations that can faciliate planning and reasoning for highly specific situated tasks. These representations vary significantly depending on the application, from two-dimensional costmaps (Elfes, 1987), volumetric 3D voxel representations (Hornung et al, 2013(Hornung et al, , 2010, primitive shape based object approximations (Miller et al, 2003;Huebner and Kragic, 2008) to more rich representations that model high level semantic properties (Galindo et al, 2005;Pronobis and Jensfelt, 2012), 6 DOF pose of the objects of interest (Hudson et al, 2012) or affordances between objects (Daniele et al, 2017). Since inferring exhaustively detailed world models is impractical, one solution is to design perception pipelines that infer task relevant world models (Eppner et al, 2016;Fallon et al, 2014).…”
Section: Introductionmentioning
confidence: 99%
“…Further approaches: Articulation motion models can be viewed as geometric constraints imposed on multiple rigid bodies. Such constraints can be learned from human demonstrations by leveraging different sensing modalities [13,[28][29][30][31]. Recently, Daniele et al [30] proposed a multimodal learning framework that incorporates both vision and natural language information for articulation model estimation.…”
Section: Related Workmentioning
confidence: 99%
“…For example, Schmidt et al [32,33] tracked deformable targets through probability inference, which requires the definition of a standard geometric structure. Daniele et al [34] combined natural motion language information and computer vision for the attitude estimation of deformable targets, but this method requires natural motion language description as an additional mode.…”
Section: Related Workmentioning
confidence: 99%