2020
DOI: 10.48550/arxiv.2008.10518
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ScrewNet: Category-Independent Articulation Model Estimation From Depth Images Using Screw Theory

Abstract: Robots in human environments will need to interact with a wide variety of articulated objects such as cabinets, drawers, and dishwashers while assisting humans in performing day-to-day tasks. Existing methods either require objects to be textured or need to know the articulation model category a priori for estimating the model parameters for an articulated object. We propose ScrewNet, a novel approach that estimates an object's articulation model directly from depth images without requiring a priori knowledge … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(26 citation statements)
references
References 30 publications
0
26
0
Order By: Relevance
“…On the other hand, methods based on passive visual observations formulate the articulation parameters inference problem as a regression task [15]- [17] and leverage large synthetic datasets [18], [19]. Abbatematteo et al [15] map a segmented depth image of an object into its kinematic model parameters and geometry by using mixture density networks.…”
Section: A Learning Articulated Modelsmentioning
confidence: 99%
“…On the other hand, methods based on passive visual observations formulate the articulation parameters inference problem as a regression task [15]- [17] and leverage large synthetic datasets [18], [19]. Abbatematteo et al [15] map a segmented depth image of an object into its kinematic model parameters and geometry by using mixture density networks.…”
Section: A Learning Articulated Modelsmentioning
confidence: 99%
“…Synthetic datasets [30,33] are used to train regression-based models to directly predict part motion parameters. However, methods train on these dataset require accurate depth information [7,9,18,34]. Thus it is not straightforward for one to directly apply them to real world scenes due to noisy depth estimation and domain gap differences.…”
Section: D Object Pose and Motion Estimationmentioning
confidence: 99%
“…Previous works [42,43,44,45] have also explored various robotic planning and control methods for manipulating 3D articulated objects. More recent works further leveraged learning techniques for better predicting articulated part configurations, parameters, and states [6,7,4,46,3,5,47], estimating kinematic structures [1,2], as well as manipulating 3D articulated objects with the learned visual knowledge [8,9,10,11,12]. While most of these works represented visual data with link poses, joint parameters, and kinematic structures, such standardized abstractions may be insufficient if fine-grained part geometry, such as drawer handles and faucet switches that exhibit rich geometric diversity among different shapes, matters for downstream robotic tasks and motion planning.…”
Section: Related Workmentioning
confidence: 99%
“…There has been a long line of research studying the perception and manipulation of 3D articulated objects in computer vision and robotics. On the perception side, researchers have developed various successful visual systems for estimating kinematic structures [1,2], articulated part poses [3,4,5], and Figure 1: Given an input 3D articulated object (a), we propose a novel perception-interaction handshaking point for robotic manipulation tasks -object-centric actionable visual priors, including per-point visual action affordance predictions (b) indicating where to interact, and diverse trajectory proposals (c) for selected contact points (marked with green dots) suggesting how to interact.…”
Section: Introductionmentioning
confidence: 99%