2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2017
DOI: 10.1109/iros.2017.8202244
|View full text |Cite
|
Sign up to set email alerts
|

Deep reinforcement learning for high precision assembly tasks

Abstract: Abstract-High precision assembly of mechanical parts requires accuracy exceeding the robot precision. Conventional part mating methods used in the current manufacturing requires tedious tuning of numerous parameters before deployment. We show how the robot can successfully perform a tight clearance peg-in-hole task through training a recurrent neural network with reinforcement learning. In addition to saving the manual effort, the proposed technique also shows robustness against position and angle errors for t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
169
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
4

Relationship

2
8

Authors

Journals

citations
Cited by 284 publications
(171 citation statements)
references
References 14 publications
1
169
0
1
Order By: Relevance
“…A recently developed model-based reinforcement learning algorithm called guided policy search (GPS) provided new insights into training end-to-end policy for solving contactrich manipulation problems [15,26,27,4,28]; however; this method is not suitable for this high-precision setting because it has no means of avoiding local optima by its formulation. There are also approaches tackling this problem by explicitly modeling contact dynamics [29,30,31,32,33] Inoue et al [34] use LSTM to learn two separate policies for finding and inserting a peg into a hole; however, their methods require several pre-defined heuristics, and also the action space is discrete.…”
Section: Problem Statement and Related Workmentioning
confidence: 99%
“…A recently developed model-based reinforcement learning algorithm called guided policy search (GPS) provided new insights into training end-to-end policy for solving contactrich manipulation problems [15,26,27,4,28]; however; this method is not suitable for this high-precision setting because it has no means of avoiding local optima by its formulation. There are also approaches tackling this problem by explicitly modeling contact dynamics [29,30,31,32,33] Inoue et al [34] use LSTM to learn two separate policies for finding and inserting a peg into a hole; however, their methods require several pre-defined heuristics, and also the action space is discrete.…”
Section: Problem Statement and Related Workmentioning
confidence: 99%
“…Reinforcement learning has been applied successfully in simulated and real-world robotic manipulation [3], [4], [5], [8], [9], [10], locomotion [6], [2] and autonomous vehicles [11]. Many of the demonstrated scenarios used tailored policy representations or discretized action spaces.…”
Section: Related Workmentioning
confidence: 99%
“…This is particularly true in the case of data-center optimization since heat distribution occurs over time (e.g., a given command may not impact sensor readings instantaneously) and can depend on non-controllable parameters (e.g., weather conditions). DRL approaches have recently provided considerable results on such problems, where the relationship between states and optimal actions is difficult to model formally, e.g., playing Atari video games directly from pixels using convolutional neural networks (CNNs) rather than handcrafted features [7], or learning complex robotic tasks in both simulated [3] and real environments [5]. In DRL, control policies π θ are typically represented by deep neural networks parameterized by vector of parameters θ (e.g., neuron weights and biases).…”
Section: Model-free Reinforcement Learningmentioning
confidence: 99%