2021
DOI: 10.1063/5.0034101
|View full text |Cite
|
Sign up to set email alerts
|

Robot grasping method optimization using improved deep deterministic policy gradient algorithm of deep reinforcement learning

Abstract: Robot grasping has become a very hot research field so that the requirements for robot operation are getting higher and higher. In previous research studies, the use of traditional target detection algorithms for grasping is often very inefficient, and this article is dedicated to improving the deep reinforcement learning algorithm to improve the grasping efficiency and solve the problem of robots dealing with the impact of unknown disturbances on grasping. Using the characteristic that deep reinforcement lear… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2025
2025

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…b) Policy-based methods are algorithms that directly optimize policy without estimating the value of states or state-action pairs by determining the actions taken by the agent as a function of the agent's state and environment [27]. Policy Gradient [44], Advantage Actor-Critic (A2C) [45], Asynchronous Advantage Actor-Critic (A3C) [46], Proximal Policy Optimization (PPO) [47] In robotic manipulation the classification of DRL is given according to the specific tasks and problem-solving approaches: Grasping and manipulation: Applications include using DRL for robots to perform ambiguous manipulation tasks such as grasping and hand manipulation [12]. Navigation and localization: Applications where DRL is not used for robots to navigate and localize different environments Figure 2 shows some examples from the studies on manipulation tasks such as robotic grasping [57], robotic hand manipulation [58], and object manipulation [59] using DRL algorithms.…”
Section: Figure 1 Drl Classificationmentioning
confidence: 99%
“…b) Policy-based methods are algorithms that directly optimize policy without estimating the value of states or state-action pairs by determining the actions taken by the agent as a function of the agent's state and environment [27]. Policy Gradient [44], Advantage Actor-Critic (A2C) [45], Asynchronous Advantage Actor-Critic (A3C) [46], Proximal Policy Optimization (PPO) [47] In robotic manipulation the classification of DRL is given according to the specific tasks and problem-solving approaches: Grasping and manipulation: Applications include using DRL for robots to perform ambiguous manipulation tasks such as grasping and hand manipulation [12]. Navigation and localization: Applications where DRL is not used for robots to navigate and localize different environments Figure 2 shows some examples from the studies on manipulation tasks such as robotic grasping [57], robotic hand manipulation [58], and object manipulation [59] using DRL algorithms.…”
Section: Figure 1 Drl Classificationmentioning
confidence: 99%
“…DDPG algorithm has been successfully applied to Robotics and motion control problems [81,82]. DDPG can be used for a variety of tasks, such as manipulation, locomotion, and navigation [83].…”
Section: Summary Of Studies Classified As Roboticsmentioning
confidence: 99%
“…Robotic manipulation (H. X. Zhang et al, 2021) [81] Techniques: Importance-Weighted Autoencoder (IWAE) and Gaussian parameter (Gaussian-DDPG) Methodology: Addition of Gaussian parameters to DDPG algorithm for better exploration and optimization of grasping position control using torque information.…”
Section: Airbornementioning
confidence: 99%
“…The effective resolution of complex manipulation tasks in an unstructured or highly variable environment remains an active field of research. Current research focuses mainly on grasping [12], picking and placing [13], and assembly tasks [14]. In particular, RL methods have shown high robustness to uncertainties in the latter, leading more and more researchers to focus on learning assembly skills.…”
Section: Contact-rich Manipulation Tasks: Assembly and Disassemblymentioning
confidence: 99%