Purpose
This paper aims to address the safety concerns of path-planning algorithms in dynamic obstacle warehouse environments. It proposes a method that uses improved artificial potential fields (IAPF) as expert knowledge for an improved deep deterministic policy gradient (IDDPG) and designs a hierarchical strategy for robots through obstacle detection methods.
Design/methodology/approach
The IAPF algorithm is used as the expert experience of reinforcement learning (RL) to reduce the useless exploration in the early stage of RL training. A strategy-switching mechanism is introduced during training to adapt to various scenarios and overcome challenges related to sparse rewards. Sensor inputs, including light detection and ranging data, are integrated to detect obstacles around waypoints, guiding the robot toward the target point.
Findings
Simulation experiments demonstrate that the integrated use of IDDPG and the IAPF method significantly enhances the safety and training efficiency of path planning for mobile robots.
Originality/value
This method enhances safety by applying safety domain judgment rules to improve APF’s security and designing an obstacle detection method for better danger anticipation. It also boosts training efficiency through using IAPF as expert experience for DDPG and the classification storage and sampling design for the RL experience pool. Additionally, adjustments to the actor network’s update frequency expedite convergence.