2021
DOI: 10.1109/access.2021.3056625
|View full text |Cite
|
Sign up to set email alerts
|

Wheel Loader Scooping Controller Using Deep Reinforcement Learning

Abstract: This article presents a deep reinforcement learning-based controller for an unmanned ground vehicle with a custom-built scooping mechanism. The robot's aim is to autonomously perform earth scooping cycles with three degrees of freedom: lift, tilt and the robot's velocity. While the majority of previous studies on automated scooping processes are based on data recorded by expert operators, we present a method to autonomously control a wheel loader to perform the scooping cycle using deep reinforcement learning … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 38 publications
(14 citation statements)
references
References 22 publications
0
14
0
Order By: Relevance
“…In future field tests, the amount of spillage on the ground and the slipping of the wheels should also be measured and taken into account when comparing the overall performance. Comparison with [21] and [18] is interesting but difficult because of the large differences in the vehicles' size and strength and the materials' properties. The reinforcement learning controller in [21] achieved a fill factor of 65%, and the energy consumption was not measured.…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…In future field tests, the amount of spillage on the ground and the slipping of the wheels should also be measured and taken into account when comparing the overall performance. Comparison with [21] and [18] is interesting but difficult because of the large differences in the vehicles' size and strength and the materials' properties. The reinforcement learning controller in [21] achieved a fill factor of 65%, and the energy consumption was not measured.…”
Section: Resultsmentioning
confidence: 99%
“…Comparison with [21] and [18] is interesting but difficult because of the large differences in the vehicles' size and strength and the materials' properties. The reinforcement learning controller in [21] achieved a fill factor of 65%, and the energy consumption was not measured. The neural network controller in [18], which was trained by learning from demonstration, reached 81% of the filling of the bucket relative to manual loading, but neither loading time nor work was reported.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…At each time-step, the agent takes action a according to the current environmental state S t and the policy π which is a mapping from perceived states to actions. Therefore, as a consequence of action, the environmental state transits from S t to S t+1 and the agent gets a reward r. The agent and environment generate the trajectories (S 1 ; A 1 ; R 1 ), (S 2 ; A 2 ; R 2 ), ..., (S T ; A T ; R T ) [23], until an episode is over. The basic architecture of RL is shown in Figure 8.…”
Section: Automatic Bucket-filling Algorithm Based On Q-learningmentioning
confidence: 99%