2020
DOI: 10.1016/j.knosys.2019.105173
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning approach for optimal control of multiple electric locomotives in a heavy-haul freight train:A Double-Switch-Q-network architecture

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 33 publications
(9 citation statements)
references
References 46 publications
0
9
0
Order By: Relevance
“…(1) Strong adaptability for different constraints and objectives: The agent of reinforcement learning performs prediction and optimization through interaction with the environment, and it learns its "knowledge" or "experience" about the environment from sampled data rather than the prior knowledge obtained from other simulation models. Therefore, this approach has strong adaptability for different constraints and objectives, such as the control of multiple electric locomotives (Tang et al, 2020), robot pathfinding (Tozer et al, 2017), and flight taxi-out time prediction (Balakrishna et al, 2010).…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…(1) Strong adaptability for different constraints and objectives: The agent of reinforcement learning performs prediction and optimization through interaction with the environment, and it learns its "knowledge" or "experience" about the environment from sampled data rather than the prior knowledge obtained from other simulation models. Therefore, this approach has strong adaptability for different constraints and objectives, such as the control of multiple electric locomotives (Tang et al, 2020), robot pathfinding (Tozer et al, 2017), and flight taxi-out time prediction (Balakrishna et al, 2010).…”
Section: Methodsmentioning
confidence: 99%
“…The action values (also called Q-values) under each state are updated when the agent comes across the corresponding state-action pairs. The principle of Q-learning (Tang et al, 2020) is presented as follows.…”
Section: Basic Principles Of Q-learning and Deep Q-networkmentioning
confidence: 99%
“…Thus, an 18-dimensional array {X 1 , X 2 , • • • X 17 , Y} consisting of feature sets and class labels is obtained. The specific processes of modeling are shown as follows [27,28]:…”
Section: Modeling Of Air Braking For Heavy-haul Trains Based On the A...mentioning
confidence: 99%
“…Based on integral reinforcement learning and parameter identification methods, an adaptive control scheme was proposed in [18] to achieve the tracking control of high-speed trains. In [19], a double-switch Q-network architecture was proposed for the optimal control of a heavyhaul freight train. In the aforementioned reinforcement learning based controller design, the reinforcement learning model was trained offline in a simulation environment generally.…”
Section: Introductionmentioning
confidence: 99%