2017
DOI: 10.1016/j.egypro.2017.12.629
|View full text |Cite
|
Sign up to set email alerts
|

Power Management for a Plug-in Hybrid Electric Vehicle Based on Reinforcement Learning with Continuous State and Action Spaces

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2019
2019
2025
2025

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(19 citation statements)
references
References 8 publications
0
19
0
Order By: Relevance
“…In the case of supervisory control strategies for an HEV, certain learning based strategies are shown to be comparable to the commonly used control strategies [31]. For continuous-spaces, the actor-critic method was used for the power management in a PHEV [36]. A qualitative study on RL techniques on HEVs and PHEVs shows potential for RL controllers to replace rule based controllers [37].…”
Section: Torque Split Controlmentioning
confidence: 98%
“…In the case of supervisory control strategies for an HEV, certain learning based strategies are shown to be comparable to the commonly used control strategies [31]. For continuous-spaces, the actor-critic method was used for the power management in a PHEV [36]. A qualitative study on RL techniques on HEVs and PHEVs shows potential for RL controllers to replace rule based controllers [37].…”
Section: Torque Split Controlmentioning
confidence: 98%
“…Due to the low computational complexity of neural network, Li et al [60] presented a power management strategy based on reinforcement learning without worrying curse of dimensionality in complex environments, where stochastic gradient descent and experience replay were adopted to guarantee the accuracy and stability of the method.…”
Section: 2mentioning
confidence: 99%
“…Next, they validated this strategy by considering the battery and ultracapacitor in the loop, where a hardware-in-loop platform was established to execute the simulations [29]. To exploit the trained value function provided by RL methods, He et al leveraged an actor-critic approach to handle the continuous action space and used stochastic gradient descent to train the state-value function [30], [31].…”
Section: Hybrid Algorithmsmentioning
confidence: 99%