2022
DOI: 10.1109/tsg.2021.3122570
|View full text |Cite
|
Sign up to set email alerts
|

Online Optimal Power Scheduling of a Microgrid via Imitation Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 55 publications
(16 citation statements)
references
References 38 publications
0
16
0
Order By: Relevance
“…Recent efforts on the application of RL for energy management are reported in economic dispatch and optimal power flow [52,53], MG operation [54][55][56][57][58][59], and VPP [60].…”
Section: Energy Managementmentioning
confidence: 99%
See 2 more Smart Citations
“…Recent efforts on the application of RL for energy management are reported in economic dispatch and optimal power flow [52,53], MG operation [54][55][56][57][58][59], and VPP [60].…”
Section: Energy Managementmentioning
confidence: 99%
“…For general energy management of MG or VPP [54,55,60], the reward is to minimize the operation cost, including the cost of DGs, energy storage, power exchange with grid, and controllable load. The action vector is the control signal of these components, and the state contains the control target and parameters or state of these components.…”
Section: Energy Managementmentioning
confidence: 99%
See 1 more Smart Citation
“…Hua recently reviewed the use of imitation and transfer learning in the field of robotics, especially for tasks such as manipulation and fine control [2]. Recently, a method to use imitation learning in the context of a building energy management problem was described by Gao et al [1], where the final agent learns to imitate the actions of a Model Predictive Controller (MPC) with a perfect forecast of the environment variables. Pezzotti et al developed MimicBot, a deep RL agent in a Fantasy Football game environment, which achieved excellent performance when RL training techniques were applied after an initial behaviour cloning step [9].…”
Section: Related Workmentioning
confidence: 99%
“…However, IL has the issue of distributional shift and tends to have poor generalization ability [25]. Moreover, IL is dependent on expert experience and prone to error, especially in off-course situations where lack of corresponding demonstrations makes the errors intensify as trajectory steps accumulate [26] while DRL will not suffer from this problem since it continuously interacts with the environment and has self-improvement ability.…”
Section: Introductionmentioning
confidence: 99%