2022
DOI: 10.1016/j.automatica.2022.110366
|View full text |Cite
|
Sign up to set email alerts
|

Resilient reinforcement learning and robust output regulation under denial-of-service attacks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 58 publications
(21 citation statements)
references
References 31 publications
0
21
0
Order By: Relevance
“…Based on accessible data, these matrices have been estimated in [54] with the help of -learning and exploration noises (guaranteeing the PE condition). Then, a -learning algorithm combined with a dropout Smith predictor has been designed in [55] to solve the optimal control problems with network-induced dropouts. Very recently, a Bernoulli-driven HJB equation has been first developed in [53] to deal with optimal control problems without both a priori knowledge of system dynamics and the probability models of packet dropouts.…”
Section: Q Qmentioning
confidence: 99%
“…Based on accessible data, these matrices have been estimated in [54] with the help of -learning and exploration noises (guaranteeing the PE condition). Then, a -learning algorithm combined with a dropout Smith predictor has been designed in [55] to solve the optimal control problems with network-induced dropouts. Very recently, a Bernoulli-driven HJB equation has been first developed in [53] to deal with optimal control problems without both a priori knowledge of system dynamics and the probability models of packet dropouts.…”
Section: Q Qmentioning
confidence: 99%
“…8 DoS attacks affect the transmission channels blocking the communication between the controller, sensors, and actuators. 9,10 On the other hand, deception attacks affect data integrity by injecting false data into some components, such as actuators or sensors. 11 In particular, control strategies concerning deception attacks are still incipient due to the difficulties in accurately modeling those attacks, which may effectively be any type of signal injection.…”
Section: Introductionmentioning
confidence: 99%
“…In particular, ADP has gained numerous attentions recently and been applied to control both continuous-time systems [11][12][13][14][15][16][17][18][19][20][21][22][23][24][25][26] and discrete-time systems. [27][28][29][30][31][32] Therefore, different studies have considered combining the theories of adaptive optimal control with the output regulation in order to achieve the adaptive optimal output regulation problem; see 13,[33][34][35][36][37][38][39][40][41][42][43][44] and references therein. Using reinforcement learning and Bellman's principle of optimality, 10 ADP methods, 16,24,34,[45][46][47][48][49][50][51][52][53]…”
Section: Introductionmentioning
confidence: 99%
“…[27][28][29][30][31][32] Therefore, different studies have considered combining the theories of adaptive optimal control with the output regulation in order to achieve the adaptive optimal output regulation problem; see 13,[33][34][35][36][37][38][39][40][41][42][43][44] and references therein. Using reinforcement learning and Bellman's principle of optimality, 10 ADP methods, 16,24,34,[45][46][47][48][49][50][51][52][53][54][55][56] which are essentially based on reinforcement learning, are developed such that the agent can learn towards the optimal control policy by interacting with its unknown environment. With this learning framework, see Fig.…”
Section: Introductionmentioning
confidence: 99%