2020
DOI: 10.1109/lcsys.2020.2979572
|View full text |Cite
|
Sign up to set email alerts
|

On-Off Adversarially Robust Q-Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 22 publications
0
3
0
Order By: Relevance
“…Other research efforts over the years have focused on addressing sequential decision-making problems under uncertainty either with direct or indirect RL methods including robust learning-based approaches in applications related to quadrotor safety and steady-state stability [64], [65], learningbased model predictive control [66]- [68] with applications on autonomous racing cars [69], real-time learning [70]- [72] of powertrain operation of vehicles with respect to the driver's driving style [73], [74], learning for planning of autonomous vehicles [75], learning for traffic control in simulation [76]- [78] in conjunction with transfer of learned policies from simulation to a scaled environment [79], [80], decentralized learning for stochastic games [81], learning for optimal social routing [82] and congestion games [83], and learning for enhanced security against replay attacks in CPS [84], [85].…”
Section: B Related Workmentioning
confidence: 99%
“…Other research efforts over the years have focused on addressing sequential decision-making problems under uncertainty either with direct or indirect RL methods including robust learning-based approaches in applications related to quadrotor safety and steady-state stability [64], [65], learningbased model predictive control [66]- [68] with applications on autonomous racing cars [69], real-time learning [70]- [72] of powertrain operation of vehicles with respect to the driver's driving style [73], [74], learning for planning of autonomous vehicles [75], learning for traffic control in simulation [76]- [78] in conjunction with transfer of learned policies from simulation to a scaled environment [79], [80], decentralized learning for stochastic games [81], learning for optimal social routing [82] and congestion games [83], and learning for enhanced security against replay attacks in CPS [84], [85].…”
Section: B Related Workmentioning
confidence: 99%
“…There have also been research efforts on developing learning approaches using Bayesian analysis to address such problems [4]. Other approaches over the years have focused on direct or indirect RL methods including robust learning-based [5], [6], learning-based model predictive control [7]- [9] on autonomous racing cars [10], real-time learning [11], [12] of powertain systems with respect to the driver's driving style [13], [14], learning for traffic control [15] for transferring optimal policies [16], [17], decentralized learning for stochastic games [18], learning for optimal social routing [19] and congestion games [20], and learning for enhanced security against replay attacks in cyber-physical systems [21], [22].…”
Section: Introductionmentioning
confidence: 99%
“…For example, Zhu [2019, 2021] investigate the impact of manipulating Q-learners through the falsification of cost signals and its effect on the algorithm's convergence. Correspondingly, robust variants of the Q-learning against such attacks have been studied extensively [Sahoo and Vamvoudakis, 2020, Nisioti et al, 2021, Wang et al, 2020.…”
Section: Introductionmentioning
confidence: 99%