2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2020
DOI: 10.1109/ipdpsw50202.2020.00024
|View full text |Cite
|
Sign up to set email alerts
|

QTAccel: A Generic FPGA based Design for Q-Table based Reinforcement Learning Accelerators

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
2

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 13 publications
(23 citation statements)
references
References 20 publications
0
21
2
Order By: Relevance
“…Instead of integrating the environment into the learning system, states and rewards are seen as input to the hardware accelerator. The architecture contains one on-chip memory block per action in [48] the action space, which stores the Q-values of the respective action for each state. For a given state, the Q-values for all actions can be read at the same time.…”
Section: A Tabular Reinforcement Learningmentioning
confidence: 99%
See 4 more Smart Citations
“…Instead of integrating the environment into the learning system, states and rewards are seen as input to the hardware accelerator. The architecture contains one on-chip memory block per action in [48] the action space, which stores the Q-values of the respective action for each state. For a given state, the Q-values for all actions can be read at the same time.…”
Section: A Tabular Reinforcement Learningmentioning
confidence: 99%
“…Current publications focus on the implementation of Q-Learning but also suggest modifications to their proposed architectures to support SARSA [67,72,48]. The following paragraphs present similarities and differences between these architectures.…”
Section: A Classical Reinforcement Learningmentioning
confidence: 99%
See 3 more Smart Citations