2017
DOI: 10.1109/tnnls.2015.2493079
|View full text |Cite
|
Sign up to set email alerts
|

Quantized Attention-Gated Kernel Reinforcement Learning for Brain–Machine Interface Decoding

Abstract: Reinforcement learning (RL)-based decoders in brain-machine interfaces (BMIs) interpret dynamic neural activity without patients' real limb movements. In conventional RL, the goal state is selected by the user or defined by the physics of the problem, and the decoder finds an optimal policy essentially by assigning credit over time, which is normally very time-consuming. However, BMI tasks require finding a good policy in very few trials, which impose a limit on the complexity of the tasks that can be learned … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

1
29
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
2
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 38 publications
(30 citation statements)
references
References 42 publications
1
29
0
Order By: Relevance
“…Some studies have implemented this method and have achieved good decoding performance; however, most of them commonly employ supervised learning and train the decoder by mapping the recorded neural activities to some kinematic outputs, such as the real movement trajectory or the movement labels [ 25 , 26 , 27 ]. In clinical applications, the kinematic outputs may be difficult to collect, particularly for paralysis or limb amputations [ 28 , 29 , 30 , 31 ]. To address this problem, reinforcement learning (RL)-based iBMIs have been developed.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Some studies have implemented this method and have achieved good decoding performance; however, most of them commonly employ supervised learning and train the decoder by mapping the recorded neural activities to some kinematic outputs, such as the real movement trajectory or the movement labels [ 25 , 26 , 27 ]. In clinical applications, the kinematic outputs may be difficult to collect, particularly for paralysis or limb amputations [ 28 , 29 , 30 , 31 ]. To address this problem, reinforcement learning (RL)-based iBMIs have been developed.…”
Section: Introductionmentioning
confidence: 99%
“…However, AGREL is sensitive to initialization and has to re-initialize numerous times to avoid becoming stuck in a poor performance, which affects the real-time capabilities and prohibits it from meeting the requirements of online decoding. Moreover, disparities between historical and new data caused by nonstationarity enlarge the state-action space and bring about more challenges to the exploration efficiency [ 29 , 30 ].…”
Section: Introductionmentioning
confidence: 99%
“…The limitation of this approach is that the large neural state-action pair gives rise to the curse of dimensionality problem leading to generalization difficulty. To overcome this problem, [17] proposed applying Attention-gated reinforcement learning (AGREL) and its variants [18], [19]. [17] reports improvement over Q-learning involving one NHP performing a center-out cursor task.…”
Section: Introductionmentioning
confidence: 99%
“…This framework enables a direct mapping of input spike timings into an RKHS. The kernel methods have already been applied in mapping binary input spike trains or spike counts into an RKHS to decode continuous movement trajectories or motor states (Bae, Chhatbar, Francis, Sanchez, & Príncipe, 2011;Wang et al, 2017;Zhang, Príncipe, & Wang, 2018), but they have not been developed to output discrete spike or designed to directly operate on input spike timings.…”
mentioning
confidence: 99%