2021
DOI: 10.1016/j.neucom.2021.06.064
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Reinforcement Learning using EEG-based implicit human feedback

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(9 citation statements)
references
References 33 publications
0
9
0
Order By: Relevance
“…When comparing the BGS tasks, BGSInt performs the best for both EEGNet (g * =0.77, CI.95[0.47, 1.08]) with a medium to large ES and SVM (g * =0.56, CI.95[0.26, 0.86]) with a small to large ES. Similarly, when comparing the OA tasks, OAOut performs best for EEGNet (g * =−1.82, CI.95[−2.16, −1.47]) 8 and SVM (g * =−1.66, CI.95[−2.0, −1.32]) with large ESs. Similar results to bACC can be observed for TNR (Fig.…”
Section: Resultsmentioning
confidence: 94%
See 1 more Smart Citation
“…When comparing the BGS tasks, BGSInt performs the best for both EEGNet (g * =0.77, CI.95[0.47, 1.08]) with a medium to large ES and SVM (g * =0.56, CI.95[0.26, 0.86]) with a small to large ES. Similarly, when comparing the OA tasks, OAOut performs best for EEGNet (g * =−1.82, CI.95[−2.16, −1.47]) 8 and SVM (g * =−1.66, CI.95[−2.0, −1.32]) with large ESs. Similar results to bACC can be observed for TNR (Fig.…”
Section: Resultsmentioning
confidence: 94%
“…ErrPs are generally characterized by a negative peak, often referred to as error-related negativity (ERN), that occurs between 80-300 ms and a subsequent positive peak (Pe) that occurs between 200-500 ms [1], [2], [4]- [6]. Recently, ErrPs have been of particular interest for BCI applications, ranging from simple detection/classification of errors [5]- [7], to training robots and agents [8], [9], to correcting mistakes made by BCI-controlled interfaces [10].…”
Section: Introductionmentioning
confidence: 99%
“…ErrP has been widely used in human-in-the-loop RL systems. In these systems, the ErrP signal is used as a positive or negative reward to accelerate the training of autonomous agents [31][32][33]. In [19], ErrPs was used as negative reinforcers of the actions to infer the optimal control strategies.…”
Section: Errp-based Human-in-the-loop Rlmentioning
confidence: 99%
“…In [2], ErrP was used to train a robot to learn human gestures through a reinforcement learning strategy based on the leap motion and ErrP features. However, when the ErrP signal was used as a reward, while the ErrPs accelerated learning, the signals operated independently of the system during testing [31][32][33].…”
Section: Errp-based Human-in-the-loop Rlmentioning
confidence: 99%
See 1 more Smart Citation