2020
DOI: 10.1007/978-3-030-53352-6_5
|View full text |Cite
|
Sign up to set email alerts
|

Personalized Dual-Hormone Control for Type 1 Diabetes Using Deep Reinforcement Learning

Abstract: We propose a dual-hormone control algorithm by exploiting deep reinforcement learning (RL) for people with Type 1 Diabetes (T1D). Specifically, double dilated recurrent neural networks are used to learn the hormone delivery strategy, trained by a variant of Q-learning, whose inputs are raw data of glucose & meal carbohydrate and outputs are the actions to deliver dual-hormone (basal insulin and glucagon). Without prior knowledge of the glucose-insulin metabolism, we develop the data-driven model in the UVA/Pad… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2
1

Relationship

3
7

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 17 publications
0
8
0
Order By: Relevance
“…However, a dataset with a much larger sample size such as the Tidepool Big Data Donation Dataset [57], may be required to capture the effects of inter-subject variability. Finally, FCNN can be used as an encoder in deep reinforcement learning to extract hidden features from physiological environment and improve decision support systems and automated insulin delivery algorithms (e.g., artificial pancreas) [66]- [68].…”
Section: Discussionmentioning
confidence: 99%
“…However, a dataset with a much larger sample size such as the Tidepool Big Data Donation Dataset [57], may be required to capture the effects of inter-subject variability. Finally, FCNN can be used as an encoder in deep reinforcement learning to extract hidden features from physiological environment and improve decision support systems and automated insulin delivery algorithms (e.g., artificial pancreas) [66]- [68].…”
Section: Discussionmentioning
confidence: 99%
“…This state was selected in place of the full sequence of blood glucose, carbohydrate and insulin data utilised in other approaches to avoid modifying the offline RL methods to incorporate recurrency and to reduce state dimensionality [13,12,44]. The reward for the agent was given by the negative of the Magni risk function, which models the clinical risk for a given blood glucose value [22].…”
Section: Problem Formulationmentioning
confidence: 99%
“…As far as difficulty is concerned, HeartPole ( Härmä et al, 2021 ), simglucose ( Zhu et al, 2021 ), and GYMIC ( Kiani et al, 2019 ) are known to be solvable with relatively small models and standard reinforcement learning algorithms like DQN ( Mnih et al, 2015 ). Thus, the only simulators difficult enough to be benchmarks for novel approaches are Virtu-ALS and Auto-ALS and Auto-ALS is the more accessible of the two.…”
Section: Effectivenessmentioning
confidence: 99%