Comparing Deterministic and Stochastic Reinforcement Learning for Glucose Regulation in Type 1 Diabetes
| dc.contributor.author | Timms, David | |
| dc.contributor.author | Hettiarachchi, Chirath | |
| dc.contributor.author | Suominen, Hanna | |
| dc.contributor.organization | fi=tietotekniikan laitos|en=Department of Computing| | |
| dc.contributor.organization-code | 1.2.246.10.2458963.20.85312822902 | |
| dc.converis.publication-id | 499745855 | |
| dc.converis.url | https://research.utu.fi/converis/portal/Publication/499745855 | |
| dc.date.accessioned | 2026-01-21T14:48:39Z | |
| dc.date.available | 2026-01-21T14:48:39Z | |
| dc.description.abstract | Type 1 Diabetes (T1D) is a chronic condition affecting millions worldwide, requiring external insulin administration to regulate blood glucose levels and prevent serious complications. Artificial Pancreas Systems (APS) for managing T1D currently rely on manual input, which adds a cognitive burden on people with T1D and their carers. Research into alleviating this burden through Reinforcement Learning (RL) explores enabling the APS to autonomously learn and adapt to the complex dynamics of blood glucose regulation, demonstrating improvements in in-silico evaluations compared to traditional clinical approaches. This evaluation study compared the primary polarities of RL for glucose regulation, namely, stochastic (e.g., Proximal Policy Optimization (PPO) and deterministic (e.g., Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithms in-silico using quantitative and qualitative methods, patient specific clinical metrics, and the adult and adolescent cohorts of the U.S. Food and Drug Administration approved UVA/PADOVA 2008 model. Although the behavior of TD3 was easier to interpret, it did not typically outperform PPO, thereby challenging assessing their safety and suitability. This conclusion highlights the importance of improving RL algorithms in APS applications for both interpretability and predictive performance in future research. | |
| dc.format.pagerange | 1039 | |
| dc.format.pagerange | 1043 | |
| dc.identifier.eisbn | 978-1-64368-608-0 | |
| dc.identifier.issn | 0926-9630 | |
| dc.identifier.jour-issn | 0926-9630 | |
| dc.identifier.olddbid | 213730 | |
| dc.identifier.oldhandle | 10024/196748 | |
| dc.identifier.uri | https://www.utupub.fi/handle/11111/55792 | |
| dc.identifier.url | https://doi.org/10.3233/shti250997 | |
| dc.identifier.urn | URN:NBN:fi-fe202601216972 | |
| dc.language.iso | en | |
| dc.okm.affiliatedauthor | Suominen, Hanna | |
| dc.okm.discipline | 113 Computer and information sciences | en_GB |
| dc.okm.discipline | 217 Medical engineering | en_GB |
| dc.okm.discipline | 3121 Internal medicine | en_GB |
| dc.okm.discipline | 113 Tietojenkäsittely ja informaatiotieteet | fi_FI |
| dc.okm.discipline | 217 Lääketieteen tekniikka | fi_FI |
| dc.okm.discipline | 3121 Sisätaudit | fi_FI |
| dc.okm.internationalcopublication | international co-publication | |
| dc.okm.internationality | International publication | |
| dc.okm.type | A4 Conference Article | |
| dc.publisher.country | Netherlands | en_GB |
| dc.publisher.country | Alankomaat | fi_FI |
| dc.publisher.country-code | NL | |
| dc.relation.conference | World Congress on Medical and Health Informatics | |
| dc.relation.doi | 10.3233/SHTI250997 | |
| dc.relation.ispartofjournal | Studies in Health Technology and Informatics | |
| dc.relation.volume | 329 | |
| dc.source.identifier | https://www.utupub.fi/handle/10024/196748 | |
| dc.title | Comparing Deterministic and Stochastic Reinforcement Learning for Glucose Regulation in Type 1 Diabetes | |
| dc.title.book | MEDINFO 2025 — Healthcare Smart × Medicine Deep: Proceedings of the 20th World Congress on Medical and Health Informatics | |
| dc.year.issued | 2025 |
Tiedostot
1 - 1 / 1