The importance of interpreting machine learning models for blood glucose prediction in diabetes: an analysis using SHAP

Sci Rep. 2023 Oct 6;13(1):16865. doi: 10.1038/s41598-023-44155-x.

Abstract

Machine learning has become a popular tool for learning models of complex dynamics from biomedical data. In Type 1 Diabetes (T1D) management, these models are increasingly been integrated in decision support systems (DSS) to forecast glucose levels and provide preventive therapeutic suggestions, like corrective insulin boluses (CIB), accordingly. Typically, models are chosen based on their prediction accuracy. However, since patient safety is a concern in this application, the algorithm should also be physiologically sound and its outcome should be explainable. This paper aims to discuss the importance of using tools to interpret the output of black-box models in T1D management by presenting a case-of-study on the selection of the best prediction algorithm to integrate in a DSS for CIB suggestion. By retrospectively "replaying" real patient data, we show that two long-short term memory neural networks (LSTM) (named p-LSTM and np-LSTM) with similar prediction accuracy could lead to different therapeutic decisions. An analysis with SHAP-a tool for explaining black-box models' output-unambiguously shows that only p-LSTM learnt the physiological relationship between inputs and glucose prediction, and should therefore be preferred. This is verified by showing that, when embedded in the DSS, only p-LSTM can improve patients' glycemic control.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Blood Glucose* / analysis
  • Diabetes Mellitus, Type 1*
  • Humans
  • Insulin / therapeutic use
  • Machine Learning
  • Neural Networks, Computer
  • Retrospective Studies

Substances

  • Blood Glucose
  • Insulin