TY - JOUR
T1 - STQS
T2 - Interpretable multi-modal Spatial-Temporal-seQuential model for automatic Sleep scoring
AU - Pathak, Shreyasi
AU - Lu, Changqing
AU - Belur Nagaraj, Sunil
AU - van Putten, Michel J.A.M.
AU - Seifert, Christin
N1 - Elsevier deal
Funding Information:
This work was supported by Pioneers in Healthcare Innovation Fund 2017 fortheproject “DeepSleep”awarded by Menzis Health Insurance and partially funded by Faculty of EEMCS, University of Twente . The authors would like to thank Mike van Klooster, Thomas Oosterveld, Mirjam Stappenbelt-Groot Kormelink and Marleen Tjepkema-Cloostermans from Medisch Spectrum Twente for very helpful discussions on sleep scoring guidelines and data collection from the hospital; Xenia Hoppenbrouwer and Ainara Garde from University of Twente for very helpful discussions related to processing the signals and for guidance at the start of the project; and Jörg Schlötterer for improving the paper during the review process.
Publisher Copyright:
© 2021 The Author(s)
PY - 2021/4/1
Y1 - 2021/4/1
N2 - Sleep scoring is an important step for the detection of sleep disorders and usually performed by visual analysis. Since manual sleep scoring is time consuming, machine-learning based approaches have been proposed. Though efficient, these algorithms are black-box in nature and difficult to interpret by clinicians. In this paper, we propose a deep learning architecture for multi-modal sleep scoring, investigate the model's decision making process, and compare the model's reasoning with the annotation guidelines in the AASM manual. Our architecture, called STQS, uses convolutional neural networks (CNN) to automatically extract spatio-temporal features from 3 modalities (EEG, EOG and EMG), a bidirectional long short-term memory (Bi-LSTM) to extract sequential information, and residual connections to combine spatio-temporal and sequential features. We evaluated our model on two large datasets, obtaining an accuracy of 85% and 77% and a macro F1 score of 79% and 73% on SHHS and an in-house dataset, respectively. We further quantify the contribution of various architectural components and conclude that adding LSTM layers improves performance over a spatio-temporal CNN, while adding residual connections does not. Our interpretability results show that the output of the model is well aligned with AASM guidelines, and therefore, the model's decisions correspond to domain knowledge. We also compare multi-modal models and single-channel models and suggest that future research should focus on improving multi-modal models.
AB - Sleep scoring is an important step for the detection of sleep disorders and usually performed by visual analysis. Since manual sleep scoring is time consuming, machine-learning based approaches have been proposed. Though efficient, these algorithms are black-box in nature and difficult to interpret by clinicians. In this paper, we propose a deep learning architecture for multi-modal sleep scoring, investigate the model's decision making process, and compare the model's reasoning with the annotation guidelines in the AASM manual. Our architecture, called STQS, uses convolutional neural networks (CNN) to automatically extract spatio-temporal features from 3 modalities (EEG, EOG and EMG), a bidirectional long short-term memory (Bi-LSTM) to extract sequential information, and residual connections to combine spatio-temporal and sequential features. We evaluated our model on two large datasets, obtaining an accuracy of 85% and 77% and a macro F1 score of 79% and 73% on SHHS and an in-house dataset, respectively. We further quantify the contribution of various architectural components and conclude that adding LSTM layers improves performance over a spatio-temporal CNN, while adding residual connections does not. Our interpretability results show that the output of the model is well aligned with AASM guidelines, and therefore, the model's decisions correspond to domain knowledge. We also compare multi-modal models and single-channel models and suggest that future research should focus on improving multi-modal models.
KW - UT-Hybrid-D
KW - Sleep stage annotation
KW - Deep learning
KW - EEG, EOG, EMG signals
KW - Post-hoc interpretability
KW - Explainable AI
KW - Sleep scoring
UR - https://github.com/ShreyasiPathak/STQS
U2 - 10.1016/j.artmed.2021.102038
DO - 10.1016/j.artmed.2021.102038
M3 - Article
SN - 0933-3657
VL - 114
JO - Artificial intelligence in medicine
JF - Artificial intelligence in medicine
M1 - 102038
ER -