1
|
Melo MC, da Silva Vallim JR, Garbuio S, Soster LA, Sousa KMM, Bonaldi RR, Pires GN. Validation of a sleep staging classification model for healthy adults based on two combinations of a single-channel EEG headband and wrist actigraphy. J Clin Sleep Med 2024; 20:983-990. [PMID: 38427322 PMCID: PMC11145037 DOI: 10.5664/jcsm.11082] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2023] [Revised: 02/16/2024] [Accepted: 02/16/2024] [Indexed: 03/02/2024]
Abstract
STUDY OBJECTIVES The aim of this study was to develop a sleep staging classification model capable of accurately performing on different wearable devices. METHODS Twenty-three healthy participants underwent a full-night type I polysomnography and used two device combinations: (A) flexible single-channel electroencephalogram (EEG) headband + actigraphy (n = 12) and (B) rigid single-channel EEG headband + actigraphy (n = 11). The signals were segmented into 30-second epochs according to polysomnographic stages (scored by a board-certified sleep technologist; model ground truth) and 18 frequency and time features were extracted. The model consisted of an ensemble of bagged decision trees. Bagging refers to bootstrap aggregation to reduce overfitting and improve generalization. To evaluate the model, a training dataset under 5-fold cross-validation and an 80-20% dataset split was used. The headbands were also evaluated without the actigraphy feature. Participants also completed a usability evaluation (comfort, pain while sleeping, and sleep disturbance). RESULTS Combination A had an F1-score of 98.4% and the flexible headband alone of 97.7% (error rate for N1: combination A = 9.8%; flexible headband alone = 15.7%). Combination B had an F1-score of 96.9% and the rigid headband alone of 95.3% (error rate for N1: combination B = 17.0%; rigid headband alone = 27.7%); in both, N1 was more confounded with N2. CONCLUSIONS We developed an accurate sleep classification model based on a single-channel EEG device, and actigraphy was not an important feature of the model. Both headbands were found to be useful, with the rigid one being more disruptive to sleep. Future research can improve our results by applying the developed model in a population with sleep disorders. CLINICAL TRIAL REGISTRATION Registry: ClinicalTrials.gov; Name: Actigraphy, Wearable EEG Band and Smartphone for Sleep Staging; URL: https://clinicaltrials.gov/study/NCT04943562; Identifier: NCT04943562. CITATION Melo MC, Vallim JRS, Garbuio S, et al. Validation of a sleep staging classification model for healthy adults based on 2 combinations of a single-channel EEG headband and wrist actigraphy. J Clin Sleep Med. 2024;20(6):983-990.
Collapse
Affiliation(s)
- Mariana Cardoso Melo
- Department of Psychobiology, Escola Paulista de Medicina, Universidade Federal de São Paulo, Sao Paulo, Brazil
- SleepUp Tecnologia em Saúde Ltda, São Caetano do Sul, Brazil
| | - Julia Ribeiro da Silva Vallim
- Department of Psychobiology, Escola Paulista de Medicina, Universidade Federal de São Paulo, Sao Paulo, Brazil
- SleepUp Tecnologia em Saúde Ltda, São Caetano do Sul, Brazil
| | | | - Leticia Azevedo Soster
- SleepUp Tecnologia em Saúde Ltda, São Caetano do Sul, Brazil
- Hospital das Clínicas, Universidade de São Paulo, Sao Paulo, Brazil
| | | | | | - Gabriel Natan Pires
- Department of Psychobiology, Escola Paulista de Medicina, Universidade Federal de São Paulo, Sao Paulo, Brazil
- SleepUp Tecnologia em Saúde Ltda, São Caetano do Sul, Brazil
| |
Collapse
|
2
|
Ge J, Xu G, Lu J, Xu X, Li L, Meng X. SensorNet: An Adaptive Attention Convolutional Neural Network for Sensor Feature Learning. SENSORS (BASEL, SWITZERLAND) 2024; 24:3274. [PMID: 38894067 PMCID: PMC11174822 DOI: 10.3390/s24113274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/18/2024] [Revised: 05/14/2024] [Accepted: 05/19/2024] [Indexed: 06/21/2024]
Abstract
This work develops a generalizable neural network, SENSORNET, for sensor feature learning across various applications. The primary challenge addressed is the poor portability of pretrained neural networks to new applications with limited sensor data. To solve this challenge, we design SensorNet, which integrates the flexibility of self-attention with multi-scale feature locality of convolution. Moreover, we invent patch-wise self-attention with stacked multi-heads to enrich the sensor feature representation. SensorNet is generalizable to pervasive applications with any number of sensor inputs, and is much smaller than the state-of-the-art self-attention and convolution hybrid baseline (0.83 M vs. 3.87 M parameters) with similar performance. The experimental results show that SensorNet is able to achieve state-of-the-art performance compared with the top five models on a competition activity recognition dataset (SHL'18). Moreover, pretrained SensorNet in a large inertial measurement unit (IMU) dataset can be fine-tuned to achieve the best accuracy on a much smaller IMU dataset (up to 5% improvement in WISDM) and to achieve the state-of-the-art performance on an EEG dataset (SLEEP-EDF-20), showing the strong generalizability of our approach.
Collapse
Affiliation(s)
- Jiaqi Ge
- Department of Computer Science and Technology, Jilin University, Changchun 130012, China; (J.G.); (G.X.); (X.X.); (L.L.)
| | - Gaochao Xu
- Department of Computer Science and Technology, Jilin University, Changchun 130012, China; (J.G.); (G.X.); (X.X.); (L.L.)
| | - Jianchao Lu
- School of Computing, Macquarie University, Sydney, NSW 2109, Australia;
| | - Xu Xu
- Department of Computer Science and Technology, Jilin University, Changchun 130012, China; (J.G.); (G.X.); (X.X.); (L.L.)
| | - Long Li
- Department of Computer Science and Technology, Jilin University, Changchun 130012, China; (J.G.); (G.X.); (X.X.); (L.L.)
| | - Xiangyu Meng
- Department of Computer Science and Technology, Jilin University, Changchun 130012, China; (J.G.); (G.X.); (X.X.); (L.L.)
| |
Collapse
|
3
|
Zaman A, Kumar S, Shatabda S, Dehzangi I, Sharma A. SleepBoost: a multi-level tree-based ensemble model for automatic sleep stage classification. Med Biol Eng Comput 2024:10.1007/s11517-024-03096-x. [PMID: 38700613 DOI: 10.1007/s11517-024-03096-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2023] [Accepted: 04/14/2024] [Indexed: 05/16/2024]
Abstract
Neurodegenerative diseases often exhibit a strong link with sleep disruption, highlighting the importance of effective sleep stage monitoring. In this light, automatic sleep stage classification (ASSC) plays a pivotal role, now more streamlined than ever due to the advancements in deep learning (DL). However, the opaque nature of DL models can be a barrier in their clinical adoption, due to trust concerns among medical practitioners. To bridge this gap, we introduce SleepBoost, a transparent multi-level tree-based ensemble model specifically designed for ASSC. Our approach includes a crafted feature engineering block (FEB) that extracts 41 time and frequency domain features, out of which 23 are selected based on their high mutual information score (> 0.23). Uniquely, SleepBoost integrates three fundamental linear models into a cohesive multi-level tree structure, further enhanced by a novel reward-based adaptive weight allocation mechanism. Tested on the Sleep-EDF-20 dataset, SleepBoost demonstrates superior performance with an accuracy of 86.3%, F1-score of 80.9%, and Cohen kappa score of 0.807, outperforming leading DL models in ASSC. An ablation study underscores the critical role of our selective feature extraction in enhancing model accuracy and interpretability, crucial for clinical settings. This innovative approach not only offers a more transparent alternative to traditional DL models but also extends potential implications for monitoring and understanding sleep patterns in the context of neurodegenerative disorders. The open-source availability of SleepBoost's implementation at https://github.com/akibzaman/SleepBoost can further facilitate its accessibility and potential for widespread clinical adoption.
Collapse
Affiliation(s)
- Akib Zaman
- Computer Science and Artificial Intelligence Laboratory (CSAIL), Electrical Engineering and Computer Science Department, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Shiu Kumar
- School of Electrical & Electronics Engineering, Fiji National University, Suva, Fiji.
| | - Swakkhar Shatabda
- Centre for Artificial Intelligence and Robotics (CAIR), United International University, Dhaka, Bangladesh
| | - Iman Dehzangi
- Department of Computer Science, Rutgers University, Camden, NJ, USA
- Center for Computational and Integrative Biology, Rutgers University, Camden, USA
| | - Alok Sharma
- Laboratory for Medical Science Mathematics, RIKEN Center for Integrative Medical Sciences, Yokohama, 230-0045, Japan
- Institute for Integrated and Intelligent Systems, Griffith University, Nathan, Brisbane, QLD, Australia
| |
Collapse
|
4
|
Jirakittayakorn N, Wongsawat Y, Mitrirattanakul S. ZleepAnlystNet: a novel deep learning model for automatic sleep stage scoring based on single-channel raw EEG data using separating training. Sci Rep 2024; 14:9859. [PMID: 38684765 PMCID: PMC11058251 DOI: 10.1038/s41598-024-60796-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 04/26/2024] [Indexed: 05/02/2024] Open
Abstract
Numerous models for sleep stage scoring utilizing single-channel raw EEG signal have typically employed CNN and BiLSTM architectures. While these models, incorporating temporal information for sequence classification, demonstrate superior overall performance, they often exhibit low per-class performance for N1-stage, necessitating an adjustment of loss function. However, the efficacy of such adjustment is constrained by the training process. In this study, a pioneering training approach called separating training is introduced, alongside a novel model, to enhance performance. The developed model comprises 15 CNN models with varying loss function weights for feature extraction and 1 BiLSTM for sequence classification. Due to its architecture, this model cannot be trained using an end-to-end approach, necessitating separate training for each component using the Sleep-EDF dataset. Achieving an overall accuracy of 87.02%, MF1 of 82.09%, Kappa of 0.8221, and per-class F1-socres (W 90.34%, N1 54.23%, N2 89.53%, N3 88.96%, and REM 87.40%), our model demonstrates promising performance. Comparison with sleep technicians reveals a Kappa of 0.7015, indicating alignment with reference sleep stags. Additionally, cross-dataset validation and adaptation through training with the SHHS dataset yield an overall accuracy of 84.40%, MF1 of 74.96% and Kappa of 0.7785 when tested with the Sleep-EDF-13 dataset. These findings underscore the generalization potential in model architecture design facilitated by our novel training approach.
Collapse
Affiliation(s)
- Nantawachara Jirakittayakorn
- Institute for Innovative Learning, Mahidol University, Nakhon Pathom, Thailand
- Faculty of Dentistry, Mahidol University, Bangkok, Thailand
| | - Yodchanan Wongsawat
- Department of Biomedical Engineering, Faculty of Engineering, Mahidol University, Nakhon Pathom, Thailand
| | - Somsak Mitrirattanakul
- Department of Masticatory Science, Faculty of Dentistry, Mahidol University, Bangkok, Thailand.
| |
Collapse
|
5
|
Radke FA, da Silva Souto CF, Pätzold W, Wolf KI. Transfer Learning for Automatic Sleep Staging Using a Pre-Gelled Electrode Grid. Diagnostics (Basel) 2024; 14:909. [PMID: 38732323 PMCID: PMC11083934 DOI: 10.3390/diagnostics14090909] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Revised: 04/18/2024] [Accepted: 04/24/2024] [Indexed: 05/13/2024] Open
Abstract
Novel sensor solutions for sleep monitoring at home could alleviate bottlenecks in sleep medical care as well as enable selective or continuous observation over long periods of time and contribute to new insights in sleep medicine and beyond. Since especially in the latter case the sensor data differ strongly in signal, number and extent of sensors from the classical polysomnography (PSG) sensor technology, an automatic evaluation is essential for the application. However, the training of an automatic algorithm is complicated by the fact that the development phase of the new sensor technology, extensive comparative measurements with standardized reference systems, is often not possible and therefore only small datasets are available. In order to circumvent high system-specific training data requirements, we employ pre-training on large datasets with finetuning on small datasets of new sensor technology to enable automatic sleep phase detection for small test series. By pre-training on publicly available PSG datasets and finetuning on 12 nights recorded with new sensor technology based on a pre-gelled electrode grid to capture electroencephalography (EEG), electrooculography (EOG) and electromyography (EMG), an F1 score across all sleep phases of 0.81 is achieved (wake 0.84, N1 0.62, N2 0.81, N3 0.87, REM 0.88), using only EEG and EOG. The analysis additionally considers the spatial distribution of the channels and an approach to approximate classical electrode positions based on specific linear combinations of the new sensor grid channels.
Collapse
Affiliation(s)
- Fabian A. Radke
- Fraunhofer Institute for Digital Media Technology IDMT, Oldenburg Branch for Hearing, Speech and Audio Technology HSA, 26129 Oldenburg, Germany
| | | | | | | |
Collapse
|
6
|
Yue H, Chen Z, Guo W, Sun L, Dai Y, Wang Y, Ma W, Fan X, Wen W, Lei W. Research and application of deep learning-based sleep staging: Data, modeling, validation, and clinical practice. Sleep Med Rev 2024; 74:101897. [PMID: 38306788 DOI: 10.1016/j.smrv.2024.101897] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Revised: 12/30/2023] [Accepted: 01/04/2024] [Indexed: 02/04/2024]
Abstract
Over the past few decades, researchers have attempted to simplify and accelerate the process of sleep stage classification through various approaches; however, only a few such approaches have gained widespread acceptance. Artificial intelligence technology, particularly deep learning, is promising for earning the trust of the sleep medicine community in automated sleep-staging systems, thus facilitating its application in clinical practice and integration into daily life. We aimed to comprehensively review the latest methods that are applying deep learning for enhancing sleep staging efficiency and accuracy. Starting from the requisite "data" for constructing deep learning algorithms, we elucidated the current landscape of this domain and summarized the fundamental modeling process, encompassing signal selection, data pre-processing, model architecture, classification tasks, and performance metrics. Furthermore, we reviewed the applications of automated sleep staging in scenarios such as sleep-disorder screening, diagnostic procedures, and health monitoring and management. Finally, we conducted an in-depth analysis and discussion of the challenges and future in intelligent sleep staging, particularly focusing on large-scale sleep datasets, interdisciplinary collaborations, and human-computer interactions.
Collapse
Affiliation(s)
- Huijun Yue
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China
| | - Zhuqi Chen
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China
| | - Wenbin Guo
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China
| | - Lin Sun
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China
| | - Yidan Dai
- School of Computer Science, South China Normal University, Guangzhou, People's Republic of China
| | - Yiming Wang
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China
| | - Wenjun Ma
- School of Computer Science, South China Normal University, Guangzhou, People's Republic of China
| | - Xiaomao Fan
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, People's Republic of China
| | - Weiping Wen
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China; Department of Otolaryngology, The Sixth Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China.
| | - Wenbin Lei
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, People's Republic of China.
| |
Collapse
|
7
|
Heremans ERM, Seedat N, Buyse B, Testelmans D, van der Schaar M, De Vos M. U-PASS: An uncertainty-guided deep learning pipeline for automated sleep staging. Comput Biol Med 2024; 171:108205. [PMID: 38401452 DOI: 10.1016/j.compbiomed.2024.108205] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Revised: 02/16/2024] [Accepted: 02/20/2024] [Indexed: 02/26/2024]
Abstract
With the increasing prevalence of machine learning in critical fields like healthcare, ensuring the safety and reliability of these systems is crucial. Estimating uncertainty plays a vital role in enhancing reliability by identifying areas of high and low confidence and reducing the risk of errors. This study introduces U-PASS, a specialized human-centered machine learning pipeline tailored for clinical applications, which effectively communicates uncertainty to clinical experts and collaborates with them to improve predictions. U-PASS incorporates uncertainty estimation at every stage of the process, including data acquisition, training, and model deployment. Training is divided into a supervised pre-training step and a semi-supervised recording-wise finetuning step. We apply U-PASS to the challenging task of sleep staging and demonstrate that it systematically improves performance at every stage. By optimizing the training dataset, actively seeking feedback from domain experts for informative samples, and deferring the most uncertain samples to experts, U-PASS achieves an impressive expert-level accuracy of 85% on a challenging clinical dataset of elderly sleep apnea patients. This represents a significant improvement over the starting point at 75% accuracy. The largest improvement gain is due to the deferral of uncertain epochs to a sleep expert. U-PASS presents a promising AI approach to incorporating uncertainty estimation in machine learning pipelines, improving their reliability and unlocking their potential in clinical settings.
Collapse
Affiliation(s)
- Elisabeth R M Heremans
- KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Kasteelpark Arenberg 10, B-3001 Leuven, Belgium.
| | | | - Bertien Buyse
- UZ Leuven, Department of Pneumology, Herestraat 49, B-3000 Leuven, Belgium
| | - Dries Testelmans
- UZ Leuven, Department of Pneumology, Herestraat 49, B-3000 Leuven, Belgium
| | | | - Maarten De Vos
- KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Kasteelpark Arenberg 10, B-3001 Leuven, Belgium.
| |
Collapse
|
8
|
Jain R, G RA. Modality-Specific Feature Selection, Data Augmentation and Temporal Context for Improved Performance in Sleep Staging. IEEE J Biomed Health Inform 2024; 28:1031-1042. [PMID: 38051608 DOI: 10.1109/jbhi.2023.3339713] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
This work attempts to design an effective sleep staging system, making the best use of the available signals, strategies, and features in the literature. It must not only perform well on different datasets comprising healthy and clinical populations but also achieve good accuracy in cross-dataset experiments. Toward this end, we propose a model comprising multiple binary classifiers in a hierarchical fashion, where, at each level, one or more of EEG, EOG, and EMG are selected to best differentiate between two sleep stages. The best set of 100 features is chosen out of all the features derived from selected signals. The class imbalance in data is addressed by random undersampling and boosting techniques with decision trees as weak learners. Temporal context and data augmentation are used to improve the performance. We also evaluate the performance of our model by training and testing on different datasets. We compare the results of five approaches: using only EEG, EEG+EOG, EEG+EMG+EOG, EEG+EMG, and selective modality with a specific combination of EEG, EMG, and/or EOG at each level. The best results are obtained by considering features from EEG+EMG+EOG at each hierarchical level. The proposed model achieves average accuracies of 83.1%, 90.0%, 84.4%, 82.1%, 81.5%, 79.9%, and 73.7% on Sleep-EDF, Exp Sleep-EDF, ISRUC-S1, S2 and S3, DRMS-SUB, and DRMS-PAT datasets, respectively. For all the datasets except DRMS-SUB, the proposed method outperforms all the state-of-the-art approaches. Cross-dataset performance exceeds 80% for all datasets except DRMS-PAT; independent of whether the test data is from normal subjects or patients.
Collapse
|
9
|
Van Der Aar JF, Van Den Ende DA, Fonseca P, Van Meulen FB, Overeem S, Van Gilst MM, Peri E. Deep transfer learning for automated single-lead EEG sleep staging with channel and population mismatches. Front Physiol 2024; 14:1287342. [PMID: 38250654 PMCID: PMC10796543 DOI: 10.3389/fphys.2023.1287342] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Accepted: 12/08/2023] [Indexed: 01/23/2024] Open
Abstract
Introduction: Automated sleep staging using deep learning models typically requires training on hundreds of sleep recordings, and pre-training on public databases is therefore common practice. However, suboptimal sleep stage performance may occur from mismatches between source and target datasets, such as differences in population characteristics (e.g., an unrepresented sleep disorder) or sensors (e.g., alternative channel locations for wearable EEG). Methods: We investigated three strategies for training an automated single-channel EEG sleep stager: pre-training (i.e., training on the original source dataset), training-from-scratch (i.e., training on the new target dataset), and fine-tuning (i.e., training on the original source dataset, fine-tuning on the new target dataset). As source dataset, we used the F3-M2 channel of healthy subjects (N = 94). Performance of the different training strategies was evaluated using Cohen's Kappa (κ) in eight smaller target datasets consisting of healthy subjects (N = 60), patients with obstructive sleep apnea (OSA, N = 60), insomnia (N = 60), and REM sleep behavioral disorder (RBD, N = 22), combined with two EEG channels, F3-M2 and F3-F4. Results: No differences in performance between the training strategies was observed in the age-matched F3-M2 datasets, with an average performance across strategies of κ = .83 in healthy, κ = .77 in insomnia, and κ = .74 in OSA subjects. However, in the RBD set, where data availability was limited, fine-tuning was the preferred method (κ = .67), with an average increase in κ of .15 to pre-training and training-from-scratch. In the presence of channel mismatches, targeted training is required, either through training-from-scratch or fine-tuning, increasing performance with κ = .17 on average. Discussion: We found that, when channel and/or population mismatches cause suboptimal sleep staging performance, a fine-tuning approach can yield similar to superior performance compared to building a model from scratch, while requiring a smaller sample size. In contrast to insomnia and OSA, RBD data contains characteristics, either inherent to the pathology or age-related, which apparently demand targeted training.
Collapse
Affiliation(s)
- Jaap F. Van Der Aar
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Philips Research, Eindhoven, Netherlands
| | | | - Pedro Fonseca
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Philips Research, Eindhoven, Netherlands
| | - Fokke B. Van Meulen
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Kempenhaeghe Center for Sleep Medicine, Heeze, Netherlands
| | - Sebastiaan Overeem
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Kempenhaeghe Center for Sleep Medicine, Heeze, Netherlands
| | - Merel M. Van Gilst
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Kempenhaeghe Center for Sleep Medicine, Heeze, Netherlands
| | - Elisabetta Peri
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
| |
Collapse
|
10
|
Anido-Alonso A, Alvarez-Estevez D. Decentralized Data-Privacy Preserving Deep-Learning Approaches for Enhancing Inter-Database Generalization in Automatic Sleep Staging. IEEE J Biomed Health Inform 2023; 27:5610-5621. [PMID: 37651482 DOI: 10.1109/jbhi.2023.3310869] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/02/2023]
Abstract
Automatic sleep staging has been an active field of development. Despite multiple efforts, the area remains a focus of research interest. Indeed, while promising results have reported in past literature, uptake of automatic sleep scoring in the clinical setting remains low. One of the current issues regards the difficulty to generalization performance results beyond the local testing scenario, i.e. across data from different clinics. Issues derived from data-privacy restrictions, that generally apply in the medical domain, pose additional difficulties in the successful development of these methods. We propose the use of several decentralized deep-learning approaches, namely ensemble models and federated learning, for robust inter-database performance generalization and data-privacy preservation in automatic sleep staging scenario. Specifically, we explore four ensemble combination strategies (max-voting, output averaging, size-proportional weighting, and Nelder-Mead) and present a new federated learning algorithm, so-called sub-sampled federated stochastic gradient descent (ssFedSGD). To evaluate generalization capabilities of such approaches, experimental procedures are carried out using a leaving-one-database-out direct-transfer scenario on six independent and heterogeneous public sleep staging databases. The resulting performance is compared with respect to two baseline approaches involving single-database and centralized multiple-database derived models. Our results show that proposed decentralized learning methods outperform baseline local approaches, and provide similar generalization results to centralized database-combined approaches. We conclude that these methods are more preferable choices, as they come with additional advantages concerning improved scalability, flexible design, and data-privacy preservation.
Collapse
|
11
|
Liu Z, Qin M, Lu Y, Luo S, Zhang Q. DenSleepNet: DenseNet based model for sleep staging with two-frequency feature fusion and coordinate attention. Biomed Eng Lett 2023; 13:751-761. [PMID: 37872995 PMCID: PMC10590351 DOI: 10.1007/s13534-023-00301-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2023] [Revised: 06/15/2023] [Accepted: 06/23/2023] [Indexed: 10/25/2023] Open
Abstract
Sleep staging is often applied to assess the quality of sleep and also be used to prevent and monitor psychiatric disorders caused by sleep. However, it remains a challenge to extract the discriminative features of salient waveforms in sleep EEG and enable the network to effectively classify sleep stages by emphasizing these crucial features, thus achieving higher accuracy. In this study, an end-to-end deep learning model based on DenseNet for automatic sleep staging is designed and constructed. In the framework, two convolutional branches are devised to extract the underlying features (Two-Frequency Feature) at various frequencies, which are then fused and input into the DenseNet module to extract salient waveform features. After that, the Coordinate Attention mechanism is employed to enhance the localization of salient waveform features by emphasizing the position of salient waveforms and the spatial relationship across the entire frequency spectrum. Finally, the obtained features are accessed to the fully connected for sleep staging. The model was validated with a 20-fold cross-validation procedure on two public available datasets, and the overall accuracy, kappa coefficient, and MF1 score reached 92.9%, 78.7, 0.86 and 90.0%, 75.8, 0.80 on Sleep-EDF-20 and Sleep-EDFx, respectively. Experimental results show that the proposed model achieves competitive performance for sleep staging compared with the reported approaches under the identical conditions.
Collapse
Affiliation(s)
- Zhi Liu
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Meiqiao Qin
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Yunhua Lu
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Sixin Luo
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Qinhan Zhang
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| |
Collapse
|
12
|
Huang X, Schmelter F, Irshad MT, Piet A, Nisar MA, Sina C, Grzegorzek M. Optimizing sleep staging on multimodal time series: Leveraging borderline synthetic minority oversampling technique and supervised convolutional contrastive learning. Comput Biol Med 2023; 166:107501. [PMID: 37742416 DOI: 10.1016/j.compbiomed.2023.107501] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2023] [Revised: 08/15/2023] [Accepted: 09/15/2023] [Indexed: 09/26/2023]
Abstract
Sleep is an important research area in nutritional medicine that plays a crucial role in human physical and mental health restoration. It can influence diet, metabolism, and hormone regulation, which can affect overall health and well-being. As an essential tool in the sleep study, the sleep stage classification provides a parsing of sleep architecture and a comprehensive understanding of sleep patterns to identify sleep disorders and facilitate the formulation of targeted sleep interventions. However, the class imbalance issue is typically salient in sleep datasets, which severely affects classification performances. To address this issue and to extract optimal multimodal features of EEG, EOG, and EMG that can improve the accuracy of sleep stage classification, a Borderline Synthetic Minority Oversampling Technique (B-SMOTE)-Based Supervised Convolutional Contrastive Learning (BST-SCCL) is proposed, which can avoid the risk of data mismatch between various sleep knowledge domains (varying health conditions and annotation rules) and strengthening learning characteristics of the N1 stage from the pair-wise segments comparison strategy. The lightweight residual network architecture with a novel truncated cross-entropy loss function is designed to accommodate multimodal time series and boost the training speed and performance stability. The proposed model has been validated on four well-known public sleep datasets (Sleep-EDF-20, Sleep-EDF-78, ISRUC-1, and ISRUC-3) and its superior performance (overall accuracy of 91.31-92.34%, MF1 of 88.21-90.08%, and Cohen's Kappa coefficient k of 0.87-0.89) has further demonstrated its effectiveness. It shows the great potential of contrastive learning for cross-domain knowledge interaction in precision medicine.
Collapse
Affiliation(s)
- Xinyu Huang
- Institute of Medical Informatics, University of Lübeck, Germany.
| | - Franziska Schmelter
- Institute of Nutritional Medicine, University of Lübeck and University Medical Center Schleswig-Holstein, Lübeck, Germany.
| | | | - Artur Piet
- Institute of Medical Informatics, University of Lübeck, Germany.
| | | | - Christian Sina
- Institute of Nutritional Medicine, University of Lübeck and University Medical Center Schleswig-Holstein, Lübeck, Germany; Fraunhofer Research Institution for Individualized and Cell-Based Medical Engineering (IMTE), Lübeck, Germany.
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, Germany; Fraunhofer Research Institution for Individualized and Cell-Based Medical Engineering (IMTE), Lübeck, Germany.
| |
Collapse
|
13
|
Li T, Gong Y, Lv Y, Wang F, Hu M, Wen Y. GAC-SleepNet: A dual-structured sleep staging method based on graph structure and Euclidean structure. Comput Biol Med 2023; 165:107477. [PMID: 37717528 DOI: 10.1016/j.compbiomed.2023.107477] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2023] [Revised: 08/16/2023] [Accepted: 09/04/2023] [Indexed: 09/19/2023]
Abstract
Sleep staging is a precondition for the diagnosis and treatment of sleep disorders. However, how to fully exploit the relationship between spatial features of the brain and sleep stages is an important task. Many current classical algorithms only extract the characteristic information of the brain in the Euclidean space without considering other spatial structures. In this study, a sleep staging network named GAC-SleepNet is designed. GAC-SleepNet uses the characteristic information in the dual structure of the graph structure and the Euclidean structure for the classification of sleep stages. In the graph structure, this study uses a graph convolutional neural network to learn the deep features of each sleep stage and converts the features in the topological structure into feature vectors by a multilayer perceptron. In the Euclidean structure, this study uses convolutional neural networks to learn the temporal features of sleep information and combine attention mechanism to portray the connection between different sleep periods and EEG signals, while enhancing the description of global features to avoid local optima. In this study, the performance of the proposed network is evaluated on two public datasets. The experimental results show that the dual spatial structure captures more adequate and comprehensive information about sleep features and shows advancement in terms of different evaluation metrics.
Collapse
Affiliation(s)
- Tianxing Li
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130000, China
| | - Yulin Gong
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130000, China.
| | - Yudan Lv
- The Department of Neurology, First Hospital of Jilin University, Changchun, 130000, China
| | - Fatong Wang
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130000, China
| | - Mingjia Hu
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130000, China
| | - Yinke Wen
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130000, China
| |
Collapse
|
14
|
Phan H, Lorenzen KP, Heremans E, Chen OY, Tran MC, Koch P, Mertins A, Baumert M, Mikkelsen KB, De Vos M. L-SeqSleepNet: Whole-cycle Long Sequence Modeling for Automatic Sleep Staging. IEEE J Biomed Health Inform 2023; 27:4748-4757. [PMID: 37552591 DOI: 10.1109/jbhi.2023.3303197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/10/2023]
Abstract
Human sleep is cyclical with a period of approximately 90 minutes, implying long temporal dependency in the sleep data. Yet, exploring this long-term dependency when developing sleep staging models has remained untouched. In this work, we show that while encoding the logic of a whole sleep cycle is crucial to improve sleep staging performance, the sequential modelling approach in existing state-of-the-art deep learning models are inefficient for that purpose. We thus introduce a method for efficient long sequence modelling and propose a new deep learning model, L-SeqSleepNet, which takes into account whole-cycle sleep information for sleep staging. Evaluating L-SeqSleepNet on four distinct databases of various sizes, we demonstrate state-of-the-art performance obtained by the model over three different EEG setups, including scalp EEG in conventional Polysomnography (PSG), in-ear EEG, and around-the-ear EEG (cEEGrid), even with a single EEG channel input. Our analyses also show that L-SeqSleepNet is able to alleviate the predominance of N2 sleep (the major class in terms of classification) to bring down errors in other sleep stages. Moreover the network becomes much more robust, meaning that for all subjects where the baseline method had exceptionally poor performance, their performance are improved significantly. Finally, the computation time only grows at a sub-linear rate when the sequence length increases.
Collapse
|
15
|
Abstract
Automatic polysomnography analysis can be leveraged to shorten scoring times, reduce associated costs, and ultimately improve the overall diagnosis of sleep disorders. Multiple and diverse strategies have been attempted for implementation of this technology at scale in the routine workflow of sleep centers. The field, however, is complex and presents unsolved challenges in a number of areas. Recent developments in computer science and artificial intelligence are nevertheless closing the gap. Technological advances are also opening new pathways for expanding our current understanding of the domain and its analysis.
Collapse
Affiliation(s)
- Diego Alvarez-Estevez
- Center for Information and Communications Technology Research (CITIC), Universidade da Coruña, 15071 A Coruña, Spain.
| |
Collapse
|
16
|
Lee M, Kwak HG, Kim HJ, Won DO, Lee SW. SeriesSleepNet: an EEG time series model with partial data augmentation for automatic sleep stage scoring. Front Physiol 2023; 14:1188678. [PMID: 37700762 PMCID: PMC10494443 DOI: 10.3389/fphys.2023.1188678] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Accepted: 08/10/2023] [Indexed: 09/14/2023] Open
Abstract
Introduction: We propose an automatic sleep stage scoring model, referred to as SeriesSleepNet, based on convolutional neural network (CNN) and bidirectional long short-term memory (bi-LSTM) with partial data augmentation. We used single-channel raw electroencephalography signals for automatic sleep stage scoring. Methods: Our framework was focused on time series information, so we applied partial data augmentation to learn the connected time information in small series. In specific, the CNN module learns the time information of one epoch (intra-epoch) whereas the bi-LSTM trains the sequential information between the adjacent epochs (inter-epoch). Note that the input of the bi-LSTM is the augmented CNN output. Moreover, the proposed loss function was used to fine-tune the model by providing additional weights. To validate the proposed framework, we conducted two experiments using the Sleep-EDF and SHHS datasets. Results and Discussion: The results achieved an overall accuracy of 0.87 and 0.84 and overall F1-score of 0.80 and 0.78 and kappa value of 0.81 and 0.78 for five-class classification, respectively. We showed that the SeriesSleepNet was superior to the baselines based on each component in the proposed framework. Our architecture also outperformed the state-of-the-art methods with overall F1-score, accuracy, and kappa value. Our framework could provide information on sleep disorders or quality of sleep to automatically classify sleep stages with high performance.
Collapse
Affiliation(s)
- Minji Lee
- Department of Biomedical Software Engineering, The Catholic University of Korea, Bucheon, Republic of Korea
| | - Heon-Gyu Kwak
- Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea
| | - Hyeong-Jin Kim
- Department of Brain and Cognitive Engineering, Korea University, Seoul, Republic of Korea
| | - Dong-Ok Won
- Department of Artificial Intelligence Convergence, Hallym University, Chuncheon, Republic of Korea
| | - Seong-Whan Lee
- Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea
| |
Collapse
|
17
|
Yao H, Liu T, Zou R, Ding S, Xu Y. A Spatial-Temporal Transformer Architecture Using Multi-Channel Signals for Sleep Stage Classification. IEEE Trans Neural Syst Rehabil Eng 2023; 31:3353-3362. [PMID: 37578925 DOI: 10.1109/tnsre.2023.3305201] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/16/2023]
Abstract
Sleep stage classification is a fundamental task in diagnosing and monitoring sleep diseases. There are 2 challenges that remain open: (1) Since most methods only rely on input from a single channel, the spatial-temporal relationship of sleep signals has not been fully explored. (2) Lack of sleep data makes models hard to train from scratch. Here, we propose a vision Transformer-based architecture to process multi-channel polysomnogram signals. The method is an end-to-end framework that consists of a spatial encoder, a temporal encoder, and an MLP head classifier. The spatial encoder using a pre-trained Vision Transformer captures spatial information from multiple PSG channels. The temporal encoder utilizing the self-attention mechanism understands transitions between nearby epochs. In addition, we introduce a tailored image generation method to extract features within multi-channel and reshape them for transfer learning. We validate our method on 3 datasets and outperform the state-of-the-art algorithms. Our method fully explores the spatial-temporal relationship among different brain regions and addresses the problem of data insufficiency in clinical environments. Benefiting from reformulating the problem as image classification, the method could be applied to other 1D-signal problems in the future.
Collapse
|
18
|
Toma TI, Choi S. An End-to-End Multi-Channel Convolutional Bi-LSTM Network for Automatic Sleep Stage Detection. SENSORS (BASEL, SWITZERLAND) 2023; 23:4950. [PMID: 37430865 DOI: 10.3390/s23104950] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Revised: 05/17/2023] [Accepted: 05/19/2023] [Indexed: 07/12/2023]
Abstract
Sleep stage detection from polysomnography (PSG) recordings is a widely used method of monitoring sleep quality. Despite significant progress in the development of machine-learning (ML)-based and deep-learning (DL)-based automatic sleep stage detection schemes focusing on single-channel PSG data, such as single-channel electroencephalogram (EEG), electrooculogram (EOG), and electromyogram (EMG), developing a standard model is still an active subject of research. Often, the use of a single source of information suffers from data inefficiency and data-skewed problems. Instead, a multi-channel input-based classifier can mitigate the aforementioned challenges and achieve better performance. However, it requires extensive computational resources to train the model, and, hence, a tradeoff between performance and computational resources cannot be ignored. In this article, we aim to introduce a multi-channel, more specifically a four-channel, convolutional bidirectional long short-term memory (Bi-LSTM) network that can effectively exploit spatiotemporal features of data collected from multiple channels of the PSG recording (e.g., EEG Fpz-Cz, EEG Pz-Oz, EOG, and EMG) for automatic sleep stage detection. First, a dual-channel convolutional Bi-LSTM network module has been designed and pre-trained utilizing data from every two distinct channels of the PSG recording. Subsequently, we have leveraged the concept of transfer learning circuitously and have fused two dual-channel convolutional Bi-LSTM network modules to detect sleep stages. In the dual-channel convolutional Bi-LSTM module, a two-layer convolutional neural network has been utilized to extract spatial features from two channels of the PSG recordings. These extracted spatial features are subsequently coupled and given as input at every level of the Bi-LSTM network to extract and learn rich temporal correlated features. Both Sleep EDF-20 and Sleep EDF-78 (expanded version of Sleep EDF-20) datasets are used in this study to evaluate the result. The model that includes an EEG Fpz-Cz + EOG module and an EEG Fpz-Cz + EMG module can classify sleep stage with the highest value of accuracy (ACC), Kappa (Kp), and F1 score (e.g., 91.44%, 0.89, and 88.69%, respectively) on the Sleep EDF-20 dataset. On the other hand, the model consisting of an EEG Fpz-Cz + EMG module and an EEG Pz-Oz + EOG module shows the best performance (e.g., the value of ACC, Kp, and F1 score are 90.21%, 0.86, and 87.02%, respectively) compared to other combinations for the Sleep EDF-78 dataset. In addition, a comparative study with respect to other existing literature has been provided and discussed in order to exhibit the efficacy of our proposed model.
Collapse
Affiliation(s)
- Tabassum Islam Toma
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| | - Sunwoong Choi
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| |
Collapse
|
19
|
Gao DR, Li J, Wang MQ, Wang LT, Zhang YQ. Automatic sleep staging of single-channel EEG based on domain adversarial neural networks and domain self-attention. Front Neurosci 2023; 17:1143495. [PMID: 37090812 PMCID: PMC10117677 DOI: 10.3389/fnins.2023.1143495] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Accepted: 03/13/2023] [Indexed: 04/09/2023] Open
Abstract
The diagnosis and management of sleep problems depend heavily on sleep staging. For autonomous sleep staging, many data-driven deep learning models have been presented by trying to construct a large-labeled auxiliary sleep dataset and test it by electroencephalograms on different subjects. These approaches suffer a significant setback cause it assumes the training and test data come from the same or similar distribution. However, this is almost impossible in scenario cross-dataset due to inherent domain shift between domains. Unsupervised domain adaption was recently created to address the domain shift issue. However, only a few customized UDA solutions for sleep staging due to two limitations in previous UDA methods. First, the domain classifier does not consider boundaries between classes. Second, they depend on a shared model to align the domain that could miss the information of domains when extracting features. Given those restrictions, we present a novel UDA approach that combines category decision boundaries and domain discriminator to align the distributions of source and target domains. Also, to keep the domain-specific features, we create an unshared attention method. In addition, we investigated effective data augmentation in cross-dataset sleep scenarios. The experimental results on three datasets validate the efficacy of our approach and show that the proposed method is superior to state-of-the-art UDA methods on accuracy and MF1-Score.
Collapse
Affiliation(s)
- Dong-Rui Gao
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
- School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, China
| | - Jing Li
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
| | - Man-Qing Wang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
- School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, China
| | - Lu-Tao Wang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
| | - Yong-Qing Zhang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
- *Correspondence: Yong-Qing Zhang
| |
Collapse
|
20
|
Huang X, Shirahama K, Irshad MT, Nisar MA, Piet A, Grzegorzek M. Sleep Stage Classification in Children Using Self-Attention and Gaussian Noise Data Augmentation. SENSORS (BASEL, SWITZERLAND) 2023; 23:3446. [PMID: 37050506 PMCID: PMC10098613 DOI: 10.3390/s23073446] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/27/2023] [Revised: 03/20/2023] [Accepted: 03/22/2023] [Indexed: 06/19/2023]
Abstract
The analysis of sleep stages for children plays an important role in early diagnosis and treatment. This paper introduces our sleep stage classification method addressing the following two challenges: the first is the data imbalance problem, i.e., the highly skewed class distribution with underrepresented minority classes. For this, a Gaussian Noise Data Augmentation (GNDA) algorithm was applied to polysomnography recordings to seek the balance of data sizes for different sleep stages. The second challenge is the difficulty in identifying a minority class of sleep stages, given their short sleep duration and similarities to other stages in terms of EEG characteristics. To overcome this, we developed a DeConvolution- and Self-Attention-based Model (DCSAM) which can inverse the feature map of a hidden layer to the input space to extract local features and extract the correlations between all possible pairs of features to distinguish sleep stages. The results on our dataset show that DCSAM based on GNDA obtains an accuracy of 90.26% and a macro F1-score of 86.51% which are higher than those of our previous method. We also tested DCSAM on a well-known public dataset-Sleep-EDFX-to prove whether it is applicable to sleep data from adults. It achieves a comparable performance to state-of-the-art methods, especially accuracies of 91.77%, 92.54%, 94.73%, and 95.30% for six-stage, five-stage, four-stage, and three-stage classification, respectively. These results imply that our DCSAM based on GNDA has a great potential to offer performance improvements in various medical domains by considering the data imbalance problems and correlations among features in time series data.
Collapse
Affiliation(s)
- Xinyu Huang
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany
| | - Kimiaki Shirahama
- Department of Informatics, Kindai University, 3-4-1 Kowakae, Higashiosaka City 577-8502, Osaka, Japan
| | - Muhammad Tausif Irshad
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany
- Department of IT, University of the Punjab, Lahore 54000, Pakistan
| | | | - Artur Piet
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany
- Department of Knowledge Engineering, University of Economics, Bogucicka 3, 40287 Katowice, Poland
| |
Collapse
|
21
|
Fiorillo L, Monachino G, van der Meer J, Pesce M, Warncke JD, Schmidt MH, Bassetti CLA, Tzovara A, Favaro P, Faraci FD. U-Sleep's resilience to AASM guidelines. NPJ Digit Med 2023; 6:33. [PMID: 36878957 PMCID: PMC9988983 DOI: 10.1038/s41746-023-00784-0] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2022] [Accepted: 02/21/2023] [Indexed: 03/08/2023] Open
Abstract
AASM guidelines are the result of decades of efforts aiming at standardizing sleep scoring procedure, with the final goal of sharing a worldwide common methodology. The guidelines cover several aspects from the technical/digital specifications, e.g., recommended EEG derivations, to detailed sleep scoring rules accordingly to age. Automated sleep scoring systems have always largely exploited the standards as fundamental guidelines. In this context, deep learning has demonstrated better performance compared to classical machine learning. Our present work shows that a deep learning-based sleep scoring algorithm may not need to fully exploit the clinical knowledge or to strictly adhere to the AASM guidelines. Specifically, we demonstrate that U-Sleep, a state-of-the-art sleep scoring algorithm, can be strong enough to solve the scoring task even using clinically non-recommended or non-conventional derivations, and with no need to exploit information about the chronological age of the subjects. We finally strengthen a well-known finding that using data from multiple data centers always results in a better performing model compared with training on a single cohort. Indeed, we show that this latter statement is still valid even by increasing the size and the heterogeneity of the single data cohort. In all our experiments we used 28528 polysomnography studies from 13 different clinical studies.
Collapse
Affiliation(s)
- Luigi Fiorillo
- Institute of Informatics, University of Bern, Bern, Switzerland.
- Institute of Digital Technologies for Personalized Healthcare ∣ MeDiTech, Department of Innovative Technologies, University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland.
| | - Giuliana Monachino
- Institute of Informatics, University of Bern, Bern, Switzerland
- Institute of Digital Technologies for Personalized Healthcare ∣ MeDiTech, Department of Innovative Technologies, University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland
| | - Julia van der Meer
- Sleep Wake Epilepsy Center ∣ NeuroTec, Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Marco Pesce
- Sleep Wake Epilepsy Center ∣ NeuroTec, Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Jan D Warncke
- Sleep Wake Epilepsy Center ∣ NeuroTec, Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Markus H Schmidt
- Sleep Wake Epilepsy Center ∣ NeuroTec, Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Claudio L A Bassetti
- Sleep Wake Epilepsy Center ∣ NeuroTec, Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Athina Tzovara
- Institute of Informatics, University of Bern, Bern, Switzerland
- Sleep Wake Epilepsy Center ∣ NeuroTec, Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Paolo Favaro
- Institute of Informatics, University of Bern, Bern, Switzerland
| | - Francesca D Faraci
- Institute of Digital Technologies for Personalized Healthcare ∣ MeDiTech, Department of Innovative Technologies, University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland
| |
Collapse
|
22
|
Do not sleep on traditional machine learning. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104429] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
23
|
Supratak A, Haddawy P. Quantifying the impact of data characteristics on the transferability of sleep stage scoring models. Artif Intell Med 2023; 139:102540. [PMID: 37100508 DOI: 10.1016/j.artmed.2023.102540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Revised: 03/18/2023] [Accepted: 03/27/2023] [Indexed: 04/03/2023]
Abstract
Deep learning models for scoring sleep stages based on single-channel EEG have been proposed as a promising method for remote sleep monitoring. However, applying these models to new datasets, particularly from wearable devices, raises two questions. First, when annotations on a target dataset are unavailable, which different data characteristics affect the sleep stage scoring performance the most and by how much? Second, when annotations are available, which dataset should be used as the source of transfer learning to optimize performance? In this paper, we propose a novel method for computationally quantifying the impact of different data characteristics on the transferability of deep learning models. Quantification is accomplished by training and evaluating two models with significant architectural differences, TinySleepNet and U-Time, under various transfer configurations in which the source and target datasets have different recording channels, recording environments, and subject conditions. For the first question, the environment had the highest impact on sleep stage scoring performance, with performance degrading by over 14% when sleep annotations were unavailable. For the second question, the most useful transfer sources for TinySleepNet and the U-Time models were MASS-SS1 and ISRUC-SG1, containing a high percentage of N1 (the rarest sleep stage) relative to the others. The frontal and central EEGs were preferred for TinySleepNet. The proposed approach enables full utilization of existing sleep datasets for training and planning model transfer to maximize the sleep stage scoring performance on a target problem when sleep annotations are limited or unavailable, supporting the realization of remote sleep monitoring.
Collapse
|
24
|
Yang CY, Chen PC, Huang WC. Cross-Domain Transfer of EEG to EEG or ECG Learning for CNN Classification Models. SENSORS (BASEL, SWITZERLAND) 2023; 23:2458. [PMID: 36904661 PMCID: PMC10007254 DOI: 10.3390/s23052458] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Revised: 02/19/2023] [Accepted: 02/20/2023] [Indexed: 06/18/2023]
Abstract
Electroencephalography (EEG) is often used to evaluate several types of neurological brain disorders because of its noninvasive and high temporal resolution. In contrast to electrocardiography (ECG), EEG can be uncomfortable and inconvenient for patients. Moreover, deep-learning techniques require a large dataset and a long time for training from scratch. Therefore, in this study, EEG-EEG or EEG-ECG transfer learning strategies were applied to explore their effectiveness for the training of simple cross-domain convolutional neural networks (CNNs) used in seizure prediction and sleep staging systems, respectively. The seizure model detected interictal and preictal periods, whereas the sleep staging model classified signals into five stages. The patient-specific seizure prediction model with six frozen layers achieved 100% accuracy for seven out of nine patients and required only 40 s of training time for personalization. Moreover, the cross-signal transfer learning EEG-ECG model for sleep staging achieved an accuracy approximately 2.5% higher than that of the ECG model; additionally, the training time was reduced by >50%. In summary, transfer learning from an EEG model to produce personalized models for a more convenient signal can both reduce the training time and increase the accuracy; moreover, challenges such as data insufficiency, variability, and inefficiency can be effectively overcome.
Collapse
|
25
|
Qiu Y, Lin F, Chen W, Xu M. Pre-training in Medical Data: A Survey. MACHINE INTELLIGENCE RESEARCH 2023. [PMCID: PMC9942039 DOI: 10.1007/s11633-022-1382-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/23/2023]
Abstract
Medical data refers to health-related information associated with regular patient care or as part of a clinical trial program. There are many categories of such data, such as clinical imaging data, bio-signal data, electronic health records (EHR), and multi-modality medical data. With the development of deep neural networks in the last decade, the emerging pre-training paradigm has become dominant in that it has significantly improved machine learning methods’ performance in a data-limited scenario. In recent years, studies of pre-training in the medical domain have achieved significant progress. To summarize these technology advancements, this work provides a comprehensive survey of recent advances for pre-training on several major types of medical data. In this survey, we summarize a large number of related publications and the existing benchmarking in the medical domain. Especially, the survey briefly describes how some pre-training methods are applied to or developed for medical data. From a data-driven perspective, we examine the extensive use of pre-training in many medical scenarios. Moreover, based on the summary of recent pre-training studies, we identify several challenges in this field to provide insights for future studies.
Collapse
Affiliation(s)
- Yixuan Qiu
- The University of Queensland, Brisbane, 4072 Australia
| | - Feng Lin
- The University of Queensland, Brisbane, 4072 Australia
| | - Weitong Chen
- The University of Adelaide, Adelaide, 5005 Australia
| | - Miao Xu
- The University of Queensland, Brisbane, 4072 Australia
| |
Collapse
|
26
|
Cheng L, Luo S, Li B, Liu R, Zhang Y, Zhang H. Multiple-instance learning for EEG based OSA event detection. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104358] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
27
|
Acien A, Morales A, Vera-Rodriguez R, Fierrez J, Mondesire-Crump I, Arroyo-Gallego T. Detection of Mental Fatigue in the General Population: Feasibility Study of Keystroke Dynamics as a Real-world Biomarker. JMIR BIOMEDICAL ENGINEERING 2022; 7:e41003. [PMID: 38875698 PMCID: PMC11041424 DOI: 10.2196/41003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2022] [Revised: 10/07/2022] [Accepted: 10/19/2022] [Indexed: 11/23/2022] Open
Abstract
BACKGROUND Mental fatigue is a common and potentially debilitating state that can affect individuals' health and quality of life. In some cases, its manifestation can precede or mask early signs of other serious mental or physiological conditions. Detecting and assessing mental fatigue can be challenging nowadays as it relies on self-evaluation and rating questionnaires, which are highly influenced by subjective bias. Introducing more objective, quantitative, and sensitive methods to characterize mental fatigue could be critical to improve its management and the understanding of its connection to other clinical conditions. OBJECTIVE This paper aimed to study the feasibility of using keystroke biometrics for mental fatigue detection during natural typing. As typing involves multiple motor and cognitive processes that are affected by mental fatigue, our hypothesis was that the information captured in keystroke dynamics can offer an interesting mean to characterize users' mental fatigue in a real-world setting. METHODS We apply domain transformation techniques to adapt and transform TypeNet, a state-of-the-art deep neural network, originally intended for user authentication, to generate a network optimized for the fatigue detection task. All experiments were conducted using 3 keystroke databases that comprise different contexts and data collection protocols. RESULTS Our preliminary results showed area under the curve performances ranging between 72.2% and 80% for fatigue versus rested sample classification, which is aligned with previously published models on daily alertness and circadian cycles. This demonstrates the potential of our proposed system to characterize mental fatigue fluctuations via natural typing patterns. Finally, we studied the performance of an active detection approach that leverages the continuous nature of keystroke biometric patterns for the assessment of users' fatigue in real time. CONCLUSIONS Our results suggest that the psychomotor patterns that characterize mental fatigue manifest during natural typing, which can be quantified via automated analysis of users' daily interaction with their device. These findings represent a step towards the development of a more objective, accessible, and transparent solution to monitor mental fatigue in a real-world environment.
Collapse
Affiliation(s)
- Alejandro Acien
- nQ Medical Inc, Cambridge, MA, United States
- School of Engineering, Universidad Autonoma de Madrid, Madrid, Spain
| | - Aythami Morales
- School of Engineering, Universidad Autonoma de Madrid, Madrid, Spain
| | | | - Julian Fierrez
- School of Engineering, Universidad Autonoma de Madrid, Madrid, Spain
| | | | | |
Collapse
|
28
|
Kim H, Lee SM, Choi S. Automatic sleep stages classification using multi-level fusion. Biomed Eng Lett 2022; 12:413-420. [PMID: 36238370 PMCID: PMC9550904 DOI: 10.1007/s13534-022-00244-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2022] [Revised: 07/12/2022] [Accepted: 07/25/2022] [Indexed: 10/15/2022] Open
Abstract
Sleep efficiency is a factor that can determine a person's healthy life. Sleep efficiency can be calculated by analyzing the results of the sleep stage classification. There have been many studies to classify sleep stages automatically using multiple signals to improve the accuracy of the sleep stage classification. The fusion method is used to process multi-signal data. Fusion methods include data-level fusion, feature-level fusion, and decision-level fusion methods. We propose a multi-level fusion method to increase the accuracy of the sleep stage classification when using multi-signal data consisting of electroencephalography and electromyography signals. First, we used feature-level fusion to fuse the extracted features using a convolutional neural network for multi-signal data. Then, after obtaining each classified result using the fused feature data, the sleep stage was derived using a decision-level fusion method that fused classified results. We used public datasets, Sleep-EDF, to measure performance; we confirmed that the proposed multi-level fusion method yielded a higher accuracy of 87.2%, respectively, compared to single-level fusion method and more existing methods. The proposed multi-level fusion method showed the most improved performance in classifying N1 stage, where existing methods had the lowest performance.
Collapse
Affiliation(s)
- Hyungjik Kim
- Department of Secured Smart Electric Vehicle, Kookmin University, 02707 Seoul, Korea
| | - Seung Min Lee
- Department of Electrical Engineering, Kookmin University, 02707 Seoul, Korea
| | - Sunwoong Choi
- Department of Electrical Engineering, Kookmin University, 02707 Seoul, Korea
| |
Collapse
|
29
|
Waters SH, Clifford GD. Comparison of deep transfer learning algorithms and transferability measures for wearable sleep staging. Biomed Eng Online 2022; 21:66. [PMID: 36096868 PMCID: PMC9465946 DOI: 10.1186/s12938-022-01033-3] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Accepted: 08/18/2022] [Indexed: 11/10/2022] Open
Abstract
Background Obtaining medical data using wearable sensors is a potential replacement for in-hospital monitoring, but the lack of data for such sensors poses a challenge for development. One solution is using in-hospital recordings to boost performance via transfer learning. While there are many possible transfer learning algorithms, few have been tested in the domain of EEG-based sleep staging. Furthermore, there are few ways for determining which transfer learning method will work best besides exhaustive testing. Measures of transferability do exist, but are typically used for selection of pre-trained models rather than algorithms and few have been tested on medical signals. We tested several supervised transfer learning algorithms on a sleep staging task using a single channel of EEG (AF7-Fpz) captured from an in-home commercial system. Results Two neural networks—one bespoke and another state-of-art open-source architecture—were pre-trained on one of six source datasets comprising 11,561 subjects undergoing clinical polysomnograms (PSGs), then re-trained on a target dataset of 75 full-night recordings from 24 subjects. Several transferability measures were then tested to determine which is most effective for assessing performance on unseen target data. Performance on the target dataset was improved using transfer learning, with re-training the head layers being the most effective in the majority of cases (up to 63.9% of cases). Transferability measures generally provided significant correlations with accuracy (up to \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$r = -0.53$$\end{document}r=-0.53). Conclusion Re-training the head layers provided the largest performance boost. Transferability measures are useful indicators of transfer learning effectiveness.
Collapse
Affiliation(s)
- Samuel H Waters
- Department of Bioengineering, Georgia Institute of Technology, Atlanta, United States.
| | - Gari D Clifford
- Department of Biomedical Informatics, Emory University, Atlanta, United States
| |
Collapse
|
30
|
Phan H, Chen OY, Tran MC, Koch P, Mertins A, De Vos M. XSleepNet: Multi-View Sequential Model for Automatic Sleep Staging. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:5903-5915. [PMID: 33788679 DOI: 10.1109/tpami.2021.3070057] [Citation(s) in RCA: 32] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Automating sleep staging is vital to scale up sleep assessment and diagnosis to serve millions experiencing sleep deprivation and disorders and enable longitudinal sleep monitoring in home environments. Learning from raw polysomnography signals and their derived time-frequency image representations has been prevalent. However, learning from multi-view inputs (e.g., both the raw signals and the time-frequency images) for sleep staging is difficult and not well understood. This work proposes a sequence-to-sequence sleep staging model, XSleepNet,1 that is capable of learning a joint representation from both raw signals and time-frequency images. Since different views may generalize or overfit at different rates, the proposed network is trained such that the learning pace on each view is adapted based on their generalization/overfitting behavior. In simple terms, the learning on a particular view is speeded up when it is generalizing well and slowed down when it is overfitting. View-specific generalization/overfitting measures are computed on-the-fly during the training course and used to derive weights to blend the gradients from different views. As a result, the network is able to retain the representation power of different views in the joint features which represent the underlying distribution better than those learned by each individual view alone. Furthermore, the XSleepNet architecture is principally designed to gain robustness to the amount of training data and to increase the complementarity between the input views. Experimental results on five databases of different sizes show that XSleepNet consistently outperforms the single-view baselines and the multi-view baseline with a simple fusion strategy. Finally, XSleepNet also outperforms prior sleep staging methods and improves previous state-of-the-art results on the experimental databases.
Collapse
|
31
|
Feature matching as improved transfer learning technique for wearable EEG. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.104009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|
32
|
Shen N, Luo T, Chen C, Zhang Y, Zhu H, Zhou Y, Wang Y, Chen W. Towards an automatic narcolepsy detection on ambiguous sleep staging and sleep transition dynamics joint model. J Neural Eng 2022; 19. [PMID: 36001951 DOI: 10.1088/1741-2552/ac8c6b] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2022] [Accepted: 08/24/2022] [Indexed: 11/11/2022]
Abstract
Objective.Mixing/dissociation of sleep stages in narcolepsy adds to the difficulty in automatic sleep staging. Moreover, automatic analytical studies for narcolepsy and multiple sleep latency test (MSLT) have only done automatic sleep staging without leveraging the sleep stage profile for further patient identification. This study aims to establish an automatic narcolepsy detection method for MSLT.Approach.We construct a two-phase model on MSLT recordings, where ambiguous sleep staging and sleep transition dynamics make joint efforts to address this issue. In phase 1, we extract representative features from electroencephalogram (EEG) and electrooculogram (EOG) signals. Then, the features are input to an EasyEnsemble classifier for automatic sleep staging. In phase 2, we investigate sleep transition dynamics, including sleep stage transitions and sleep stages, and output likelihood of narcolepsy by virtue of principal component analysis (PCA) and a logistic regression classifier. To demonstrate the proposed framework in clinical application, we conduct experiments on 24 participants from our hospital, considering ten patients with narcolepsy and fourteen patients with MSLT negative.Main results.Applying the two-phase leave-one-subject-out testing scheme, the model reaches an accuracy, sensitivity, and specificity of 87.5%, 80.0%, and 92.9% for narcolepsy detection. Influenced by disease pathology, accuracy of automatic sleep staging in narcolepsy appears to decrease compared to that in the non-narcoleptic population.Significance.This method can automatically and efficiently distinguish patients with narcolepsy based on MSLT. It probes into the amalgamation of automatic sleep staging and sleep transition dynamics for narcolepsy detection, which would assist clinic and neuroelectrophysiology specialists in visual interpretation and diagnosis.
Collapse
Affiliation(s)
- Ning Shen
- Fudan University School of Information Science and Engineering, 220 Handan Road, Yangpu District, Shanghai, China, 2005 Songhu Road, Yangpu District, Shanghai, China, Shanghai, 200433, CHINA
| | - Tian Luo
- Department of Neurology, Children's Hospital of Fudan University, 399 Wanyuan Road, Minhang District, Shanghai, China, Shanghai, 201102, CHINA
| | - Chen Chen
- Fudan University Human Phenome Institute, 825 Zhangheng Road, Pudong District, Shanghai, China, Shanghai, 201203, CHINA
| | - Yanjiong Zhang
- Department of Neurology, Children's Hospital of Fudan University, 399 Wanyuan Road, Minhang District, Shanghai, China, Shanghai, 201102, CHINA
| | - Hangyu Zhu
- Fudan University School of Information Science and Engineering, 220 Handan Road, Yangpu District, Shanghai, China, 2005 Songhu Road, Yangpu District, Shanghai, China, Shanghai, 200433, CHINA
| | - Yuanfeng Zhou
- Department of Neurology, Children's Hospital of Fudan University, 399 Wanyuan Road, Minhang District, Shanghai, China, Shanghai, 201102, CHINA
| | - Yi Wang
- Department of Neurology, Children's Hospital of Fudan University, 399 Wanyuan Road, Minhang District, Shanghai, China, Shanghai, 201102, CHINA
| | - Wei Chen
- Department of Electronic Engineering, Fudan University, 220 Handan Road, Yangpu District, Shanghai, China, 2005 Songhu Road, Yangpu District, Shanghai, China, Shanghai, Shanghai, 200433, CHINA
| |
Collapse
|
33
|
Heremans ERM, Phan H, Borzée P, Buyse B, Testelmans D, De Vos M. From unsupervised to semi-supervised adversarial domain adaptation in EEG-based sleep staging. J Neural Eng 2022; 19. [PMID: 35508121 DOI: 10.1088/1741-2552/ac6ca8] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2022] [Accepted: 05/04/2022] [Indexed: 10/18/2022]
Abstract
OBJECTIVE The recent breakthrough of wearable sleep monitoring devices results in large amounts of sleep data. However, as limited labels are available, interpreting these data requires automated sleep stage classification methods with a small need for labeled training data. Transfer learning and domain adaptation offer possible solutions by enabling models to learn on a source dataset and adapt to a target dataset. APPROACH In this paper, we investigate adversarial domain adaptation applied to real use cases with wearable sleep datasets acquired from diseased patient populations. Different practical aspects of the adversarial domain adaptation framework \hl{are examined}, including the added value of (pseudo-)labels from the target dataset and the influence of domain mismatch between the source and target data. The method is also implemented for personalization to specific patients. MAIN RESULTS The results show that adversarial domain adaptation is effective in the application of sleep staging on wearable data. When compared to a model applied on a target dataset without any adaptation, the domain adaptation method in its simplest form achieves relative gains of 7%-27% in accuracy. The performance on the target domain is further boosted by adding pseudo-labels and real target domain labels when available, and by choosing an appropriate source dataset. Furthermore, unsupervised adversarial domain adaptation can also personalize a model, improving the performance by 1%-2% compared to a non-personal model. SIGNIFICANCE In conclusion, adversarial domain adaptation provides a flexible framework for semi-supervised and unsupervised transfer learning. This is particularly useful in sleep staging and other wearable EEG applications.
Collapse
Affiliation(s)
- Elisabeth Roxane Marie Heremans
- Department of Electrical Engineering, KU Leuven Science Engineering and Technology Group, Kasteelpark Arenberg 10, Leuven, 3001, BELGIUM
| | - Huy Phan
- School of Electronic Engineering and Computer Science, Queen Mary University of London, Mile End Rd, Bethnal Green, London, E1 4NS, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
| | - Pascal Borzée
- Department of Pneumology, KU Leuven University Hospitals Leuven, Herestraat 49, Leuven, 3000, BELGIUM
| | - Bertien Buyse
- Department of Pneumology, KU Leuven University Hospitals Leuven, Herestraat 49, Leuven, Flanders, 3000, BELGIUM
| | - Dries Testelmans
- Department of Pneumology, KU Leuven University Hospitals Leuven, Herestraat 49, Leuven, 3000, BELGIUM
| | - Maarten De Vos
- Department of Electrical Engineering, KU Leuven Science Engineering and Technology Group, Kasteelpark Arenberg 10, Leuven, 3000, BELGIUM
| |
Collapse
|
34
|
Phan H, Mikkelsen K. Automatic sleep staging of EEG signals: recent development, challenges, and future directions. Physiol Meas 2022; 43. [PMID: 35320788 DOI: 10.1088/1361-6579/ac6049] [Citation(s) in RCA: 34] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2021] [Accepted: 03/23/2022] [Indexed: 11/11/2022]
Abstract
Modern deep learning holds a great potential to transform clinical practice on human sleep. Teaching a machine to carry out routine tasks would be a tremendous reduction in workload for clinicians. Sleep staging, a fundamental step in sleep practice, is a suitable task for this and will be the focus in this article. Recently, automatic sleep staging systems have been trained to mimic manual scoring, leading to similar performance to human sleep experts, at least on scoring of healthy subjects. Despite tremendous progress, we have not seen automatic sleep scoring adopted widely in clinical environments. This review aims to give a shared view of the authors on the most recent state-of-the-art development in automatic sleep staging, the challenges that still need to be addressed, and the future directions for automatic sleep scoring to achieve clinical value.
Collapse
Affiliation(s)
- Huy Phan
- School of Electronic Engineering and Computer Science, Queen Mary University of London, Mile End Rd, London, E1 4NS, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
| | - Kaare Mikkelsen
- Department of Electrical and Computer Engineering, Aarhus Universitet, Finlandsgade 22, Aarhus, 8000, DENMARK
| |
Collapse
|
35
|
Ebbehoj A, Thunbo MØ, Andersen OE, Glindtvad MV, Hulman A. Transfer learning for non-image data in clinical research: A scoping review. PLOS DIGITAL HEALTH 2022; 1:e0000014. [PMID: 36812540 PMCID: PMC9931256 DOI: 10.1371/journal.pdig.0000014] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Accepted: 12/15/2021] [Indexed: 01/14/2023]
Abstract
BACKGROUND Transfer learning is a form of machine learning where a pre-trained model trained on a specific task is reused as a starting point and tailored to another task in a different dataset. While transfer learning has garnered considerable attention in medical image analysis, its use for clinical non-image data is not well studied. Therefore, the objective of this scoping review was to explore the use of transfer learning for non-image data in the clinical literature. METHODS AND FINDINGS We systematically searched medical databases (PubMed, EMBASE, CINAHL) for peer-reviewed clinical studies that used transfer learning on human non-image data. We included 83 studies in the review. More than half of the studies (63%) were published within 12 months of the search. Transfer learning was most often applied to time series data (61%), followed by tabular data (18%), audio (12%) and text (8%). Thirty-three (40%) studies applied an image-based model to non-image data after transforming data into images (e.g. spectrograms). Twenty-nine (35%) studies did not have any authors with a health-related affiliation. Many studies used publicly available datasets (66%) and models (49%), but fewer shared their code (27%). CONCLUSIONS In this scoping review, we have described current trends in the use of transfer learning for non-image data in the clinical literature. We found that the use of transfer learning has grown rapidly within the last few years. We have identified studies and demonstrated the potential of transfer learning in clinical research in a wide range of medical specialties. More interdisciplinary collaborations and the wider adaption of reproducible research principles are needed to increase the impact of transfer learning in clinical research.
Collapse
Affiliation(s)
- Andreas Ebbehoj
- Department of Endocrinology and Internal Medicine, Aarhus University Hospital, Denmark
- Department of Clinical Medicine, Aarhus University, Denmark
| | | | | | | | - Adam Hulman
- Steno Diabetes Center Aarhus, Aarhus University Hospital, Denmark
- * E-mail:
| |
Collapse
|
36
|
Phan H, Mikkelsen K, Chen OY, Koch P, Mertins A, De Vos M. SleepTransformer: Automatic Sleep Staging with Interpretability and Uncertainty Quantification. IEEE Trans Biomed Eng 2022; 69:2456-2467. [PMID: 35100107 DOI: 10.1109/tbme.2022.3147187] [Citation(s) in RCA: 39] [Impact Index Per Article: 19.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
BACKGROUND Black-box skepticism is one of the main hindrances impeding deep-learning-based automatic sleep scoring from being used in clinical environments. METHODS Towards interpretability, this work proposes a sequence-to-sequence sleep staging model, namely SleepTransformer. It is based on the transformer backbone and offers interpretability of the models decisions at both the epoch and sequence level. We further propose a simple yet efficient method to quantify uncertainty in the models decisions. The method, which is based on entropy, can serve as a metric for deferring low-confidence epochs to a human expert for further inspection. RESULTS Making sense of the transformers self-attention scores for interpretability, at the epoch level, the attention scores are encoded as a heat map to highlight sleep-relevant features captured from the input EEG signal. At the sequence level, the attention scores are visualized as the influence of different neighboring epochs in an input sequence (i.e. the context) to recognition of a target epoch, mimicking the way manual scoring is done by human experts. CONCLUSION Additionally, we demonstrate that SleepTransformer performs on par with existing methods on two databases of different sizes. SIGNIFICANCE Equipped with interpretability and the ability of uncertainty quantification, SleepTransformer holds promise for being integrated into clinical settings.
Collapse
|
37
|
Fan J, Zhu H, Jiang X, Meng L, Chen C, Fu C, Yu H, Dai C, Chen W. Unsupervised Domain Adaptation by Statistics Alignment for Deep Sleep Staging Networks. IEEE Trans Neural Syst Rehabil Eng 2022; 30:205-216. [PMID: 35041607 DOI: 10.1109/tnsre.2022.3144169] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Deep sleep staging networks have reached top performance on large-scale datasets. However, these models perform poorer when training and testing on small sleep cohorts due to data inefficiency. Transferring well-trained models from large-scale datasets (source domain) to small sleep cohorts (target domain) is a promising solution but still remains challenging due to the domain-shift issue. In this work, an unsupervised domain adaptation approach, domain statistics alignment (DSA), is developed to bridge the gap between the data distribution of source and target domains. DSA adapts the source models on the target domain by modulating the domain-specific statistics of deep features stored in the Batch Normalization (BN) layers. Furthermore, we have extended DSA by introducing cross-domain statistics in each BN layer to perform DSA adaptively (AdaDSA). The proposed methods merely need the well-trained source model without access to the source data, which may be proprietary and inaccessible. DSA and AdaDSA are universally applicable to various deep sleep staging networks that have BN layers. We have validated the proposed methods by extensive experiments on two state-of-the-art deep sleep staging networks, DeepSleepNet+ and U-time. The performance was evaluated by conducting various transfer tasks on six sleep databases, including two large-scale databases, MASS and SHHS, as the source domain, four small sleep databases as the target domain. Thereinto, clinical sleep records acquired in Huashan Hospital, Shanghai, were used. The results show that both DSA and AdaDSA could significantly improve the performance of source models on target domains, providing novel insights into the domain generalization problem in sleep staging tasks.
Collapse
|
38
|
Fiorillo L, Favaro P, Faraci FD. DeepSleepNet-Lite: A Simplified Automatic Sleep Stage Scoring Model With Uncertainty Estimates. IEEE Trans Neural Syst Rehabil Eng 2021; 29:2076-2085. [PMID: 34648450 DOI: 10.1109/tnsre.2021.3117970] [Citation(s) in RCA: 27] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Deep learning is widely used in the most recent automatic sleep scoring algorithms. Its popularity stems from its excellent performance and from its ability to process raw signals and to learn feature directly from the data. Most of the existing scoring algorithms exploit very computationally demanding architectures, due to their high number of training parameters, and process lengthy time sequences in input (up to 12 minutes). Only few of these architectures provide an estimate of the model uncertainty. In this study we propose DeepSleepNet-Lite, a simplified and lightweight scoring architecture, processing only 90-seconds EEG input sequences. We exploit, for the first time in sleep scoring, the Monte Carlo dropout technique to enhance the performance of the architecture and to also detect the uncertain instances. The evaluation is performed on a single-channel EEG Fpz-Cz from the open source Sleep-EDF expanded database. DeepSleepNet-Lite achieves slightly lower performance, if not on par, compared to the existing state-of-the-art architectures, in overall accuracy, macro F1-score and Cohen's kappa (on Sleep-EDF v1-2013 ±30mins: 84.0%, 78.0%, 0.78; on Sleep-EDF v2-2018 ±30mins: 80.3%, 75.2%, 0.73). Monte Carlo dropout enables the estimate of the uncertain predictions. By rejecting the uncertain instances, the model achieves higher performance on both versions of the database (on Sleep-EDF v1-2013 ±30mins: 86.1.0%, 79.6%, 0.81; on Sleep-EDF v2-2018 ±30mins: 82.3%, 76.7%, 0.76). Our lighter sleep scoring approach paves the way to the application of scoring algorithms for sleep analysis in real-time.
Collapse
|
39
|
Mikkelsen KB, Phan H, Rank ML, Hemmsen MC, de Vos M, Kidmose P. Sleep monitoring using ear-centered setups: Investigating the influence from electrode configurations. IEEE Trans Biomed Eng 2021; 69:1564-1572. [PMID: 34587000 DOI: 10.1109/tbme.2021.3116274] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Modern sleep monitoring development is shifting towards the use of unobtrusive sensors combined with algorithms for automatic sleep scoring. Many different combinations of wet and dry electrodes, ear-centered, forehead-mounted or headband-inspired designs have been proposed, alongside an ever growing variety of machine learning algorithms for automatic sleep scoring. OBJECTIVE Among candidate positions, those in the facial area and around the ears have the benefit of being relatively hairless, and in our view deserve extra attention. In this paper, we seek to determine the limits to sleep monitoring quality within this spatial constraint. METHODS We compare 13 different, realistic sensor setups derived from the same data set and analysed with the same pipeline. RESULTS All setups which include both a lateral and an EOG derivation show similar, state-of-the-art performance, with average Cohen's kappa values of at least 0.80. CONCLUSION If large electrode distances are used, positioning is not critical for achieving large sleep-related signal-to-noise-ratio, and hence accurate sleep scoring. SIGNIFICANCE We argue that with the current competitive performance of automated staging approaches, there is a need for establishing an improved benchmark beyond current single human rater scoring.
Collapse
|
40
|
Yoo C, Lee HW, Kang JW. Transferring Structured Knowledge in Unsupervised Domain Adaptation of a Sleep Staging Network. IEEE J Biomed Health Inform 2021; 26:1273-1284. [PMID: 34388101 DOI: 10.1109/jbhi.2021.3103614] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Automatic sleep staging based on deep learning (DL) has been attracting attention for analyzing sleep quality and determining treatment effects. It is challenging to acquire long-term sleep data from numerous subjects and manually labeling them even though most DL-based models are trained using large-scale sleep data to provide state-of-the-art performance. One way to overcome this data shortage is to create a pre-trained network with an existing large-scale dataset (source domain) that is applicable to small cohorts of datasets (target domain); however, discrepancies in data distribution between the domains prevent successful refinement of this approach. In this paper, we propose an unsupervised domain adaptation method for sleep staging networks to reduce discrepancies by realigning the domains in the same space and producing domain-invariant features. Specifically, in addition to a classical domain discriminator, we introduce local dis-criminators-subject and stage-to maintain the intrinsic structure of sleep data to decrease local misalignments while using adversarial learning to play a minimax game between the feature extractor and discriminators. Moreover, we present several optimization schemes during training because the conventional adversarial learning is not effective to our training scheme. We evaluate the performance of the proposed method by examining the staging performances of a baseline network compared with direct transfer (DT) learning in various conditions. The experimental results demonstrate that the proposed domain adaptation significantly improves the performance though it needs no labeled sleep data in target domain.
Collapse
|
41
|
Creagh AP, Lipsmeier F, Lindemann M, Vos MD. Interpretable deep learning for the remote characterisation of ambulation in multiple sclerosis using smartphones. Sci Rep 2021; 11:14301. [PMID: 34253769 PMCID: PMC8275610 DOI: 10.1038/s41598-021-92776-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2021] [Accepted: 06/14/2021] [Indexed: 12/04/2022] Open
Abstract
The emergence of digital technologies such as smartphones in healthcare applications have demonstrated the possibility of developing rich, continuous, and objective measures of multiple sclerosis (MS) disability that can be administered remotely and out-of-clinic. Deep Convolutional Neural Networks (DCNN) may capture a richer representation of healthy and MS-related ambulatory characteristics from the raw smartphone-based inertial sensor data than standard feature-based methodologies. To overcome the typical limitations associated with remotely generated health data, such as low subject numbers, sparsity, and heterogeneous data, a transfer learning (TL) model from similar large open-source datasets was proposed. Our TL framework leveraged the ambulatory information learned on human activity recognition (HAR) tasks collected from wearable smartphone sensor data. It was demonstrated that fine-tuning TL DCNN HAR models towards MS disease recognition tasks outperformed previous Support Vector Machine (SVM) feature-based methods, as well as DCNN models trained end-to-end, by upwards of 8-15%. A lack of transparency of "black-box" deep networks remains one of the largest stumbling blocks to the wider acceptance of deep learning for clinical applications. Ensuing work therefore aimed to visualise DCNN decisions attributed by relevance heatmaps using Layer-Wise Relevance Propagation (LRP). Through the LRP framework, the patterns captured from smartphone-based inertial sensor data that were reflective of those who are healthy versus people with MS (PwMS) could begin to be established and understood. Interpretations suggested that cadence-based measures, gait speed, and ambulation-related signal perturbations were distinct characteristics that distinguished MS disability from healthy participants. Robust and interpretable outcomes, generated from high-frequency out-of-clinic assessments, could greatly augment the current in-clinic assessment picture for PwMS, to inform better disease management techniques, and enable the development of better therapeutic interventions.
Collapse
Affiliation(s)
- Andrew P Creagh
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK.
| | | | | | - Maarten De Vos
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK
- Department of Electrical Engineering, KU Leuven, Leuven, Belgium
- Department of Development and Regeneration, KU Leuven, Leuven, Belgium
| |
Collapse
|
42
|
Vandecasteele K, De Cooman T, Chatzichristos C, Cleeren E, Swinnen L, Macea Ortiz J, Van Huffel S, Dümpelmann M, Schulze-Bonhage A, De Vos M, Van Paesschen W, Hunyadi B. The power of ECG in multimodal patient-specific seizure monitoring: Added value to an EEG-based detector using limited channels. Epilepsia 2021; 62:2333-2343. [PMID: 34240748 PMCID: PMC8518059 DOI: 10.1111/epi.16990] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2021] [Revised: 06/17/2021] [Accepted: 06/17/2021] [Indexed: 11/27/2022]
Abstract
Objective Wearable seizure detection devices could provide more reliable seizure documentation outside the hospital compared to seizure self‐reporting by patients, which is the current standard. Previously, during the SeizeIT1 project, we studied seizure detection based on behind‐the‐ear electroencephalography (EEG). However, the obtained sensitivities were too low for practical use, because not all seizures are associated with typical ictal EEG patterns. Therefore, in this paper, we aim to develop a multimodal automated seizure detection algorithm integrating behind‐the‐ear EEG and electrocardiography (ECG) for detecting focal seizures. In this framework, we quantified the added value of ECG to behind‐the‐ear EEG. Methods This study analyzed three multicenter databases consisting of 135 patients having focal epilepsy and a total of 896 seizures. A patient‐specific multimodal automated seizure detection algorithm was developed using behind‐the‐ear/temporal EEG and single‐lead ECG. The EEG and ECG data were processed separately using machine learning methods. A late integration approach was applied for fusing those predictions. Results The multimodal algorithm outperformed the EEG‐based algorithm in two of three databases, with an increase of 11% and 8% in sensitivity for the same false alarm rate. Significance ECG can be of added value to an EEG‐based seizure detection algorithm using only behind‐the‐ear/temporal lobe electrodes for patients with focal epilepsy.
Collapse
Affiliation(s)
- Kaat Vandecasteele
- Department of Electrical Engineering, STADIUS Center for Dynamic Systems, Signal Processing and Data Analytics, KU Leuven, Leuven, Belgium
| | - Thomas De Cooman
- Department of Electrical Engineering, STADIUS Center for Dynamic Systems, Signal Processing and Data Analytics, KU Leuven, Leuven, Belgium
| | - Christos Chatzichristos
- Department of Electrical Engineering, STADIUS Center for Dynamic Systems, Signal Processing and Data Analytics, KU Leuven, Leuven, Belgium
| | - Evy Cleeren
- Department of Neurology, University Hospitals Leuven, Leuven, Belgium
| | - Lauren Swinnen
- Laboratory for Epilepsy Research, Department of Neurology, University Hospital, KU Leuven, Leuven, Belgium
| | - Jaiver Macea Ortiz
- Laboratory for Epilepsy Research, Department of Neurology, University Hospital, KU Leuven, Leuven, Belgium
| | - Sabine Van Huffel
- Department of Electrical Engineering, STADIUS Center for Dynamic Systems, Signal Processing and Data Analytics, KU Leuven, Leuven, Belgium
| | - Matthias Dümpelmann
- Faculty of Medicine, Department of Neurosurgery, Epilepsy Center, University of Freiburg, Freiburg, Germany
| | - Andreas Schulze-Bonhage
- Faculty of Medicine, Department of Neurosurgery, Epilepsy Center, University of Freiburg, Freiburg, Germany
| | - Maarten De Vos
- Department of Electrical Engineering, STADIUS Center for Dynamic Systems, Signal Processing and Data Analytics, KU Leuven, Leuven, Belgium.,Department of Development and Regeneration, KU Leuven, Leuven, Belgium
| | - Wim Van Paesschen
- Laboratory for Epilepsy Research, Department of Neurology, University Hospital, KU Leuven, Leuven, Belgium
| | - Borbála Hunyadi
- Department of Microelectronics, TU Delft, Delft, Netherlands
| |
Collapse
|
43
|
Cesari M, Stefani A, Penzel T, Ibrahim A, Hackner H, Heidbreder A, Szentkirályi A, Stubbe B, Völzke H, Berger K, Högl B. Interrater sleep stage scoring reliability between manual scoring from two European sleep centers and automatic scoring performed by the artificial intelligence-based Stanford-STAGES algorithm. J Clin Sleep Med 2021; 17:1237-1247. [PMID: 33599203 DOI: 10.5664/jcsm.9174] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
Abstract
STUDY OBJECTIVES The objective of this study was to evaluate interrater reliability between manual sleep stage scoring performed in 2 European sleep centers and automatic sleep stage scoring performed by the previously validated artificial intelligence-based Stanford-STAGES algorithm. METHODS Full night polysomnographies of 1,066 participants were included. Sleep stages were manually scored in Berlin and Innsbruck sleep centers and automatically scored with the Stanford-STAGES algorithm. For each participant, we compared (1) Innsbruck to Berlin scorings (INN vs BER); (2) Innsbruck to automatic scorings (INN vs AUTO); (3) Berlin to automatic scorings (BER vs AUTO); (4) epochs where scorers from Innsbruck and Berlin had consensus to automatic scoring (CONS vs AUTO); and (5) both Innsbruck and Berlin manual scorings (MAN) to the automatic ones (MAN vs AUTO). Interrater reliability was evaluated with several measures, including overall and sleep stage-specific Cohen's κ. RESULTS Overall agreement across participants was substantial for INN vs BER (κ = 0.66 ± 0.13), INN vs AUTO (κ = 0.68 ± 0.14), CONS vs AUTO (κ = 0.73 ± 0.14), and MAN vs AUTO (κ = 0.61 ± 0.14), and moderate for BER vs AUTO (κ = 0.55 ± 0.15). Human scorers had the highest disagreement for N1 sleep (κN1 = 0.40 ± 0.16 for INN vs BER). Automatic scoring had lowest agreement with manual scorings for N1 and N3 sleep (κN1 = 0.25 ± 0.14 and κN3 = 0.42 ± 0.32 for MAN vs AUTO). CONCLUSIONS Interrater reliability for sleep stage scoring between human scorers was in line with previous findings, and the algorithm achieved an overall substantial agreement with manual scoring. In this cohort, the Stanford-STAGES algorithm showed similar performances to the ones achieved in the original study, suggesting that it is generalizable to new cohorts. Before its integration in clinical practice, future independent studies should further evaluate it in other cohorts.
Collapse
Affiliation(s)
- Matteo Cesari
- Department of Neurology, Medical University of Innsbruck, Innsbruck, Austria
| | - Ambra Stefani
- Department of Neurology, Medical University of Innsbruck, Innsbruck, Austria
| | - Thomas Penzel
- Interdisciplinary Sleep Medicine Center, Charité-Universitätsmedizin Berlin, Berlin, Germany.,Saratov State University, Saratov, Russian Federation
| | - Abubaker Ibrahim
- Department of Neurology, Medical University of Innsbruck, Innsbruck, Austria
| | - Heinz Hackner
- Department of Neurology, Medical University of Innsbruck, Innsbruck, Austria
| | - Anna Heidbreder
- Department of Neurology, Medical University of Innsbruck, Innsbruck, Austria
| | - András Szentkirályi
- Institute of Epidemiology and Social Medicine, University of Münster, Münster, Germany
| | - Beate Stubbe
- Department of Internal Medicine B, University Medicine Greifswald, Greifswald, Germany
| | - Henry Völzke
- Institute for Community Medicine, University Medicine Greifswald, Greifswald, Germany
| | - Klaus Berger
- Institute of Epidemiology and Social Medicine, University of Münster, Münster, Germany
| | - Birgit Högl
- Department of Neurology, Medical University of Innsbruck, Innsbruck, Austria
| |
Collapse
|
44
|
Banluesombatkul N, Ouppaphan P, Leelaarporn P, Lakhan P, Chaitusaney B, Jaimchariyatam N, Chuangsuwanich E, Chen W, Phan H, Dilokthanakul N, Wilaiprasitporn T. MetaSleepLearner: A Pilot Study on Fast Adaptation of Bio-Signals-Based Sleep Stage Classifier to New Individual Subject Using Meta-Learning. IEEE J Biomed Health Inform 2021; 25:1949-1963. [PMID: 33180737 DOI: 10.1109/jbhi.2020.3037693] [Citation(s) in RCA: 31] [Impact Index Per Article: 10.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Identifying bio-signals based-sleep stages requires time-consuming and tedious labor of skilled clinicians. Deep learning approaches have been introduced in order to challenge the automatic sleep stage classification conundrum. However, the difficulties can be posed in replacing the clinicians with the automatic system due to the differences in many aspects found in individual bio-signals, causing the inconsistency in the performance of the model on every incoming individual. Thus, we aim to explore the feasibility of using a novel approach, capable of assisting the clinicians and lessening the workload. We propose the transfer learning framework, entitled MetaSleepLearner, based on Model Agnostic Meta-Learning (MAML), in order to transfer the acquired sleep staging knowledge from a large dataset to new individual subjects (source code is available at https://github.com/IoBT-VISTEC/MetaSleepLearner). The framework was demonstrated to require the labelling of only a few sleep epochs by the clinicians and allow the remainder to be handled by the system. Layer-wise Relevance Propagation (LRP) was also applied to understand the learning course of our approach. In all acquired datasets, in comparison to the conventional approach, MetaSleepLearner achieved a range of 5.4% to 17.7% improvement with statistical difference in the mean of both approaches. The illustration of the model interpretation after the adaptation to each subject also confirmed that the performance was directed towards reasonable learning. MetaSleepLearner outperformed the conventional approaches as a result from the fine-tuning using the recordings of both healthy subjects and patients. This is the first work that investigated a non-conventional pre-training method, MAML, resulting in a possibility for human-machine collaboration in sleep stage classification and easing the burden of the clinicians in labelling the sleep stages through only several epochs rather than an entire recording.
Collapse
|
45
|
Ko W, Jeon E, Jeong S, Phyo J, Suk HI. A Survey on Deep Learning-Based Short/Zero-Calibration Approaches for EEG-Based Brain-Computer Interfaces. Front Hum Neurosci 2021; 15:643386. [PMID: 34140883 PMCID: PMC8204721 DOI: 10.3389/fnhum.2021.643386] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2020] [Accepted: 04/27/2021] [Indexed: 11/28/2022] Open
Abstract
Brain-computer interfaces (BCIs) utilizing machine learning techniques are an emerging technology that enables a communication pathway between a user and an external system, such as a computer. Owing to its practicality, electroencephalography (EEG) is one of the most widely used measurements for BCI. However, EEG has complex patterns and EEG-based BCIs mostly involve a cost/time-consuming calibration phase; thus, acquiring sufficient EEG data is rarely possible. Recently, deep learning (DL) has had a theoretical/practical impact on BCI research because of its use in learning representations of complex patterns inherent in EEG. Moreover, algorithmic advances in DL facilitate short/zero-calibration in BCI, thereby suppressing the data acquisition phase. Those advancements include data augmentation (DA), increasing the number of training samples without acquiring additional data, and transfer learning (TL), taking advantage of representative knowledge obtained from one dataset to address the so-called data insufficiency problem in other datasets. In this study, we review DL-based short/zero-calibration methods for BCI. Further, we elaborate methodological/algorithmic trends, highlight intriguing approaches in the literature, and discuss directions for further research. In particular, we search for generative model-based and geometric manipulation-based DA methods. Additionally, we categorize TL techniques in DL-based BCIs into explicit and implicit methods. Our systematization reveals advances in the DA and TL methods. Among the studies reviewed herein, ~45% of DA studies used generative model-based techniques, whereas ~45% of TL studies used explicit knowledge transferring strategy. Moreover, based on our literature review, we recommend an appropriate DA strategy for DL-based BCIs and discuss trends of TLs used in DL-based BCIs.
Collapse
Affiliation(s)
- Wonjun Ko
- Department of Brain and Cognitive Engineering, Korea University, Seoul, South Korea
| | - Eunjin Jeon
- Department of Brain and Cognitive Engineering, Korea University, Seoul, South Korea
| | - Seungwoo Jeong
- Department of Artificial Intelligence, Korea University, Seoul, South Korea
| | - Jaeun Phyo
- Department of Brain and Cognitive Engineering, Korea University, Seoul, South Korea
| | - Heung-Il Suk
- Department of Brain and Cognitive Engineering, Korea University, Seoul, South Korea
- Department of Artificial Intelligence, Korea University, Seoul, South Korea
| |
Collapse
|
46
|
Perslev M, Darkner S, Kempfner L, Nikolic M, Jennum PJ, Igel C. U-Sleep: resilient high-frequency sleep staging. NPJ Digit Med 2021; 4:72. [PMID: 33859353 PMCID: PMC8050216 DOI: 10.1038/s41746-021-00440-5] [Citation(s) in RCA: 84] [Impact Index Per Article: 28.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2020] [Accepted: 03/10/2021] [Indexed: 02/02/2023] Open
Abstract
Sleep disorders affect a large portion of the global population and are strong predictors of morbidity and all-cause mortality. Sleep staging segments a period of sleep into a sequence of phases providing the basis for most clinical decisions in sleep medicine. Manual sleep staging is difficult and time-consuming as experts must evaluate hours of polysomnography (PSG) recordings with electroencephalography (EEG) and electrooculography (EOG) data for each patient. Here, we present U-Sleep, a publicly available, ready-to-use deep-learning-based system for automated sleep staging ( sleep.ai.ku.dk ). U-Sleep is a fully convolutional neural network, which was trained and evaluated on PSG recordings from 15,660 participants of 16 clinical studies. It provides accurate segmentations across a wide range of patient cohorts and PSG protocols not considered when building the system. U-Sleep works for arbitrary combinations of typical EEG and EOG channels, and its special deep learning architecture can label sleep stages at shorter intervals than the typical 30 s periods used during training. We show that these labels can provide additional diagnostic information and lead to new ways of analyzing sleep. U-Sleep performs on par with state-of-the-art automatic sleep staging systems on multiple clinical datasets, even if the other systems were built specifically for the particular data. A comparison with consensus-scores from a previously unseen clinic shows that U-Sleep performs as accurately as the best of the human experts. U-Sleep can support the sleep staging workflow of medical experts, which decreases healthcare costs, and can provide highly accurate segmentations when human expertize is lacking.
Collapse
Affiliation(s)
- Mathias Perslev
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark
| | - Sune Darkner
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark
| | - Lykke Kempfner
- Danish Center for Sleep Medicine, Rigshospitalet, Copenhagen, Denmark
| | - Miki Nikolic
- Danish Center for Sleep Medicine, Rigshospitalet, Copenhagen, Denmark
| | | | - Christian Igel
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.
| |
Collapse
|
47
|
Raurale SA, Boylan GB, Mathieson S, Marnane WP, Lightbody G, O'Toole JM. Grading hypoxic-ischemic encephalopathy in neonatal EEG with convolutional neural networks and quadratic time-frequency distributions. J Neural Eng 2021; 18. [PMID: 33618337 PMCID: PMC8208632 DOI: 10.1088/1741-2552/abe8ae] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2020] [Accepted: 02/22/2021] [Indexed: 12/12/2022]
Abstract
OBJECTIVE To develop an automated system to classify the severity of hypoxic-ischaemic encephalopathy injury (HIE) in neonates from the background electroencephalogram (EEG). METHOD By combining a quadratic time{frequency distribution (TFD) with a convolutional neural network, we develop a system that classifies 4 EEG grades of HIE. The network learns directly from the two- dimensional TFD through 3 independent layers with convolution in the time, frequency, and time{frequency directions. Computationally efficient algorithms make it feasible to transform each 5 minute epoch to the time-frequency domain by controlling for oversampling to reduce both computation and computer memory. The system is developed on EEG recordings from 54 neonates. Then the system is validated on a large unseen dataset of 338 hours of EEG recordings from 91 neonates obtained across multiple international centres. RESULTS The proposed EEG HIE-grading system achieves a leave-one-subject-out testing accuracy of 88.9% and kappa of 0.84 on the development dataset. Accuracy for the large unseen test dataset is 69.5% (95% confidence interval, CI: 65.3 to 73.6%) and kappa of 0.54, which is a significant (P < 0.001) improvement over a state-of-the-art feature-based method with an accuracy of 56.8% (95% CI: 51.4 to 61.7%) and kappa of 0.39. Performance of the proposed system was unaffected when the number of channels in testing was reduced from 8 to 2|accuracy for large validation dataset remained at 69.5% (95% CI: 65.5 to 74.0%). SIGNIFICANCE The proposed system outperforms the state-of-the-art machine learning algorithms for EEG grade classification on a large multi-centre unseen dataset, indicating the potential to assist clinical decision making for neonates with HIE.
Collapse
Affiliation(s)
- Sumit Arun Raurale
- Pediatrics and child health, INFANT Centre, University College Cork, Cork, Cork, T12 DC4A, IRELAND
| | - Geraldine B Boylan
- Department of Paediatrics and Child Health, University College Cork, University College Cork,, Cork, IRELAND
| | - Sean Mathieson
- Podiatric and Child Health, INFANT Centre, University College Cork, Wilton, Cork, T12 DC4A, IRELAND
| | - W P Marnane
- Department of Electrical Engineering and Microelectronics, University College Cork, College Road, Cork, T12 DC4A, IRELAND
| | - Gordon Lightbody
- Department of Electrical Engineering and Microelectronics, University College Cork, College Road, Cork, T12 DC4A, IRELAND
| | - John M O'Toole
- Irish Centre for Fetal and Neonatal Translational Research, Dept. ofPaediatrics and Child Health, University College Cork National University of Ireland, Western Gateway Building, Western Road, Cork, IRELAND
| |
Collapse
|