Toward Interpretable Sleep Stage Classification Using Cross-Modal Transformers

Accurate sleep stage classification is significant for sleep health assessment. In recent years, several machine-learning based sleep staging algorithms have been developed, and in particular, deep-learning based algorithms have achieved performance on par with human annotation. Despite improved per...

Full description

Saved in:
Bibliographic Details
Main Authors: Jathurshan Pradeepkumar (Author), Mithunjha Anandakumar (Author), Vinith Kugathasan (Author), Dhinesh Suntharalingham (Author), Simon L. Kappel (Author), Anjula C. De Silva (Author), Chamira U. S. Edussooriya (Author)
Format: Book
Published: IEEE, 2024-01-01T00:00:00Z.
Subjects:
Online Access:Connect to this object online.
Tags: Add Tag
No Tags, Be the first to tag this record!

MARC

LEADER 00000 am a22000003u 4500
001 doaj_941ab13af7b84b45915e3cc2f938b29d
042 |a dc 
100 1 0 |a Jathurshan Pradeepkumar  |e author 
700 1 0 |a Mithunjha Anandakumar  |e author 
700 1 0 |a Vinith Kugathasan  |e author 
700 1 0 |a Dhinesh Suntharalingham  |e author 
700 1 0 |a Simon L. Kappel  |e author 
700 1 0 |a Anjula C. De Silva  |e author 
700 1 0 |a Chamira U. S. Edussooriya  |e author 
245 0 0 |a Toward Interpretable Sleep Stage Classification Using Cross-Modal Transformers 
260 |b IEEE,   |c 2024-01-01T00:00:00Z. 
500 |a 1534-4320 
500 |a 1558-0210 
500 |a 10.1109/TNSRE.2024.3438610 
520 |a Accurate sleep stage classification is significant for sleep health assessment. In recent years, several machine-learning based sleep staging algorithms have been developed, and in particular, deep-learning based algorithms have achieved performance on par with human annotation. Despite improved performance, a limitation of most deep-learning based algorithms is their black-box behavior, which have limited their use in clinical settings. Here, we propose a cross-modal transformer, which is a transformer-based method for sleep stage classification. The proposed cross-modal transformer consists of a cross-modal transformer encoder architecture along with a multi-scale one-dimensional convolutional neural network for automatic representation learning. The performance of our method is on-par with the state-of-the-art methods and eliminates the black-box behavior of deep-learning models by utilizing the interpretability aspect of the attention modules. Furthermore, our method provides considerable reductions in the number of parameters and training time compared to the state-of-the-art methods. Our code is available at <uri>https://github.com/Jathurshan0330/Cross-Modal-Transformer</uri>. A demo of our work can be found at <uri>https://bit.ly/Cross_modal_transformer_demo</uri>. 
546 |a EN 
690 |a Automatic sleep stage classification 
690 |a interpretable deep learning 
690 |a transformers 
690 |a deep neural networks 
690 |a Medical technology 
690 |a R855-855.5 
690 |a Therapeutics. Pharmacology 
690 |a RM1-950 
655 7 |a article  |2 local 
786 0 |n IEEE Transactions on Neural Systems and Rehabilitation Engineering, Vol 32, Pp 2893-2904 (2024) 
787 0 |n https://ieeexplore.ieee.org/document/10623416/ 
787 0 |n https://doaj.org/toc/1534-4320 
787 0 |n https://doaj.org/toc/1558-0210 
856 4 1 |u https://doaj.org/article/941ab13af7b84b45915e3cc2f938b29d  |z Connect to this object online.