A4 Vertaisarvioitu artikkeli konferenssijulkaisussa

Multimodal Sleep Stage and Sleep Apnea Classification Using Vision Transformer: A Multitask Explainable Learning Approach




TekijätKazemi, Kianoosh; Azimi, Iman; Khine, Michelle; Khayat, Rami N.; Rahmani, Amir M.; Liljeberg, Pasi

ToimittajaN/A

Konferenssin vakiintunut nimiAnnual International Conference of the IEEE Engineering in Medicine and Biology Society

Julkaisuvuosi2025

Lehti: Annual International Conference of the IEEE Engineering in Medicine and Biology Society

Kokoomateoksen nimi2025 47th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)

Vuosikerta47

ISBN979-8-3315-8619-5

eISBN979-8-3315-8618-8

ISSN2375-7477

eISSN2694-0604

DOIhttps://doi.org/10.1109/EMBC58623.2025.11252880

Julkaisun avoimuus kirjaamishetkelläEi avoimesti saatavilla

Julkaisukanavan avoimuus Ei avoin julkaisukanava

Verkko-osoitehttps://ieeexplore.ieee.org/document/11252880


Tiivistelmä

Sleep is an essential component of human physiology, contributing significantly to overall health and quality of life. Accurate sleep staging and disorder detection are crucial for assessing sleep quality. Studies in the literature have proposed PSG-based approaches and machine-learning methods utilizing single-modality signals. However, existing methods often lack multimodal, multilabel frameworks and address sleep stages and disorders classification separately. In this paper, we propose a 1D-Vision Transformer for simultaneous classification of sleep stages and sleep disorders. Our method exploits the sleep disorders’ correlation with specific sleep stage patterns and performs a simultaneous identification of a sleep stage and sleep disorder. The model is trained and tested using multimodal-multilabel sensory data (including photoplethysmogram, respiratory flow, and respiratory effort signals). The proposed method shows an overall accuracy (cohen’s Kappa) of 78% (0.66) for five-stage sleep classification and 74% (0.58) for sleep apnea classification. Moreover, we analyzed the encoder attention weights to clarify our models’ predictions and investigate the influence different features have on the models’ outputs. The result shows that identified patterns, such as respiratory troughs and peaks, make a higher contribution to the final classification process.


Julkaisussa olevat rahoitustiedot
This work was partially supported by the Finnish Foundation for Technology Promotion and the Nokia Foundation.


Last updated on 2025-04-12 at 12:36