Adoption of Transformer Neural Network to Improve the Diagnostic Performance of Oximetry for Obstructive Sleep Apnea


Creative Commons License

Almarshad M. A., Al-Ahmadi S., Islam S., BaHammam A. S., Soudani A.

Sensors, cilt.23, sa.18, 2023 (SCI-Expanded) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 23 Sayı: 18
  • Basım Tarihi: 2023
  • Doi Numarası: 10.3390/s23187924
  • Dergi Adı: Sensors
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus, Academic Search Premier, Aerospace Database, Aqualine, Aquatic Science & Fisheries Abstracts (ASFA), Biotechnology Research Abstracts, CAB Abstracts, Communication Abstracts, Compendex, INSPEC, MEDLINE, Metadex, Veterinary Science Database, Directory of Open Access Journals, Civil Engineering Abstracts
  • Anahtar Kelimeler: artificial intelligence (AI), autoscoring, deep learning (DL), obstructive sleep apnea (OSA), oxygen saturation (SpO2), polysomnography (PSG), transformer neural networks
  • TED Üniversitesi Adresli: Evet

Özet

Scoring polysomnography for obstructive sleep apnea diagnosis is a laborious, long, and costly process. Machine learning approaches, such as deep neural networks, can reduce scoring time and costs. However, most methods require prior filtering and preprocessing of the raw signal. Our work presents a novel method for diagnosing obstructive sleep apnea using a transformer neural network with learnable positional encoding, which outperforms existing state-of-the-art solutions. This approach has the potential to improve the diagnostic performance of oximetry for obstructive sleep apnea and reduce the time and costs associated with traditional polysomnography. Contrary to existing approaches, our approach performs annotations at one-second granularity. Allowing physicians to interpret the model’s outcome. In addition, we tested different positional encoding designs as the first layer of the model, and the best results were achieved using a learnable positional encoding based on an autoencoder with structural novelty. In addition, we tried different temporal resolutions with various granularity levels from 1 to 360 s. All experiments were carried out on an independent test set from the public OSASUD dataset and showed that our approach outperforms current state-of-the-art solutions with a satisfactory AUC of 0.89, accuracy of 0.80, and F1-score of 0.79.