Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Standard

Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks. / Hanif, Umaer; Kezirian, Eric; Kiar, Eva Kirkegaard; Mignot, Emmanuel; Sorensen, Helge B.D.; Jennum, Poul.

2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021. IEEE, 2021. p. 3957-3960 (Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS).

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Harvard

Hanif, U, Kezirian, E, Kiar, EK, Mignot, E, Sorensen, HBD & Jennum, P 2021, Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks. in 2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021. IEEE, Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS, pp. 3957-3960, 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021, Virtual, Online, Mexico, 01/11/2021. https://doi.org/10.1109/EMBC46164.2021.9630098

APA

Hanif, U., Kezirian, E., Kiar, E. K., Mignot, E., Sorensen, H. B. D., & Jennum, P. (2021). Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks. In 2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021 (pp. 3957-3960). IEEE. Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS https://doi.org/10.1109/EMBC46164.2021.9630098

Vancouver

Hanif U, Kezirian E, Kiar EK, Mignot E, Sorensen HBD, Jennum P. Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks. In 2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021. IEEE. 2021. p. 3957-3960. (Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS). https://doi.org/10.1109/EMBC46164.2021.9630098

Author

Hanif, Umaer ; Kezirian, Eric ; Kiar, Eva Kirkegaard ; Mignot, Emmanuel ; Sorensen, Helge B.D. ; Jennum, Poul. / Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks. 2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021. IEEE, 2021. pp. 3957-3960 (Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS).

Bibtex

@inproceedings{9e3e26cd640840169d4b3daf7e974a27,
title = "Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks",
abstract = "Assessing the upper airway (UA) of obstructive sleep apnea patients using drug-induced sleep endoscopy (DISE) before potential surgery is standard practice in clinics to determine the location of UA collapse. According to the VOTE classification system, UA collapse can occur at the velum (V), oropharynx (O), tongue (T), and/or epiglottis (E). Analyzing DISE videos is not trivial due to anatomical variation, simultaneous UA collapse in several locations, and video distortion caused by mucus or saliva. The first step towards automated analysis of DISE videos is to determine which UA region the endoscope is in at any time throughout the video: V (velum) or OTE (oropharynx, tongue, or epiglottis). An additional class denoted X is introduced for times when the video is distorted to an extent where it is impossible to determine the region. This paper is a proof of concept for classifying UA regions using 24 annotated DISE videos. We propose a convolutional recurrent neural network using a ResNet18 architecture combined with a two-layer bidirectional long short-term memory network. The classifications were performed on a sequence of 5 seconds of video at a time. The network achieved an overall accuracy of 82% and F1-score of 79% for the three-class problem, showing potential for recognition of regions across patients despite anatomical variation. Results indicate that large-scale training on videos can be used to further predict the location(s), type(s), and degree(s) of UA collapse, showing potential for derivation of automatic diagnoses from DISE videos eventually.",
author = "Umaer Hanif and Eric Kezirian and Kiar, {Eva Kirkegaard} and Emmanuel Mignot and Sorensen, {Helge B.D.} and Poul Jennum",
note = "Publisher Copyright: {\textcopyright} 2021 IEEE.; 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021 ; Conference date: 01-11-2021 Through 05-11-2021",
year = "2021",
doi = "10.1109/EMBC46164.2021.9630098",
language = "English",
series = "Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS",
publisher = "IEEE",
pages = "3957--3960",
booktitle = "2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021",

}

RIS

TY - GEN

T1 - Upper Airway Classification in Sleep Endoscopy Examinations using Convolutional Recurrent Neural Networks

AU - Hanif, Umaer

AU - Kezirian, Eric

AU - Kiar, Eva Kirkegaard

AU - Mignot, Emmanuel

AU - Sorensen, Helge B.D.

AU - Jennum, Poul

N1 - Publisher Copyright: © 2021 IEEE.

PY - 2021

Y1 - 2021

N2 - Assessing the upper airway (UA) of obstructive sleep apnea patients using drug-induced sleep endoscopy (DISE) before potential surgery is standard practice in clinics to determine the location of UA collapse. According to the VOTE classification system, UA collapse can occur at the velum (V), oropharynx (O), tongue (T), and/or epiglottis (E). Analyzing DISE videos is not trivial due to anatomical variation, simultaneous UA collapse in several locations, and video distortion caused by mucus or saliva. The first step towards automated analysis of DISE videos is to determine which UA region the endoscope is in at any time throughout the video: V (velum) or OTE (oropharynx, tongue, or epiglottis). An additional class denoted X is introduced for times when the video is distorted to an extent where it is impossible to determine the region. This paper is a proof of concept for classifying UA regions using 24 annotated DISE videos. We propose a convolutional recurrent neural network using a ResNet18 architecture combined with a two-layer bidirectional long short-term memory network. The classifications were performed on a sequence of 5 seconds of video at a time. The network achieved an overall accuracy of 82% and F1-score of 79% for the three-class problem, showing potential for recognition of regions across patients despite anatomical variation. Results indicate that large-scale training on videos can be used to further predict the location(s), type(s), and degree(s) of UA collapse, showing potential for derivation of automatic diagnoses from DISE videos eventually.

AB - Assessing the upper airway (UA) of obstructive sleep apnea patients using drug-induced sleep endoscopy (DISE) before potential surgery is standard practice in clinics to determine the location of UA collapse. According to the VOTE classification system, UA collapse can occur at the velum (V), oropharynx (O), tongue (T), and/or epiglottis (E). Analyzing DISE videos is not trivial due to anatomical variation, simultaneous UA collapse in several locations, and video distortion caused by mucus or saliva. The first step towards automated analysis of DISE videos is to determine which UA region the endoscope is in at any time throughout the video: V (velum) or OTE (oropharynx, tongue, or epiglottis). An additional class denoted X is introduced for times when the video is distorted to an extent where it is impossible to determine the region. This paper is a proof of concept for classifying UA regions using 24 annotated DISE videos. We propose a convolutional recurrent neural network using a ResNet18 architecture combined with a two-layer bidirectional long short-term memory network. The classifications were performed on a sequence of 5 seconds of video at a time. The network achieved an overall accuracy of 82% and F1-score of 79% for the three-class problem, showing potential for recognition of regions across patients despite anatomical variation. Results indicate that large-scale training on videos can be used to further predict the location(s), type(s), and degree(s) of UA collapse, showing potential for derivation of automatic diagnoses from DISE videos eventually.

UR - http://www.scopus.com/inward/record.url?scp=85122550872&partnerID=8YFLogxK

U2 - 10.1109/EMBC46164.2021.9630098

DO - 10.1109/EMBC46164.2021.9630098

M3 - Article in proceedings

C2 - 34892097

AN - SCOPUS:85122550872

T3 - Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS

SP - 3957

EP - 3960

BT - 2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021

PB - IEEE

T2 - 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2021

Y2 - 1 November 2021 through 5 November 2021

ER -

ID: 304302369