1
|
Miller ME, Witte D, Lina I, Walsh J, Rameau A, Bhatti NI. Development of Machine Learning Copilot to Assist Novices in Learning Flexible Laryngoscopy. Laryngoscope 2024. [PMID: 39363661 DOI: 10.1002/lary.31812] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2024] [Revised: 09/07/2024] [Accepted: 09/17/2024] [Indexed: 10/05/2024]
Abstract
OBJECTIVES Here we describe the development and pilot testing of the first artificial intelligence (AI) software "copilot" to help train novices to competently perform flexible fiberoptic laryngoscopy (FFL) on a mannikin and improve their uptake of FFL skills. METHODS Supervised machine learning was used to develop an image classifier model, dubbed the "anatomical region classifier," responsible for predicting the location of camera in the upper aerodigestive tract and an object detection model, dubbed the "anatomical structure detector," responsible for locating and identifying key anatomical structures in images. Training data were collected by performing FFL on an AirSim Combo Bronchi X mannikin (United Kingdom, TruCorp Ltd) using an Ambu aScope 4 RhinoLaryngo Slim connected to an Ambu® aView™ 2 Advance Displaying Unit (Ballerup, Ambu A/S). Medical students were prospectively recruited to try the FFL copilot and rate its ease of use and self-rate their skills with and without the copilot. RESULTS This model classified anatomical regions with an overall accuracy of 91.9% on the validation set and 80.1% on the test set. The model detected anatomical structures with overall mean average precision of 0.642. Through various optimizations, we were able to run the AI copilot at approximately 28 frames per second (FPS), which is imperceptible from real time and nearly matches the video frame rate of 30 FPS. Sixty-four novice medical students were recruited for feedback on the copilot. Although 90.9% strongly agreed/agreed that the AI copilot was easy to use, their self-rating of FFL skills following use of the copilot were overall equivocal to their self-rating without the copilot. CONCLUSIONS The AI copilot tracked successful capture of diagnosable views of key anatomical structures effectively guiding users through FFL to ensure all anatomical structures are sufficiently captured. This tool has the potential to assist novices in efficiently gaining competence in FFL. LEVEL OF EVIDENCE NA Laryngoscope, 2024.
Collapse
Affiliation(s)
- Mattea E Miller
- Department of Otolaryngology-Head and Neck Surgery, Johns Hopkins School of Medicine, Baltimore, Maryland, USA
| | - Dan Witte
- Perceptron Health, Inc, New York, New York, U.S.A
| | - Ioan Lina
- Department of Otolaryngology-Head and Neck Surgery, Johns Hopkins School of Medicine, Baltimore, Maryland, USA
| | - Jonathan Walsh
- Department of Otolaryngology-Head and Neck Surgery, Johns Hopkins School of Medicine, Baltimore, Maryland, USA
| | - Anaïs Rameau
- Department of Otolaryngology - Head and Neck Surgery, Sean Parker Institute for the Voice, Weill Cornell Medical College, New York, New York, U.S.A
| | - Nasir I Bhatti
- Department of Otolaryngology-Head and Neck Surgery, Johns Hopkins School of Medicine, Baltimore, Maryland, USA
| |
Collapse
|
2
|
Chang KM, Surapaneni SS, Shaikh N, Marston AP, Vecchiotti MA, Rangarajan N, Hill CA, Scott AR. Pediatric tympanostomy tube assessment via deep learning. Am J Otolaryngol 2024; 45:104334. [PMID: 38723380 DOI: 10.1016/j.amjoto.2024.104334] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2024] [Accepted: 04/21/2024] [Indexed: 06/14/2024]
Abstract
PURPOSE Tympanostomy tube (TT) placement is the most frequently performed ambulatory surgery in children under 15. After the procedure it is recommended that patients follow up regularly for "tube checks" until TT extrusion. Such visits incur direct and indirect costs to families in the form of days off from work, copays, and travel expenses. This pilot study aims to compare the efficacy of tympanic membrane (TM) evaluation by an artificial intelligence algorithm with that of clinical staff for determining presence or absence of a tympanostomy tube within the TM. METHODS Using a digital otoscope, we performed a prospective study in children (ages 10 months-10 years) with a history of TTs who were being seen for follow up in a pediatric otolaryngology clinic. A smartphone otoscope was used by study personnel who were not physicians to take ear exam images, then through conventional otoscopic exam, ears were assessed by a clinician for tubes being in place or tubes having extruded from the TM. We trained and tested a deep learning (artificial intelligence) algorithm to assess the images and compared that with the clinician's assessment. RESULTS A total of 123 images were obtained from 28 subjects. The algorithm classified images as TM with or without tube in place. Overall classification accuracy was 97.7 %. Recall and precision were 100 % and 96 %, respectively, for TM without a tube present, and 95 % and 100 %, respectively, for TM with a tube in place. DISCUSSION This is a promising deep learning algorithm for classifying ear tube presence in the TM utilizing images obtained in awake children using an over-the-counter otoscope available to the lay population. We are continuing enrollment, with the goal of building an algorithm to assess tube patency and extrusion.
Collapse
Affiliation(s)
- K M Chang
- Tufts University School of Medicine, Boston, MA, United States of America
| | | | - N Shaikh
- Tufts Medical Center, Boston, MA, United States of America
| | - A P Marston
- Tufts Medical Center, Boston, MA, United States of America
| | - M A Vecchiotti
- Tufts Medical Center, Boston, MA, United States of America
| | - N Rangarajan
- COHI Group, St. Paul, MN, United States of America
| | - C A Hill
- COHI Group, St. Paul, MN, United States of America
| | - A R Scott
- Tufts University School of Medicine, Boston, MA, United States of America; Tufts Medical Center, Boston, MA, United States of America.
| |
Collapse
|
3
|
Shim JH, Sunwoo W, Choi BY, Kim KG, Kim YJ. Improving the Accuracy of Otitis Media with Effusion Diagnosis in Pediatric Patients Using Deep Learning. Bioengineering (Basel) 2023; 10:1337. [PMID: 38002461 PMCID: PMC10669592 DOI: 10.3390/bioengineering10111337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 11/09/2023] [Accepted: 11/13/2023] [Indexed: 11/26/2023] Open
Abstract
Otitis media with effusion (OME), primarily seen in children aged 2 years and younger, is characterized by the presence of fluid in the middle ear, often resulting in hearing loss and aural fullness. While deep learning networks have been explored to aid OME diagnosis, prior work did not often specify if pediatric images were used for training, causing uncertainties about their clinical relevance, especially due to important distinctions between the tympanic membranes of small children and adults. We trained cross-validated ResNet50, DenseNet201, InceptionV3, and InceptionResNetV2 models on 1150 pediatric tympanic membrane images from otoendoscopes to classify OME. When assessed using a separate dataset of 100 pediatric tympanic membrane images, the models achieved mean accuracies of 92.9% (ResNet50), 97.2% (DenseNet201), 96.0% (InceptionV3), and 94.8% (InceptionResNetV2), compared to the seven otolaryngologists that achieved accuracies between 84.0% and 69.0%. The results showed that even the worst-performing model trained on fold 3 of InceptionResNetV2 with an accuracy of 88.0% exceeded the accuracy of the highest-performing otolaryngologist at 84.0%. Our findings suggest that these specifically trained deep learning models can potentially enhance the clinical diagnosis of OME using pediatric otoendoscopic tympanic membrane images.
Collapse
Affiliation(s)
- Jae-Hyuk Shim
- Department of Biomedical Engineering, Gil Medical Center, Gachon University College of Medicine, Incheon 21565, Republic of Korea
| | - Woongsang Sunwoo
- Department of Otorhinolaryngology-Head and Neck Surgery, Gil Medical Center, Gachon University College of Medicine, Incheon 21565, Republic of Korea
| | - Byung Yoon Choi
- Department of Otorhinolaryngology-Head and Neck Surgery, Seoul National University Bundang Hospital, Seongnam 13620, Republic of Korea
| | - Kwang Gi Kim
- Department of Biomedical Engineering, Gil Medical Center, Gachon University College of Medicine, Incheon 21565, Republic of Korea
| | - Young Jae Kim
- Department of Biomedical Engineering, Gil Medical Center, Gachon University College of Medicine, Incheon 21565, Republic of Korea
| |
Collapse
|
4
|
Tsilivigkos C, Athanasopoulos M, Micco RD, Giotakis A, Mastronikolis NS, Mulita F, Verras GI, Maroulis I, Giotakis E. Deep Learning Techniques and Imaging in Otorhinolaryngology-A State-of-the-Art Review. J Clin Med 2023; 12:6973. [PMID: 38002588 PMCID: PMC10672270 DOI: 10.3390/jcm12226973] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2023] [Revised: 11/02/2023] [Accepted: 11/06/2023] [Indexed: 11/26/2023] Open
Abstract
Over the last decades, the field of medicine has witnessed significant progress in artificial intelligence (AI), the Internet of Medical Things (IoMT), and deep learning (DL) systems. Otorhinolaryngology, and imaging in its various subspecialties, has not remained untouched by this transformative trend. As the medical landscape evolves, the integration of these technologies becomes imperative in augmenting patient care, fostering innovation, and actively participating in the ever-evolving synergy between computer vision techniques in otorhinolaryngology and AI. To that end, we conducted a thorough search on MEDLINE for papers published until June 2023, utilizing the keywords 'otorhinolaryngology', 'imaging', 'computer vision', 'artificial intelligence', and 'deep learning', and at the same time conducted manual searching in the references section of the articles included in our manuscript. Our search culminated in the retrieval of 121 related articles, which were subsequently subdivided into the following categories: imaging in head and neck, otology, and rhinology. Our objective is to provide a comprehensive introduction to this burgeoning field, tailored for both experienced specialists and aspiring residents in the domain of deep learning algorithms in imaging techniques in otorhinolaryngology.
Collapse
Affiliation(s)
- Christos Tsilivigkos
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| | - Michail Athanasopoulos
- Department of Otolaryngology, University Hospital of Patras, 265 04 Patras, Greece; (M.A.); (N.S.M.)
| | - Riccardo di Micco
- Department of Otolaryngology and Head and Neck Surgery, Medical School of Hannover, 30625 Hannover, Germany;
| | - Aris Giotakis
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| | - Nicholas S. Mastronikolis
- Department of Otolaryngology, University Hospital of Patras, 265 04 Patras, Greece; (M.A.); (N.S.M.)
| | - Francesk Mulita
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Georgios-Ioannis Verras
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Ioannis Maroulis
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Evangelos Giotakis
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| |
Collapse
|
5
|
Taylor A, Habib AR, Kumar A, Wong E, Hasan Z, Singh N. An artificial intelligence algorithm for the classification of sphenoid sinus pneumatisation on sinus computed tomography scans. Clin Otolaryngol 2023; 48:888-894. [PMID: 37488094 DOI: 10.1111/coa.14088] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2022] [Revised: 02/17/2023] [Accepted: 06/18/2023] [Indexed: 07/26/2023]
Abstract
BACKGROUND Classifying sphenoid pneumatisation is an important but often overlooked task in reporting sinus CT scans. Artificial intelligence (AI) and one of its key methods, convolutional neural networks (CNNs), can create algorithms that can learn from data without being programmed with explicit rules and have shown utility in radiological image classification. OBJECTIVE To determine if a trained CNN can accurately classify sphenoid sinus pneumatisation on CT sinus imaging. METHODS Sagittal slices through the natural ostium of the sphenoid sinus were extracted from retrospectively collected bone-window CT scans of the paranasal sinuses for consecutive patients over 6 years. Two blinded Otolaryngology residents reviewed each image and classified the sphenoid sinus pneumatisation as either conchal, presellar or sellar. An AI algorithm was developed using the Microsoft Azure Custom Vision deep learning platform to classify the pattern of pneumatisation. RESULTS Seven hundred eighty images from 400 patients were used to train the algorithm, which was then tested on a further 118 images from 62 patients. The algorithm achieved an accuracy of 93.2% (95% confidence interval [CI] 87.1-97.0), 87.3% (95% CI 79.9-92.7) and 85.6% (95% CI 78.0-91.4) in correctly identifying conchal, presellar and sellar sphenoid pneumatisation, respectively. The overall weighted accuracy of the CNN was 85.9%. CONCLUSION The CNN described demonstrated a moderately accurate classification of sphenoid pneumatisation subtypes on CT scans. The use of CNN-based assistive tools may enable surgeons to achieve safer operative planning through routine automated reporting allowing greater resources to be directed towards the identification of pathology.
Collapse
Affiliation(s)
- Alon Taylor
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Westmead, New South Wales, Australia
| | - Al-Rahim Habib
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Westmead, New South Wales, Australia
- Faculty of Medicine and Health, University of Sydney, Sydney, New South Wales, Australia
| | - Ashnil Kumar
- School of Biomedical Engineering, Faculty of Engineering, University of Sydney, Sydney, New South Wales, Australia
- ARC Training Centre for Innovative BioEngineering, Sydney, New South Wales, Australia
| | - Eugene Wong
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Westmead, New South Wales, Australia
- Westmead Clinical School, Faculty of Medicine and Health, University of Sydney, Sydney, New South Wales, Australia
| | - Zubair Hasan
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Westmead, New South Wales, Australia
| | - Narinder Singh
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Westmead, New South Wales, Australia
- Westmead Clinical School, Faculty of Medicine and Health, University of Sydney, Sydney, New South Wales, Australia
| |
Collapse
|
6
|
Wu Q, Wang X, Liang G, Luo X, Zhou M, Deng H, Zhang Y, Huang X, Yang Q. Advances in Image-Based Artificial Intelligence in Otorhinolaryngology-Head and Neck Surgery: A Systematic Review. Otolaryngol Head Neck Surg 2023; 169:1132-1142. [PMID: 37288505 DOI: 10.1002/ohn.391] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2023] [Revised: 04/27/2023] [Accepted: 05/13/2023] [Indexed: 06/09/2023]
Abstract
OBJECTIVE To update the literature and provide a systematic review of image-based artificial intelligence (AI) applications in otolaryngology, highlight its advances, and propose future challenges. DATA SOURCES Web of Science, Embase, PubMed, and Cochrane Library. REVIEW METHODS Studies written in English, published between January 2020 and December 2022. Two independent authors screened the search results, extracted data, and assessed studies. RESULTS Overall, 686 studies were identified. After screening titles and abstracts, 325 full-text studies were assessed for eligibility, and 78 studies were included in this systematic review. The studies originated from 16 countries. Among these countries, the top 3 were China (n = 29), Korea (n = 8), the United States, and Japan (n = 7 each). The most common area was otology (n = 35), followed by rhinology (n = 20), pharyngology (n = 18), and head and neck surgery (n = 5). Most applications of AI in otology, rhinology, pharyngology, and head and neck surgery mainly included chronic otitis media (n = 9), nasal polyps (n = 4), laryngeal cancer (n = 12), and head and neck squamous cell carcinoma (n = 3), respectively. The overall performance of AI in accuracy, the area under the curve, sensitivity, and specificity were 88.39 ± 9.78%, 91.91 ± 6.70%, 86.93 ± 11.59%, and 88.62 ± 14.03%, respectively. CONCLUSION This state-of-the-art review aimed to highlight the increasing applications of image-based AI in otorhinolaryngology head and neck surgery. The following steps will entail multicentre collaboration to ensure data reliability, ongoing optimization of AI algorithms, and integration into real-world clinical practice. Future studies should consider 3-dimensional (3D)-based AI, such as 3D surgical AI.
Collapse
Affiliation(s)
- Qingwu Wu
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xinyue Wang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Guixian Liang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xin Luo
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Min Zhou
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Huiyi Deng
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Yana Zhang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xuekun Huang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Qintai Yang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| |
Collapse
|
7
|
Song D, Kim T, Lee Y, Kim J. Image-Based Artificial Intelligence Technology for Diagnosing Middle Ear Diseases: A Systematic Review. J Clin Med 2023; 12:5831. [PMID: 37762772 PMCID: PMC10531728 DOI: 10.3390/jcm12185831] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Revised: 08/27/2023] [Accepted: 08/29/2023] [Indexed: 09/29/2023] Open
Abstract
Otolaryngological diagnoses, such as otitis media, are traditionally performed using endoscopy, wherein diagnostic accuracy can be subjective and vary among clinicians. The integration of objective tools, like artificial intelligence (AI), could potentially improve the diagnostic process by minimizing the influence of subjective biases and variability. We systematically reviewed the AI techniques using medical imaging in otolaryngology. Relevant studies related to AI-assisted otitis media diagnosis were extracted from five databases: Google Scholar, PubMed, Medline, Embase, and IEEE Xplore, without date restrictions. Publications that did not relate to AI and otitis media diagnosis or did not utilize medical imaging were excluded. Of the 32identified studies, 26 used tympanic membrane images for classification, achieving an average diagnosis accuracy of 86% (range: 48.7-99.16%). Another three studies employed both segmentation and classification techniques, reporting an average diagnosis accuracy of 90.8% (range: 88.06-93.9%). These findings suggest that AI technologies hold promise for improving otitis media diagnosis, offering benefits for telemedicine and primary care settings due to their high diagnostic accuracy. However, to ensure patient safety and optimal outcomes, further improvements in diagnostic performance are necessary.
Collapse
Affiliation(s)
- Dahye Song
- Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea; (D.S.); (T.K.)
| | - Taewan Kim
- Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea; (D.S.); (T.K.)
| | - Yeonjoon Lee
- Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea; (D.S.); (T.K.)
| | - Jaeyoung Kim
- Department of Dermatology and Skin Sciences, University of British Columbia, Vancouver, BC V6T 1Z1, Canada;
- Core Research & Development Center, Korea University Ansan Hospital, Ansan 15355, Republic of Korea
| |
Collapse
|
8
|
Ding X, Huang Y, Tian X, Zhao Y, Feng G, Gao Z. Diagnosis, Treatment, and Management of Otitis Media with Artificial Intelligence. Diagnostics (Basel) 2023; 13:2309. [PMID: 37443702 DOI: 10.3390/diagnostics13132309] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2023] [Revised: 06/04/2023] [Accepted: 06/14/2023] [Indexed: 07/15/2023] Open
Abstract
A common infectious disease, otitis media (OM) has a low rate of early diagnosis, which significantly increases the difficulty of treating the disease and the likelihood of serious complications developing including hearing loss, speech impairment, and even intracranial infection. Several areas of healthcare have shown great promise in the application of artificial intelligence (AI) systems, such as the accurate detection of diseases, the automated interpretation of images, and the prediction of patient outcomes. Several articles have reported some machine learning (ML) algorithms such as ResNet, InceptionV3 and Unet, were applied to the diagnosis of OM successfully. The use of these techniques in the OM is still in its infancy, but their potential is enormous. We present in this review important concepts related to ML and AI, describe how these technologies are currently being applied to diagnosing, treating, and managing OM, and discuss the challenges associated with developing AI-assisted OM technologies in the future.
Collapse
Affiliation(s)
- Xin Ding
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Yu Huang
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Xu Tian
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Yang Zhao
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Guodong Feng
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Zhiqiang Gao
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| |
Collapse
|
9
|
Cao C, Song J, Su R, Wu X, Wang Z, Hou M. Structure-constrained deep feature fusion for chronic otitis media and cholesteatoma identification. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1-21. [PMID: 37362730 PMCID: PMC10157598 DOI: 10.1007/s11042-023-15425-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/18/2022] [Revised: 03/19/2023] [Accepted: 04/18/2023] [Indexed: 06/28/2023]
Abstract
Chronic suppurative otitis media (CSOM) and middle ear cholesteatoma (MEC) were two most common chronic middle ear disease(MED) clinically. Accurate differential diagnosis between these two diseases is of high clinical importance given the difference in etiologies, lesion manifestations and treatments. The high-resolution computed tomography (CT) scanning of the temporal bone presents a better view of auditory structures, which is currently regarded as the first-line diagnostic imaging modality in the case of MED. In this paper, we first used a region-of-interest (ROI) network to find the area of the middle ear in the entire temporal bone CT image and segment it to a size of 100*100 pixels. Then, we used a structure-constrained deep feature fusion algorithm to convert different characteristic features of the middle ear in three groups as suppurative otitis media (CSOM), middle ear cholesteatoma (MEC) and normal patches. To fuse structure information, we introduced a graph isomorphism network that implements a feature vector from neighbourhoods and the coordinate distance between vertices. Finally, we construct a classifier named the "otitis media, cholesteatoma and normal identification classifier" (OMCNIC). The experimental results achieved by the graph isomorphism network revealed a 96.36% accuracy in all CSOM and MEC classifications. The experimental results indicate that our structure-constrained deep feature fusion algorithm can quickly and effectively classify CSOM and MEC. It will help otologist in the selection of the most appropriate treatment, and the complications can also be reduced.
Collapse
Affiliation(s)
- Cong Cao
- School of Mathematics and Statistics, Central South University, Changsha, 410083 China
| | - Jian Song
- Department of Otorhinolaryngology of Xiangya Hospital, Central South University, Changsha, 410008 China
- Key Laboratory of Otolaryngology Major Disease Research of Hunan Province, Changsha, 410008 China
- National Clinical Research Centre for Geriatric Disorders, Department of Geriatrics, Xiangya Hospital, Central South University, Changsha, 410008 China
| | - Ri Su
- School of Mathematics and Statistics, Central South University, Changsha, 410083 China
| | - Xuewen Wu
- Department of Otorhinolaryngology of Xiangya Hospital, Central South University, Changsha, 410008 China
- Key Laboratory of Otolaryngology Major Disease Research of Hunan Province, Changsha, 410008 China
- National Clinical Research Centre for Geriatric Disorders, Department of Geriatrics, Xiangya Hospital, Central South University, Changsha, 410008 China
| | - Zheng Wang
- School of Computer Science, Hunan First Normal University, Changsha, 410205 China
| | - Muzhou Hou
- School of Mathematics and Statistics, Central South University, Changsha, 410083 China
| |
Collapse
|
10
|
Ngombu S, Binol H, Gurcan MN, Moberly AC. Advances in Artificial Intelligence to Diagnose Otitis Media: State of the Art Review. Otolaryngol Head Neck Surg 2023; 168:635-642. [PMID: 35290142 DOI: 10.1177/01945998221083502] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2021] [Accepted: 02/09/2022] [Indexed: 11/15/2022]
Abstract
OBJECTIVE Otitis media (OM) is a model disease for developing, validating, and implementing artificial intelligence (AI) techniques. We aim to review the state of the art applications of AI used to diagnose OM in pediatric and adult populations. DATA SOURCES Several comprehensive databases were searched to identify all articles that applied AI technologies to diagnose OM. REVIEW METHODS Relevant articles from January 2010 through May 2021 were identified by title and abstract. Articles were excluded if they did not discuss AI in conjunction with diagnosing OM. References of included studies and relevant review articles were cross-referenced to identify any additional studies. CONCLUSION Title and abstract screening resulted in full-text retrieval of 40 articles that met initial screening parameters. Of this total, secondary review articles (n = 7) and commentary-based articles (n = 2) were removed, as were articles that did not specifically discuss AI and OM diagnosis (n = 5), leaving 25 articles for review. Applications of AI technologies specific to diagnosing OM included machine learning and natural language processing (n = 23) and prototype approaches (n = 2). IMPLICATIONS FOR PRACTICE This review emphasizes the utility of AI techniques to automate and aid in diagnosing OM. Although these techniques are still in the development and testing stages, AI has the potential to improve the practice of otolaryngologists and primary care clinicians by increasing the efficiency and accuracy of diagnoses.
Collapse
Affiliation(s)
- Stephany Ngombu
- Department of Otolaryngology-Head and Neck Surgery, Wexner Medical Center at The Ohio State University, Columbus, Ohio, USA
| | - Hamidullah Binol
- Center for Biomedical Informatics, Wake Forest School of Medicine, Winston-Salem, North Carolina, USA
| | - Metin N Gurcan
- Center for Biomedical Informatics, Wake Forest School of Medicine, Winston-Salem, North Carolina, USA
| | - Aaron C Moberly
- Department of Otolaryngology-Head and Neck Surgery, Wexner Medical Center at The Ohio State University, Columbus, Ohio, USA
| |
Collapse
|
11
|
Cao Z, Chen F, Grais EM, Yue F, Cai Y, Swanepoel DW, Zhao F. Machine Learning in Diagnosing Middle Ear Disorders Using Tympanic Membrane Images: A Meta-Analysis. Laryngoscope 2023; 133:732-741. [PMID: 35848851 DOI: 10.1002/lary.30291] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2022] [Revised: 06/18/2022] [Accepted: 06/21/2022] [Indexed: 11/12/2022]
Abstract
OBJECTIVE To systematically evaluate the development of Machine Learning (ML) models and compare their diagnostic accuracy for the classification of Middle Ear Disorders (MED) using Tympanic Membrane (TM) images. METHODS PubMed, EMBASE, CINAHL, and CENTRAL were searched up until November 30, 2021. Studies on the development of ML approaches for diagnosing MED using TM images were selected according to the inclusion criteria. PRISMA guidelines were followed with study design, analysis method, and outcomes extracted. Sensitivity, specificity, and area under the curve (AUC) were used to summarize the performance metrics of the meta-analysis. Risk of Bias was assessed using the Quality Assessment of Diagnostic Accuracy Studies-2 tool in combination with the Prediction Model Risk of Bias Assessment Tool. RESULTS Sixteen studies were included, encompassing 20254 TM images (7025 normal TM and 13229 MED). The sample size ranged from 45 to 6066 per study. The accuracy of the 25 included ML approaches ranged from 76.00% to 98.26%. Eleven studies (68.8%) were rated as having a low risk of bias, with the reference standard as the major domain of high risk of bias (37.5%). Sensitivity and specificity were 93% (95% CI, 90%-95%) and 85% (95% CI, 82%-88%), respectively. The AUC of total TM images was 94% (95% CI, 91%-96%). The greater AUC was found using otoendoscopic images than otoscopic images. CONCLUSIONS ML approaches perform robustly in distinguishing between normal ears and MED, however, it is proposed that a standardized TM image acquisition and annotation protocol should be developed. LEVEL OF EVIDENCE NA Laryngoscope, 133:732-741, 2023.
Collapse
Affiliation(s)
- Zuwei Cao
- Center for Rehabilitative Auditory Research, Guizhou Provincial People's Hospital, Guiyang City, China
| | - Feifan Chen
- Centre for Speech and Language Therapy and Hearing Science, Cardiff School of Sport and Health Sciences, Cardiff Metropolitan University, Cardiff, UK
| | - Emad M Grais
- Centre for Speech and Language Therapy and Hearing Science, Cardiff School of Sport and Health Sciences, Cardiff Metropolitan University, Cardiff, UK
| | - Fengjuan Yue
- Medical Examination Center, Guizhou Provincial People's Hospital, Guiyang City, China
| | - Yuexin Cai
- Department of Otolaryngology, Sun Yat-sen Memorial Hospital, Sun Yat-sen University, Guangzhou City, China
| | - De Wet Swanepoel
- Department of Speech-Language Pathology and Audiology, University of Pretoria, Pretoria, South Africa
| | - Fei Zhao
- Centre for Speech and Language Therapy and Hearing Science, Cardiff School of Sport and Health Sciences, Cardiff Metropolitan University, Cardiff, UK
| |
Collapse
|
12
|
Habib AR, Xu Y, Bock K, Mohanty S, Sederholm T, Weeks WB, Dodhia R, Ferres JL, Perry C, Sacks R, Singh N. Evaluating the generalizability of deep learning image classification algorithms to detect middle ear disease using otoscopy. Sci Rep 2023; 13:5368. [PMID: 37005441 PMCID: PMC10067817 DOI: 10.1038/s41598-023-31921-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2022] [Accepted: 03/20/2023] [Indexed: 04/04/2023] Open
Abstract
To evaluate the generalizability of artificial intelligence (AI) algorithms that use deep learning methods to identify middle ear disease from otoscopic images, between internal to external performance. 1842 otoscopic images were collected from three independent sources: (a) Van, Turkey, (b) Santiago, Chile, and (c) Ohio, USA. Diagnostic categories consisted of (i) normal or (ii) abnormal. Deep learning methods were used to develop models to evaluate internal and external performance, using area under the curve (AUC) estimates. A pooled assessment was performed by combining all cohorts together with fivefold cross validation. AI-otoscopy algorithms achieved high internal performance (mean AUC: 0.95, 95%CI: 0.80-1.00). However, performance was reduced when tested on external otoscopic images not used for training (mean AUC: 0.76, 95%CI: 0.61-0.91). Overall, external performance was significantly lower than internal performance (mean difference in AUC: -0.19, p ≤ 0.04). Combining cohorts achieved a substantial pooled performance (AUC: 0.96, standard error: 0.01). Internally applied algorithms for otoscopy performed well to identify middle ear disease from otoscopy images. However, external performance was reduced when applied to new test cohorts. Further efforts are required to explore data augmentation and pre-processing techniques that might improve external performance and develop a robust, generalizable algorithm for real-world clinical applications.
Collapse
Affiliation(s)
- Al-Rahim Habib
- Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia.
- Department of Otolaryngology, Head and Neck Surgery, Westmead Hospital, Sydney, NSW, Australia.
| | - Yixi Xu
- AI for Good Lab, Microsoft, Redmond, WA, USA
| | - Kris Bock
- Azure FastTrack Engineering, Brisbane, QLD, Australia
| | | | | | | | | | | | - Chris Perry
- University of Queensland Medical School, Brisbane, QLD, Australia
| | - Raymond Sacks
- Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia
| | - Narinder Singh
- Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia
- Department of Otolaryngology, Head and Neck Surgery, Westmead Hospital, Sydney, NSW, Australia
| |
Collapse
|
13
|
Tseng CC, Lim V, Jyung RW. Use of artificial intelligence for the diagnosis of cholesteatoma. Laryngoscope Investig Otolaryngol 2023; 8:201-211. [PMID: 36846416 PMCID: PMC9948563 DOI: 10.1002/lio2.1008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Revised: 12/07/2022] [Accepted: 12/30/2022] [Indexed: 01/19/2023] Open
Abstract
Objectives Accurate diagnosis of cholesteatomas is crucial. However, cholesteatomas can easily be missed in routine otoscopic exams. Convolutional neural networks (CNNs) have performed well in medical image classification, so we evaluated their use for detecting cholesteatomas in otoscopic images. Study Design Design and evaluation of artificial intelligence driven workflow for cholesteatoma diagnosis. Methods Otoscopic images collected from the faculty practice of the senior author were deidentified and labeled by the senior author as cholesteatoma, abnormal non-cholesteatoma, or normal. An image classification workflow was developed to automatically differentiate cholesteatomas from other possible tympanic membrane appearances. Eight pretrained CNNs were trained on our otoscopic images, then tested on a withheld subset of images to evaluate their final performance. CNN intermediate activations were also extracted to visualize important image features. Results A total of 834 otoscopic images were collected, further categorized into 197 cholesteatoma, 457 abnormal non-cholesteatoma, and 180 normal. Final trained CNNs demonstrated strong performance, achieving accuracies of 83.8%-98.5% for differentiating cholesteatoma from normal, 75.6%-90.1% for differentiating cholesteatoma from abnormal non-cholesteatoma, and 87.0%-90.4% for differentiating cholesteatoma from non-cholesteatoma (abnormal non-cholesteatoma + normal). DenseNet201 (100% sensitivity, 97.1% specificity), NASNetLarge (100% sensitivity, 88.2% specificity), and MobileNetV2 (94.1% sensitivity, 100% specificity) were among the best performing CNNs in distinguishing cholesteatoma versus normal. Visualization of intermediate activations showed robust detection of relevant image features by the CNNs. Conclusion While further refinement and more training images are needed to improve performance, artificial intelligence-driven analysis of otoscopic images shows great promise as a diagnostic tool for detecting cholesteatomas. Level of Evidence 3.
Collapse
Affiliation(s)
- Christopher C. Tseng
- Department of Otolaryngology – Head and Neck SurgeryRutgers New Jersey Medical SchoolNewarkNew JerseyUSA
| | - Valerie Lim
- Department of Otolaryngology – Head and Neck SurgeryRutgers New Jersey Medical SchoolNewarkNew JerseyUSA
| | - Robert W. Jyung
- Department of Otolaryngology – Head and Neck SurgeryRutgers New Jersey Medical SchoolNewarkNew JerseyUSA
| |
Collapse
|
14
|
Habib AR, Crossland G, Patel H, Wong E, Kong K, Gunasekera H, Richards B, Caffery L, Perry C, Sacks R, Kumar A, Singh N. An Artificial Intelligence Computer-vision Algorithm to Triage Otoscopic Images From Australian Aboriginal and Torres Strait Islander Children. Otol Neurotol 2022; 43:481-488. [PMID: 35239622 DOI: 10.1097/mao.0000000000003484] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVE To develop an artificial intelligence image classification algorithm to triage otoscopic images from rural and remote Australian Aboriginal and Torres Strait Islander children. STUDY DESIGN Retrospective observational study. SETTING Tertiary referral center. PATIENTS Rural and remote Aboriginal and Torres Strait Islander children who underwent tele-otology ear health screening in the Northern Territory, Australia between 2010 and 2018. INTERVENTIONS Otoscopic images were labeled by otolaryngologists to classify the ground truth. Deep and transfer learning methods were used to develop an image classification algorithm. MAIN OUTCOME MEASURES Accuracy, sensitivity, specificity, positive predictive value, negative predictive value, area under the curve (AUC) of the resultant algorithm compared with the ground truth. RESULTS Six thousand five hundred twenty seven images were used (5927 images for training and 600 for testing). The algorithm achieved an accuracy of 99.3% for acute otitis media, 96.3% for chronic otitis media, 77.8% for otitis media with effusion (OME), and 98.2% to classify wax/obstructed canal. To differentiate between multiple diagnoses, the algorithm achieved 74.4 to 92.8% accuracy and an AUC of 0.963 to 0.997. The most common incorrect classification pattern was OME misclassified as normal tympanic membranes. CONCLUSIONS The paucity of access to tertiary otolaryngology care for rural and remote Aboriginal and Torres Strait Islander communities may contribute to an under-identification of ear disease. Computer vision image classification algorithms can accurately classify ear disease from otoscopic images of Indigenous Australian children. In the future, a validated algorithm may integrate with existing telemedicine initiatives to support effective triage and facilitate early treatment and referral.
Collapse
Affiliation(s)
- Al-Rahim Habib
- Sydney Medical School, Faculty of Medicine and Health, University of Sydney, Camperdown, New South Wales, Australia
- Department of Otolaryngology-Head and Neck Surgery, Princess Alexandra Hospital, Brisbane, Queensland, Australia
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Sydney, New South Wales, Australia
| | - Graeme Crossland
- Department of Otolaryngology - Head and Neck Surgery, Royal Darwin Hospital, Darwin, Northern Territory, Australia
| | - Hemi Patel
- Department of Otolaryngology - Head and Neck Surgery, Royal Darwin Hospital, Darwin, Northern Territory, Australia
| | - Eugene Wong
- Sydney Medical School, Faculty of Medicine and Health, University of Sydney, Camperdown, New South Wales, Australia
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Sydney, New South Wales, Australia
| | - Kelvin Kong
- School of Medicine and Public Health, University of Newcastle, Newcastle, New South Wales, Australia
- Department of Linguistics, Faculty of Medicine, Macquarie University, Sydney, New South Wales, Australia
- School of Population Health, Faculty of Medicine, University of New South Wales, Sydney, Australia
| | - Hasantha Gunasekera
- Sydney Medical School, Faculty of Medicine and Health, University of Sydney, Camperdown, New South Wales, Australia
- The Children's Hospital at Westmead, Sydney, New South Wales, Australia
| | - Brent Richards
- Division of Medical Services, Gold Coast University Hospital, Gold Coast, Queensland, Australia
- Griffith Health, Griffith University Queensland, Australia
| | - Liam Caffery
- Centre for Online Health, University of Queensland, Australia
| | - Chris Perry
- Centre for Online Health, University of Queensland, Australia
| | - Raymond Sacks
- Sydney Medical School, Faculty of Medicine and Health, University of Sydney, Camperdown, New South Wales, Australia
| | - Ashnil Kumar
- School of Biomedical Engineering, Faculty of Engineering, University of Sydney, Camperdown, New South Wales, Australia
| | - Narinder Singh
- Sydney Medical School, Faculty of Medicine and Health, University of Sydney, Camperdown, New South Wales, Australia
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, Sydney, New South Wales, Australia
| |
Collapse
|
15
|
Chawdhary G, Shoman N. Emerging artificial intelligence applications in otological imaging. Curr Opin Otolaryngol Head Neck Surg 2021; 29:357-364. [PMID: 34459798 DOI: 10.1097/moo.0000000000000754] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
PURPOSE OF REVIEW To highlight the recent literature on artificial intelligence (AI) pertaining to otological imaging and to discuss future directions, obstacles and opportunities. RECENT FINDINGS The main themes in the recent literature centre around automated otoscopic image diagnosis and automated image segmentation for application in virtual reality surgical simulation and planning. Other applications that have been studied include identification of tinnitus MRI biomarkers, facial palsy analysis, intraoperative augmented reality systems, vertigo diagnosis and endolymphatic hydrops ratio calculation in Meniere's disease. Studies are presently at a preclinical, proof-of-concept stage. SUMMARY The recent literature on AI in otological imaging is promising and demonstrates the future potential of this technology in automating certain imaging tasks in a healthcare environment of ever-increasing demand and workload. Some studies have shown equivalence or superiority of the algorithm over physicians, albeit in narrowly defined realms. Future challenges in developing this technology include the compilation of large high quality annotated datasets, fostering strong collaborations between the health and technology sectors, testing the technology within real-world clinical pathways and bolstering trust among patients and physicians in this new method of delivering healthcare.
Collapse
Affiliation(s)
- Gaurav Chawdhary
- ENT Department, Royal Hallamshire Hospital, Broomhall, Sheffield, UK
| | - Nael Shoman
- ENT Department, Queen Elizabeth II Health Sciences Centre, Halifax, Nova Scotia, Canada
| |
Collapse
|
16
|
Canares TL, Wang W, Unberath M, Clark JH. Artificial intelligence to diagnose ear disease using otoscopic image analysis: a review. J Investig Med 2021; 70:354-362. [PMID: 34521730 DOI: 10.1136/jim-2021-001870] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/27/2021] [Indexed: 12/22/2022]
Abstract
AI relates broadly to the science of developing computer systems to imitate human intelligence, thus allowing for the automation of tasks that would otherwise necessitate human cognition. Such technology has increasingly demonstrated capacity to outperform humans for functions relating to image recognition. Given the current lack of cost-effective confirmatory testing, accurate diagnosis and subsequent management depend on visual detection of characteristic findings during otoscope examination. The aim of this manuscript is to perform a comprehensive literature review and evaluate the potential application of artificial intelligence for the diagnosis of ear disease from otoscopic image analysis.
Collapse
Affiliation(s)
- Therese L Canares
- Johns Hopkins University School of Medicine, Baltimore, Maryland, USA
| | - Weiyao Wang
- Johns Hopkins University Whiting School of Engineering, Baltimore, Maryland, USA
| | - Mathias Unberath
- Johns Hopkins University Whiting School of Engineering, Baltimore, Maryland, USA
| | - James H Clark
- Otolaryngology-HNS, Johns Hopkins Medicine School of Medicine, Baltimore, Maryland, USA
| |
Collapse
|
17
|
Byun H, Yu S, Oh J, Bae J, Yoon MS, Lee SH, Chung JH, Kim TH. An Assistive Role of a Machine Learning Network in Diagnosis of Middle Ear Diseases. J Clin Med 2021; 10:jcm10153198. [PMID: 34361982 PMCID: PMC8347824 DOI: 10.3390/jcm10153198] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2021] [Revised: 07/16/2021] [Accepted: 07/19/2021] [Indexed: 02/07/2023] Open
Abstract
The present study aimed to develop a machine learning network to diagnose middle ear diseases with tympanic membrane images and to identify its assistive role in the diagnostic process. The medical records of subjects who underwent ear endoscopy tests were reviewed. From these records, 2272 diagnostic tympanic membranes images were appropriately labeled as normal, otitis media with effusion (OME), chronic otitis media (COM), or cholesteatoma and were used for training. We developed the “ResNet18 + Shuffle” network and validated the model performance. Seventy-one representative cases were selected to test the final accuracy of the network and resident physicians. We asked 10 resident physicians to make diagnoses from tympanic membrane images with and without the help of the machine learning network, and the change of the diagnostic performance of resident physicians with the aid of the answers from the machine learning network was assessed. The devised network showed a highest accuracy of 97.18%. A five-fold validation showed that the network successfully diagnosed ear diseases with an accuracy greater than 93%. All resident physicians were able to diagnose middle ear diseases more accurately with the help of the machine learning network. The increase in diagnostic accuracy was up to 18% (1.4% to 18.4%). The machine learning network successfully classified middle ear diseases and was assistive to clinicians in the interpretation of tympanic membrane images.
Collapse
Affiliation(s)
- Hayoung Byun
- Department of Otolaryngology & Head and Neck Surgery, College of Medicine, Hanyang University, Seoul 04763, Korea; (H.B.); (S.H.L.)
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
| | - Sangjoon Yu
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
- Department of Computer Science, Hanyang University, Seoul 04763, Korea
| | - Jaehoon Oh
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
- Department of Emergency Medicine, College of Medicine, Hanyang University, Seoul 04763, Korea
| | - Junwon Bae
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
- Department of Emergency Medicine, College of Medicine, Hanyang University, Seoul 04763, Korea
| | - Myeong Seong Yoon
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
- Department of Emergency Medicine, College of Medicine, Hanyang University, Seoul 04763, Korea
| | - Seung Hwan Lee
- Department of Otolaryngology & Head and Neck Surgery, College of Medicine, Hanyang University, Seoul 04763, Korea; (H.B.); (S.H.L.)
| | - Jae Ho Chung
- Department of Otolaryngology & Head and Neck Surgery, College of Medicine, Hanyang University, Seoul 04763, Korea; (H.B.); (S.H.L.)
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
- Department of HY-KIST Bio-Convergence, College of Medicine, Hanyang University, Seoul 04763, Korea
- Correspondence: (J.H.C.); (T.H.K.)
| | - Tae Hyun Kim
- Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea; (S.Y.); (J.O.); (J.B.); (M.S.Y.)
- Department of Computer Science, Hanyang University, Seoul 04763, Korea
- Correspondence: (J.H.C.); (T.H.K.)
| |
Collapse
|
18
|
García-Domínguez A, Galván-Tejada CE, Brena RF, Aguileta AA, Galván-Tejada JI, Gamboa-Rosales H, Celaya-Padilla JM, Luna-García H. Children's Activity Classification for Domestic Risk Scenarios Using Environmental Sound and a Bayesian Network. Healthcare (Basel) 2021; 9:healthcare9070884. [PMID: 34356262 PMCID: PMC8307924 DOI: 10.3390/healthcare9070884] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2021] [Revised: 06/26/2021] [Accepted: 07/06/2021] [Indexed: 11/16/2022] Open
Abstract
Children’s healthcare is a relevant issue, especially the prevention of domestic accidents, since it has even been defined as a global health problem. Children’s activity classification generally uses sensors embedded in children’s clothing, which can lead to erroneous measurements for possible damage or mishandling. Having a non-invasive data source for a children’s activity classification model provides reliability to the monitoring system where it is applied. This work proposes the use of environmental sound as a data source for the generation of children’s activity classification models, implementing feature selection methods and classification techniques based on Bayesian networks, focused on the recognition of potentially triggering activities of domestic accidents, applicable in child monitoring systems. Two feature selection techniques were used: the Akaike criterion and genetic algorithms. Likewise, models were generated using three classifiers: naive Bayes, semi-naive Bayes and tree-augmented naive Bayes. The generated models, combining the methods of feature selection and the classifiers used, present accuracy of greater than 97% for most of them, with which we can conclude the efficiency of the proposal of the present work in the recognition of potentially detonating activities of domestic accidents.
Collapse
Affiliation(s)
- Antonio García-Domínguez
- Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147, Centro 98000, Zacatecas, Mexico; (A.G.-D.); (J.I.G.-T.); (H.G.-R.); (J.M.C.-P.); (H.L.-G.)
| | - Carlos E. Galván-Tejada
- Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147, Centro 98000, Zacatecas, Mexico; (A.G.-D.); (J.I.G.-T.); (H.G.-R.); (J.M.C.-P.); (H.L.-G.)
- Correspondence:
| | - Ramón F. Brena
- Tecnológico de Monterrey, School of Engineering and Sciences, Av. Eugenio Garza Sada 2501 Sur, Monterrey 64849, Nuevo León, Mexico;
| | - Antonio A. Aguileta
- Facultad de Matemáticas, Universidad Autónoma de Yucatán, Anillo Periférico Norte, Tablaje Cat. 13615, Colonia Chuburná Hidalgo Inn, Mérida 97110, Yucatan, Mexico;
| | - Jorge I. Galván-Tejada
- Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147, Centro 98000, Zacatecas, Mexico; (A.G.-D.); (J.I.G.-T.); (H.G.-R.); (J.M.C.-P.); (H.L.-G.)
| | - Hamurabi Gamboa-Rosales
- Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147, Centro 98000, Zacatecas, Mexico; (A.G.-D.); (J.I.G.-T.); (H.G.-R.); (J.M.C.-P.); (H.L.-G.)
| | - José M. Celaya-Padilla
- Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147, Centro 98000, Zacatecas, Mexico; (A.G.-D.); (J.I.G.-T.); (H.G.-R.); (J.M.C.-P.); (H.L.-G.)
| | - Huizilopoztli Luna-García
- Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147, Centro 98000, Zacatecas, Mexico; (A.G.-D.); (J.I.G.-T.); (H.G.-R.); (J.M.C.-P.); (H.L.-G.)
| |
Collapse
|
19
|
Kashani RG, Młyńczak MC, Zarabanda D, Solis-Pazmino P, Huland DM, Ahmad IN, Singh SP, Valdez TA. Shortwave infrared otoscopy for diagnosis of middle ear effusions: a machine-learning-based approach. Sci Rep 2021; 11:12509. [PMID: 34131163 PMCID: PMC8206083 DOI: 10.1038/s41598-021-91736-9] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2021] [Accepted: 05/04/2021] [Indexed: 02/05/2023] Open
Abstract
Otitis media, a common disease marked by the presence of fluid within the middle ear space, imparts a significant global health and economic burden. Identifying an effusion through the tympanic membrane is critical to diagnostic success but remains challenging due to the inherent limitations of visible light otoscopy and user interpretation. Here we describe a powerful diagnostic approach to otitis media utilizing advancements in otoscopy and machine learning. We developed an otoscope that visualizes middle ear structures and fluid in the shortwave infrared region, holding several advantages over traditional approaches. Images were captured in vivo and then processed by a novel machine learning based algorithm. The model predicts the presence of effusions with greater accuracy than current techniques, offering specificity and sensitivity over 90%. This platform has the potential to reduce costs and resources associated with otitis media, especially as improvements are made in shortwave imaging and machine learning.
Collapse
Affiliation(s)
- Rustin G. Kashani
- grid.168010.e0000000419368956Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, 801 Welch Road, Palo Alto, CA 94304 USA
| | - Marcel C. Młyńczak
- grid.1035.70000000099214842Institute of Metrology and Biomedical Engineering, Faculty of Mechatronics, Warsaw University of Technology, Warsaw, Poland
| | - David Zarabanda
- grid.168010.e0000000419368956Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, 801 Welch Road, Palo Alto, CA 94304 USA
| | - Paola Solis-Pazmino
- grid.168010.e0000000419368956Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, 801 Welch Road, Palo Alto, CA 94304 USA
| | - David M. Huland
- grid.168010.e0000000419368956Department of Radiology, Stanford University School of Medicine, Palo Alto, CA USA
| | - Iram N. Ahmad
- grid.168010.e0000000419368956Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, 801 Welch Road, Palo Alto, CA 94304 USA ,grid.414123.10000 0004 0450 875XLucile Packard Children’s Hospital, Palo Alto, CA USA
| | - Surya P. Singh
- grid.495560.b0000 0004 6003 8393Department of Biosciences and Bioengineering, Indian Institute of Technology Dharwad, Dharwad, Karnataka India
| | - Tulio A. Valdez
- grid.168010.e0000000419368956Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, 801 Welch Road, Palo Alto, CA 94304 USA ,grid.414123.10000 0004 0450 875XLucile Packard Children’s Hospital, Palo Alto, CA USA
| |
Collapse
|