1
|
Edalati S, Vasan V, Cheng CP, Patel Z, Govindaraj S, Iloreta AM. Can GPT-4 revolutionize otolaryngology? Navigating opportunities and ethical considerations. Am J Otolaryngol 2024; 45:104303. [PMID: 38678799 DOI: 10.1016/j.amjoto.2024.104303] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2024] [Accepted: 04/14/2024] [Indexed: 05/01/2024]
Abstract
Otolaryngologists can enhance workflow efficiency, provide better patient care, and advance medical research and education by integrating artificial intelligence (AI) into their practices. GPT-4 technology is a revolutionary and contemporary example of AI that may apply to otolaryngology. The knowledge of otolaryngologists should be supplemented, not replaced when using GPT-4 to make critical medical decisions and provide individualized patient care. In our thorough examination, we explore the potential uses of the groundbreaking GPT-4 technology in the field of otolaryngology, covering aspects such as potential outcomes and technical boundaries. Additionally, we delve into the intricate and intellectually challenging dilemmas that emerge when incorporating GPT-4 into otolaryngology, considering the ethical considerations inherent in its implementation. Our stance is that GPT-4 has the potential to be very helpful. Its capabilities, which include aid in clinical decision-making, patient care, and administrative job automation, present exciting possibilities for enhancing patient outcomes, boosting the efficiency of healthcare delivery, and enhancing patient experiences. Even though there are still certain obstacles and limitations, the progress made so far shows that GPT-4 can be a valuable tool for modern medicine. GPT-4 may play a more significant role in clinical practice as technology develops, helping medical professionals deliver high-quality care tailored to every patient's unique needs.
Collapse
Affiliation(s)
- Shaun Edalati
- Department of Otolaryngology-Head and Neck Surgery, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
| | - Vikram Vasan
- Department of Otolaryngology-Head and Neck Surgery, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Christopher P Cheng
- Department of Otolaryngology-Head and Neck Surgery, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Zara Patel
- Department of Otolaryngology-Head & Neck Surgery, Stanford University School of Medicine, Stanford, CA, USA
| | - Satish Govindaraj
- Department of Otolaryngology-Head and Neck Surgery, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Alfred Marc Iloreta
- Department of Otolaryngology-Head and Neck Surgery, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| |
Collapse
|
2
|
Cacace AT, Berri B. Blast Overpressures as a Military and Occupational Health Concern. Am J Audiol 2023; 32:779-792. [PMID: 37713532 DOI: 10.1044/2023_aja-23-00125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/17/2023] Open
Abstract
PURPOSE This tutorial reviews effects of environmental stressors like blast overpressures and other well-known acoustic contaminants (continuous, intermittent, and impulsive noise) on hearing, tinnitus, vestibular, and balance-related functions. Based on the overall outcome of these effects, detailed consideration is given to the health and well-being of individuals. METHOD Because hearing loss and tinnitus are consequential in affecting quality of life, novel neuromodulation paradigms are reviewed for their positive abatement and treatment-related effects. Examples of clinical data, research strategies, and methodological approaches focus on repetitive transcranial magnetic stimulation (rTMS) and electrical stimulation of the vagus nerve paired with tones (VNSt) for their unique contributions to this area. RESULTS Acoustic toxicants transmitted through the atmosphere are noteworthy for their propensity to induce hearing loss and tinnitus. Mounting evidence also indicates that high-level rapid onset changes in atmospheric sound pressure can significantly impact vestibular and balance function. Indeed, the risk of falling secondary to loss of, or damage to, sensory receptor cells in otolith organs (utricle and saccule) is a primary reason for this concern. As part of the complexities involved in VNSt treatment strategies, vocal dysfunction may also manifest. In addition, evaluation of temporospatial gait parameters is worthy of consideration based on their ability to detect and monitor incipient neurological disease, cognitive decline, and mortality. CONCLUSION Highlighting these respective areas underscores the need to enhance information exchange among scientists, clinicians, and caregivers on the benefits and complications of these outcomes.
Collapse
Affiliation(s)
- Anthony T Cacace
- Department of Communication Sciences & Disorders, Wayne State University, Detroit, MI
| | - Batoul Berri
- Department of Communication Sciences & Disorders, Wayne State University, Detroit, MI
- Department of Otolaryngology, University of Michigan, Ann Arbor
| |
Collapse
|
3
|
Liu GS, Hodges JM, Yu J, Sung CK, Erickson‐DiRenzo E, Doyle PC. End-to-end deep learning classification of vocal pathology using stacked vowels. Laryngoscope Investig Otolaryngol 2023; 8:1312-1318. [PMID: 37899847 PMCID: PMC10601590 DOI: 10.1002/lio2.1144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2023] [Accepted: 08/13/2023] [Indexed: 10/31/2023] Open
Abstract
Objectives Advances in artificial intelligence (AI) technology have increased the feasibility of classifying voice disorders using voice recordings as a screening tool. This work develops upon previous models that take in single vowel recordings by analyzing multiple vowel recordings simultaneously to enhance prediction of vocal pathology. Methods Voice samples from the Saarbruecken Voice Database, including three sustained vowels (/a/, /i/, /u/) from 687 healthy human participants and 334 dysphonic patients, were used to train 1-dimensional convolutional neural network models for multiclass classification of healthy, hyperfunctional dysphonia, and laryngitis voice recordings. Three models were trained: (1) a baseline model that analyzed individual vowels in isolation, (2) a stacked vowel model that analyzed three vowels (/a/, /i/, /u/) in the neutral pitch simultaneously, and (3) a stacked pitch model that analyzed the /a/ vowel in three pitches (low, neutral, and high) simultaneously. Results For multiclass classification of healthy, hyperfunctional dysphonia, and laryngitis voice recordings, the stacked vowel model demonstrated higher performance compared with the baseline and stacked pitch models (F1 score 0.81 vs. 0.77 and 0.78, respectively). Specifically, the stacked vowel model achieved higher performance for class-specific classification of hyperfunctional dysphonia voice samples compared with the baseline and stacked pitch models (F1 score 0.56 vs. 0.49 and 0.50, respectively). Conclusions This study demonstrates the feasibility and potential of analyzing multiple sustained vowel recordings simultaneously to improve AI-driven screening and classification of vocal pathology. The stacked vowel model architecture in particular offers promise to enhance such an approach. Lay Summary AI analysis of multiple vowel recordings can improve classification of voice pathologies compared with models using a single sustained vowel and offer a strategy to enhance AI-driven screening of voice disorders. Level of Evidence 3.
Collapse
Affiliation(s)
- George S. Liu
- Department of Otolaryngology Head and Neck SurgeryStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
- Division of LaryngologyStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
| | - Jordan M. Hodges
- Computer Science DepartmentSchool of Engineering, Stanford UniversityStanfordCaliforniaUSA
| | - Jingzhi Yu
- Biomedical Informatics, Department of Biomedical Data ScienceStanford University School of MedicineStanfordCaliforniaUSA
| | - C. Kwang Sung
- Department of Otolaryngology Head and Neck SurgeryStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
- Division of LaryngologyStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
| | - Elizabeth Erickson‐DiRenzo
- Department of Otolaryngology Head and Neck SurgeryStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
- Division of LaryngologyStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
| | - Philip C. Doyle
- Department of Otolaryngology Head and Neck SurgeryStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
- Division of LaryngologyStanford University School of Medicine, Stanford UniversityStanfordCaliforniaUSA
| |
Collapse
|
4
|
Liu GS, Yang A, Kim D, Hojel A, Voevodsky D, Wang J, Tong CCL, Ungerer H, Palmer JN, Kohanski MA, Nayak JV, Hwang PH, Adappa ND, Patel ZM. Deep learning classification of inverted papilloma malignant transformation using 3D convolutional neural networks and magnetic resonance imaging. Int Forum Allergy Rhinol 2022; 12:1025-1033. [PMID: 34989484 DOI: 10.1002/alr.22958] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2021] [Revised: 11/30/2021] [Accepted: 12/24/2021] [Indexed: 01/23/2023]
Abstract
BACKGROUND Distinguishing benign inverted papilloma (IP) tumors from those that have undergone malignant transformation to squamous cell carcinoma (IP-SCC) is important but challenging to do preoperatively. Magnetic resonance imaging (MRI) can help differentiate these 2 entities, but no established method exists that can automatically synthesize all potentially relevant MRI image features to distinguish IP and IP-SCC. We explored a deep learning approach, using 3-dimensional convolutional neural networks (CNNs), to address this challenge. METHODS Retrospective chart reviews were performed at 2 institutions to create a data set of preoperative MRIs with corresponding surgical pathology reports. The MRI data set included all available MRI sequences in the axial plane, which were used to train, validate, and test 3 CNN models. Saliency maps were generated to visualize areas of MRIs with greatest influence on predictions. RESULTS A total of 90 patients with IP (n = 64) or IP-SCC (n = 26) tumors were identified, with a total of 446 images of distinct MRI sequences for IP (n = 329) or IP-SCC (n = 117). The best CNN model, All-Net, demonstrated a sensitivity of 66.7%, specificity of 81.5%, overall accuracy of 77.9%, and receiver-operating characteristic area under the curve of 0.80 (95% confidence interval, 0.682-0.898) for test classification performance. The other 2 models, Small-All-Net and Elastic-All-Net, showed similar performance levels. CONCLUSION A deep learning approach with 3-dimensional CNNs can distinguish IP and IP-SCC with moderate test classification performance. Although CNNs demonstrate promise to enhance the prediction of IP-SCC using MRIs, more data are needed before they can reach the predictive value already established by human MRI evaluation.
Collapse
Affiliation(s)
- George S Liu
- Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, Stanford, CA
| | - Angela Yang
- Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, Stanford, CA
| | - Dayoung Kim
- Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, Stanford, CA
| | - Andrew Hojel
- Department of Computer Science, Stanford University, Stanford, CA
| | - Diana Voevodsky
- Department of Mathematics, Stanford University, Stanford, CA
| | - Julia Wang
- Department of Electrical Engineering, Stanford University, Stanford, CA
| | - Charles C L Tong
- Department of Otolaryngology-Head and Neck Surgery, University of Pennsylvania School of Medicine, Philadelphia, PA
| | - Heather Ungerer
- Department of Otolaryngology-Head and Neck Surgery, University of Pennsylvania School of Medicine, Philadelphia, PA
| | - James N Palmer
- Department of Otolaryngology-Head and Neck Surgery, University of Pennsylvania School of Medicine, Philadelphia, PA
| | - Michael A Kohanski
- Department of Otolaryngology-Head and Neck Surgery, University of Pennsylvania School of Medicine, Philadelphia, PA
| | - Jayakar V Nayak
- Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, Stanford, CA
| | - Peter H Hwang
- Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, Stanford, CA
| | - Nithin D Adappa
- Department of Otolaryngology-Head and Neck Surgery, University of Pennsylvania School of Medicine, Philadelphia, PA
| | - Zara M Patel
- Department of Otolaryngology-Head and Neck Surgery, Stanford University School of Medicine, Stanford, CA
| |
Collapse
|
5
|
George MM, Tolley NS. AIM in Otolaryngology and Head and Neck Surgery. Artif Intell Med 2022. [DOI: 10.1007/978-3-030-64573-1_198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/09/2022]
|
6
|
Badash I, Applegate BE, Oghalai JS. In Vivo Cochlear imaging provides a tool to study endolymphatic hydrops. J Vestib Res 2021; 31:269-276. [PMID: 33136083 DOI: 10.3233/ves-200718] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Exposure to noise trauma, such as that from improvised explosive devices, can lead to sensorineural hearing loss and a reduced quality of life. In order to elucidate the mechanisms underlying noise-induced hearing loss, we have adapted optical coherence tomography (OCT) for real-time cochlear visualization in live mice after blast exposure. We demonstrated that endolymphatic hydrops develops following blast injury, and that this phenomenon may be associated with glutamate excitotoxicity and cochlear synaptopathy. Additionally, osmotic stabilization of endolymphatic hydrops partially rescues cochlear synapses after blast trauma. OCT is thus a valuable research tool for investigating the mechanisms underlying acoustic trauma and dynamic changes in endolymph volume. It may also help with the diagnosis and treatment of human hearing loss and/or vertigo in the near future.
Collapse
Affiliation(s)
- Ido Badash
- Caruso Department of Otolaryngology-Head and Neck Surgery, University of Southern California, Los Angeles, CA, USA
| | - Brian E Applegate
- Caruso Department of Otolaryngology-Head and Neck Surgery, University of Southern California, Los Angeles, CA, USA
| | - John S Oghalai
- Caruso Department of Otolaryngology-Head and Neck Surgery, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
7
|
George MM, Tolley NS. AIM in Otolaryngology and Head & Neck Surgery. Artif Intell Med 2021. [DOI: 10.1007/978-3-030-58080-3_198-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
8
|
Butola A, Prasad DK, Ahmad A, Dubey V, Qaiser D, Srivastava A, Senthilkumaran P, Ahluwalia BS, Mehta DS. Deep learning architecture "LightOCT" for diagnostic decision support using optical coherence tomography images of biological samples. BIOMEDICAL OPTICS EXPRESS 2020; 11:5017-5031. [PMID: 33014597 PMCID: PMC7510870 DOI: 10.1364/boe.395487] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/16/2020] [Revised: 06/16/2020] [Accepted: 07/06/2020] [Indexed: 05/06/2023]
Abstract
Optical coherence tomography (OCT) is being increasingly adopted as a label-free and non-invasive technique for biomedical applications such as cancer and ocular disease diagnosis. Diagnostic information for these tissues is manifest in textural and geometric features of the OCT images, which are used by human expertise to interpret and triage. However, it suffers delays due to the long process of the conventional diagnostic procedure and shortage of human expertise. Here, a custom deep learning architecture, LightOCT, is proposed for the classification of OCT images into diagnostically relevant classes. LightOCT is a convolutional neural network with only two convolutional layers and a fully connected layer, but it is shown to provide excellent training and test results for diverse OCT image datasets. We show that LightOCT provides 98.9% accuracy in classifying 44 normal and 44 malignant (invasive ductal carcinoma) breast tissue volumetric OCT images. Also, >96% accuracy in classifying public datasets of ocular OCT images as normal, age-related macular degeneration and diabetic macular edema. Additionally, we show ∼96% test accuracy for classifying retinal images as belonging to choroidal neovascularization, diabetic macular edema, drusen, and normal samples on a large public dataset of more than 100,000 images. The performance of the architecture is compared with transfer learning based deep neural networks. Through this, we show that LightOCT can provide significant diagnostic support for a variety of OCT images with sufficient training and minimal hyper-parameter tuning. The trained LightOCT networks for the three-classification problem will be released online to support transfer learning on other datasets.
Collapse
Affiliation(s)
- Ankit Butola
- Bio-photonics Laboratory, Department of Physics, Indian Institute of Technology Delhi, Hauz-Khas, New Delhi 110016, India
| | - Dilip K. Prasad
- School of Computer Science & Engineering, Nanyang Technological University, Singapore 639798, Singapore
| | - Azeem Ahmad
- Department of Physics and Technology, UiT The Arctic University of Norway, Norway
| | - Vishesh Dubey
- Bio-photonics Laboratory, Department of Physics, Indian Institute of Technology Delhi, Hauz-Khas, New Delhi 110016, India
| | - Darakhshan Qaiser
- Department of Surgical Disciplines, All India Institute of Medical Science, Ansari Nagar, New Delhi 110029, India
| | - Anurag Srivastava
- Department of Surgical Disciplines, All India Institute of Medical Science, Ansari Nagar, New Delhi 110029, India
| | | | | | - Dalip Singh Mehta
- Bio-photonics Laboratory, Department of Physics, Indian Institute of Technology Delhi, Hauz-Khas, New Delhi 110016, India
| |
Collapse
|
9
|
Shenson JA, Liu GS, Farrell J, Blevins NH. Multispectral Imaging for Automated Tissue Identification of Normal Human Surgical Specimens. Otolaryngol Head Neck Surg 2020; 164:328-335. [PMID: 32838646 DOI: 10.1177/0194599820941013] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
OBJECTIVE Safe surgery requires the accurate discrimination of tissue intraoperatively. We assess the feasibility of using multispectral imaging and deep learning to enhance surgical vision by automated identification of normal human head and neck tissues. STUDY DESIGN Construction and feasibility testing of novel multispectral imaging system for surgery. SETTING Academic university hospital. SUBJECTS AND METHODS Multispectral images of fresh-preserved human cadaveric tissues were captured with our adapted digital operating microscope. Eleven tissue types were sampled, each sequentially exposed to 6 lighting conditions. Two convolutional neural network machine learning models were developed to classify tissues based on multispectral and white-light color images (ARRInet-M and ARRInet-W, respectively). Blinded otolaryngology residents were asked to identify tissue specimens from white-light color images, and their performance was compared with that of the ARRInet models. RESULTS A novel multispectral imaging system was developed with minimal adaptation to an existing digital operating microscope. With 81.8% accuracy in tissue identification of full-size images, the multispectral ARRInet-M classifier outperformed the white-light-only ARRInet-W model (45.5%) and surgical residents (69.7%). Challenges with discrimination occurred with parotid vs fat and blood vessels vs nerve. CONCLUSIONS A deep learning model using multispectral imaging outperformed a similar model and surgical residents using traditional white-light imaging at the task of classifying normal human head and neck tissue ex vivo. These results suggest that multispectral imaging can enhance surgical vision and augment surgeons' ability to identify tissues during a procedure.
Collapse
Affiliation(s)
- Jared A Shenson
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
| | - George S Liu
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
| | - Joyce Farrell
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Nikolas H Blevins
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
| |
Collapse
|
10
|
Optical coherence tomography: current and future clinical applications in otology. Curr Opin Otolaryngol Head Neck Surg 2020; 28:296-301. [PMID: 32833887 DOI: 10.1097/moo.0000000000000654] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
PURPOSE OF REVIEW This article reviews literature on the use of optical coherence tomography (OCT) in otology and provides the reader with a timely update on its current clinical and research applications. The discussion focuses on the principles of OCT, the use of the technology for the diagnosis of middle ear disease and for the delineation of in-vivo cochlear microarchitecture and function. RECENT FINDINGS Recent advances in OCT include the measurement of structural and vibratory properties of the tympanic membrane, ossicles and inner ear in healthy and diseased states. Accurate, noninvasive diagnosis of middle ear disease, such as otosclerosis and acute otitis media using OCT, has been validated in clinical studies, whereas inner ear OCT imaging remains at the preclinical stage. The development of recent microscopic, otoscopic and endoscopic systems to address clinical and research problems is reviewed. SUMMARY OCT is a real-time, noninvasive, nonionizing, point-of-care imaging modality capable of imaging ear structures in vivo. Although current clinical systems are mainly focused on middle ear imaging, OCT has also been shown to have the ability to identify inner ear disease, an exciting possibility that will become increasingly relevant with the advent of targeted inner ear therapies.
Collapse
|
11
|
Artificial Intelligence Applications in Otology: A State of the Art Review. Otolaryngol Head Neck Surg 2020; 163:1123-1133. [DOI: 10.1177/0194599820931804] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
Objective Recent advances in artificial intelligence (AI) are driving innovative new health care solutions. We aim to review the state of the art of AI in otology and provide a discussion of work underway, current limitations, and future directions. Data Sources Two comprehensive databases, MEDLINE and EMBASE, were mined using a directed search strategy to identify all articles that applied AI to otology. Review Methods An initial abstract and title screening was completed. Exclusion criteria included nonavailable abstract and full text, language, and nonrelevance. References of included studies and relevant review articles were cross-checked to identify additional studies. Conclusion The database search identified 1374 articles. Abstract and title screening resulted in full-text retrieval of 96 articles. A total of N = 38 articles were retained. Applications of AI technologies involved the optimization of hearing aid technology (n = 5; 13% of all articles), speech enhancement technologies (n = 4; 11%), diagnosis and management of vestibular disorders (n = 11; 29%), prediction of sensorineural hearing loss outcomes (n = 9; 24%), interpretation of automatic brainstem responses (n = 5; 13%), and imaging modalities and image-processing techniques (n = 4; 10%). Publication counts of the included articles from each decade demonstrated a marked increase in interest in AI in recent years. Implications for Practice This review highlights several applications of AI that otologists and otolaryngologists alike should be aware of given the possibility of implementation in mainstream clinical practice. Although there remain significant ethical and regulatory challenges, AI powered systems offer great potential to shape how healthcare systems of the future operate and clinicians are key stakeholders in this process.
Collapse
|
12
|
Artificial intelligence to detect tympanic membrane perforations. The Journal of Laryngology & Otology 2020; 134:311-315. [PMID: 32238202 DOI: 10.1017/s0022215120000717] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
Abstract
OBJECTIVE To explore the feasibility of constructing a proof-of-concept artificial intelligence algorithm to detect tympanic membrane perforations, for future application in under-resourced rural settings. METHODS A retrospective review was conducted of otoscopic images analysed using transfer learning with Google's Inception-V3 convolutional neural network architecture. The 'gold standard' 'ground truth' was defined by otolaryngologists. Perforation size was categorised as less than one-third (small), one-third to two-thirds (medium), or more than two-thirds (large) of the total tympanic membrane diameter. RESULTS A total of 233 tympanic membrane images were used (183 for training, 50 for testing). The algorithm correctly identified intact and perforated tympanic membranes (overall accuracy = 76.0 per cent, 95 per cent confidence interval = 62.1-86.0 per cent); the area under the curve was 0.867 (95 per cent confidence interval = 0.771-0.963). CONCLUSION A proof-of-concept image-classification artificial intelligence algorithm can be used to detect tympanic membrane perforations and, with further development, may prove to be a valuable tool for ear disease screening. Future endeavours are warranted to develop a point-of-care tool for healthcare workers in areas distant from otolaryngology.
Collapse
|
13
|
Wu C, Qiao Z, Zhang N, Li X, Fan J, Song H, Ai D, Yang J, Huang Y. Phase unwrapping based on a residual en-decoder network for phase images in Fourier domain Doppler optical coherence tomography. BIOMEDICAL OPTICS EXPRESS 2020; 11:1760-1771. [PMID: 32341846 PMCID: PMC7173896 DOI: 10.1364/boe.386101] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/16/2019] [Revised: 02/19/2020] [Accepted: 02/27/2020] [Indexed: 06/01/2023]
Abstract
To solve the phase unwrapping problem for phase images in Fourier domain Doppler optical coherence tomography (DOCT), we propose a deep learning-based residual en-decoder network (REDN) method. In our approach, we reformulate the definition for obtaining the true phase as obtaining an integer multiple of 2π at each pixel by semantic segmentation. The proposed REDN architecture can provide recognition performance with pixel-level accuracy. To address the lack of phase images that are noise and wrapping free from DOCT systems for training, we used simulated images synthesized with DOCT phase image background noise features. An evaluation study on simulated images, DOCT phase images of phantom milk flowing in a plastic tube and a mouse artery, was performed. Meanwhile, a comparison study with recently proposed deep learning-based DeepLabV3+ and PhaseNet methods for signal phase unwrapping and traditional modified networking programming (MNP) method was also performed. Both visual inspection and quantitative metrical evaluation based on accuracy, specificity, sensitivity, root-mean-square-error, total-variation, and processing time demonstrate the robustness, effectiveness and superiority of our method. The proposed REDN method will benefit accurate and fast DOCT phase image-based diagnosis and evaluation when the detected phase is wrapped and will enrich the deep learning-based image processing platform for DOCT images.
Collapse
Affiliation(s)
- Chuanchao Wu
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Zhengyu Qiao
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Nan Zhang
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Xiaochen Li
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Jingfan Fan
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Hong Song
- School of Computer Science and Technology, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Danni Ai
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Jian Yang
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| | - Yong Huang
- School of Optics and Photonics, Beijing Institute of Technology, No. 5 South Zhongguancun Street, Haidian, Beijing 100081, China
| |
Collapse
|
14
|
Viscaino M, Maass JC, Delano PH, Torrente M, Stott C, Auat Cheein F. Computer-aided diagnosis of external and middle ear conditions: A machine learning approach. PLoS One 2020; 15:e0229226. [PMID: 32163427 PMCID: PMC7067442 DOI: 10.1371/journal.pone.0229226] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2019] [Accepted: 01/31/2020] [Indexed: 12/27/2022] Open
Abstract
In medicine, a misdiagnosis or the absence of specialists can affect the patient’s health, leading to unnecessary tests and increasing the costs of healthcare. In particular, the lack of specialists in otolaryngology in third world countries forces patients to seek medical attention from general practitioners, whom might not have enough training and experience for making correct diagnosis in this field. To tackle this problem, we propose and test a computer-aided system based on machine learning models and image processing techniques for otoscopic examination, as a support for a more accurate diagnosis of ear conditions at primary care before specialist referral; in particular, for myringosclerosis, earwax plug, and chronic otitis media. To characterize the tympanic membrane and ear canal for each condition, we implemented three different feature extraction methods: color coherence vector, discrete cosine transform, and filter bank. We also considered three machine learning algorithms: support vector machine (SVM), k-nearest neighbor (k-NN) and decision trees to develop the ear condition predictor model. To conduct the research, our database included 160 images as testing set and 720 images as training and validation sets of 180 patients. We repeatedly trained the learning models using the training dataset and evaluated them using the validation dataset to thus obtain the best feature extraction method and learning model that produce the highest validation accuracy. The results showed that the SVM and k-NN presented the best performance followed by decision trees model. Finally, we performed a classification stage –i.e., diagnosis– using testing data, where the SVM model achieved an average classification accuracy of 93.9%, average sensitivity of 87.8%, average specificity of 95.9%, and average positive predictive value of 87.7%. The results show that this system might be used for general practitioners as a reference to make better decisions in the ear pathologies diagnosis.
Collapse
Affiliation(s)
- Michelle Viscaino
- Department of Electronic Engineering, Universidad Técnica Federico Santa María, Valparaíso, Chile
| | - Juan C. Maass
- Interdisciplinary Program of Phisiology and Biophisics, Facultad de Medicina, Instituto de Ciencias Biomedicas, Universidad de Chile, Santiago, Chile
- Department of Otolaryngology, Hospital Clínico de la Universidad de Chile, Santiago, Chile
| | - Paul H. Delano
- Department of Neuroscience, Facultad de Medicina, Universidad de Chile, Santiago, Chile
- Department of Otolaryngology, Hospital Clínico de la Universidad de Chile, Santiago, Chile
| | - Mariela Torrente
- Department of Otolaryngology, Hospital Clínico de la Universidad de Chile, Santiago, Chile
| | - Carlos Stott
- Department of Otolaryngology, Hospital Clínico de la Universidad de Chile, Santiago, Chile
| | - Fernando Auat Cheein
- Department of Electronic Engineering, Universidad Técnica Federico Santa María, Valparaíso, Chile
- * E-mail:
| |
Collapse
|
15
|
Heidari AE, Pham TT, Ifegwu I, Burwell R, Armstrong WB, Tjoson T, Whyte S, Giorgioni C, Wang B, Wong BJ, Chen Z. The use of optical coherence tomography and convolutional neural networks to distinguish normal and abnormal oral mucosa. JOURNAL OF BIOPHOTONICS 2020; 13:e201900221. [PMID: 31710775 PMCID: PMC7250484 DOI: 10.1002/jbio.201900221] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/05/2019] [Revised: 11/08/2019] [Accepted: 11/10/2019] [Indexed: 06/10/2023]
Abstract
Incomplete surgical resection of head and neck squamous cell carcinoma (HNSCC) is the most common cause of local HNSCC recurrence. Currently, surgeons rely on preoperative imaging, direct visualization, palpation and frozen section to determine the extent of tissue resection. It has been demonstrated that optical coherence tomography (OCT), a minimally invasive, nonionizing near infrared mesoscopic imaging modality can resolve subsurface differences between normal and abnormal head and neck mucosa. Previous work has utilized two-dimensional OCT imaging which is limited to the evaluation of small regions of interest generated frame by frame. OCT technology is capable of performing rapid volumetric imaging, but the capacity and expertise to analyze this massive amount of image data is lacking. In this study, we evaluate the ability of a retrained convolutional neural network to classify three-dimensional OCT images of head and neck mucosa to differentiate normal and abnormal tissues with sensitivity and specificity of 100% and 70%, respectively. This method has the potential to serve as a real-time analytic tool in the assessment of surgical margins.
Collapse
Affiliation(s)
- Andrew E. Heidari
- Beckman Laser Institute & Medical Clinic, Irvine, CA 92612, USA
- Department of Biomedical Engineering, University of California - Irvine, Irvine, CA 92697, USA
| | - Tiffany T. Pham
- Beckman Laser Institute & Medical Clinic, Irvine, CA 92612, USA
- University of California - Irvine, School of Medicine, Irvine, CA 92617, USA
| | - Ibe Ifegwu
- Department of Pathology, University of California – Irvine, Irvine, CA 92697, USA
| | - Ross Burwell
- Department of Pathology, University of California – Irvine, Irvine, CA 92697, USA
| | - William B. Armstrong
- Department of Otolaryngology - Head and Neck Surgery, University of California - Irvine, School of Medicine, Orange, CA 92868, USA
| | - Tjoa Tjoson
- Department of Otolaryngology - Head and Neck Surgery, University of California - Irvine, School of Medicine, Orange, CA 92868, USA
| | - Stephanie Whyte
- Department of Pathology, University of California – Irvine, Irvine, CA 92697, USA
| | - Carmen Giorgioni
- Department of Pathology, University of California – Irvine, Irvine, CA 92697, USA
| | - Beverly Wang
- Department of Pathology, University of California – Irvine, Irvine, CA 92697, USA
| | - Brian J.F. Wong
- Beckman Laser Institute & Medical Clinic, Irvine, CA 92612, USA
- Department of Biomedical Engineering, University of California - Irvine, Irvine, CA 92697, USA
- Department of Otolaryngology - Head and Neck Surgery, University of California - Irvine, School of Medicine, Orange, CA 92868, USA
| | - Zhongping Chen
- Beckman Laser Institute & Medical Clinic, Irvine, CA 92612, USA
- Department of Biomedical Engineering, University of California - Irvine, Irvine, CA 92697, USA
| |
Collapse
|
16
|
Affiliation(s)
- Turgut Karlıdağ
- Department of Otorhinolaryngology, Fırat University School of Medicine, Elazığ, Turkey
| |
Collapse
|
17
|
Crowson MG, Ranisau J, Eskander A, Babier A, Xu B, Kahmke RR, Chen JM, Chan TCY. A contemporary review of machine learning in otolaryngology-head and neck surgery. Laryngoscope 2019; 130:45-51. [PMID: 30706465 DOI: 10.1002/lary.27850] [Citation(s) in RCA: 65] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2018] [Accepted: 01/11/2019] [Indexed: 11/07/2022]
Abstract
One of the key challenges with big data is leveraging the complex network of information to yield useful clinical insights. The confluence of massive amounts of health data and a desire to make inferences and insights on these data has produced a substantial amount of interest in machine-learning analytic methods. There has been a drastic increase in the otolaryngology literature volume describing novel applications of machine learning within the past 5 years. In this timely contemporary review, we provide an overview of popular machine-learning techniques, and review recent machine-learning applications in otolaryngology-head and neck surgery including neurotology, head and neck oncology, laryngology, and rhinology. Investigators have realized significant success in validated models with model sensitivities and specificities approaching 100%. Challenges remain in the implementation of machine-learning algorithms. This may be in part the unfamiliarity of these techniques to clinician leaders on the front lines of patient care. Spreading awareness and confidence in machine learning will follow with further validation and proof-of-value analyses that demonstrate model performance superiority over established methods. We are poised to see a greater influx of machine-learning applications to clinical problems in otolaryngology-head and neck surgery, and it is prudent for providers to understand the potential benefits and limitations of these technologies. Laryngoscope, 130:45-51, 2020.
Collapse
Affiliation(s)
- Matthew G Crowson
- Department of Otolaryngology-Head and Neck Surgery, Sunnybrook Health Sciences Center, Toronto, Ontario, Canada
| | - Jonathan Ranisau
- Department of Mechanical and Industrial Engineering, University of Toronto, Toronto, Ontario, Canada
| | - Antoine Eskander
- Department of Otolaryngology-Head and Neck Surgery, Sunnybrook Health Sciences Center, Toronto, Ontario, Canada
- Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada
| | - Aaron Babier
- Department of Mechanical and Industrial Engineering, University of Toronto, Toronto, Ontario, Canada
| | - Bin Xu
- Department of Pathology, Sunnybrook Health Sciences Center, Toronto, Ontario, Canada
| | - Russel R Kahmke
- Division of Otolaryngology-Head and Neck Surgery, Duke University Medical Center, Durham, North Carolina, U.S.A
| | - Joseph M Chen
- Department of Otolaryngology-Head and Neck Surgery, Sunnybrook Health Sciences Center, Toronto, Ontario, Canada
| | - Timothy C Y Chan
- Department of Mechanical and Industrial Engineering, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
18
|
Chowdhury NI, Smith TL, Chandra RK, Turner JH. Automated classification of osteomeatal complex inflammation on computed tomography using convolutional neural networks. Int Forum Allergy Rhinol 2019; 9:46-52. [PMID: 30098123 PMCID: PMC6318014 DOI: 10.1002/alr.22196] [Citation(s) in RCA: 42] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2018] [Revised: 06/08/2018] [Accepted: 07/09/2018] [Indexed: 12/13/2022]
Abstract
BACKGROUND Convolutional neural networks (CNNs) are advanced artificial intelligence algorithms well suited to image classification tasks with variable features. These have been used to great effect in various real-world applications including handwriting recognition, face detection, image search, and fraud prevention. We sought to retrain a robust CNN with coronal computed tomography (CT) images to classify osteomeatal complex (OMC) occlusion and assess the performance of this technology with rhinologic data. METHODS The Google Inception-V3 CNN trained with 1.28 million images was used as the base model. Preoperative coronal sections through the OMC were obtained from 239 patients enrolled in 2 prospective chronic rhinosinusitis (CRS) outcomes studies, labeled according to OMC status, and mirrored to obtain a set of 956 images. Using this data, the classification layer of Inception-V3 was retrained in Python using a transfer learning method to adapt the CNN to the task of interpreting sinonasal CT images. RESULTS The retrained neural network achieved 85% classification accuracy for OMC occlusion, with a 95% confidence interval for algorithm accuracy of 78% to 92%. Receiver operating characteristic (ROC) curve analysis on the test set confirmed good classification ability of the CNN with an area under the ROC curve (AUC) of 0.87, significantly different than both random guessing and a dominant classifier that predicts the most common class (p < 0.0001). CONCLUSION Current state-of-the-art CNNs may be able to learn clinically relevant information from 2-dimensional sinonasal CT images with minimal supervision. Future work will extend this approach to 3-dimensional images in order to further refine this technology for possible clinical applications.
Collapse
Affiliation(s)
- Naweed I. Chowdhury
- Vanderbilt University School of Medicine, Otolaryngology & Head and Neck Surgery, Nashville, TN., USA
| | - Timothy L. Smith
- Oregon Health & Science University, Department of Otolaryngology-Head & Neck Surgery, Portland, OR., USA
| | - Rakesh K. Chandra
- Vanderbilt University School of Medicine, Otolaryngology & Head and Neck Surgery, Nashville, TN., USA
| | - Justin H. Turner
- Vanderbilt University School of Medicine, Otolaryngology & Head and Neck Surgery, Nashville, TN., USA
| |
Collapse
|
19
|
Sawyer TW, Rice PFS, Sawyer DM, Koevary JW, Barton JK. Evaluation of segmentation algorithms for optical coherence tomography images of ovarian tissue. J Med Imaging (Bellingham) 2019; 6:014002. [PMID: 30746391 PMCID: PMC6350616 DOI: 10.1117/1.jmi.6.1.014002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Accepted: 12/27/2018] [Indexed: 12/31/2022] Open
Abstract
Ovarian cancer has the lowest survival rate among all gynecologic cancers predominantly due to late diagnosis. Early detection of ovarian cancer can increase 5-year survival rates from 40% up to 92%, yet no reliable early detection techniques exist. Optical coherence tomography (OCT) is an emerging technique that provides depth-resolved, high-resolution images of biological tissue in real-time and demonstrates great potential for imaging of ovarian tissue. Mouse models are crucial to quantitatively assess the diagnostic potential of OCT for ovarian cancer imaging; however, due to small organ size, the ovaries must first be separated from the image background using the process of segmentation. Manual segmentation is time-intensive, as OCT yields three-dimensional data. Furthermore, speckle noise complicates OCT images, frustrating many processing techniques. While much work has investigated noise-reduction and automated segmentation for retinal OCT imaging, little has considered the application to the ovaries, which exhibit higher variance and inhomogeneity than the retina. To address these challenges, we evaluate a set of algorithms to segment OCT images of mouse ovaries. We examine five preprocessing techniques and seven segmentation algorithms. While all preprocessing methods improve segmentation, Gaussian filtering is most effective, showing an improvement of 32 % ± 1.2 % . Of the segmentation algorithms, active contours performs best, segmenting with an accuracy of 94.8 % ± 1.2 % compared with manual segmentation. Even so, further optimization could lead to maximizing the performance for segmenting OCT images of the ovaries.
Collapse
Affiliation(s)
- Travis W. Sawyer
- University of Arizona, College of Optical Sciences, Tucson, Arizona, United States
| | - Photini F. S. Rice
- University of Arizona, Department of Biomedical Engineering, Tucson, Arizona, United States
| | | | - Jennifer W. Koevary
- University of Arizona, Department of Biomedical Engineering, Tucson, Arizona, United States
| | - Jennifer K. Barton
- University of Arizona, College of Optical Sciences, Tucson, Arizona, United States
- University of Arizona, Department of Biomedical Engineering, Tucson, Arizona, United States
| |
Collapse
|
20
|
Loo J, Fang L, Cunefare D, Jaffe GJ, Farsiu S. Deep longitudinal transfer learning-based automatic segmentation of photoreceptor ellipsoid zone defects on optical coherence tomography images of macular telangiectasia type 2. BIOMEDICAL OPTICS EXPRESS 2018; 9:2681-2698. [PMID: 30258683 PMCID: PMC6154208 DOI: 10.1364/boe.9.002681] [Citation(s) in RCA: 37] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/26/2018] [Revised: 05/10/2018] [Accepted: 05/11/2018] [Indexed: 05/20/2023]
Abstract
Photoreceptor ellipsoid zone (EZ) defects visible on optical coherence tomography (OCT) are important imaging biomarkers for the onset and progression of macular diseases. As such, accurate quantification of EZ defects is paramount to monitor disease progression and treatment efficacy over time. We developed and trained a novel deep learning-based method called Deep OCT Atrophy Detection (DOCTAD) to automatically segment EZ defect areas by classifying 3-dimensional A-scan clusters as normal or defective. Furthermore, we introduce a longitudinal transfer learning paradigm in which the algorithm learns from segmentation errors on images obtained at one time point to segment subsequent images with higher accuracy. We evaluated the performance of this method on 134 eyes of 67 subjects enrolled in a clinical trial of a novel macular telangiectasia type 2 (MacTel2) therapeutic agent. Our method compared favorably to other deep learning-based and non-deep learning-based methods in matching expert manual segmentations. To the best of our knowledge, this is the first automatic segmentation method developed for EZ defects on OCT images of MacTel2.
Collapse
Affiliation(s)
- Jessica Loo
- Department of Biomedical Engineering, Duke University, Durham, NC 27708, USA
| | - Leyuan Fang
- Department of Biomedical Engineering, Duke University, Durham, NC 27708, USA
| | - David Cunefare
- Department of Biomedical Engineering, Duke University, Durham, NC 27708, USA
| | - Glenn J. Jaffe
- Department of Ophthalmology, Duke University, Durham, NC 27708, USA
| | - Sina Farsiu
- Department of Biomedical Engineering, Duke University, Durham, NC 27708, USA
- Department of Ophthalmology, Duke University, Durham, NC 27708, USA
| |
Collapse
|