1
|
Łajczak P, Matyja J, Jóźwik K, Nawrat Z. Accuracy of vestibular schwannoma segmentation using deep learning models - a systematic review & meta-analysis. Neuroradiology 2024:10.1007/s00234-024-03449-1. [PMID: 39179652 DOI: 10.1007/s00234-024-03449-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Accepted: 08/09/2024] [Indexed: 08/26/2024]
Abstract
Vestibular Schwannoma (VS) is a rare tumor with varied incidence rates, predominantly affecting the 60-69 age group. In the era of artificial intelligence (AI), deep learning (DL) algorithms show promise in automating diagnosis. However, a knowledge gap exists in the automated segmentation of VS using DL. To address this gap, this meta-analysis aims to provide insights into the current state of DL algorithms applied to MR images of VS. METHODOLOGY Following 2020 PRISMA guidelines, a search across four databases was conducted. Inclusion criteria focused on articles using DL for VS MR image segmentation. The primary metric was the Dice score, supplemented by relative volume error (RVE) and average symmetric surface distance (ASSD). RESULTS The search process identified 752 articles, leading to 11 studies for meta-analysis. A QUADAS- 2 analysis revealed varying biases. The overall Dice score for 56 models was 0.89 (CI: 0.88-0.90), with high heterogeneity (I2 = 95.9%). Subgroup analyses based on DL architecture, MRI inputs, and testing set sizes revealed performance variations. 2.5D DL networks demonstrated comparable efficacy to 3D networks. Imaging input analyses highlighted the superiority of contrast-enhanced T1-weighted imaging and mixed MRI inputs. DISCUSSION This study fills a gap in systematic review in the automated segmentation of VS using DL techniques. Despite promising results, limitations include publication bias and high heterogeneity. Future research should focus on standardized designs, larger testing sets, and addressing biases for more reliable results. DL have promising efficacy in VS diagnosis, however further validation and standardization is needed. CONCLUSION In conclusion, this meta-analysis provides comprehensive review into the current landscape of automated VS segmentation using DL. The high Dice score indicates promising agreement in segmentation, yet challenges like bias and heterogeneity must be addressed in the future research.
Collapse
Affiliation(s)
- Paweł Łajczak
- Department of Biophysics, Faculty of Medical Sciences in Zabrze, Medical University of Silesia in Katowice, Jordana 18, Mekelweg 5, Zabrze, 40-043,, Poland.
| | - Jakub Matyja
- TU Delft, Mekelweg 5,, Delft 2628 CD,, Netherlands
| | - Kamil Jóźwik
- Department of Biophysics, Faculty of Medical Sciences in Zabrze, Medical University of Silesia in Katowice, Jordana 18, Mekelweg 5, Zabrze, 40-043,, Poland
| | - Zbigniew Nawrat
- Department of Biophysics, Faculty of Medical Sciences in Zabrze, Medical University of Silesia in Katowice, Jordana 18, Mekelweg 5, Zabrze, 40-043,, Poland
- Foundation of Cardiac Surgery Development, Zabrze, 41-808, Poland
| |
Collapse
|
2
|
Nernekli K, Persad AR, Hori YS, Yener U, Celtikci E, Sahin MC, Sozer A, Sozer B, Park DJ, Chang SD. Automatic Segmentation of Vestibular Schwannomas: A Systematic Review. World Neurosurg 2024; 188:35-44. [PMID: 38685346 DOI: 10.1016/j.wneu.2024.04.145] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 04/23/2024] [Indexed: 05/02/2024]
Abstract
BACKGROUND Vestibular schwannomas (VSs) are benign tumors often monitored over time, with measurement techniques for assessing growth rates subject to significant interobserver variability. Automatic segmentation of these tumors could provide a more reliable and efficient for tracking their progression, especially given the irregular shape and growth patterns of VS. METHODS Various studies and segmentation techniques employing different Convolutional Neural Network architectures and models, such as U-Net and convolutional-attention transformer segmentation, were analyzed. Models were evaluated based on their performance across diverse datasets, and challenges, including domain shift and data sharing, were scrutinized. RESULTS Automatic segmentation methods offer a promising alternative to conventional measurement techniques, offering potential benefits in precision and efficiency. However, these methods are not without challenges, notably the "domain shift" that occurs when models trained on specific datasets underperform when applied to different datasets. Techniques such as domain adaptation, domain generalization, and data diversity were discussed as potential solutions. CONCLUSIONS Accurate measurement of VS growth is a complex process, with volumetric analysis currently appearing more reliable than linear measurements. Automatic segmentation, despite its challenges, offers a promising avenue for future investigation. Robust well-generalized models could potentially improve the efficiency of tracking tumor growth, thereby augmenting clinical decision-making. Further work needs to be done to develop more robust models, address the domain shift, and enable secure data sharing for wider applicability.
Collapse
Affiliation(s)
- Kerem Nernekli
- Department of Radiology, Stanford University School of Medicine, Stanford, California, USA
| | - Amit R Persad
- Department of Neurosurgery, Stanford University School of Medicine, Stanford, California, USA
| | - Yusuke S Hori
- Department of Neurosurgery, Stanford University School of Medicine, Stanford, California, USA
| | - Ulas Yener
- Department of Neurosurgery, Stanford University School of Medicine, Stanford, California, USA
| | - Emrah Celtikci
- Department of Neurosurgery, Gazi University, Ankara, Turkey
| | | | - Alperen Sozer
- Department of Neurosurgery, Gazi University, Ankara, Turkey
| | - Batuhan Sozer
- Department of Neurosurgery, Gazi University, Ankara, Turkey
| | - David J Park
- Department of Neurosurgery, Stanford University School of Medicine, Stanford, California, USA.
| | - Steven D Chang
- Department of Neurosurgery, Stanford University School of Medicine, Stanford, California, USA
| |
Collapse
|
3
|
Cornelissen S, Schouten SM, Langenhuizen PPJH, Lie ST, Kunst HPM, de With PHN, Verheul JB. Defining tumor growth in vestibular schwannomas: a volumetric inter-observer variability study in contrast-enhanced T1-weighted MRI. Neuroradiology 2024:10.1007/s00234-024-03416-w. [PMID: 38980343 DOI: 10.1007/s00234-024-03416-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2024] [Accepted: 06/25/2024] [Indexed: 07/10/2024]
Abstract
PURPOSE For patients with vestibular schwannomas (VS), a conservative observational approach is increasingly used. Therefore, the need for accurate and reliable volumetric tumor monitoring is important. Currently, a volumetric cutoff of 20% increase in tumor volume is widely used to define tumor growth in VS. The study investigates the tumor volume dependency on the limits of agreement (LoA) for volumetric measurements of VS by means of an inter-observer study. METHODS This retrospective study included 100 VS patients who underwent contrast-enhanced T1-weighted MRI. Five observers volumetrically annotated the images. Observer agreement and reliability was measured using the LoA, estimated using the limits of agreement with the mean (LOAM) method, and the intraclass correlation coefficient (ICC). RESULTS The 100 patients had a median average tumor volume of 903 mm3 (IQR: 193-3101). Patients were divided into four volumetric size categories based on tumor volume quartile. The smallest tumor volume quartile showed a LOAM relative to the mean of 26.8% (95% CI: 23.7-33.6), whereas for the largest tumor volume quartile this figure was found to be 7.3% (95% CI: 6.5-9.7) and when excluding peritumoral cysts: 4.8% (95% CI: 4.2-6.2). CONCLUSION Agreement limits within volumetric annotation of VS are affected by tumor volume, since the LoA improves with increasing tumor volume. As a result, for tumors larger than 200 mm3, growth can reliably be detected at an earlier stage, compared to the currently widely used cutoff of 20%. However, for very small tumors, growth should be assessed with higher agreement limits than previously thought.
Collapse
Affiliation(s)
- Stefan Cornelissen
- Gamma Knife Center, Department of Neurosurgery, Elisabeth-TweeSteden Hospital, Tilburg, The Netherlands.
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands.
| | - Sammy M Schouten
- Gamma Knife Center, Department of Neurosurgery, Elisabeth-TweeSteden Hospital, Tilburg, The Netherlands
- Department of Otolaryngology, Radboud University Medical Center, Nijmegen, The Netherlands
- Department of Otolaryngology, Maastricht University Medical Centre+, Maastricht, The Netherlands
| | - Patrick P J H Langenhuizen
- Gamma Knife Center, Department of Neurosurgery, Elisabeth-TweeSteden Hospital, Tilburg, The Netherlands
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Suan Te Lie
- Gamma Knife Center, Department of Neurosurgery, Elisabeth-TweeSteden Hospital, Tilburg, The Netherlands
| | - Henricus P M Kunst
- Department of Otolaryngology, Radboud University Medical Center, Nijmegen, The Netherlands
- Department of Otolaryngology, Maastricht University Medical Centre+, Maastricht, The Netherlands
| | - Peter H N de With
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Jeroen B Verheul
- Gamma Knife Center, Department of Neurosurgery, Elisabeth-TweeSteden Hospital, Tilburg, The Netherlands
| |
Collapse
|
4
|
Heman-Ackah SM, Blue R, Quimby AE, Abdallah H, Sweeney EM, Chauhan D, Hwa T, Brant J, Ruckenstein MJ, Bigelow DC, Jackson C, Zenonos G, Gardner P, Briggs SE, Cohen Y, Lee JYK. A multi-institutional machine learning algorithm for prognosticating facial nerve injury following microsurgical resection of vestibular schwannoma. Sci Rep 2024; 14:12963. [PMID: 38839778 PMCID: PMC11153496 DOI: 10.1038/s41598-024-63161-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2023] [Accepted: 05/26/2024] [Indexed: 06/07/2024] Open
Abstract
Vestibular schwannomas (VS) are the most common tumor of the skull base with available treatment options that carry a risk of iatrogenic injury to the facial nerve, which can significantly impact patients' quality of life. As facial nerve outcomes remain challenging to prognosticate, we endeavored to utilize machine learning to decipher predictive factors relevant to facial nerve outcomes following microsurgical resection of VS. A database of patient-, tumor- and surgery-specific features was constructed via retrospective chart review of 242 consecutive patients who underwent microsurgical resection of VS over a 7-year study period. This database was then used to train non-linear supervised machine learning classifiers to predict facial nerve preservation, defined as House-Brackmann (HB) I vs. facial nerve injury, defined as HB II-VI, as determined at 6-month outpatient follow-up. A random forest algorithm demonstrated 90.5% accuracy, 90% sensitivity and 90% specificity in facial nerve injury prognostication. A random variable (rv) was generated by randomly sampling a Gaussian distribution and used as a benchmark to compare the predictiveness of other features. This analysis revealed age, body mass index (BMI), case length and the tumor dimension representing tumor growth towards the brainstem as prognosticators of facial nerve injury. When validated via prospective assessment of facial nerve injury risk, this model demonstrated 84% accuracy. Here, we describe the development of a machine learning algorithm to predict the likelihood of facial nerve injury following microsurgical resection of VS. In addition to serving as a clinically applicable tool, this highlights the potential of machine learning to reveal non-linear relationships between variables which may have clinical value in prognostication of outcomes for high-risk surgical procedures.
Collapse
Affiliation(s)
- Sabrina M Heman-Ackah
- Department of Neurosurgery, Perelman Center for Advanced Medicine, University of Pennsylvania, 3400 Civic Center Boulevard, 15th Floor, Philadelphia, PA, 19104, USA.
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, USA.
| | - Rachel Blue
- Department of Neurosurgery, Perelman Center for Advanced Medicine, University of Pennsylvania, 3400 Civic Center Boulevard, 15th Floor, Philadelphia, PA, 19104, USA
| | - Alexandra E Quimby
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
- Department of Otolaryngology and Communication Sciences, SUNY Upstate Medical University Hospital, Syracuse, NY, USA
| | - Hussein Abdallah
- School of Medicine, University of Pittsburgh, Pittsburgh, PA, USA
| | - Elizabeth M Sweeney
- Department of Biostatistics, Epidemiology and Informatics, University of Pennsylvania, Philadelphia, PA, USA
| | - Daksh Chauhan
- University of Pennsylvania, Perelman School of Medicine, Philadelphia, PA, USA
| | - Tiffany Hwa
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
| | - Jason Brant
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
- Corporal Michael J. Crescenz VAMC, Philadelphia, PA, USA
| | - Michael J Ruckenstein
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
| | - Douglas C Bigelow
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
| | - Christina Jackson
- Department of Neurosurgery, Perelman Center for Advanced Medicine, University of Pennsylvania, 3400 Civic Center Boulevard, 15th Floor, Philadelphia, PA, 19104, USA
| | - Georgios Zenonos
- Center for Cranial Base Surgery, University of Pittsburgh, Pittsburgh, PA, USA
| | - Paul Gardner
- Center for Cranial Base Surgery, University of Pittsburgh, Pittsburgh, PA, USA
| | - Selena E Briggs
- Department of Otolaryngology, MedStar Washington Hospital Center, Washington, DC, USA
- Department of Otolaryngology, Georgetown University, Washington, DC, USA
| | - Yale Cohen
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, USA
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
- University of Pennsylvania, Perelman School of Medicine, Philadelphia, PA, USA
| | - John Y K Lee
- Department of Neurosurgery, Perelman Center for Advanced Medicine, University of Pennsylvania, 3400 Civic Center Boulevard, 15th Floor, Philadelphia, PA, 19104, USA
- Department of Otorhinolaryngology, University of Pennsylvania, Philadelphia, PA, USA
| |
Collapse
|
5
|
Kujawa A, Dorent R, Connor S, Thomson S, Ivory M, Vahedi A, Guilhem E, Wijethilake N, Bradford R, Kitchen N, Bisdas S, Ourselin S, Vercauteren T, Shapey J. Deep learning for automatic segmentation of vestibular schwannoma: a retrospective study from multi-center routine MRI. Front Comput Neurosci 2024; 18:1365727. [PMID: 38784680 PMCID: PMC11111906 DOI: 10.3389/fncom.2024.1365727] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2024] [Accepted: 04/17/2024] [Indexed: 05/25/2024] Open
Abstract
Automatic segmentation of vestibular schwannoma (VS) from routine clinical MRI has potential to improve clinical workflow, facilitate treatment decisions, and assist patient management. Previous work demonstrated reliable automatic segmentation performance on datasets of standardized MRI images acquired for stereotactic surgery planning. However, diagnostic clinical datasets are generally more diverse and pose a larger challenge to automatic segmentation algorithms, especially when post-operative images are included. In this work, we show for the first time that automatic segmentation of VS on routine MRI datasets is also possible with high accuracy. We acquired and publicly release a curated multi-center routine clinical (MC-RC) dataset of 160 patients with a single sporadic VS. For each patient up to three longitudinal MRI exams with contrast-enhanced T1-weighted (ceT1w) (n = 124) and T2-weighted (T2w) (n = 363) images were included and the VS manually annotated. Segmentations were produced and verified in an iterative process: (1) initial segmentations by a specialized company; (2) review by one of three trained radiologists; and (3) validation by an expert team. Inter- and intra-observer reliability experiments were performed on a subset of the dataset. A state-of-the-art deep learning framework was used to train segmentation models for VS. Model performance was evaluated on a MC-RC hold-out testing set, another public VS datasets, and a partially public dataset. The generalizability and robustness of the VS deep learning segmentation models increased significantly when trained on the MC-RC dataset. Dice similarity coefficients (DSC) achieved by our model are comparable to those achieved by trained radiologists in the inter-observer experiment. On the MC-RC testing set, median DSCs were 86.2(9.5) for ceT1w, 89.4(7.0) for T2w, and 86.4(8.6) for combined ceT1w+T2w input images. On another public dataset acquired for Gamma Knife stereotactic radiosurgery our model achieved median DSCs of 95.3(2.9), 92.8(3.8), and 95.5(3.3), respectively. In contrast, models trained on the Gamma Knife dataset did not generalize well as illustrated by significant underperformance on the MC-RC routine MRI dataset, highlighting the importance of data variability in the development of robust VS segmentation models. The MC-RC dataset and all trained deep learning models were made available online.
Collapse
Affiliation(s)
- Aaron Kujawa
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
| | - Reuben Dorent
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
| | - Steve Connor
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
- Department of Neuroradiology, King's College Hospital, London, United Kingdom
- Department of Radiology, Guy's and St Thomas' Hospital, London, United Kingdom
| | - Suki Thomson
- Department of Neuroradiology, King's College Hospital, London, United Kingdom
| | - Marina Ivory
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
| | - Ali Vahedi
- Department of Neuroradiology, King's College Hospital, London, United Kingdom
| | - Emily Guilhem
- Department of Neuroradiology, King's College Hospital, London, United Kingdom
| | - Navodini Wijethilake
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
| | - Robert Bradford
- Queen Square Radiosurgery Centre (Gamma Knife), National Hospital for Neurology and Neurosurgery, London, United Kingdom
- Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, United Kingdom
| | - Neil Kitchen
- Queen Square Radiosurgery Centre (Gamma Knife), National Hospital for Neurology and Neurosurgery, London, United Kingdom
- Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, United Kingdom
| | - Sotirios Bisdas
- Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, London, United Kingdom
| | - Sebastien Ourselin
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
| | - Tom Vercauteren
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
| | - Jonathan Shapey
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, United Kingdom
- Department of Neurosurgery, King's College Hospital, London, United Kingdom
| |
Collapse
|
6
|
Suresh K, Luo G, Bartholomew RA, Brown A, Juliano AF, Lee DJ, Welling DB, Cai W, Crowson MG. An External Validation Study for Automated Segmentation of Vestibular Schwannoma. Otol Neurotol 2024; 45:e193-e197. [PMID: 38361299 DOI: 10.1097/mao.0000000000004125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/17/2024]
Abstract
OBJECTIVE To validate how an automated model for vestibular schwannoma (VS) segmentation developed on an external homogeneous dataset performs when applied to internal heterogeneous data. PATIENTS The external dataset comprised 242 patients with previously untreated, sporadic unilateral VS undergoing Gamma Knife radiosurgery, with homogeneous magnetic resonance imaging (MRI) scans. The internal dataset comprised 10 patients from our institution, with heterogeneous MRI scans. INTERVENTIONS An automated VS segmentation model was developed on the external dataset. The model was tested on the internal dataset. MAIN OUTCOME MEASURE Dice score, which measures agreement between ground truth and predicted segmentations. RESULTS When applied to the internal patient scans, the automated model achieved a mean Dice score of 61% across all 10 images. There were three tumors that were not detected. These tumors were 0.01 ml on average (SD = 0.00 ml). The mean Dice score for the seven tumors that were detected was 87% (SD = 14%). There was one outlier with Dice of 55%-on further review of this scan, it was discovered that hyperintense petrous bone had been included in the tumor segmentation. CONCLUSIONS We show that an automated segmentation model developed using a restrictive set of siloed institutional data can be successfully adapted for data from different imaging systems and patient populations. This is an important step toward the validation of automated VS segmentation. However, there are significant shortcomings that likely reflect limitations of the data used to train the model. Further validation is needed to make automated segmentation for VS generalizable.
Collapse
Affiliation(s)
- Krish Suresh
- Department of Otolaryngology-Head and Neck Surgery, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| | - Guibo Luo
- Department of Radiology, Massachusetts General Hospital, Harvard University, Boston, Massachusetts
| | - Ryan A Bartholomew
- Department of Otolaryngology-Head and Neck Surgery, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| | - Alyssa Brown
- Department of Otolaryngology-Head and Neck Surgery, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| | - Amy F Juliano
- Department of Radiology, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| | - Daniel J Lee
- Department of Otolaryngology-Head and Neck Surgery, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| | - D Bradley Welling
- Department of Otolaryngology-Head and Neck Surgery, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| | - Wenli Cai
- Department of Radiology, Massachusetts General Hospital, Harvard University, Boston, Massachusetts
| | - Matthew G Crowson
- Department of Otolaryngology-Head and Neck Surgery, Massachusetts Eye and Ear, Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
7
|
Neve OM, Romeijn SR, Chen Y, Nagtegaal L, Grootjans W, Jansen JC, Staring M, Verbist BM, Hensen EF. Automated 2-Dimensional Measurement of Vestibular Schwannoma: Validity and Accuracy of an Artificial Intelligence Algorithm. Otolaryngol Head Neck Surg 2023; 169:1582-1589. [PMID: 37555251 DOI: 10.1002/ohn.470] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2023] [Revised: 06/12/2023] [Accepted: 07/11/2023] [Indexed: 08/10/2023]
Abstract
OBJECTIVE Validation of automated 2-dimensional (2D) diameter measurements of vestibular schwannomas on magnetic resonance imaging (MRI). STUDY DESIGN Retrospective validation study using 2 data sets containing MRIs of vestibular schwannoma patients. SETTING University Hospital in The Netherlands. METHODS Two data sets were used, 1 containing 1 scan per patient (n = 134) and the other containing at least 3 consecutive MRIs of 51 patients, all with contrast-enhanced T1 or high-resolution T2 sequences. 2D measurements of the maximal extrameatal diameters in the axial plane were automatically derived from a 3D-convolutional neural network compared to manual measurements by 2 human observers. Intra- and interobserver variabilities were calculated using the intraclass correlation coefficient (ICC), agreement on tumor progression using Cohen's kappa. RESULTS The human intra- and interobserver variability showed a high correlation (ICC: 0.98-0.99) and limits of agreement of 1.7 to 2.1 mm. Comparing the automated to human measurements resulted in ICC of 0.98 (95% confidence interval [CI]: 0.974; 0.987) and 0.97 (95% CI: 0.968; 0.984), with limits of agreement of 2.2 and 2.1 mm for diameters parallel and perpendicular to the posterior side of the temporal bone, respectively. There was satisfactory agreement on tumor progression between automated measurements and human observers (Cohen's κ = 0.77), better than the agreement between the human observers (Cohen's κ = 0.74). CONCLUSION Automated 2D diameter measurements and growth detection of vestibular schwannomas are at least as accurate as human 2D measurements. In clinical practice, measurements of the maximal extrameatal tumor (2D) diameters of vestibular schwannomas provide important complementary information to total tumor volume (3D) measurements. Combining both in an automated measurement algorithm facilitates clinical adoption.
Collapse
Affiliation(s)
- Olaf M Neve
- Department of Otorhinolaryngology-Head and Neck Surgery, Leiden University Medical Center, Leiden, The Netherlands
| | - Stephan R Romeijn
- Department of Radiology, Leiden University Medical Center, Leiden, The Netherlands
| | - Yunjie Chen
- Department of Radiology, Division of Image Processing, Leiden University Medical Center, Leiden, The Netherlands
| | - Larissa Nagtegaal
- Department of Otorhinolaryngology-Head and Neck Surgery, Leiden University Medical Center, Leiden, The Netherlands
- Department of Radiology, Leiden University Medical Center, Leiden, The Netherlands
| | - Willem Grootjans
- Department of Radiology, Leiden University Medical Center, Leiden, The Netherlands
| | - Jeroen C Jansen
- Department of Otorhinolaryngology-Head and Neck Surgery, Leiden University Medical Center, Leiden, The Netherlands
| | - Marius Staring
- Department of Radiology, Division of Image Processing, Leiden University Medical Center, Leiden, The Netherlands
| | - Berit M Verbist
- Department of Radiology, Leiden University Medical Center, Leiden, The Netherlands
| | - Erik F Hensen
- Department of Otorhinolaryngology-Head and Neck Surgery, Leiden University Medical Center, Leiden, The Netherlands
| |
Collapse
|
8
|
Fujima N, Kamagata K, Ueda D, Fujita S, Fushimi Y, Yanagawa M, Ito R, Tsuboyama T, Kawamura M, Nakaura T, Yamada A, Nozaki T, Fujioka T, Matsui Y, Hirata K, Tatsugami F, Naganawa S. Current State of Artificial Intelligence in Clinical Applications for Head and Neck MR Imaging. Magn Reson Med Sci 2023; 22:401-414. [PMID: 37532584 PMCID: PMC10552661 DOI: 10.2463/mrms.rev.2023-0047] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2023] [Accepted: 07/09/2023] [Indexed: 08/04/2023] Open
Abstract
Due primarily to the excellent soft tissue contrast depictions provided by MRI, the widespread application of head and neck MRI in clinical practice serves to assess various diseases. Artificial intelligence (AI)-based methodologies, particularly deep learning analyses using convolutional neural networks, have recently gained global recognition and have been extensively investigated in clinical research for their applicability across a range of categories within medical imaging, including head and neck MRI. Analytical approaches using AI have shown potential for addressing the clinical limitations associated with head and neck MRI. In this review, we focus primarily on the technical advancements in deep-learning-based methodologies and their clinical utility within the field of head and neck MRI, encompassing aspects such as image acquisition and reconstruction, lesion segmentation, disease classification and diagnosis, and prognostic prediction for patients presenting with head and neck diseases. We then discuss the limitations of current deep-learning-based approaches and offer insights regarding future challenges in this field.
Collapse
Affiliation(s)
- Noriyuki Fujima
- Department of Diagnostic and Interventional Radiology, Hokkaido University Hospital, Sapporo, Hokkaido, Japan
| | - Koji Kamagata
- Department of Radiology, Juntendo University Graduate School of Medicine, Tokyo, Japan
| | - Daiju Ueda
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Osaka, Japan
| | - Shohei Fujita
- Department of Radiology, University of Tokyo, Tokyo, Japan
| | - Yasutaka Fushimi
- Department of Diagnostic Imaging and Nuclear Medicine, Kyoto University Graduate School of Medicine, Kyoto, Kyoto, Japan
| | - Masahiro Yanagawa
- Department of Radiology, Osaka University Graduate School of Medicine, Suita, Osaka, Japan
| | - Rintaro Ito
- Department of Radiology, Nagoya University Graduate School of Medicine, Nagoya, Aichi, Japan
| | - Takahiro Tsuboyama
- Department of Radiology, Osaka University Graduate School of Medicine, Suita, Osaka, Japan
| | - Mariko Kawamura
- Department of Radiology, Nagoya University Graduate School of Medicine, Nagoya, Aichi, Japan
| | - Takeshi Nakaura
- Department of Diagnostic Radiology, Kumamoto University Graduate School of Medicine, Kumamoto, Kumamoto, Japan
| | - Akira Yamada
- Department of Radiology, Shinshu University School of Medicine, Matsumoto, Nagano, Japan
| | - Taiki Nozaki
- Department of Radiology, Keio University School of Medicine, Tokyo, Japan
| | - Tomoyuki Fujioka
- Department of Diagnostic Radiology, Tokyo Medical and Dental University, Tokyo, Japan
| | - Yusuke Matsui
- Department of Radiology, Faculty of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Okayama, Japan
| | - Kenji Hirata
- Department of Diagnostic Imaging, Graduate School of Medicine, Hokkaido University, Sapporo, Hokkaido, Japan
| | - Fuminari Tatsugami
- Department of Diagnostic Radiology, Hiroshima University, Hiroshima, Hiroshima, Japan
| | - Shinji Naganawa
- Department of Radiology, Nagoya University Graduate School of Medicine, Nagoya, Aichi, Japan
| |
Collapse
|
9
|
Wang K, George-Jones NA, Chen L, Hunter JB, Wang J. Joint Vestibular Schwannoma Enlargement Prediction and Segmentation Using a Deep Multi-task Model. Laryngoscope 2023; 133:2754-2760. [PMID: 36495306 PMCID: PMC10256836 DOI: 10.1002/lary.30516] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2022] [Revised: 11/17/2022] [Accepted: 11/20/2022] [Indexed: 12/14/2022]
Abstract
OBJECTIVE To develop a deep-learning-based multi-task (DMT) model for joint tumor enlargement prediction (TEP) and automatic tumor segmentation (TS) for vestibular schwannoma (VS) patients using their initial diagnostic contrast-enhanced T1-weighted (ceT1) magnetic resonance images (MRIs). METHODS Initial ceT1 MRIs for VS patients meeting the inclusion/exclusion criteria of this study were retrospectively collected. VSs on the initial MRIs and their first follow-up scans were manually contoured. Tumor volume and enlargement ratio were measured based on expert contours. A DMT model was constructed for jointly TS and TEP. The manually segmented VS volume on the initial scan and the tumor enlargement label (≥20% volumetric growth) were used as the ground truth for training and evaluating the TS and TEP modules, respectively. RESULTS We performed 5-fold cross-validation with the eligible patients (n = 103). Median segmentation dice coefficient, prediction sensitivity, specificity, accuracy, and area under the receiver operating characteristic curve (AUC) were measured and achieved the following values: 84.20%, 0.68, 0.78, 0.72, and 0.77, respectively. The segmentation result is significantly better than the separate TS network (dice coefficient of 83.13%, p = 0.03) and marginally lower than the state-of-the-art segmentation model nnU-Net (dice coefficient of 86.45%, p = 0.16). The TEP performance is significantly better than the single-task prediction model (AUC = 0.60, p = 0.01) and marginally better than a radiomics-based prediction model (AUC = 0.70, p = 0.17). CONCLUSION The proposed DMT model is of higher learning efficiency and achieves promising performance on TEP and TS. The proposed technology has the potential to improve VS patient management. LEVEL OF EVIDENCE NA Laryngoscope, 133:2754-2760, 2023.
Collapse
Affiliation(s)
- Kai Wang
- The Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - Nicholas A George-Jones
- The Department of Otolaryngology-Head and Neck Surgery, University of Texas Southwestern Medical Center, Dallas, Texas, USA
- The Department of Otolaryngology-Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, Iowa, USA
| | - Liyuan Chen
- The Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - Jacob B Hunter
- The Department of Otolaryngology-Head and Neck Surgery, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - Jing Wang
- The Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| |
Collapse
|
10
|
Petsiou DP, Martinos A, Spinos D. Applications of Artificial Intelligence in Temporal Bone Imaging: Advances and Future Challenges. Cureus 2023; 15:e44591. [PMID: 37795060 PMCID: PMC10545916 DOI: 10.7759/cureus.44591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/02/2023] [Indexed: 10/06/2023] Open
Abstract
The applications of artificial intelligence (AI) in temporal bone (TB) imaging have gained significant attention in recent years, revolutionizing the field of otolaryngology and radiology. Accurate interpretation of imaging features of TB conditions plays a crucial role in diagnosing and treating a range of ear-related pathologies, including middle and inner ear diseases, otosclerosis, and vestibular schwannomas. According to multiple clinical studies published in the literature, AI-powered algorithms have demonstrated exceptional proficiency in interpreting imaging findings, not only saving time for physicians but also enhancing diagnostic accuracy by reducing human error. Although several challenges remain in routinely relying on AI applications, the collaboration between AI and healthcare professionals holds the key to better patient outcomes and significantly improved patient care. This overview delivers a comprehensive update on the advances of AI in the field of TB imaging, summarizes recent evidence provided by clinical studies, and discusses future insights and challenges in the widespread integration of AI in clinical practice.
Collapse
Affiliation(s)
- Dioni-Pinelopi Petsiou
- Otolaryngology-Head and Neck Surgery, National and Kapodistrian University of Athens, School of Medicine, Athens, GRC
| | - Anastasios Martinos
- Otolaryngology-Head and Neck Surgery, National and Kapodistrian University of Athens, School of Medicine, Athens, GRC
| | - Dimitrios Spinos
- Otolaryngology-Head and Neck Surgery, Gloucestershire Hospitals NHS Foundation Trust, Gloucester, GBR
| |
Collapse
|
11
|
Neves CA, Liu GS, El Chemaly T, Bernstein IA, Fu F, Blevins NH. Automated Radiomic Analysis of Vestibular Schwannomas and Inner Ears Using Contrast-Enhanced T1-Weighted and T2-Weighted Magnetic Resonance Imaging Sequences and Artificial Intelligence. Otol Neurotol 2023; 44:e602-e609. [PMID: 37464458 DOI: 10.1097/mao.0000000000003959] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/20/2023]
Abstract
OBJECTIVE To objectively evaluate vestibular schwannomas (VSs) and their spatial relationships with the ipsilateral inner ear (IE) in magnetic resonance imaging (MRI) using deep learning. STUDY DESIGN Cross-sectional study. PATIENTS A total of 490 adults with VS, high-resolution MRI scans, and no previous neurotologic surgery. INTERVENTIONS MRI studies of VS patients were split into training (390 patients) and test (100 patients) sets. A three-dimensional convolutional neural network model was trained to segment VS and IE structures using contrast-enhanced T1-weighted and T2-weighted sequences, respectively. Manual segmentations were used as ground truths. Model performance was evaluated on the test set and on an external set of 100 VS patients from a public data set (Vestibular-Schwannoma-SEG). MAIN OUTCOME MEASURES Dice score, relative volume error, average symmetric surface distance, 95th-percentile Hausdorff distance, and centroid locations. RESULTS Dice scores for VS and IE volume segmentations were 0.91 and 0.90, respectively. On the public data set, the model segmented VS tumors with a Dice score of 0.89 ± 0.06 (mean ± standard deviation), relative volume error of 9.8 ± 9.6%, average symmetric surface distance of 0.31 ± 0.22 mm, and 95th-percentile Hausdorff distance of 1.26 ± 0.76 mm. Predicted VS segmentations overlapped with ground truth segmentations in all test subjects. Mean errors of predicted VS volume, VS centroid location, and IE centroid location were 0.05 cm 3 , 0.52 mm, and 0.85 mm, respectively. CONCLUSIONS A deep learning system can segment VS and IE structures in high-resolution MRI scans with excellent accuracy. This technology offers promise to improve the clinical workflow for assessing VS radiomics and enhance the management of VS patients.
Collapse
Affiliation(s)
| | - George S Liu
- Department of Otolaryngology-Head and Neck Surgery, Stanford University
| | | | - Isaac A Bernstein
- Department of Otolaryngology-Head and Neck Surgery, Stanford University
| | - Fanrui Fu
- Department of Otolaryngology-Head and Neck Surgery, Stanford University
| | - Nikolas H Blevins
- Department of Otolaryngology-Head and Neck Surgery, Stanford University
| |
Collapse
|
12
|
Wang H, Qu T, Bernstein K, Barbee D, Kondziolka D. Automatic segmentation of vestibular schwannomas from T1-weighted MRI with a deep neural network. Radiat Oncol 2023; 18:78. [PMID: 37158968 PMCID: PMC10169364 DOI: 10.1186/s13014-023-02263-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2022] [Accepted: 04/12/2023] [Indexed: 05/10/2023] Open
Abstract
BACKGROUND Long-term follow-up using volumetric measurement could significantly assist in the management of vestibular schwannomas (VS). Manual segmentation of VS from MRI for treatment planning and follow-up assessment is labor-intensive and time-consuming. This study aims to develop a deep learning technique to fully automatically segment VS from MRI. METHODS This study retrospectively analyzed MRI data of 737 patients who received gamma knife radiosurgery for VS. Treatment planning T1-weighted isotropic MR and manually contoured gross tumor volumes (GTV) were used for model development. A 3D convolutional neural network (CNN) was built on ResNet blocks. Spatial attenuation and deep supervision modules were integrated in each decoder level to enhance the training for the small tumor volume on brain MRI. The model was trained and tested on 587 and 150 patient data, respectively, from this institution (n = 495) and a publicly available dataset (n = 242). The model performance were assessed by the Dice similarity coefficient (DSC), 95% Hausdorff distance (HD95), average symmetric surface (ASSD) and relative absolute volume difference (RAVD) of the model segmentation results against the GTVs. RESULTS Measured on combined testing data from two institutions, the proposed method achieved mean DSC of 0.91 ± 0.08, ASSD of 0.3 ± 0.4 mm, HD95 of 1.3 ± 1.6 mm, and RAVD of 0.09 ± 0.15. The DSCs were 0.91 ± 0.09 and 0.92 ± 0.06 on 100 testing patients of this institution and 50 of the public data, respectively. CONCLUSIONS A CNN model was developed for fully automated segmentation of VS on T1-Weighted isotropic MRI. The model achieved good performance compared with physician clinical delineations on a sizeable dataset from two institutions. The proposed method potentially facilitates clinical workflow of radiosurgery for VS patient management.
Collapse
Affiliation(s)
- Hesheng Wang
- Department of Radiation Oncology, NYU Grossman School of Medicine, New York, NY, 10016, USA.
| | - Tanxia Qu
- Department of Radiation Oncology, NYU Grossman School of Medicine, New York, NY, 10016, USA
| | - Kenneth Bernstein
- Department of Radiation Oncology, NYU Grossman School of Medicine, New York, NY, 10016, USA
| | - David Barbee
- Department of Radiation Oncology, NYU Grossman School of Medicine, New York, NY, 10016, USA
| | - Douglas Kondziolka
- Department of Radiation Oncology, NYU Grossman School of Medicine, New York, NY, 10016, USA
- Department of Neurosurgery, NYU Grossman School of Medicine, New York, NY, 10016, USA
| |
Collapse
|