1
|
Neves CA, Chemaly TE, Fu F, Blevins NH. Deep Learning Method for Rapid Simultaneous Multistructure Temporal Bone Segmentation. Otolaryngol Head Neck Surg 2024; 170:1570-1580. [PMID: 38769857 DOI: 10.1002/ohn.764] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 02/25/2024] [Accepted: 03/19/2024] [Indexed: 05/22/2024]
Abstract
OBJECTIVE To develop and validate a deep learning algorithm for the automated segmentation of key temporal bone structures from clinical computed tomography (CT) data sets. STUDY DESIGN Cross-sectional study. SETTING A total of 325 CT scans from a clinical database. METHOD A state-of-the-art deep learning (DL) algorithm (SwinUNETR) was used to train a prediction model for rapid segmentation of 9 key temporal bone structures in a data set of 325 clinical CTs. The data set was manually annotated by a specialist to serve as the ground truth. The data set was randomly split into training (n = 260) and testing (n = 65) sets. The model's performance was objectively assessed through external validation on the test set using metrics including Dice, Balanced accuracy, Hausdorff distances, and processing time. RESULTS The model achieved an average Dice coefficient of 0.87 for all structures, an average balanced accuracy of 0.94, an average Hausdorff distance of 0.79 mm, and an average processing time of 9.1 seconds per CT. CONCLUSION The present DL model for the automated simultaneous segmentation of multiple structures within the temporal bone from CTs achieved high accuracy according to currently commonly employed objective analysis. The results demonstrate the potential of the method to improve preoperative evaluation and intraoperative guidance in otologic surgery.
Collapse
Affiliation(s)
- Caio A Neves
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
- Faculty of Medicine, University of Brasilia UnB, Brasilia, Brazil
| | - Trishia El Chemaly
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
| | - Fanrui Fu
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
| | - Nikolas H Blevins
- Department of Otolaryngology-Head and Neck Surgery, Stanford University, Stanford, California, USA
| |
Collapse
|
2
|
Boubaker F, Puel U, Eliezer M, Hossu G, Assabah B, Haioun K, Blum A, Gondim-Teixeira PA, Parietti-Winkler C, Gillet R. Radiation dose reduction and image quality improvement with ultra-high resolution temporal bone CT using deep learning-based reconstruction: An anatomical study. Diagn Interv Imaging 2024:S2211-5684(24)00119-0. [PMID: 38744577 DOI: 10.1016/j.diii.2024.05.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2024] [Revised: 04/30/2024] [Accepted: 05/02/2024] [Indexed: 05/16/2024]
Abstract
PURPOSE The purpose of this study was to evaluate the achievable radiation dose reduction of an ultra-high resolution computed tomography (UHR-CT) scanner using deep learning reconstruction (DLR) while maintaining temporal bone image quality equal to or better than high-resolution CT (HR-CT). MATERIALS AND METHODS UHR-CT acquisitions were performed with variable tube voltages and currents at eight different dose levels (volumic CT dose index [CTDIvol] range: 4.6-79 mGy), 10242 matrix, and 0.25 mm slice thickness and reconstructed using DLR and hybrid iterative reconstruction (HIR) algorithms. HR-CT images were acquired using a standard protocol (120 kV/220 mAs; CTDI vol, 54.2 mGy, 5122 matrix, and 0.5 mm slice thickness). Two radiologists rated the image quality of seven structures using a five point confidence scale on six cadaveric temporal bone CTs. A global image quality score was obtained for each CT protocol by summing the image quality scores of all structures. RESULTS With DLR, UHR-CT at 120 kV/220 mAs (CTDIvol, 50.9 mGy) and 140 kV/220 mAs (CTDIvol, 79 mGy) received the highest global image quality scores (4.88 ± 0.32 [standard deviation (SD)] [range: 4-5] and 4.85 ± 0.35 [range: 4-5], respectively; P = 0.31), while HR-CT at 120 kV/220 mAs and UHR-CT at 120 kV/20 mAs received the lowest (i.e., 3.14 ± 0.75 [SD] [range: 2-5] and 2.97 ± 0.86 [SD] [range: 1-5], respectively; P = 0.14). All the DLR protocols had better image quality scores than HR-CT with HIR. CONCLUSION UHR-CT with DLR can be performed with up to a tenfold reduction in radiation dose compared to HR-CT with HIR while maintaining or improving image quality.
Collapse
Affiliation(s)
- Fatma Boubaker
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France
| | - Ulysse Puel
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Michael Eliezer
- Department of Radiology, Hôpital des 15-20, 75571 Paris, France
| | - Gabriela Hossu
- Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Bouchra Assabah
- Department of Anatomy, University Hospital Center of Nancy, 54000, Nancy, France
| | - Karim Haioun
- Canon Medical Systems Corporation, Kawasaki-shi, 212-0015 Kanagawa, Japan
| | - Alain Blum
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Pedro Augusto Gondim-Teixeira
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Cécile Parietti-Winkler
- ENT Surgery Department, Central Hospital, University Hospital Center of Nancy, 54000 Nancy, France
| | - Romain Gillet
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France.
| |
Collapse
|
3
|
Quatre R, Schmerber S, Attyé A. Improving rehabilitation of deaf patients by advanced imaging before cochlear implantation. J Neuroradiol 2024; 51:145-154. [PMID: 37806523 DOI: 10.1016/j.neurad.2023.10.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2023] [Revised: 10/05/2023] [Accepted: 10/05/2023] [Indexed: 10/10/2023]
Abstract
INTRODUCTION Cochlear implants have advanced the management of severe to profound deafness. However, there is a strong disparity in hearing performance after implantation from one patient to another. Moreover, there are several advanced kinds of imaging assessment before cochlear implantation. Microstructural white fiber degeneration can be studied with Diffusion weighted MRI (DWI) or tractography of the central auditory pathways. Functional MRI (fMRI) allows us to evaluate brain function, and CT or MRI segmentation to better detect inner ear anomalies. OBJECTIVE This literature review aims to evaluate how helpful pre-implantation anatomic imaging can be to predict hearing rehabilitation outcomes in deaf patients. These techniques include DWI and fMRI of the central auditory pathways, and automated labyrinth segmentation by CT scan, cone beam CT and MRI. DESIGN This systematic review was performed according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. Studies were selected by searching in PubMed and by checking the reference lists of relevant articles. Inclusion criteria were adults over 18, with unilateral or bilateral hearing loss, who had DWI acquisition or fMRI or CT/ Cone Beam CT/ MRI image segmentation. RESULTS After reviewing 172 articles, we finally included 51. Studies on DWI showed changes in the central auditory pathways affecting the white matter, extending to the primary and non-primary auditory cortices, even in sudden and mild hearing impairment. Hearing loss patients show a reorganization of brain activity in various areas, such as the auditory and visual cortices, as well as regions involved in language and emotions, according to fMRI studies. Deep Learning's automatic segmentation produces the best CT segmentation in just a few seconds. MRI segmentation is mainly used to evaluate fluid space of the inner ear and determine the presence of an endolymphatic hydrops. CONCLUSION Before cochlear implantation, a DWI with tractography can evaluate the central auditory pathways up to the primary and non-primary auditory cortices. This data is then used to generate predictions on the auditory rehabilitation of patients. A CT segmentation with systematic 3D reconstruction allow a better evaluation of cochlear malformations and predictable difficulties during surgery.
Collapse
Affiliation(s)
- Raphaële Quatre
- Department of Oto-Rhino-Laryngology, Head and Neck Surgery, University Hospital, Grenoble, France; BrainTech Lab INSERM UMR 2015, Grenoble, France; GeodAIsics, Grenoble, France.
| | - Sébastien Schmerber
- Department of Oto-Rhino-Laryngology, Head and Neck Surgery, University Hospital, Grenoble, France; BrainTech Lab INSERM UMR 2015, Grenoble, France
| | - Arnaud Attyé
- Department of Neuroradiology, University Hospital, Grenoble, France; GeodAIsics, Grenoble, France
| |
Collapse
|
4
|
Lee JW, Andersen SAW, Hittle B, Powell KA, Al-Fartoussi H, Banks L, Brannen Z, Lahchich M, Wiet GJ. Variability in Manual Segmentation of Temporal Bone Structures in Cone Beam CT Images. Otol Neurotol 2024; 45:e137-e141. [PMID: 38361290 DOI: 10.1097/mao.0000000000004119] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/17/2024]
Abstract
PURPOSE Manual segmentation of anatomical structures is the accepted "gold standard" for labeling structures in clinical images. However, the variability in manual segmentation of temporal bone structures in CBCT images of the temporal bone has not been systematically evaluated using multiple reviewers. Therefore, we evaluated the intravariability and intervariability of manual segmentation of inner ear structures in CBCT images of the temporal bone. METHODS Preoperative CBCTs scans of the inner ear were obtained from 10 patients who had undergone cochlear implant surgery. The cochlea, facial nerve, chorda tympani, mid-modiolar (MM) axis, and round window (RW) were manually segmented by five reviewers in two separate sessions that were at least 1 month apart. Interreviewer and intrareviewer variabilities were assessed using the Dice coefficient (DICE), volume similarity, mean Hausdorff Distance metrics, and visual review. RESULTS Manual segmentation of the cochlea was the most consistent within and across reviewers with a mean DICE of 0.91 (SD = 0.02) and 0.89 (SD = 0.01) respectively, followed by the facial nerve with a mean DICE of 0.83 (SD = 0.02) and 0.80 (SD = 0.03), respectively. The chorda tympani had the greatest amount of reviewer variability due to its thin size, and the location of the centroid of the RW and the MM axis were also quite variable between and within reviewers. CONCLUSIONS We observed significant variability in manual segmentation of some of the temporal bone structures across reviewers. This variability needs to be considered when interpreting the results in studies using one manual reviewer.
Collapse
Affiliation(s)
- Julian W Lee
- Ohio State University College of Medicine, Columbus, Ohio
| | - Steven Arild Wuyts Andersen
- Copenhagen Hearing and Balance Center, Department of Otorhinolaryngology, Head and Neck Surgery and Audiology, Rigshospitalet, Copenhagen, Denmark
| | - Bradley Hittle
- Department of Biomedical Informatics, Ohio State University Wexner Medical Center, Columbus, Ohio
| | - Kimerly A Powell
- Department of Biomedical Informatics, Ohio State University Wexner Medical Center, Columbus, Ohio
| | - Hagar Al-Fartoussi
- Copenhagen Hearing and Balance Center, Department of Otorhinolaryngology, Head and Neck Surgery and Audiology, Rigshospitalet, Copenhagen, Denmark
| | - Laura Banks
- Ohio State University College of Medicine, Columbus, Ohio
| | | | - Mariam Lahchich
- Copenhagen Hearing and Balance Center, Department of Otorhinolaryngology, Head and Neck Surgery and Audiology, Rigshospitalet, Copenhagen, Denmark
| | | |
Collapse
|
5
|
Stritzel J, Ebrahimzadeh AH, Büchner A, Lanfermann H, Marschollek M, Wolff D. Landmark-based registration of a cochlear model to a human cochlea using conventional CT scans. Sci Rep 2024; 14:1115. [PMID: 38212412 PMCID: PMC10784596 DOI: 10.1038/s41598-023-50632-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Accepted: 12/22/2023] [Indexed: 01/13/2024] Open
Abstract
Cochlear implants can provide an advanced treatment option to restore hearing. In standard pre-implant procedures, many factors are already considered, but it seems that not all underlying factors have been identified yet. One reason is the low quality of the conventional computed tomography images taken before implantation, making it difficult to assess these parameters. A novel method is presented that uses the Pietsch Model, a well-established model of the human cochlea, as well as landmark-based registration to address these challenges. Different landmark numbers and placements are investigated by visually comparing the mean error per landmark and the registrations' results. The landmarks on the first cochlear turn and the apex are difficult to discern on a low-resolution CT scan. It was possible to achieve a mean error markedly smaller than the image resolution while achieving a good visual fit on a cochlear segment and directly in the conventional computed tomography image. The employed cochlear model adjusts image resolution problems, while the effort of setting landmarks is markedly less than the segmentation of the whole cochlea. As a next step, the specific parameters of the patient could be extracted from the adapted model, which enables a more personalized implantation with a presumably better outcome.
Collapse
Affiliation(s)
- Jenny Stritzel
- Peter L. Reichertz Institute for Medical Informatics of TU Braunschweig and Hannover Medical School, Hannover, Germany.
| | - Amir Hossein Ebrahimzadeh
- Peter L. Reichertz Institute for Medical Informatics of TU Braunschweig and Hannover Medical School, Hannover, Germany
| | - Andreas Büchner
- German Hearing Center, Hannover Medical School, Hannover, Germany
- Department of Otorhinolaryngology, Hannover Medical School, Hannover, Germany
| | - Heinrich Lanfermann
- Institute of Diagnostic and Interventional Neuroradiology, Hannover Medical School, Hannover, Germany
| | - Michael Marschollek
- Peter L. Reichertz Institute for Medical Informatics of TU Braunschweig and Hannover Medical School, Hannover, Germany
| | - Dominik Wolff
- Peter L. Reichertz Institute for Medical Informatics of TU Braunschweig and Hannover Medical School, Hannover, Germany
| |
Collapse
|
6
|
Boubaker F, Teixeira PAG, Hossu G, Douis N, Gillet P, Blum A, Gillet R. In vivo depiction of cortical bone vascularization with ultra-high resolution-CT and deep learning algorithm reconstruction using osteoid osteoma as a model. Diagn Interv Imaging 2024; 105:26-32. [PMID: 37482455 DOI: 10.1016/j.diii.2023.07.001] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2023] [Revised: 06/24/2023] [Accepted: 07/12/2023] [Indexed: 07/25/2023]
Abstract
PURPOSE The purpose of this study was to evaluate the ability to depict in vivo bone vascularization using ultra-high-resolution (UHR) computed tomography (CT) with deep learning reconstruction (DLR) and hybrid iterative reconstruction algorithm, compared to simulated conventional CT, using osteoid osteoma as a model. MATERIALS AND METHODS Patients with histopathologically proven cortical osteoid osteoma who underwent UHR-CT between October 2019 and October 2022 were retrospectively included. Images were acquired with a 1024 × 1024 matrix and reconstructed with DLR and hybrid iterative reconstruction algorithm. To simulate conventional CT, images with a 512 × 512 matrix were also reconstructed. Two radiologists (R1, R2) independently evaluated the number of blood vessels entering the nidus and crossing the bone cortex, as well as vessel identification and image quality with a 5-point scale. Standard deviation (SD) of attenuation in the adjacent muscle and that of air were used as image noise and recorded. RESULTS Thirteen patients with 13 osteoid osteomas were included. There were 11 men and two women with a mean age of 21.8 ± 9.1 (SD) years. For both readers, UHR-CT with DLR depicted more nidus vessels (11.5 ± 4.3 [SD] (R1) and 11.9 ± 4.6 [SD] (R2)) and cortical vessels (4 ± 3.8 [SD] and 4.3 ± 4.1 [SD], respectively) than UHR-CT with hybrid iterative reconstruction (10.5 ± 4.3 [SD] and 10.4 ± 4.6 [SD], and 4.1 ± 3.8 [SD] and 4.3 ± 3.8 [SD], respectively) and simulated conventional CT (5.3 ± 2.2 [SD] and 6.4 ± 2.5 [SD], 2 ± 1.2 [SD] and 2.4 ± 1.6 [SD], respectively) (P < 0.05). UHR-CT with DLR provided less image noise than simulated conventional CT and UHR-CT with hybrid iterative reconstruction (P < 0.05). UHR-CT with DLR received the greatest score and simulated conventional CT the lowest score for vessel identification and image quality. CONCLUSION UHR-CT with DLR shows less noise than UHR-CT with hybrid iterative reconstruction and significantly improves cortical bone vascularization depiction compared to simulated conventional CT.
Collapse
Affiliation(s)
- Fatma Boubaker
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France
| | - Pedro Augusto Gondim Teixeira
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Gabriela Hossu
- Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Nicolas Douis
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France
| | - Pierre Gillet
- Université de Lorraine, CNRS, IMoPA, 54000, Nancy, France
| | - Alain Blum
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France
| | - Romain Gillet
- Guilloz Imaging Department, Central Hospital, University Hospital Center of Nancy, 54000, Nancy, France; Université de Lorraine, INSERM, IADI, 54000, Nancy, France; Université de Lorraine, CIC, Innovation Technologique, University Hospital Center of Nancy, 54000, Nancy, France.
| |
Collapse
|
7
|
Musleh A. Machine learning framework for simulation of artifacts in paranasal sinuses diagnosis using CT images. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:839-855. [PMID: 38393882 DOI: 10.3233/xst-230284] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/25/2024]
Abstract
In the medical field, diagnostic tools that make use of deep neural networks have reached a level of performance never before seen. A proper diagnosis of a patient's condition is crucial in modern medicine since it determines whether or not the patient will receive the care they need. Data from a sinus CT scan is uploaded to a computer and displayed on a high-definition monitor to give the surgeon a clear anatomical orientation before endoscopic sinus surgery. In this study, a unique method is presented for detecting and diagnosing paranasal sinus disorders using machine learning. The researchers behind the current study designed their own approach. To speed up diagnosis, one of the primary goals of our study is to create an algorithm that can accurately evaluate the paranasal sinuses in CT scans. The proposed technology makes it feasible to automatically cut down on the number of CT scan images that require investigators to manually search through them all. In addition, the approach offers an automatic segmentation that may be used to locate the paranasal sinus region and crop it accordingly. As a result, the suggested method dramatically reduces the amount of data that is necessary during the training phase. As a result, this results in an increase in the efficiency of the computer while retaining a high degree of performance accuracy. The suggested method not only successfully identifies sinus irregularities but also automatically executes the necessary segmentation without requiring any manual cropping. This eliminates the need for time-consuming and error-prone human labor. When tested with actual CT scans, the method in question was discovered to have an accuracy of 95.16 percent while retaining a sensitivity of 99.14 percent throughout.
Collapse
Affiliation(s)
- Abdullah Musleh
- Department of Surgery, King Khalid University, Abha, Saudi Arabia
| |
Collapse
|
8
|
Tang R, Li J, Zhao P, Zhang Z, Yin H, Ding H, Xu N, Yang Z, Wang Z. Utility of machine learning for identifying stapes fixation on ultra-high-resolution CT. Jpn J Radiol 2024; 42:69-77. [PMID: 37561264 DOI: 10.1007/s11604-023-01475-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Accepted: 07/18/2023] [Indexed: 08/11/2023]
Abstract
PURPOSE Imaging diagnosis of stapes fixation (SF) is challenging owing to a lack of definite evidence. We developed a comprehensive machine learning (ML) model to identify SF on ultra-high-resolution CT. MATERIALS AND METHODS We retrospectively enrolled 109 participants (143 ears) and divided them into the training set (115 ears) and test set (28 ears). Stapes mobility (SF or non-SF) was determined by surgical inspection. In the ML analysis, rectangular regions of interest were placed on consecutive axial slices in the training set. Radiomic features were extracted and fed into the training session. The test set was analyzed using 7 ML models (support vector machine, k nearest neighbor, decision tree, random forest, extra trees, eXtreme Gradient Boosting, and Light Gradient Boosting Machine) and by 2 dedicated neuroradiologists. Diagnostic performance (sensitivity, specificity and accuracy, with surgical findings as the reference) was compared between the radiologists and the optimal ML model by using the McNemar test. RESULTS The mean age of the participants was 42.3 ± 17.5 years. The Light Gradient Boosting Machine (LightGBM) model showed the highest sensitivity (0.83), specificity (0.81), accuracy (0.82) and area under the curve (0.88) for detecting SF among the 7 ML models. The neuroradiologists achieved good sensitivities (0.75 and 0.67), moderate-to-good specificities (0.63 and 0.56) and good accuracies (0.68 and 0.61). This model showed no statistical differences with the neuroradiologists (P values 0.289-1.000). CONCLUSIONS Compared to the neuroradiologists, the LightGBM model achieved competitive diagnostic performance in identifying SF, and has the potential to be a supportive tool in clinical practice.
Collapse
Affiliation(s)
- Ruowei Tang
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Jia Li
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Pengfei Zhao
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China.
| | - Zhengyu Zhang
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Hongxia Yin
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Heyu Ding
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Ning Xu
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Zhenghan Yang
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China
| | - Zhenchang Wang
- Department of Radiology, Beijing Friendship Hospital, Capital Medical University, 95 Yong'an Road, Xicheng District, Beijing, 100050, People's Republic of China.
| |
Collapse
|
9
|
Stebani J, Blaimer M, Zabler S, Neun T, Pelt DM, Rak K. Towards fully automated inner ear analysis with deep-learning-based joint segmentation and landmark detection framework. Sci Rep 2023; 13:19057. [PMID: 37925540 PMCID: PMC10625555 DOI: 10.1038/s41598-023-45466-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2022] [Accepted: 10/19/2023] [Indexed: 11/06/2023] Open
Abstract
Automated analysis of the inner ear anatomy in radiological data instead of time-consuming manual assessment is a worthwhile goal that could facilitate preoperative planning and clinical research. We propose a framework encompassing joint semantic segmentation of the inner ear and anatomical landmark detection of helicotrema, oval and round window. A fully automated pipeline with a single, dual-headed volumetric 3D U-Net was implemented, trained and evaluated using manually labeled in-house datasets from cadaveric specimen ([Formula: see text]) and clinical practice ([Formula: see text]). The model robustness was further evaluated on three independent open-source datasets ([Formula: see text] scans) consisting of cadaveric specimen scans. For the in-house datasets, Dice scores of [Formula: see text], intersection-over-union scores of [Formula: see text] and average Hausdorff distances of [Formula: see text] and [Formula: see text] voxel units were achieved. The landmark localization task was performed automatically with an average localization error of [Formula: see text] voxel units. A robust, albeit reduced performance could be attained for the catalogue of three open-source datasets. Results of the ablation studies with 43 mono-parametric variations of the basal architecture and training protocol provided task-optimal parameters for both categories. Ablation studies against single-task variants of the basal architecture showed a clear performance benefit of coupling landmark localization with segmentation and a dataset-dependent performance impact on segmentation ability.
Collapse
Affiliation(s)
- Jannik Stebani
- Magnetic Resonance and X-Ray Imaging Department, Fraunhofer Institute for Integrated Circuits IIS, 97074, Würzburg, Germany.
- Universität Würzburg, Experimentelle Physik V, 97074, Würzburg, Germany.
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, Universitätsklinikum Würzburg, 97080, Würzburg, Germany.
| | - Martin Blaimer
- Magnetic Resonance and X-Ray Imaging Department, Fraunhofer Institute for Integrated Circuits IIS, 97074, Würzburg, Germany
| | - Simon Zabler
- Magnetic Resonance and X-Ray Imaging Department, Fraunhofer Institute for Integrated Circuits IIS, 97074, Würzburg, Germany
- Faculty of Computer Science, Deggendorf Institute of Technology, Deggendorf, Germany
| | - Tilmann Neun
- Institute for Diagnostic and Interventional Neuroradiology, Universitätsklinikum Würzburg, 97080, Würzburg, Germany
| | - Daniël M Pelt
- Leiden Institute of Advanced Computer Science (LIACS), Universiteit Leiden, Leiden, CA, 2333, The Netherlands
| | - Kristen Rak
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, Universitätsklinikum Würzburg, 97080, Würzburg, Germany
| |
Collapse
|
10
|
Al-Dhamari I, Helal R, Abdelaziz T, Waldeck S, Paulus D. Automatic cochlear multimodal 3D image segmentation and analysis using atlas-model-based method. Cochlear Implants Int 2023:1-13. [PMID: 37922404 DOI: 10.1080/14670100.2023.2274199] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2023]
Abstract
Objectives: To propose an automated fast cochlear segmentation, length, and volume estimation method from clinical 3D multimodal images which has a potential role in the choice of cochlear implant type, surgery planning, and robotic surgeries. Methods: Two datasets from different countries were used. These datasets include 219 clinical 3D images of cochlea from 3 modalities: CT, CBCT, and MR. The datasets include different ages, genders, and types of cochlear implants. We propose an atlas-model-based method for cochlear segmentation and measurement based on high-resolution μCT model and A-value. The method was evaluated using 3D landmarks located by two experts. Results: The average error was 0.61 ± 0.22 mm and the average time required to process an image was 5.21 ± 0.93 seconds (P<0.001). The volume of the cochlea ranged from 73.96 mm3 to 106.97 mm3 , the cochlear length ranged from 36.69 to 45.91 mm at the lateral wall and from 29.12 to 39.05 mm at the organ of Corti. Discussion: We propose a method that produces nine different automated measurements of the cochlea: volume of scala tympani, volume of scala vestibuli, central lengths of the two scalae, the scala tympani lateral wall length, and the organ of Corti length in addition to three measurements related to A-value. Conclusion: This automatic cochlear image segmentation and analysis method can help clinician process multimodal cochlear images in approximately 5 seconds using a simple computer. The proposed method is publicly available for free download as an extension for 3D Slicer software.
Collapse
Affiliation(s)
- Ibraheem Al-Dhamari
- Medical Informatics Group, Berlin Institute of Health at Charité - Universitätsmedizin, Berlin, Germany
| | | | | | | | | |
Collapse
|
11
|
Petsiou DP, Martinos A, Spinos D. Applications of Artificial Intelligence in Temporal Bone Imaging: Advances and Future Challenges. Cureus 2023; 15:e44591. [PMID: 37795060 PMCID: PMC10545916 DOI: 10.7759/cureus.44591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/02/2023] [Indexed: 10/06/2023] Open
Abstract
The applications of artificial intelligence (AI) in temporal bone (TB) imaging have gained significant attention in recent years, revolutionizing the field of otolaryngology and radiology. Accurate interpretation of imaging features of TB conditions plays a crucial role in diagnosing and treating a range of ear-related pathologies, including middle and inner ear diseases, otosclerosis, and vestibular schwannomas. According to multiple clinical studies published in the literature, AI-powered algorithms have demonstrated exceptional proficiency in interpreting imaging findings, not only saving time for physicians but also enhancing diagnostic accuracy by reducing human error. Although several challenges remain in routinely relying on AI applications, the collaboration between AI and healthcare professionals holds the key to better patient outcomes and significantly improved patient care. This overview delivers a comprehensive update on the advances of AI in the field of TB imaging, summarizes recent evidence provided by clinical studies, and discusses future insights and challenges in the widespread integration of AI in clinical practice.
Collapse
Affiliation(s)
- Dioni-Pinelopi Petsiou
- Otolaryngology-Head and Neck Surgery, National and Kapodistrian University of Athens, School of Medicine, Athens, GRC
| | - Anastasios Martinos
- Otolaryngology-Head and Neck Surgery, National and Kapodistrian University of Athens, School of Medicine, Athens, GRC
| | - Dimitrios Spinos
- Otolaryngology-Head and Neck Surgery, Gloucestershire Hospitals NHS Foundation Trust, Gloucester, GBR
| |
Collapse
|
12
|
Curtis DP, Baumann AN, Jeyakumar A. Variation in cochlear size: A systematic review. Int J Pediatr Otorhinolaryngol 2023; 171:111659. [PMID: 37459768 DOI: 10.1016/j.ijporl.2023.111659] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Revised: 06/22/2023] [Accepted: 07/09/2023] [Indexed: 07/26/2023]
Abstract
BACKGROUND Advancements in imaging and implantation technology have invited reexamination of the classic teaching that the human cochlea maintains uniform size across demographics. Yet, studies yield conflicting results and relatively few broad systematic reviews have examined cochlear size variation. PURPOSE The purpose of this study is to quantify cochlear variability across eight different measurement categories and suggest normative values and ranges for each with consideration of disease state and gender where possible. METHODS A systematic search was conducted up to October 1, 2022, using the search terms "Cochlea/anatomy and histology"[Mesh]) AND 'size'" with filters "Humans" and "English" across three databases (PubMed, CINAHL, Medline). Further inclusion criteria involved reporting of numerical measurements in any of the eight included categories. RESULTS Of the 625 articles manually reviewed for relevance by title and abstract, 91 were selected for full-text review and 33 met all eligibility criteria. 5,791 cochleae were included and weighted means and ranges were calculated: "A" value (defined as the distance from the round window, through the modiolus, to the oppsite lateral wall) = 9.23 mm (8.43-10.4 mm, n = 2559); cochlear duct length (CDL) = 33.04 mm (range 28.2-36.4 mm, n = 2252); cochlear height = 5.14 mm (2.8-6.9 mm, n = 2098); the basal turn lumen diameter = 2.09 mm (1.7-2.2 mm, n = 617); "B" value (defined as perpendicular to "A" value and in the same plane) = 6.52 mm (5.73-6.9 mm, n = 908); width of the basal turn = 6.4 mm (6.22-6.86 mm, n = 356); height of the basal turn = 1.96 mm (1.77-2.56 mm, n = 204); length of the basal turn 21.87 mm (21.03-22.5 mm, n = 384). CONCLUSION A notable size range exists across the eight different cochlear parameters considered and we provide normative values for each measurement. Females tend to have smaller CDL and "A" value than males and the sensorineural hearing loss patients had smaller CDL and "A" value but larger cochlear height than the general population.
Collapse
Affiliation(s)
| | | | - Anita Jeyakumar
- Department of Otolaryngology, Mercy Bon Secours, Youngstown, OH, 44512, USA
| |
Collapse
|
13
|
Li X, Zhu Z, Yin H, Zhao P, Lv H, Tang R, Qin Y, Zhuo L, Wang Z. Labyrinth morphological modeling and its application on unreferenced segmentation assessment. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104891] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/30/2023]
|
14
|
Wong KKL, Xu W, Ayoub M, Fu YL, Xu H, Shi R, Zhang M, Su F, Huang Z, Chen W. Brain image segmentation of the corpus callosum by combining Bi-Directional Convolutional LSTM and U-Net using multi-slice CT and MRI. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 238:107602. [PMID: 37244234 DOI: 10.1016/j.cmpb.2023.107602] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/12/2022] [Revised: 05/12/2023] [Accepted: 05/14/2023] [Indexed: 05/29/2023]
Abstract
BACKGROUND AND OBJECTIVE Traditional disease diagnosis is usually performed by experienced physicians, but misdiagnosis or missed diagnosis still exists. Exploring the relationship between changes in the corpus callosum and multiple brain infarcts requires extracting corpus callosum features from brain image data, which requires addressing three key issues. (1) automation, (2) completeness, and (3) accuracy. Residual learning can facilitate network training, Bi-Directional Convolutional LSTM (BDC-LSTM) can exploit interlayer spatial dependencies, and HDC can expand the receptive domain without losing resolution. METHODS In this paper, we propose a segmentation method by combining BDC-LSTM and U-Net to segment the corpus callosum from multiple angles of brain images based on computed tomography (CT) and magnetic resonance imaging (MRI) in which two types of sequence, namely T2-weighted imaging as well as the Fluid Attenuated Inversion Recovery (Flair), were utilized. The two-dimensional slice sequences are segmented in the cross-sectional plane, and the segmentation results are combined to obtain the final results. Encoding, BDC- LSTM, and decoding include convolutional neural networks. The coding part uses asymmetric convolutional layers of different sizes and dilated convolutions to get multi-slice information and extend the convolutional layers' perceptual field. RESULTS This paper uses BDC-LSTM between the encoding and decoding parts of the algorithm. On the image segmentation of the brain in multiple cerebral infarcts dataset, accuracy rates of 0.876, 0.881, 0.887, and 0.912 were attained for the intersection of union (IOU), dice similarity coefficient (DS), sensitivity (SE), and predictive positivity value (PPV). The experimental findings demonstrate that the algorithm outperforms its rivals in accuracy. CONCLUSION This paper obtained segmentation results for three images using three models, ConvLSTM, Pyramid-LSTM, and BDC-LSTM, and compared them to verify that BDC-LSTM is the best method to perform the segmentation task for faster and more accurate detection of 3D medical images. We improve the convolutional neural network segmentation method to obtain medical images with high segmentation accuracy by solving the over-segmentation problem.
Collapse
Affiliation(s)
- Kelvin K L Wong
- School of Information and Electronics, Hunan City University, Yiyang 413000, China.
| | - Wanni Xu
- College of Technology and Engineering, National Taiwan Normal University, Taipei 106, Taiwan; Department of Computer Information Engineering, Nanchang Institute of Technology, Nanchang 330044, China.
| | - Muhammad Ayoub
- School of Computer Science and Engineering, Central South University, Changsha, Hunan 410083, China
| | - You-Lei Fu
- College of Technology and Engineering, National Taiwan Normal University, Taipei 106, Taiwan; Department of Computer Information Engineering, Nanchang Institute of Technology, Nanchang 330044, China
| | - Huasen Xu
- Department of Civil Engineering, Shanghai Normal University, Shanghai 201418, China
| | - Ruizheng Shi
- National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, Hunan 410008, China
| | - Mu Zhang
- Department of Emergency, Xiangya Hospital, Central South University, Changsha, Hunan 410008, China
| | - Feng Su
- Department of Emergency, Xiangya Hospital, Central South University, Changsha, Hunan 410008, China
| | - Zhiguo Huang
- Department of Emergency, Xiangya Hospital, Central South University, Changsha, Hunan 410008, China
| | - Weimin Chen
- School of Information and Electronics, Hunan City University, Yiyang 413000, China
| |
Collapse
|
15
|
Müller-Graff FT, Voelker J, Kurz A, Hagen R, Neun T, Rak K. Accuracy of radiological prediction of electrode position with otological planning software and implications of high-resolution imaging. Cochlear Implants Int 2023; 24:144-154. [PMID: 36617441 DOI: 10.1080/14670100.2022.2159128] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
Abstract
OBJECTIVES In cochlear implantation, preoperative prediction of electrode position has recently gained increasing attention. Currently, planning is usually done by multislice CT (MSCT). However, flat-panel volume CT (fpVCT) and its secondary reconstructions (fpVCTSECO) allow for more precise visualization of the cochlea. Combined with a newly developed otological planning software, the position of every single contact can be effectively predicted. In this study it was investigated how accurately radiological prediction forecasts the postoperative electrode localization and whether higher image resolution is advantageous. METHODS Utilizing otological planning software (OTOPLAN®) and different clinical imaging modalities (MSCT, fpVCT and fpVCTSECO) the electrode localization [angular insertion depth (AID)] and respective contact frequencies were predicted preoperatively and examined postoperatively. Furthermore, inter-electrode-distance (IED) and inter-electrode-frequency difference (IEFD) were evaluated postoperatively. RESULTS Measurements revealed a preoperative overestimation of AID. Corresponding frequencies were also miscalculated. Determination of IED and IEFD revealed discrepancies at the transition from the basal to the middle turn and round window to the basal turn. All predictions and discrepancies were lowest when using fpVCTSECO. CONCLUSION The postoperative electrode position can be predicted quite accurately using otological planning software. However, because of several potential misjudgments, high-resolution imaging, such as offered by fpVCTSECO, should be used pre- and postoperatively.
Collapse
Affiliation(s)
- Franz-Tassilo Müller-Graff
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University of Wuerzburg, Wuerzburg, Germany
| | - Johannes Voelker
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University of Wuerzburg, Wuerzburg, Germany
| | - Anja Kurz
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University of Wuerzburg, Wuerzburg, Germany
| | - Rudolf Hagen
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University of Wuerzburg, Wuerzburg, Germany
| | - Tilmann Neun
- Institute for Diagnostic and Interventional Neuroradiology, University of Wuerzburg, Wuerzburg, Germany
| | - Kristen Rak
- Department of Oto-Rhino-Laryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University of Wuerzburg, Wuerzburg, Germany
| |
Collapse
|
16
|
Li Z, Zhou L, Tan S, Tang A. Application of UNETR for automatic cochlear segmentation in temporal bone CTs. Auris Nasus Larynx 2023; 50:212-217. [PMID: 35970625 DOI: 10.1016/j.anl.2022.06.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2022] [Revised: 06/02/2022] [Accepted: 06/30/2022] [Indexed: 11/29/2022]
Abstract
OBJECTIVE To investigate the feasibility of a deep learning method based on a UNETR model for fully automatic segmentation of the cochlea in temporal bone CT images. METHODS The normal temporal bone CTs of 77 patients were used in 3D U-Net and UNETR model automatic cochlear segmentation. Tests were performed on two types of CT datasets and cochlear deformity datasets. RESULTS Through training the UNETR model, when batch_size=1, the Dice coefficient of the normal cochlear test set was 0.92, which was higher than that of the 3D U-Net model; on the GE 256 CT, SE-DS CT and Cochlear Deformity CT dataset tests, the Dice coefficients were 0.91, 0.93, 0 93, respectively. CONCLUSION According to the anatomical characteristics of the temporal bone, the use of the UNETR model can achieve fully automatic segmentation of the cochlea and obtain an accuracy close to manual segmentation. This method is feasible and has high accuracy.
Collapse
Affiliation(s)
- Zhenhua Li
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Guangxi Medical University, Nanning, Guangxi 530000, China
| | - Langtao Zhou
- School of Computer Science and Cyber Engineering, Guangzhou University, Guangzhou, China
| | - Songhua Tan
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Guangxi Medical University, Nanning, Guangxi 530000, China
| | - Anzhou Tang
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Guangxi Medical University, Nanning, Guangxi 530000, China.
| |
Collapse
|
17
|
Wu W, Xia W, Jun Z, Saghatchi S, Lavasani SN, Mohagheghi S, Ahmadian A, Gao X. Coordinate-based fast lightweight path search algorithm for electromagnetic navigation bronchoscopy. Med Biol Eng Comput 2023; 61:699-708. [PMID: 36585561 DOI: 10.1007/s11517-022-02740-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2022] [Accepted: 12/07/2022] [Indexed: 01/01/2023]
Abstract
Electromagnetic navigation bronchoscopy (ENB) uses electromagnetic positioning technology to guide the bronchoscope to accurately and quickly reach the lesion along the planned path. However, enormous data in high-resolution lung computed tomography (CT) and the complex structure of multilevel branching bronchial tree make fast path search challenging for path planning. We propose a coordinate-based fast lightweight path search (CPS) algorithm for ENB. First, the centerline is extracted from the bronchial tree by applying topological thinning. Then, Euclidean-distance-based coordinate search is applied. The centerline points are represented by their coordinates, and adjacent points along the navigation path are selected considering the shortest Euclidean distance to the target on the centerline nearest the lesion. From the top of the trachea centerline, search is repeated until reaching the target. In 50 high-resolution lung CT images acquired from five scanners, the CPS algorithm achieves accuracy, average search time, and average memory consumption of 100%, 88.5 ms, and 166.0 MB, respectively, reducing search time by 74.3% and 73.1% and memory consumption by 83.3% and 83.0% compared with Dijkstra and A* algorithms, respectively. CPS algorithm is suitable for path search in multilevel branching bronchial tree navigation based on high-resolution lung CT images.
Collapse
Affiliation(s)
- Wenbin Wu
- School of Biomedical Engineering (Suzhou), Division of Life Science and Medicine, University of Science and Technology of China, Hefei, 230026, China
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, No. 88, Keling Road, Suzhou New District, Suzhou, 215163, China
| | - Wei Xia
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, No. 88, Keling Road, Suzhou New District, Suzhou, 215163, China
- Jinan Guoke Medical Engineering and Technology Development Co., Ltd., Pharmaceutical Valley New Drug Creation Platform, Jinan, 250109, Shandong, China
| | - Zhong Jun
- Gaochun District, Nanjing Zhongao Jingzhong Medical Technology Co., LTD., No. 205, Shuanggao Road, Nanjing, 211300, China
| | - Samaneh Saghatchi
- Image Guided Surgery Lab, Research Centre of Biomedical Technology and Robotics, RCBTR, Tehran University of Medical Sciences, Tehran, 1416753955, Iran
| | - Saeedeh Navaei Lavasani
- Image Guided Surgery Lab, Research Centre of Biomedical Technology and Robotics, RCBTR, Tehran University of Medical Sciences, Tehran, 1416753955, Iran
| | - Saeed Mohagheghi
- Image Guided Surgery Lab, Research Centre of Biomedical Technology and Robotics, RCBTR, Tehran University of Medical Sciences, Tehran, 1416753955, Iran
| | - Alireza Ahmadian
- Department of Medical Physics & Biomedical Engineering & Research Centre for Biomedical Technology and Robotics, RCBTR, Tehran University of Medical Sciences, TUMS, Tehran, 1416753955, Iran
| | - Xin Gao
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, No. 88, Keling Road, Suzhou New District, Suzhou, 215163, China.
- Jinan Guoke Medical Engineering and Technology Development Co., Ltd., Pharmaceutical Valley New Drug Creation Platform, Jinan, 250109, Shandong, China.
| |
Collapse
|
18
|
Dieterich M, Hergenroeder T, Boegle R, Gerb J, Kierig E, Stöcklein S, Kirsch V. Endolymphatic space is age-dependent. J Neurol 2023; 270:71-81. [PMID: 36197569 PMCID: PMC9813103 DOI: 10.1007/s00415-022-11400-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2022] [Revised: 09/20/2022] [Accepted: 09/23/2022] [Indexed: 01/09/2023]
Abstract
Knowledge of the physiological endolymphatic space (ELS) is necessary to estimate endolymphatic hydrops (ELH) in patients with vestibulocochlear syndromes. Therefore, the current study investigated age-dependent changes in the ELS of participants with normal vestibulocochlear testing. Sixty-four ears of 32 participants with normal vestibulocochlear testing aged between 21 and 75 years (45.8 ± 17.2 years, 20 females, 30 right-handed, two left-handed) were examined by intravenous delayed gadolinium-enhanced magnetic resonance imaging of the inner ear (iMRI). Clinical diagnostics included neuro-otological assessment, video-oculography during caloric stimulation, and head-impulse test. iMRI data analysis provided semi-quantitative visual grading and automatic algorithmic quantitative segmentation of ELS volume (3D, mm3) using a deep learning-based segmentation of the inner ear's total fluid space (TFS) and volumetric local thresholding, as described earlier. As a result, following a 4-point ordinal scale, a mild ELH (grade 1) was found in 21/64 (32.8%) ears uni- or bilaterally in either cochlear, vestibulum, or both. Age and ELS were found to be positively correlated for the inner ear (r(64) = 0.33, p < 0.01), and vestibulum (r(64) = 0.25, p < 0.05). For the cochlea, the values correlated positively without reaching significance (r(64) = 0.21). In conclusion, age-dependent increases of the ELS should be considered when evaluating potential ELH in single subjects and statistical group comparisons.
Collapse
Affiliation(s)
- Marianne Dieterich
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany.,German Center for Vertigo and Balance Disorders-IFB, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany.,Graduate School of Systemic Neuroscience (GSN), Ludwig-Maximilians-Universität, Munich, Germany.,Munich Cluster for Systems Neurology (SyNergy), Munich, Germany
| | - Tatjana Hergenroeder
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
| | - Rainer Boegle
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany.,German Center for Vertigo and Balance Disorders-IFB, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany.,Graduate School of Systemic Neuroscience (GSN), Ludwig-Maximilians-Universität, Munich, Germany
| | - Johannes Gerb
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany.,German Center for Vertigo and Balance Disorders-IFB, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
| | - Emilie Kierig
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany.,German Center for Vertigo and Balance Disorders-IFB, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
| | - Sophia Stöcklein
- Department of Radiology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
| | - Valerie Kirsch
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany. .,German Center for Vertigo and Balance Disorders-IFB, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany. .,Graduate School of Systemic Neuroscience (GSN), Ludwig-Maximilians-Universität, Munich, Germany.
| |
Collapse
|
19
|
A Web-Based Automated Image Processing Research Platform for Cochlear Implantation-Related Studies. J Clin Med 2022; 11:jcm11226640. [PMID: 36431117 PMCID: PMC9699139 DOI: 10.3390/jcm11226640] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2022] [Revised: 10/27/2022] [Accepted: 10/28/2022] [Indexed: 11/11/2022] Open
Abstract
The robust delineation of the cochlea and its inner structures combined with the detection of the electrode of a cochlear implant within these structures is essential for envisaging a safer, more individualized, routine image-guided cochlear implant therapy. We present Nautilus-a web-based research platform for automated pre- and post-implantation cochlear analysis. Nautilus delineates cochlear structures from pre-operative clinical CT images by combining deep learning and Bayesian inference approaches. It enables the extraction of electrode locations from a post-operative CT image using convolutional neural networks and geometrical inference. By fusing pre- and post-operative images, Nautilus is able to provide a set of personalized pre- and post-operative metrics that can serve the exploration of clinically relevant questions in cochlear implantation therapy. In addition, Nautilus embeds a self-assessment module providing a confidence rating on the outputs of its pipeline. We present a detailed accuracy and robustness analyses of the tool on a carefully designed dataset. The results of these analyses provide legitimate grounds for envisaging the implementation of image-guided cochlear implant practices into routine clinical workflows.
Collapse
|
20
|
Lin Q, Wu HJ, Song QS, Tang YK. CT-based radiomics in predicting pathological response in non-small cell lung cancer patients receiving neoadjuvant immunotherapy. Front Oncol 2022; 12:937277. [PMID: 36267975 PMCID: PMC9577189 DOI: 10.3389/fonc.2022.937277] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2022] [Accepted: 08/30/2022] [Indexed: 11/13/2022] Open
Abstract
Objectives In radiomics, high-throughput algorithms extract objective quantitative features from medical images. In this study, we evaluated CT-based radiomics features, clinical features, in-depth learning features, and a combination of features for predicting a good pathological response (GPR) in non-small cell lung cancer (NSCLC) patients receiving immunotherapy-based neoadjuvant therapy (NAT). Materials and methods We reviewed 62 patients with NSCLC who received surgery after immunotherapy-based NAT and collected clinicopathological data and CT images before and after immunotherapy-based NAT. A series of image preprocessing was carried out on CT scanning images: tumor segmentation, conventional radiomics feature extraction, deep learning feature extraction, and normalization. Spearman correlation coefficient, principal component analysis (PCA), and least absolute shrinkage and selection operator (LASSO) were used to screen features. The pretreatment traditional radiomics combined with clinical characteristics (before_rad_cil) model and pretreatment deep learning characteristics (before_dl) model were constructed according to the data collected before treatment. The data collected after NAT created the after_rad_cil model and after_dl model. The entire model was jointly constructed by all clinical features, conventional radiomics features, and deep learning features before and after neoadjuvant treatment. Finally, according to the data obtained before and after treatment, the before_nomogram and after_nomogram were constructed. Results In the before_rad_cil model, four traditional radiomics features (“original_shape_flatness,” “wavelet hhl_firer_skewness,” “wavelet hlh_firer_skewness,” and “wavelet lll_glcm_correlation”) and two clinical features (“gender” and “N stage”) were screened out to predict a GPR. The average prediction accuracy (ACC) after modeling with k-nearest neighbor (KNN) was 0.707. In the after_rad_cil model, nine features predictive of GPR were obtained after feature screening, among which seven were traditional radiomics features: “exponential_firer_skewness,” “exponential_glrlm_runentropy,” “log- sigma-5-0-mm-3d_firer_kurtosis,” “logarithm_skewness,” “original_shape_elongation,” “original_shape_brilliance,” and “wavelet llh_glcm_clustershade”; two were clinical features: “after_CRP” and “after lymphocyte percentage.” The ACC after modeling with support vector machine (SVM) was 0.682. The before_dl model and after_dl model were modeled by SVM, and the ACC was 0.629 and 0.603, respectively. After feature screening, the entire model was constructed by multilayer perceptron (MLP), and the ACC of the GPR was the highest, 0.805. The calibration curve showed that the predictions of the GPR by the before_nomogram and after_nomogram were in consensus with the actual GPR. Conclusion CT-based radiomics has a good predictive ability for a GPR in NSCLC patients receiving immunotherapy-based NAT. Among the radiomics features combined with the clinicopathological information model, deep learning feature model, and the entire model, the entire model had the highest prediction accuracy.
Collapse
|
21
|
Xu J, Zeng B, Egger J, Wang C, Smedby Ö, Jiang X, Chen X. A review on AI-based medical image computing in head and neck surgery. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac840f] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2022] [Accepted: 07/25/2022] [Indexed: 11/11/2022]
Abstract
Abstract
Head and neck surgery is a fine surgical procedure with a complex anatomical space, difficult operation and high risk. Medical image computing (MIC) that enables accurate and reliable preoperative planning is often needed to reduce the operational difficulty of surgery and to improve patient survival. At present, artificial intelligence, especially deep learning, has become an intense focus of research in MIC. In this study, the application of deep learning-based MIC in head and neck surgery is reviewed. Relevant literature was retrieved on the Web of Science database from January 2015 to May 2022, and some papers were selected for review from mainstream journals and conferences, such as IEEE Transactions on Medical Imaging, Medical Image Analysis, Physics in Medicine and Biology, Medical Physics, MICCAI, etc. Among them, 65 references are on automatic segmentation, 15 references on automatic landmark detection, and eight references on automatic registration. In the elaboration of the review, first, an overview of deep learning in MIC is presented. Then, the application of deep learning methods is systematically summarized according to the clinical needs, and generalized into segmentation, landmark detection and registration of head and neck medical images. In segmentation, it is mainly focused on the automatic segmentation of high-risk organs, head and neck tumors, skull structure and teeth, including the analysis of their advantages, differences and shortcomings. In landmark detection, the focus is mainly on the introduction of landmark detection in cephalometric and craniomaxillofacial images, and the analysis of their advantages and disadvantages. In registration, deep learning networks for multimodal image registration of the head and neck are presented. Finally, their shortcomings and future development directions are systematically discussed. The study aims to serve as a reference and guidance for researchers, engineers or doctors engaged in medical image analysis of head and neck surgery.
Collapse
|
22
|
Minnema J, Ernst A, van Eijnatten M, Pauwels R, Forouzanfar T, Batenburg KJ, Wolff J. A review on the application of deep learning for CT reconstruction, bone segmentation and surgical planning in oral and maxillofacial surgery. Dentomaxillofac Radiol 2022; 51:20210437. [PMID: 35532946 PMCID: PMC9522976 DOI: 10.1259/dmfr.20210437] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Revised: 04/21/2022] [Accepted: 04/25/2022] [Indexed: 12/11/2022] Open
Abstract
Computer-assisted surgery (CAS) allows clinicians to personalize treatments and surgical interventions and has therefore become an increasingly popular treatment modality in maxillofacial surgery. The current maxillofacial CAS consists of three main steps: (1) CT image reconstruction, (2) bone segmentation, and (3) surgical planning. However, each of these three steps can introduce errors that can heavily affect the treatment outcome. As a consequence, tedious and time-consuming manual post-processing is often necessary to ensure that each step is performed adequately. One way to overcome this issue is by developing and implementing neural networks (NNs) within the maxillofacial CAS workflow. These learning algorithms can be trained to perform specific tasks without the need for explicitly defined rules. In recent years, an extremely large number of novel NN approaches have been proposed for a wide variety of applications, which makes it a difficult task to keep up with all relevant developments. This study therefore aimed to summarize and review all relevant NN approaches applied for CT image reconstruction, bone segmentation, and surgical planning. After full text screening, 76 publications were identified: 32 focusing on CT image reconstruction, 33 focusing on bone segmentation and 11 focusing on surgical planning. Generally, convolutional NNs were most widely used in the identified studies, although the multilayer perceptron was most commonly applied in surgical planning tasks. Moreover, the drawbacks of current approaches and promising research avenues are discussed.
Collapse
Affiliation(s)
- Jordi Minnema
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Anne Ernst
- Institute for Medical Systems Biology, University Hospital Hamburg-Eppendorf, Hamburg, Germany
| | - Maureen van Eijnatten
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Ruben Pauwels
- Aarhus Institute of Advanced Studies, Aarhus University, Aarhus, Denmark
| | - Tymour Forouzanfar
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Kees Joost Batenburg
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Jan Wolff
- Department of Dentistry and Oral Health, Aarhus University, Vennelyst Boulevard, Aarhus, Denmark
| |
Collapse
|
23
|
Deep Learning Approaches for Automatic Localization in Medical Images. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:6347307. [PMID: 35814554 PMCID: PMC9259335 DOI: 10.1155/2022/6347307] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/18/2022] [Accepted: 05/23/2022] [Indexed: 12/21/2022]
Abstract
Recent revolutionary advances in deep learning (DL) have fueled several breakthrough achievements in various complicated computer vision tasks. The remarkable successes and achievements started in 2012 when deep learning neural networks (DNNs) outperformed the shallow machine learning models on a number of significant benchmarks. Significant advances were made in computer vision by conducting very complex image interpretation tasks with outstanding accuracy. These achievements have shown great promise in a wide variety of fields, especially in medical image analysis by creating opportunities to diagnose and treat diseases earlier. In recent years, the application of the DNN for object localization has gained the attention of researchers due to its success over conventional methods, especially in object localization. As this has become a very broad and rapidly growing field, this study presents a short review of DNN implementation for medical images and validates its efficacy on benchmarks. This study presents the first review that focuses on object localization using the DNN in medical images. The key aim of this study was to summarize the recent studies based on the DNN for medical image localization and to highlight the research gaps that can provide worthwhile ideas to shape future research related to object localization tasks. It starts with an overview on the importance of medical image analysis and existing technology in this space. The discussion then proceeds to the dominant DNN utilized in the current literature. Finally, we conclude by discussing the challenges associated with the application of the DNN for medical image localization which can drive further studies in identifying potential future developments in the relevant field of study.
Collapse
|
24
|
Ahmadi SA, Frei J, Vivar G, Dieterich M, Kirsch V. IE-Vnet: Deep Learning-Based Segmentation of the Inner Ear's Total Fluid Space. Front Neurol 2022; 13:663200. [PMID: 35645963 PMCID: PMC9130477 DOI: 10.3389/fneur.2022.663200] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2021] [Accepted: 04/04/2022] [Indexed: 12/30/2022] Open
Abstract
Background In-vivo MR-based high-resolution volumetric quantification methods of the endolymphatic hydrops (ELH) are highly dependent on a reliable segmentation of the inner ear's total fluid space (TFS). This study aimed to develop a novel open-source inner ear TFS segmentation approach using a dedicated deep learning (DL) model. Methods The model was based on a V-Net architecture (IE-Vnet) and a multivariate (MR scans: T1, T2, FLAIR, SPACE) training dataset (D1, 179 consecutive patients with peripheral vestibulocochlear syndromes). Ground-truth TFS masks were generated in a semi-manual, atlas-assisted approach. IE-Vnet model segmentation performance, generalizability, and robustness to domain shift were evaluated on four heterogenous test datasets (D2-D5, n = 4 × 20 ears). Results The IE-Vnet model predicted TFS masks with consistently high congruence to the ground-truth in all test datasets (Dice overlap coefficient: 0.9 ± 0.02, Hausdorff maximum surface distance: 0.93 ± 0.71 mm, mean surface distance: 0.022 ± 0.005 mm) without significant difference concerning side (two-sided Wilcoxon signed-rank test, p>0.05), or dataset (Kruskal-Wallis test, p>0.05; post-hoc Mann-Whitney U, FDR-corrected, all p>0.2). Prediction took 0.2 s, and was 2,000 times faster than a state-of-the-art atlas-based segmentation method. Conclusion IE-Vnet TFS segmentation demonstrated high accuracy, robustness toward domain shift, and rapid prediction times. Its output works seamlessly with a previously published open-source pipeline for automatic ELS segmentation. IE-Vnet could serve as a core tool for high-volume trans-institutional studies of the inner ear. Code and pre-trained models are available free and open-source under https://github.com/pydsgz/IEVNet.
Collapse
Affiliation(s)
- Seyed-Ahmad Ahmadi
- German Center for Vertigo and Balance Disorders, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- NVIDIA GmbH, Munich, Germany
| | - Johann Frei
- IT-Infrastructure for Translational Medical Research, University of Augsburg, Augsburg, Germany
| | - Gerome Vivar
- German Center for Vertigo and Balance Disorders, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- Computer Aided Medical Procedures (CAMP), Technical University of Munich (TUM), Munich, Germany
| | - Marianne Dieterich
- German Center for Vertigo and Balance Disorders, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- Graduate School of Systemic Neuroscience (GSN), Ludwig-Maximilians-Universität, Munich, Germany
- Munich Cluster for Systems Neurology (SyNergy), Munich, Germany
| | - Valerie Kirsch
- German Center for Vertigo and Balance Disorders, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- Department of Neurology, University Hospital, Ludwig-Maximilians-Universität, Munich, Germany
- Graduate School of Systemic Neuroscience (GSN), Ludwig-Maximilians-Universität, Munich, Germany
| |
Collapse
|
25
|
Fan Y, Zhang D, Banalagay R, Wang J, Noble JH, Dawant BM. Hybrid active shape and deep learning method for the accurate and robust segmentation of the intracochlear anatomy in clinical head CT and CBCT images. J Med Imaging (Bellingham) 2021; 8:064002. [PMID: 34853805 DOI: 10.1117/1.jmi.8.6.064002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2020] [Accepted: 11/08/2021] [Indexed: 11/14/2022] Open
Abstract
Purpose: Robust and accurate segmentation methods for the intracochlear anatomy (ICA) are a critical step in the image-guided cochlear implant programming process. We have proposed an active shape model (ASM)-based method and a deep learning (DL)-based method for this task, and we have observed that the DL method tends to be more accurate than the ASM method while the ASM method tends to be more robust. Approach: We propose a DL-based U-Net-like architecture that incorporates ASM segmentation into the network. A quantitative analysis is performed on a dataset that consists of 11 cochlea specimens for which a segmentation ground truth is available. To qualitatively evaluate the robustness of the method, an experienced expert is asked to visually inspect and grade the segmentation results on a clinical dataset made of 138 image volumes acquired with conventional CT scanners and of 39 image volumes acquired with cone beam CT (CBCT) scanners. Finally, we compare training the network (1) first with the ASM results, and then fine-tuning it with the ground truth segmentation and (2) directly with the specimens with ground truth segmentation. Results: Quantitative and qualitative results show that the proposed method increases substantially the robustness of the DL method while having only a minor detrimental effect (though not significant) on its accuracy. Expert evaluation of the clinical dataset shows that by incorporating the ASM segmentation into the DL network, the proportion of good segmentation cases increases from 60/177 to 119/177 when training only with the specimens and increases from 129/177 to 151/177 when pretraining with the ASM results. Conclusions: A hybrid ASM and DL-based segmentation method is proposed to segment the ICA in CT and CBCT images. Our results show that combining DL and ASM methods leads to a solution that is both robust and accurate.
Collapse
Affiliation(s)
- Yubo Fan
- Vanderbilt University, Department of Computer Science, Nashville, Tennessee, United States
| | | | - Rueben Banalagay
- Vanderbilt University, Department of Electrical and Computer Engineering, Nashville, Tennessee, United States
| | - Jianing Wang
- Vanderbilt University, Department of Electrical and Computer Engineering, Nashville, Tennessee, United States
| | - Jack H Noble
- Vanderbilt University, Department of Electrical and Computer Engineering, Nashville, Tennessee, United States
| | - Benoit M Dawant
- Vanderbilt University, Department of Electrical and Computer Engineering, Nashville, Tennessee, United States
| |
Collapse
|
26
|
Eser MB, Atalay B, Dogan MB, Gündüz N, Kalcioglu MT. Measuring 3D Cochlear Duct Length on MRI: Is It Accurate and Reliable? AJNR Am J Neuroradiol 2021; 42:2016-2022. [PMID: 34593380 DOI: 10.3174/ajnr.a7287] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2021] [Accepted: 07/25/2021] [Indexed: 11/07/2022]
Abstract
BACKGROUND AND PURPOSE Prior studies have evaluated cochlear length using CT to select the most suitable cochlear implants and obtain patient-specific anatomy. This study aimed to test the accuracy and reliability of cochlear lateral wall length measurements using 3D MR imaging. MATERIALS AND METHODS Two observers measured the cochlear lateral wall length of 35 patients (21 men) with postlingual hearing loss using CT and MR imaging. The intraclass correlation coefficient (with 95% confidence intervals) was used to evaluate intraobserver and interobserver reliability for the 3D cochlear measurements. RESULTS The mean age of the participants was 39.85 (SD, 16.60) years. Observer 1 measured the mean lateral wall length as 41.52 (SD, 2.25) mm on CT and 41.44 (SD, 2.18) mm on MR imaging, with a mean difference of 0.08 mm (95% CI, -0.11 to 0.27 mm), while observer 2 measured the mean lateral wall length as 41.74 (SD, 2.69) mm on CT and 42.34 (SD, 2.53) mm on MR imaging, with a mean difference of -0.59 mm (95% CI, -1.00 to -0.20 mm). An intraclass correlation coefficient value of 0.90 (95% CI, 0.84-0.94) for CT and 0.69 (95% CI, 0.46-0.82) for MR imaging was obtained for the interobserver reliability for the full-turn cochlear lateral wall length. CONCLUSIONS CT-based 3D cochlear measurements show excellent intraobserver and interobserver reliability, while MR imaging-based lateral wall length measurements have good-to-excellent intraobserver reliability and moderate interobserver reliability. These results corroborate the use of CT for 3D cochlear measurements as a reference method and demonstrate MR imaging to be an alternative acquisition technique with comparably reliable results.
Collapse
Affiliation(s)
- M B Eser
- From the Departments of Radiology (M.B.E., B.A., M.B.D., N.G.)
| | - B Atalay
- From the Departments of Radiology (M.B.E., B.A., M.B.D., N.G.)
| | - M B Dogan
- From the Departments of Radiology (M.B.E., B.A., M.B.D., N.G.)
| | - N Gündüz
- From the Departments of Radiology (M.B.E., B.A., M.B.D., N.G.)
| | - M T Kalcioglu
- Otorhinolaryngology-Head and Neck Surgery (M.T.K.), Faculty of Medicine, Goztepe Prof. Dr. Suleyman Yalcin City Hospital, Istanbul, Turkey
| |
Collapse
|
27
|
Wasmann JWA, Lanting CP, Huinck WJ, Mylanus EAM, van der Laak JWM, Govaerts PJ, Swanepoel DW, Moore DR, Barbour DL. Computational Audiology: New Approaches to Advance Hearing Health Care in the Digital Age. Ear Hear 2021; 42:1499-1507. [PMID: 33675587 PMCID: PMC8417156 DOI: 10.1097/aud.0000000000001041] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
The global digital transformation enables computational audiology for advanced clinical applications that can reduce the global burden of hearing loss. In this article, we describe emerging hearing-related artificial intelligence applications and argue for their potential to improve access, precision, and efficiency of hearing health care services. Also, we raise awareness of risks that must be addressed to enable a safe digital transformation in audiology. We envision a future where computational audiology is implemented via interoperable systems using shared data and where health care providers adopt expanded roles within a network of distributed expertise. This effort should take place in a health care system where privacy, responsibility of each stakeholder, and patients' safety and autonomy are all guarded by design.
Collapse
Affiliation(s)
- Jan-Willem A Wasmann
- Department of Otorhinolaryngology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center Nijmegen, the Netherlands
| | - Cris P Lanting
- Department of Otorhinolaryngology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center Nijmegen, the Netherlands
| | - Wendy J Huinck
- Department of Otorhinolaryngology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center Nijmegen, the Netherlands
| | - Emmanuel A M Mylanus
- Department of Otorhinolaryngology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center Nijmegen, the Netherlands
| | - Jeroen W M van der Laak
- Department of Pathology, Radboud University Medical Center Nijmegen, the Netherlands
- Center for Medical Image Science and Visualization, Linköping University, Sweden
| | | | - De Wet Swanepoel
- Department of Speech-Language Pathology and Audiology, University of Pretoria, South Africa
| | - David R Moore
- Communication Sciences Research Center, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio, USA
- Department of Otolaryngology, University of Cincinnati, Cincinnati, Ohio, USA
- Manchester Centre for Audiology and Deafness, University of Manchester, Manchester, United Kingdom
| | - Dennis L Barbour
- Department of Biomedical Engineering. Washington University in St. Louis, St. Louis, Missouri, USA
| |
Collapse
|
28
|
Wang Z, Demarcy T, Vandersteen C, Gnansia D, Raffaelli C, Guevara N, Delingette H. Bayesian logistic shape model inference: Application to cochlear image segmentation. Med Image Anal 2021; 75:102268. [PMID: 34710654 DOI: 10.1016/j.media.2021.102268] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2021] [Revised: 09/01/2021] [Accepted: 10/08/2021] [Indexed: 11/28/2022]
Abstract
Incorporating shape information is essential for the delineation of many organs and anatomical structures in medical images. While previous work has mainly focused on parametric spatial transformations applied to reference template shapes, in this paper, we address the Bayesian inference of parametric shape models for segmenting medical images with the objective of providing interpretable results. The proposed framework defines a likelihood appearance probability and a prior label probability based on a generic shape function through a logistic function. A reference length parameter defined in the sigmoid controls the trade-off between shape and appearance information. The inference of shape parameters is performed within an Expectation-Maximisation approach in which a Gauss-Newton optimization stage provides an approximation of the posterior probability of the shape parameters. This framework is applied to the segmentation of cochlear structures from clinical CT images constrained by a 10-parameter shape model. It is evaluated on three different datasets, one of which includes more than 200 patient images. The results show performances comparable to supervised methods and better than previously proposed unsupervised ones. It also enables an analysis of parameter distributions and the quantification of segmentation uncertainty, including the effect of the shape model.
Collapse
Affiliation(s)
- Zihao Wang
- Inria, Epione Team, Université Côte d'Azur, Sophia Antipolis, France.
| | - Thomas Demarcy
- Oticon Medical, 14 Chemin de Saint-Bernard Porte, Vallauris 06220, France
| | - Clair Vandersteen
- Inria, Epione Team, Université Côte d'Azur, Sophia Antipolis, France; Head and Neck University Institute, Nice University Hospital, 31 Avenue de Valombrose, Nice 06100, France
| | - Dan Gnansia
- Oticon Medical, 14 Chemin de Saint-Bernard Porte, Vallauris 06220, France
| | - Charles Raffaelli
- Department of Radiology, Centre Hospitalier Universitaire de Nice, 31 Avenue de Valombrose, Nice 06100, France
| | - Nicolas Guevara
- Head and Neck University Institute, Nice University Hospital, 31 Avenue de Valombrose, Nice 06100, France
| | - Hervé Delingette
- Inria, Epione Team, Université Côte d'Azur, Sophia Antipolis, France
| |
Collapse
|
29
|
Chawdhary G, Shoman N. Emerging artificial intelligence applications in otological imaging. Curr Opin Otolaryngol Head Neck Surg 2021; 29:357-364. [PMID: 34459798 DOI: 10.1097/moo.0000000000000754] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
PURPOSE OF REVIEW To highlight the recent literature on artificial intelligence (AI) pertaining to otological imaging and to discuss future directions, obstacles and opportunities. RECENT FINDINGS The main themes in the recent literature centre around automated otoscopic image diagnosis and automated image segmentation for application in virtual reality surgical simulation and planning. Other applications that have been studied include identification of tinnitus MRI biomarkers, facial palsy analysis, intraoperative augmented reality systems, vertigo diagnosis and endolymphatic hydrops ratio calculation in Meniere's disease. Studies are presently at a preclinical, proof-of-concept stage. SUMMARY The recent literature on AI in otological imaging is promising and demonstrates the future potential of this technology in automating certain imaging tasks in a healthcare environment of ever-increasing demand and workload. Some studies have shown equivalence or superiority of the algorithm over physicians, albeit in narrowly defined realms. Future challenges in developing this technology include the compilation of large high quality annotated datasets, fostering strong collaborations between the health and technology sectors, testing the technology within real-world clinical pathways and bolstering trust among patients and physicians in this new method of delivering healthcare.
Collapse
Affiliation(s)
- Gaurav Chawdhary
- ENT Department, Royal Hallamshire Hospital, Broomhall, Sheffield, UK
| | - Nael Shoman
- ENT Department, Queen Elizabeth II Health Sciences Centre, Halifax, Nova Scotia, Canada
| |
Collapse
|
30
|
Neves CA, Tran ED, Cooperman SP, Blevins NH. Fully Automated Measurement of Cochlear Duct Length From Clinical Temporal Bone Computed Tomography. Laryngoscope 2021; 132:449-458. [PMID: 34536238 DOI: 10.1002/lary.29869] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2021] [Revised: 08/24/2021] [Accepted: 08/31/2021] [Indexed: 11/06/2022]
Abstract
OBJECTIVES/HYPOTHESIS To present and validate a novel fully automated method to measure cochlear dimensions, including cochlear duct length (CDL). STUDY DESIGN Cross-sectional study. METHODS The computational method combined 1) a deep learning (DL) algorithm to segment the cochlea and otic capsule and 2) geometric analysis to measure anti-modiolar distances from the round window to the apex. The algorithm was trained using 165 manually segmented clinical computed tomography (CT). A Testing group of 159 CTs were then measured for cochlear diameter and width (A- and B-values) and CDL using the automated system and compared against manual measurements. The results were also compared with existing approaches and historical data. In addition, pre- and post-implantation scans from 27 cochlear implant recipients were studied to compare predicted versus actual array insertion depth. RESULTS Measurements were successfully obtained in 98.1% of scans. The mean CDL to 900° was 35.52 mm (SD, 2.06; range, [30.91-40.50]), the mean A-value was 8.88 mm (0.47; [7.67-10.49]), and mean B-value was 6.38 mm (0.42; [5.16-7.38]). The R2 fit of the automated to manual measurements was 0.87 for A-value, 0.70 for B-value, and 0.71 for CDL. For anti-modiolar arrays, the distance between the imaged and predicted array tip location was 0.57 mm (1.25; [0.13-5.28]). CONCLUSION Our method provides a fully automated means of cochlear analysis from clinical CTs. The distribution of CDL, dimensions, and cochlear quadrant lengths is similar to those from historical data. This approach requires no radiographic experience and is free from user-related variation. LEVEL OF EVIDENCE 3 Laryngoscope, 2021.
Collapse
Affiliation(s)
- Caio A Neves
- Faculty of Medicine, University of Brasilia, Brasilia, Brazil
| | - Emma D Tran
- Department of Otolaryngology-Head & Neck Surgery, Stanford University School of Medicine, Stanford, California, U.S.A
| | - Shayna P Cooperman
- Department of Otolaryngology-Head & Neck Surgery, Stanford University School of Medicine, Stanford, California, U.S.A
| | - Nikolas H Blevins
- Department of Otolaryngology-Head & Neck Surgery, Stanford University School of Medicine, Stanford, California, U.S.A
| |
Collapse
|
31
|
Li S, Zheng J, Li D. Precise segmentation of non-enhanced computed tomography in patients with ischemic stroke based on multi-scale U-Net deep network model. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2021; 208:106278. [PMID: 34274610 DOI: 10.1016/j.cmpb.2021.106278] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/27/2020] [Accepted: 07/04/2021] [Indexed: 06/13/2023]
Abstract
BACKGROUND AND OBJECTIVE Acute ischemic stroke requires timely diagnosis and thrombolytic therapy, but it is difficult to locate and quantify the lesion site manually. The purpose of this study was to explore a more rapid and effective method for automatic image segmentation of acute ischemic stroke. METHODS The image features of 30 stroke patients were segmented from non-enhanced computed tomography (CT) images using a multi-scale U-Net deep network model. The Dice loss function training model was used to counter the similar imbalance problem in the data. The difference was compared between manual segmentation and automatic segmentation. RESULTS The Dice similarity coefficient based on multi-scale convolution U-Net network segmentation was 0.86±0.04, higher than the Dice based on classic U-Net (0.81±0.07, P=0.001). The lesion contour of automatic segmentation based on multi-scale U-Net was very close to manual segmentation. The error of lesion area is 1.28±0.59 mm2, and the Pearson correlation coefficient was r=0.986 (P<0.01). The motion time of automatic segmentation is less than 20 ms. CONCLUSIONS Multi-scale U-Net deep network model can effectively segment ischemic stroke lesions in non-enhanced CT and meet real-time clinical requirements.
Collapse
Affiliation(s)
- Shaoquan Li
- Department of Neurosurgery, Cangzhou Central Hospital, Hebei 061000, China.
| | - Jianye Zheng
- Department of Neurosurgery, Cangzhou Central Hospital, Hebei 061000, China
| | - Dongjiao Li
- Department of Neurosurgery, Cangzhou Central Hospital, Hebei 061000, China
| |
Collapse
|
32
|
Atalay B, Eser MB, Kalcioglu MT, Ankarali H. Comprehensive Analysis of Factors Affecting Cochlear Size: A Systematic Review and Meta-analysis. Laryngoscope 2021; 132:188-197. [PMID: 33764541 DOI: 10.1002/lary.29532] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2020] [Revised: 03/09/2021] [Accepted: 03/16/2021] [Indexed: 12/14/2022]
Abstract
OBJECTIVES To determine the cochlea's average size in humans and evaluate the relationships between certain covariates and cochlear size. METHODS A systematic search on articles on cochlear size and published in English was conducted using Cochrane, PubMed, Web of Science, and Scopus databases up to September 15, 2020. Data were pooled using random-effects with three models. The effect of demographic, clinical, and measurement-related parameters was specifically analyzed. Meta-regression and subgroup analyses were conducted. The overall effect estimation was made for outcomes. RESULTS The meta-analysis included 4,708 cochleae from 56 studies. The overall length of the organ of Corti was 32.94 mm (95% confidence interval [CI]: 32.51-33.38). The first and second models revealed that age, gender, country, continent, measurement method (direct, indirect), measured structure ("A" value, cochlear lateral wall), origin (in vivo, in vitro), and type (histology specimens, plastic casts, imaging) of the cochlear material did not affect the cochlear size. However, study populations (general population, patients with a cochlear implant, and patients with congenital sensorineural hearing loss [CSNHL]) were found to affect the outcomes. Compared to the other populations, patients with CSNHL had shorter cochleae. Therefore, we developed a third model and found that the general population and patients with cochlear implants did not differ in cochlear size. CONCLUSION This meta-analysis investigated the factors that could affect the cochlear size and found that patients with CSNHL had significantly shorter cochleae, whereas other covariates had no significant effect. Laryngoscope, 2021.
Collapse
Affiliation(s)
- Basak Atalay
- Faculty of Medicine, Department of Radiology, Goztepe Training and Research Hospital, Istanbul Medeniyet University, Istanbul, Turkey
| | - Mehmet Bilgin Eser
- Faculty of Medicine, Department of Radiology, Goztepe Training and Research Hospital, Istanbul Medeniyet University, Istanbul, Turkey
| | - Mahmut Tayyar Kalcioglu
- Faculty of Medicine, Department of Otorhinolaryngology-Head and Neck Surgery, Goztepe Training and Research Hospital, Istanbul Medeniyet University, Istanbul, Turkey
| | - Handan Ankarali
- Faculty of Medicine, Department of Biostatistics and Medical Informatics, Istanbul Medeniyet University, Istanbul, Turkey
| |
Collapse
|
33
|
Mlynski R, Lüsebrink A, Oberhoffner T, Langner S, Weiss NM. Mapping Cochlear Duct Length to Electrically Evoked Compound Action Potentials in Cochlear Implantation. Otol Neurotol 2021; 42:e254-e260. [PMID: 33273309 DOI: 10.1097/mao.0000000000002957] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
Abstract
OBJECTIVE Objective measurements may assist in indicating cochlear implants and in predicting outcomes of cochlear implantation surgery. Using electrically evoked compound action potentials (ECAP), information about the function of the auditory nerve can be obtained by analyzing responses to electrical stimulation transmitted and derived by the recording electrode. The aim of this study was to determine whether ECAP characteristics differ depending on the stimulated intracochlear region and the size of the cochlea. STUDY DESIGN Retrospective cohort study. SETTING University Medical center, tertiary academic referral center. PATIENTS Patients undergoing cochlear implant surgery between 2015 and 2018. INTERVENTION Cochlear implantation with FLEXsoft electrode arrays (length 31.5 mm, 12 stimulating channels). MAIN OUTCOME MEASURES The cochlear duct length (CDL) and the cochlear coverage (CC) were measured using a new computed tomography-based software and correlated to the postoperative speech performance. Additionally, ECAP were measured and associated to the CDL. RESULTS A total of 59 ears of 53 cochlear implant users with a mean age of 63.6 (SD 14.9) years were included. The mean estimated CDL was 35.0 (SD 2.2) mm. The mean CC was 90.3% (SD 5.5%). A total of 4,873 ECAP were measured. A statistically significant, moderate, negative correlation between the ECAP slope and the site of stimulation was found (r = -0.29, 95% confidence interval: -0.32 to -0.27, p < 0.0001). No correlation between the CC and the speech performance was found (r = -0.08, 95% confidence interval: -0.33 to 0.18 p = 0.52). CONCLUSION ECAP slopes seem to be a reliable tool to identify the electrode's position inside the cochlea and also showed correlations to the anatomy of the patient. A combination of objective measurements such as anatomical parameters and ECAPs are helpful to assist the postoperative fitting and are promising tools to improve patient care.
Collapse
Affiliation(s)
- Robert Mlynski
- Department of Otorhinolaryngology, Head and Neck Surgery "Otto Körner"
| | - Adele Lüsebrink
- Department of Otorhinolaryngology, Head and Neck Surgery "Otto Körner"
| | | | - Soenke Langner
- Department of Radiology, Rostock University Medical Center, Rostock, Germany
| | - Nora M Weiss
- Department of Otorhinolaryngology, Head and Neck Surgery "Otto Körner"
| |
Collapse
|
34
|
Practicable assessment of cochlear size and shape from clinical CT images. Sci Rep 2021; 11:3448. [PMID: 33568727 PMCID: PMC7876007 DOI: 10.1038/s41598-021-83059-6] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2020] [Accepted: 01/25/2021] [Indexed: 11/08/2022] Open
Abstract
There is considerable interpersonal variation in the size and shape of the human cochlea, with evident consequences for cochlear implantation. The ability to characterize a specific cochlea, from preoperative computed tomography (CT) images, would allow the clinician to personalize the choice of electrode, surgical approach and postoperative programming. In this study, we present a fast, practicable and freely available method for estimating cochlear size and shape from clinical CT. The approach taken is to fit a template surface to the CT data, using either a statistical shape model or a locally affine deformation (LAD). After fitting, we measure cochlear size, duct length and a novel measure of basal turn non-planarity, which we suggest might correlate with the risk of insertion trauma. Gold-standard measurements from a convenience sample of 18 micro-CT scans are compared with the same quantities estimated from low-resolution, noisy, pseudo-clinical data synthesized from the same micro-CT scans. The best results were obtained using the LAD method, with an expected error of 8-17% of the gold-standard sample range for non-planarity, cochlear size and duct length.
Collapse
|
35
|
Lv Y, Ke J, Xu Y, Shen Y, Wang J, Wang J. Automatic segmentation of temporal bone structures from clinical conventional CT using a CNN approach. Int J Med Robot 2021; 17:e2229. [PMID: 33462998 DOI: 10.1002/rcs.2229] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2020] [Revised: 01/13/2021] [Accepted: 01/14/2021] [Indexed: 12/15/2022]
Abstract
BACKGROUND Automatic segmentation of temporal bone structures from patients' conventional computed tomography (CT) data plays an important role in the image-guided cochlear implant surgery. Existing convolutional neural network approaches have difficulties in segmenting such small tubular structures. METHODS We propose a light-weight three-dimensional convolutional neural network referred to as W-Net to achieve multiobjective segmentation of temporal bone structures including the cochlear labyrinth, ossicular chain and facial nerve from conventional temporal bone CT images. Data augmentation with morphological enhancement is proposed to increase the segmentation accuracy of small tubular structures. Evaluation against the state-of-the-art methods is performed. RESULTS Our method achieved mean Dice similarity coefficients (DSCs) of 0.90, 0.85 and 0.77 for the cochlear labyrinth, ossicular chain and facial nerve, respectively. These results were also close to the DSCs between human expert annotators (0.91, 0.91 and 0.72). CONCLUSIONS Our method achieves human-level accuracy in the segmentation of the cochlear labyrinth, ossicular chain and facial nerve.
Collapse
Affiliation(s)
- Yi Lv
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Jia Ke
- Department of Otorhinolaryngology-Head and Neck Surgery, Peking University Third Hospital, Beijing, China
| | - Ying Xu
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Yu Shen
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Junchen Wang
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China.,Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, China
| | - Jiang Wang
- Department of Otorhinolaryngology-Head and Neck Surgery, Peking University Third Hospital, Beijing, China
| |
Collapse
|
36
|
Deep learning for the fully automated segmentation of the inner ear on MRI. Sci Rep 2021; 11:2885. [PMID: 33536451 PMCID: PMC7858625 DOI: 10.1038/s41598-021-82289-y] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2020] [Accepted: 01/19/2021] [Indexed: 12/15/2022] Open
Abstract
Segmentation of anatomical structures is valuable in a variety of tasks, including 3D visualization, surgical planning, and quantitative image analysis. Manual segmentation is time-consuming and deals with intra and inter-observer variability. To develop a deep-learning approach for the fully automated segmentation of the inner ear in MRI, a 3D U-net was trained on 944 MRI scans with manually segmented inner ears as reference standard. The model was validated on an independent, multicentric dataset consisting of 177 MRI scans from three different centers. The model was also evaluated on a clinical validation set containing eight MRI scans with severe changes in the morphology of the labyrinth. The 3D U-net model showed precise Dice Similarity Coefficient scores (mean DSC-0.8790) with a high True Positive Rate (91.5%) and low False Discovery Rate and False Negative Rates (14.8% and 8.49% respectively) across images from three different centers. The model proved to perform well with a DSC of 0.8768 on the clinical validation dataset. The proposed auto-segmentation model is equivalent to human readers and is a reliable, consistent, and efficient method for inner ear segmentation, which can be used in a variety of clinical applications such as surgical planning and quantitative image analysis.
Collapse
|
37
|
Fully automated preoperative segmentation of temporal bone structures from clinical CT scans. Sci Rep 2021; 11:116. [PMID: 33420386 PMCID: PMC7794235 DOI: 10.1038/s41598-020-80619-0] [Citation(s) in RCA: 31] [Impact Index Per Article: 10.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2020] [Accepted: 12/23/2020] [Indexed: 11/11/2022] Open
Abstract
Middle- and inner-ear surgery is a vital treatment option in hearing loss, infections, and tumors of the lateral skull base. Segmentation of otologic structures from computed tomography (CT) has many potential applications for improving surgical planning but can be an arduous and time-consuming task. We propose an end-to-end solution for the automated segmentation of temporal bone CT using convolutional neural networks (CNN). Using 150 manually segmented CT scans, a comparison of 3 CNN models (AH-Net, U-Net, ResNet) was conducted to compare Dice coefficient, Hausdorff distance, and speed of segmentation of the inner ear, ossicles, facial nerve and sigmoid sinus. Using AH-Net, the Dice coefficient was 0.91 for the inner ear; 0.85 for the ossicles; 0.75 for the facial nerve; and 0.86 for the sigmoid sinus. The average Hausdorff distance was 0.25, 0.21, 0.24 and 0.45 mm, respectively. Blinded experts assessed the accuracy of both techniques, and there was no statistical difference between the ratings for the two methods (p = 0.93). Objective and subjective assessment confirm good correlation between automated segmentation of otologic structures and manual segmentation performed by a specialist. This end-to-end automated segmentation pipeline can help to advance the systematic application of augmented reality, simulation, and automation in otologic procedures.
Collapse
|
38
|
Effect of Filtered Back-Projection Filters to Low-Contrast Object Imaging in Ultra-High-Resolution (UHR) Cone-Beam Computed Tomography (CBCT). SENSORS 2020; 20:s20226416. [PMID: 33182640 PMCID: PMC7697695 DOI: 10.3390/s20226416] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/22/2020] [Revised: 10/23/2020] [Accepted: 11/06/2020] [Indexed: 01/09/2023]
Abstract
In this study, the effect of filter schemes on several low-contrast materials was compared using standard and ultra-high-resolution (UHR) cone-beam computed tomography (CBCT) imaging. The performance of the UHR-CBCT was quantified by measuring the modulation transfer function (MTF) and the noise power spectrum (NPS). The MTF was measured at the radial location around the cylindrical phantom, whereas the NPS was measured in the eight different homogeneous regions of interest. Six different filter schemes were designed and implemented in the CT sinogram from each imaging configuration. The experimental results indicated that the filter with smaller smoothing window preserved the MTF up to the highest spatial frequency, but larger NPS. In addition, the UHR imaging protocol provided 1.77 times better spatial resolution than the standard acquisition by comparing the specific spatial frequency (f50) under the same conditions. The f50s with the flat-top window in UHR mode was 1.86, 0.94, 2.52, 2.05, and 1.86 lp/mm for Polyethylene (Material 1, M1), Polystyrene (M2), Nylon (M3), Acrylic (M4), and Polycarbonate (M5), respectively. The smoothing window in the UHR protocol showed a clearer performance in the MTF according to the low-contrast objects, showing agreement with the relative contrast of materials in order of M3, M4, M1, M5, and M2. In conclusion, although the UHR-CBCT showed the disadvantages of acquisition time and radiation dose, it could provide greater spatial resolution with smaller noise property compared to standard imaging; moreover, the optimal window function should be considered in advance for the best UHR performance.
Collapse
|