1
|
Divya R, Shantha Selva Kumari R. Multi-instance learning attention model for amyloid quantification of brain sub regions in longitudinal cognitive decline. Brain Res 2024; 1842:149103. [PMID: 38955250 DOI: 10.1016/j.brainres.2024.149103] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2023] [Revised: 05/21/2024] [Accepted: 06/26/2024] [Indexed: 07/04/2024]
Abstract
Amyloid PET scans help in identifying the beta-amyloid deposition in different brain regions. The purpose of this study is to develop a deep learning model that can automate the task of finding amyloid deposition in different regions of the brain only by using PET scan and without the corresponding MRI scan. 2647 18F-Florbetapir PET scans are collected from Alzheimer's Disease Neuroimaging Initiative (ADNI) from multiple centres taken over a period. A deep learning model based on multi-instance learning and attention is proposed which is trained and validated using 80% of the scans and the remaining 20% of the scans are used for testing the model. The performance of the model is validated using Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). The proposed model is further tested upon an external dataset consisting of 1413 18F-Florbetapir PET scans from the Anti-Amyloid Treatment in Asymptomatic Alzheimer's (A4) study. The proposed model achieves MAE of 0.0243 and RMSE of 0.0320 for summary Standardized Uptake Value Ratio (SUVR) based on composite reference region for ADNI test set. When tested on the A4-study dataset, the proposed model achieves MAE of 0.038 and RMSE of 0.0495 for summary SUVR based on the composite region. The results show that the proposed model provides less MAE and RMSE when compared with existing models. A graphical user interface is developed based on the proposed model where the predictions are made by selecting the files of 18F-Florbetapir PET scans.
Collapse
Affiliation(s)
- R Divya
- Department of Electronics and Communication Engineering, Mepco Schlenk Engineering College, Sivakasi 626 005, Tamil Nadu, India.
| | - R Shantha Selva Kumari
- Department of Electronics and Communication Engineering, Mepco Schlenk Engineering College, Sivakasi 626 005, Tamil Nadu, India.
| |
Collapse
|
2
|
Yu B, Whitmarsh T, Riede P, McDonald S, Kaggie JD, Cox TM, Poole KES, Deegan P. Deep learning-based quantification of osteonecrosis using magnetic resonance images in Gaucher disease. Bone 2024; 186:117142. [PMID: 38834102 DOI: 10.1016/j.bone.2024.117142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/11/2024] [Revised: 05/28/2024] [Accepted: 05/31/2024] [Indexed: 06/06/2024]
Abstract
Gaucher disease is one of the most common lysosomal storage disorders. Osteonecrosis is a principal clinical manifestation of Gaucher disease and often leads to joint collapse and fractures. T1-weighted (T1w) modality in MRI is widely used to monitor bone involvement in Gaucher disease and to diagnose osteonecrosis. However, objective and quantitative methods for characterizing osteonecrosis are still limited. In this work, we present a deep learning-based quantification approach for the segmentation of osteonecrosis and the extraction of characteristic parameters. We first constructed two independent U-net models to segment the osteonecrosis and bone marrow unaffected by osteonecrosis (UBM) in spine and femur respectively, based on T1w images from patients in the UK national Gaucherite study database. We manually delineated parcellation maps including osteonecrosis and UBM from 364 T1w images (176 for spine, 188 for femur) as the training datasets, and the trained models were subsequently applied to all the 917 T1w images in the database. To quantify the segmentation, we calculated morphological parameters including the volume of osteonecrosis, the volume of UBM, and the fraction of total marrow occupied by osteonecrosis. Then, we examined the correlation between calculated features and the bone marrow burden score for marrow infiltration of the corresponding image, and no strong correlation was found. In addition, we analyzed the influence of splenectomy and the interval between the age at first symptom and the age of onset of treatment on the quantitative measurements of osteonecrosis. The results are consistent with previous studies, showing that prior splenectomy is closely associated with the fractional volume of osteonecrosis, and there is a positive relationship between the duration of untreated disease and the quantifications of osteonecrosis. We propose this technique as an efficient and reliable tool for assessing the extent of osteonecrosis in MR images of patients and improving prediction of clinically important adverse events.
Collapse
Affiliation(s)
- Boliang Yu
- Department of Medicine, University of Cambridge, Cambridge, UK.
| | | | - Philipp Riede
- Department of Radiology, University of Cambridge, Cambridge, UK
| | - Scott McDonald
- Department of Radiology, University of Cambridge, Cambridge, UK
| | - Joshua D Kaggie
- Department of Radiology, University of Cambridge, Cambridge, UK
| | - Timothy M Cox
- Department of Medicine, University of Cambridge, Cambridge, UK
| | | | - Patrick Deegan
- Department of Medicine, University of Cambridge, Cambridge, UK
| |
Collapse
|
3
|
d'Albenzio G, Kamkova Y, Naseem R, Ullah M, Colonnese S, Cheikh FA, Kumar RP. A dual-encoder double concatenation Y-shape network for precise volumetric liver and lesion segmentation. Comput Biol Med 2024; 179:108870. [PMID: 39024904 DOI: 10.1016/j.compbiomed.2024.108870] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2023] [Revised: 07/07/2024] [Accepted: 07/08/2024] [Indexed: 07/20/2024]
Abstract
Accurate segmentation of the liver and tumors from CT volumes is crucial for hepatocellular carcinoma diagnosis and pre-operative resection planning. Despite advances in deep learning-based methods for abdominal CT images, fully-automated segmentation remains challenging due to class imbalance and structural variations, often requiring cascaded approaches that incur significant computational costs. In this paper, we present the Dual-Encoder Double Concatenation Network (DEDC-Net) for simultaneous segmentation of the liver and its tumors. DEDC-Net leverages both residual and skip connections to enhance feature reuse and optimize performance in liver and tumor segmentation tasks. Extensive qualitative and quantitative experiments on the LiTS dataset demonstrate that DEDC-Net outperforms existing state-of-the-art liver segmentation methods. An ablation study was conducted to evaluate different encoder backbones - specifically VGG19 and ResNet - and the impact of incorporating an attention mechanism. Our results indicate that DEDC-Net, without any additional attention gates, achieves a superior mean Dice Score (DS) of 0.898 for liver segmentation. Moreover, integrating residual connections into one encoder yielded the highest DS for tumor segmentation tasks. The robustness of our proposed network was further validated on two additional, unseen CT datasets: IDCARDb-01 and COMET. Our model demonstrated superior lesion segmentation capabilities, particularly on IRCADb-01, achieving a DS of 0.629. The code implementation is publicly available at this website.
Collapse
Affiliation(s)
- Gabriella d'Albenzio
- The Intervention Center, Oslo University Hospital, 0slo, Norway; Department of Informatics, University of Oslo, Oslo, Norway.
| | - Yuliia Kamkova
- Department of Informatics, University of Oslo, Oslo, Norway; Department of Research and Development, Division of Emergencies and Critical Care, Oslo University Hospital, Oslo, Norway
| | - Rabia Naseem
- COMSATS, University Islamabad, Islamabad, Pakistan
| | - Mohib Ullah
- Department of Computer Science, Norwegian University of Science and Technology, Gjøvik, Norway
| | - Stefania Colonnese
- Department of Information Engineering, Electronics and Telecommunications (DIET), La Sapienza University of Rome, Rome, Italy
| | - Faouzi Alaya Cheikh
- Department of Computer Science, Norwegian University of Science and Technology, Gjøvik, Norway
| | | |
Collapse
|
4
|
Ghosh Moulic A, Gaurkar SS, Deshmukh PT. Artificial Intelligence in Otology, Rhinology, and Laryngology: A Narrative Review of Its Current and Evolving Picture. Cureus 2024; 16:e66036. [PMID: 39224718 PMCID: PMC11366564 DOI: 10.7759/cureus.66036] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2024] [Accepted: 08/02/2024] [Indexed: 09/04/2024] Open
Abstract
With technological advancements, artificial intelligence (AI) has progressed to become a ubiquitous part of human life. Its aspects in otorhinolaryngology are varied and are continuously evolving. Currently, AI has applications in hearing aids, imaging technologies, interpretation of auditory brain stem systems, and many more in otology. In rhinology, AI is seen to impact navigation, robotic surgeries, and the determination of various anomalies. Detection of voice pathologies and imaging are some areas of laryngology where AI is being used. This review gives an outlook on the diverse elements, applications, and advancements of AI in otorhinolaryngology. The various subfields of AI including machine learning, neural networks, and deep learning are also discussed. Clinical integration of AI and otorhinolaryngology has immense potential to revolutionize the healthcare system and improve the standards of patient care. The current applications of AI and its future scopes in developing this field are highlighted in this review.
Collapse
Affiliation(s)
- Ayushi Ghosh Moulic
- Otolaryngology - Head and Neck Surgery, Jawaharlal Nehru Medical College, Datta Meghe Institute of Higher Education and Research, Wardha, IND
| | - Sagar S Gaurkar
- Otolaryngology - Head and Neck Surgery, Jawaharlal Nehru Medical College, Datta Meghe Institute of Higher Education and Research, Wardha, IND
| | - Prasad T Deshmukh
- Otolaryngology - Head and Neck Surgery, Jawaharlal Nehru Medical College, Datta Meghe Institute of Higher Education and Research, Wardha, IND
| |
Collapse
|
5
|
Pham TV, Vu TN, Le HMQ, Pham VT, Tran TT. CapNet: An Automatic Attention-Based with Mixer Model for Cardiovascular Magnetic Resonance Image Segmentation. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01191-x. [PMID: 38980628 DOI: 10.1007/s10278-024-01191-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/24/2023] [Revised: 05/21/2024] [Accepted: 05/22/2024] [Indexed: 07/10/2024]
Abstract
Deep neural networks have shown excellent performance in medical image segmentation, especially for cardiac images. Transformer-based models, though having advantages over convolutional neural networks due to the ability of long-range dependence learning, still have shortcomings such as having a large number of parameters and and high computational cost. Additionally, for better results, they are often pretrained on a larger data, thus requiring large memory size and increasing resource expenses. In this study, we propose a new lightweight but efficient model, namely CapNet, based on convolutions and mixing modules for cardiac segmentation from magnetic resonance images (MRI) that can be trained from scratch with a small amount of parameters. To handle varying sizes and shapes which often occur in cardiac systolic and diastolic phases, we propose attention modules for pooling, spatial, and channel information. We also propose a novel loss called the Tversky Shape Power Distance function based on the shape dissimilarity between labels and predictions that shows promising performances compared to other losses. Experiments on three public datasets including ACDC benchmark, Sunnybrook data, and MS-CMR challenge are conducted and compared with other state of the arts (SOTA). For binary segmentation, the proposed CapNet obtained the Dice similarity coefficient (DSC) of 94% and 95.93% for respectively the Endocardium and Epicardium regions with Sunnybrook dataset, 94.49% for Endocardium, and 96.82% for Epicardium with the ACDC data. Regarding the multiclass case, the average DSC by CapNet is 93.05% for the ACDC data; and the DSC scores for the MS-CMR are 94.59%, 92.22%, and 93.99% for respectively the bSSFP, T2-SPAIR, and LGE sequences of the MS-CMR. Moreover, the statistical significance analysis tests with p-value < 0.05 compared with transformer-based methods and some CNN-based approaches demonstrated that the CapNet, though having fewer training parameters, is statistically significant. The promising evaluation metrics show comparative results in both Dice and IoU indices compared to SOTA CNN-based and Transformer-based architectures.
Collapse
Affiliation(s)
- Tien Viet Pham
- Department of Automation Engineering, School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, Hanoi, Vietnam
| | - Tu Ngoc Vu
- Department of Automation Engineering, School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, Hanoi, Vietnam
| | - Hoang-Minh-Quang Le
- Department of Automation Engineering, School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, Hanoi, Vietnam
| | - Van-Truong Pham
- Department of Automation Engineering, School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, Hanoi, Vietnam
| | - Thi-Thao Tran
- Department of Automation Engineering, School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, Hanoi, Vietnam.
| |
Collapse
|
6
|
Fang TY, Lin TY, Shen CM, Hsu SY, Lin SH, Kuo YJ, Chen MH, Yin TK, Liu CH, Lo MT, Wang PC. Algorithm-Driven Tele-otoscope for Remote Care for Patients With Otitis Media. Otolaryngol Head Neck Surg 2024; 170:1590-1597. [PMID: 38545686 DOI: 10.1002/ohn.738] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Revised: 02/05/2024] [Accepted: 02/29/2024] [Indexed: 05/31/2024]
Abstract
OBJECTIVE The COVID-19 pandemic has spurred a growing demand for telemedicine. Artificial intelligence and image processing systems with wireless transmission functionalities can facilitate remote care for otitis media (OM). Accordingly, this study developed and validated an algorithm-driven tele-otoscope system equipped with Wi-Fi transmission and a cloud-based automatic OM diagnostic algorithm. STUDY DESIGN Prospective, cross-sectional, diagnostic study. SETTING Tertiary Academic Medical Center. METHODS We designed a tele-otoscope (Otiscan, SyncVision Technology Corp) equipped with digital imaging and processing modules, Wi-Fi transmission capabilities, and an automatic OM diagnostic algorithm. A total of 1137 otoscopic images, comprising 987 images of normal cases and 150 images of cases of acute OM and OM with effusion, were used as the dataset for image classification. Two convolutional neural network models, trained using our dataset, were used for raw image segmentation and OM classification. RESULTS The tele-otoscope delivered images with a resolution of 1280 × 720 pixels. Our tele-otoscope effectively differentiated OM from normal images, achieving a classification accuracy rate of up to 94% (sensitivity, 80%; specificity, 96%). CONCLUSION Our study demonstrated that the developed tele-otoscope has acceptable accuracy in diagnosing OM. This system can assist health care professionals in early detection and continuous remote monitoring, thus mitigating the consequences of OM.
Collapse
Affiliation(s)
- Te-Yung Fang
- Department of Otolaryngology, Cathay General Hospital, Taipei, Taiwan
- School of Medicine, Fu-Jen Catholic University, New Taipei City, Taiwan
- Department of Otolaryngology, Sijhih Cathay General Hospital, New Taipei City, Taiwan
| | - Tse-Yu Lin
- Department of Biomedical Sciences and Engineering, National Central University, Taoyuan, Taiwan
| | - Chung-Min Shen
- School of Medicine, Fu-Jen Catholic University, New Taipei City, Taiwan
- Department of Pediatric, Cathay General Hospital, Taipei, Taiwan
| | - Su-Yi Hsu
- Department of Otolaryngology, Cathay General Hospital, Taipei, Taiwan
- School of Medicine, Fu-Jen Catholic University, New Taipei City, Taiwan
| | - Shing-Huey Lin
- School of Medicine, Fu-Jen Catholic University, New Taipei City, Taiwan
- Department of Family and Community Medicine, Cathay General Hospital, Taipei, Taiwan
| | - Yu-Jung Kuo
- Department of Biomedical Sciences and Engineering, National Central University, Taoyuan, Taiwan
| | - Ming-Hsu Chen
- Department of Otolaryngology, Cathay General Hospital, Taipei, Taiwan
| | - Tan-Kuei Yin
- Department of Otolaryngology, Cathay General Hospital, Taipei, Taiwan
| | - Chih-Hsien Liu
- Department of Otolaryngology, Cathay General Hospital, Taipei, Taiwan
| | - Men-Tzung Lo
- Department of Biomedical Sciences and Engineering, National Central University, Taoyuan, Taiwan
| | - Pa-Chun Wang
- Department of Otolaryngology, Cathay General Hospital, Taipei, Taiwan
- School of Medicine, Fu-Jen Catholic University, New Taipei City, Taiwan
- Department of Biomedical Sciences and Engineering, National Central University, Taoyuan, Taiwan
- Department of Medical Research, China Medical University Hospital, China Medical University, Taichung, Taiwan
| |
Collapse
|
7
|
Tsilivigkos C, Athanasopoulos M, Micco RD, Giotakis A, Mastronikolis NS, Mulita F, Verras GI, Maroulis I, Giotakis E. Deep Learning Techniques and Imaging in Otorhinolaryngology-A State-of-the-Art Review. J Clin Med 2023; 12:6973. [PMID: 38002588 PMCID: PMC10672270 DOI: 10.3390/jcm12226973] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2023] [Revised: 11/02/2023] [Accepted: 11/06/2023] [Indexed: 11/26/2023] Open
Abstract
Over the last decades, the field of medicine has witnessed significant progress in artificial intelligence (AI), the Internet of Medical Things (IoMT), and deep learning (DL) systems. Otorhinolaryngology, and imaging in its various subspecialties, has not remained untouched by this transformative trend. As the medical landscape evolves, the integration of these technologies becomes imperative in augmenting patient care, fostering innovation, and actively participating in the ever-evolving synergy between computer vision techniques in otorhinolaryngology and AI. To that end, we conducted a thorough search on MEDLINE for papers published until June 2023, utilizing the keywords 'otorhinolaryngology', 'imaging', 'computer vision', 'artificial intelligence', and 'deep learning', and at the same time conducted manual searching in the references section of the articles included in our manuscript. Our search culminated in the retrieval of 121 related articles, which were subsequently subdivided into the following categories: imaging in head and neck, otology, and rhinology. Our objective is to provide a comprehensive introduction to this burgeoning field, tailored for both experienced specialists and aspiring residents in the domain of deep learning algorithms in imaging techniques in otorhinolaryngology.
Collapse
Affiliation(s)
- Christos Tsilivigkos
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| | - Michail Athanasopoulos
- Department of Otolaryngology, University Hospital of Patras, 265 04 Patras, Greece; (M.A.); (N.S.M.)
| | - Riccardo di Micco
- Department of Otolaryngology and Head and Neck Surgery, Medical School of Hannover, 30625 Hannover, Germany;
| | - Aris Giotakis
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| | - Nicholas S. Mastronikolis
- Department of Otolaryngology, University Hospital of Patras, 265 04 Patras, Greece; (M.A.); (N.S.M.)
| | - Francesk Mulita
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Georgios-Ioannis Verras
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Ioannis Maroulis
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Evangelos Giotakis
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| |
Collapse
|
8
|
Wu Q, Wang X, Liang G, Luo X, Zhou M, Deng H, Zhang Y, Huang X, Yang Q. Advances in Image-Based Artificial Intelligence in Otorhinolaryngology-Head and Neck Surgery: A Systematic Review. Otolaryngol Head Neck Surg 2023; 169:1132-1142. [PMID: 37288505 DOI: 10.1002/ohn.391] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2023] [Revised: 04/27/2023] [Accepted: 05/13/2023] [Indexed: 06/09/2023]
Abstract
OBJECTIVE To update the literature and provide a systematic review of image-based artificial intelligence (AI) applications in otolaryngology, highlight its advances, and propose future challenges. DATA SOURCES Web of Science, Embase, PubMed, and Cochrane Library. REVIEW METHODS Studies written in English, published between January 2020 and December 2022. Two independent authors screened the search results, extracted data, and assessed studies. RESULTS Overall, 686 studies were identified. After screening titles and abstracts, 325 full-text studies were assessed for eligibility, and 78 studies were included in this systematic review. The studies originated from 16 countries. Among these countries, the top 3 were China (n = 29), Korea (n = 8), the United States, and Japan (n = 7 each). The most common area was otology (n = 35), followed by rhinology (n = 20), pharyngology (n = 18), and head and neck surgery (n = 5). Most applications of AI in otology, rhinology, pharyngology, and head and neck surgery mainly included chronic otitis media (n = 9), nasal polyps (n = 4), laryngeal cancer (n = 12), and head and neck squamous cell carcinoma (n = 3), respectively. The overall performance of AI in accuracy, the area under the curve, sensitivity, and specificity were 88.39 ± 9.78%, 91.91 ± 6.70%, 86.93 ± 11.59%, and 88.62 ± 14.03%, respectively. CONCLUSION This state-of-the-art review aimed to highlight the increasing applications of image-based AI in otorhinolaryngology head and neck surgery. The following steps will entail multicentre collaboration to ensure data reliability, ongoing optimization of AI algorithms, and integration into real-world clinical practice. Future studies should consider 3-dimensional (3D)-based AI, such as 3D surgical AI.
Collapse
Affiliation(s)
- Qingwu Wu
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xinyue Wang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Guixian Liang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xin Luo
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Min Zhou
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Huiyi Deng
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Yana Zhang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xuekun Huang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Qintai Yang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| |
Collapse
|
9
|
Zhou T, Liu F, Ye X, Wang H, Lu H. CCGL-YOLOV5:A cross-modal cross-scale global-local attention YOLOV5 lung tumor detection model. Comput Biol Med 2023; 165:107387. [PMID: 37659112 DOI: 10.1016/j.compbiomed.2023.107387] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2023] [Revised: 07/29/2023] [Accepted: 08/25/2023] [Indexed: 09/04/2023]
Abstract
BACKGROUND Multimodal medical image detection is a key technology in medical image analysis, which plays an important role in tumor diagnosis. There are different sizes lesions and different shapes lesions in multimodal lung tumor images, which makes it difficult to effectively extract key features of lung tumor lesions. METHODS A Cross-modal Cross-scale Clobal-Local Attention YOLOV5 Lung Tumor Detection Model (CCGL-YOLOV5) is proposed in this paper. The main works are as follows: Firstly, the Cross-Modal Fusion Transformer Module (CMFTM) is designed to improve the multimodal key lesion feature extraction ability and fusion ability through the interactive assisted fusion of multimodal features; Secondly, the Global-Local Feature Interaction Module (GLFIM) is proposed to enhance the interaction ability between multimodal global features and multimodal local features through bidirectional interactive branches. Thirdly, the Cross-Scale Attention Fusion Module (CSAFM) is designed to obtain rich multi-scale features through grouping multi-scale attention for feature fusion. RESULTS The comparison experiments with advanced networks are done. The Acc, Rec, mAP, F1 score and FPS of CCGL-YOLOV5 model on multimodal lung tumor PET/CT dataset are 97.83%, 97.39%, 96.67%, 97.61% and 98.59, respectively; The experimental results show that the performance of CCGL-YOLOV5 model in this paper are better than other typical models. CONCLUSION The CCGL-YOLOV5 model can effectively use the multimodal feature information. There are important implications for multimodal medical image research and clinical disease diagnosis in CCGL-YOLOV5 model.
Collapse
Affiliation(s)
- Tao Zhou
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China
| | - Fengzhen Liu
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China.
| | - Xinyu Ye
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China
| | - Hongwei Wang
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China
| | - Huiling Lu
- School of Medical Information and Engineering, Ningxia Medical University, Yinchuan, 750004, China.
| |
Collapse
|
10
|
Song D, Kim T, Lee Y, Kim J. Image-Based Artificial Intelligence Technology for Diagnosing Middle Ear Diseases: A Systematic Review. J Clin Med 2023; 12:5831. [PMID: 37762772 PMCID: PMC10531728 DOI: 10.3390/jcm12185831] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Revised: 08/27/2023] [Accepted: 08/29/2023] [Indexed: 09/29/2023] Open
Abstract
Otolaryngological diagnoses, such as otitis media, are traditionally performed using endoscopy, wherein diagnostic accuracy can be subjective and vary among clinicians. The integration of objective tools, like artificial intelligence (AI), could potentially improve the diagnostic process by minimizing the influence of subjective biases and variability. We systematically reviewed the AI techniques using medical imaging in otolaryngology. Relevant studies related to AI-assisted otitis media diagnosis were extracted from five databases: Google Scholar, PubMed, Medline, Embase, and IEEE Xplore, without date restrictions. Publications that did not relate to AI and otitis media diagnosis or did not utilize medical imaging were excluded. Of the 32identified studies, 26 used tympanic membrane images for classification, achieving an average diagnosis accuracy of 86% (range: 48.7-99.16%). Another three studies employed both segmentation and classification techniques, reporting an average diagnosis accuracy of 90.8% (range: 88.06-93.9%). These findings suggest that AI technologies hold promise for improving otitis media diagnosis, offering benefits for telemedicine and primary care settings due to their high diagnostic accuracy. However, to ensure patient safety and optimal outcomes, further improvements in diagnostic performance are necessary.
Collapse
Affiliation(s)
- Dahye Song
- Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea; (D.S.); (T.K.)
| | - Taewan Kim
- Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea; (D.S.); (T.K.)
| | - Yeonjoon Lee
- Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University, Ansan 15588, Republic of Korea; (D.S.); (T.K.)
| | - Jaeyoung Kim
- Department of Dermatology and Skin Sciences, University of British Columbia, Vancouver, BC V6T 1Z1, Canada;
- Core Research & Development Center, Korea University Ansan Hospital, Ansan 15355, Republic of Korea
| |
Collapse
|
11
|
Ding X, Huang Y, Tian X, Zhao Y, Feng G, Gao Z. Diagnosis, Treatment, and Management of Otitis Media with Artificial Intelligence. Diagnostics (Basel) 2023; 13:2309. [PMID: 37443702 DOI: 10.3390/diagnostics13132309] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2023] [Revised: 06/04/2023] [Accepted: 06/14/2023] [Indexed: 07/15/2023] Open
Abstract
A common infectious disease, otitis media (OM) has a low rate of early diagnosis, which significantly increases the difficulty of treating the disease and the likelihood of serious complications developing including hearing loss, speech impairment, and even intracranial infection. Several areas of healthcare have shown great promise in the application of artificial intelligence (AI) systems, such as the accurate detection of diseases, the automated interpretation of images, and the prediction of patient outcomes. Several articles have reported some machine learning (ML) algorithms such as ResNet, InceptionV3 and Unet, were applied to the diagnosis of OM successfully. The use of these techniques in the OM is still in its infancy, but their potential is enormous. We present in this review important concepts related to ML and AI, describe how these technologies are currently being applied to diagnosing, treating, and managing OM, and discuss the challenges associated with developing AI-assisted OM technologies in the future.
Collapse
Affiliation(s)
- Xin Ding
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Yu Huang
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Xu Tian
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Yang Zhao
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Guodong Feng
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| | - Zhiqiang Gao
- Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
| |
Collapse
|
12
|
An EffcientNet-encoder U-Net Joint Residual Refinement Module with Tversky–Kahneman Baroni–Urbani–Buser loss for biomedical image Segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104631] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/07/2023]
|
13
|
Luo S, Jiang H, Wang M. C 2BA-UNet: A context-coordination multi-atlas boundary-aware UNet-like method for PET/CT images based tumor segmentation. Comput Med Imaging Graph 2023; 103:102159. [PMID: 36549193 DOI: 10.1016/j.compmedimag.2022.102159] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Revised: 11/11/2022] [Accepted: 12/05/2022] [Indexed: 12/13/2022]
Abstract
Tumor segmentation is a necessary step in clinical processing that can help doctors diagnose tumors and plan surgical treatments. Since tumors are usually small, the locations and appearances vary substantially across individuals, and the contrast between tumors and adjacent normal tissues is low, tumor segmentation is still a challenging task. Although convolutional neural networks (CNNs) have achieved good results in tumor segmentation, the information about tumor boundaries has been rarely explored. To solve the problem, this paper proposes a new method for automatic tumor segmentation in PET/CT images based on context-coordination and boundary-aware, termed as C2BA-UNet. We employ a UNet-like backbone network and replace the encoder with EfficientNet-B0 for efficiency. To acquire potential tumor boundaries, we propose a new multi-atlas boundary-aware (MABA) module based on gradient atlas, uncertainty atlas, and level set atlas, that focuses on uncertain regions between tumors and adjacent tissues. Furthermore, we propose a new context coordination module (CCM) to combine multi-scale context information with attention mechanism to optimize skip connection in high-level layers. To validate the superiority of our method, we conduct experiments on a publicly available soft tissue sarcoma (STS) dataset and a lymphoma dataset, and the results show our method is competitive with other comparison methods.
Collapse
Affiliation(s)
- Shijie Luo
- Software College, Northeastern University, Shenyang 110819, China
| | - Huiyan Jiang
- Software College, Northeastern University, Shenyang 110819, China; Key Laboratory of Intelligent Computing in Biomedical Image, Ministry of Education, Northeastern University, Shenyang 110819, China.
| | - Meng Wang
- Software College, Northeastern University, Shenyang 110819, China
| |
Collapse
|
14
|
Ben Hamida A, Devanne M, Weber J, Truntzer C, Derangère V, Ghiringhelli F, Forestier G, Wemmert C. Weakly Supervised Learning using Attention gates for colon cancer histopathological image segmentation. Artif Intell Med 2022; 133:102407. [PMID: 36328667 DOI: 10.1016/j.artmed.2022.102407] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2022] [Revised: 09/07/2022] [Accepted: 09/15/2022] [Indexed: 02/08/2023]
Abstract
Recently, Artificial Intelligence namely Deep Learning methods have revolutionized a wide range of domains and applications. Besides, Digital Pathology has so far played a major role in the diagnosis and the prognosis of tumors. However, the characteristics of the Whole Slide Images namely the gigapixel size, high resolution and the shortage of richly labeled samples have hindered the efficiency of classical Machine Learning methods. That goes without saying that traditional methods are poor in generalization to different tasks and data contents. Regarding the success of Deep learning when dealing with Large Scale applications, we have resorted to the use of such models for histopathological image segmentation tasks. First, we review and compare the classical UNet and Att-UNet models for colon cancer WSI segmentation in a sparsely annotated data scenario. Then, we introduce novel enhanced models of the Att-UNet where different schemes are proposed for the skip connections and spatial attention gates positions in the network. In fact, spatial attention gates assist the training process and enable the model to avoid irrelevant feature learning. Alternating the presence of such modules namely in our Alter-AttUNet model adds robustness and ensures better image segmentation results. In order to cope with the lack of richly annotated data in our AiCOLO colon cancer dataset, we suggest the use of a multi-step training strategy that also deals with the WSI sparse annotations and unbalanced class issues. All proposed methods outperform state-of-the-art approaches but Alter-AttUNet generates the best compromise between accurate results and light network. The model achieves 95.88% accuracy with our sparse AiCOLO colon cancer datasets. Finally, to evaluate and validate our proposed architectures we resort to publicly available WSI data: the NCT-CRC-HE-100K, the CRC-5000 and the Warwick colon cancer histopathological dataset. Respective accuracies of 99.65%, 99.73% and 79.03% were reached. A comparison with state-of-art approaches is established to view and compare the key solutions for histopathological image segmentation.
Collapse
Affiliation(s)
| | - M Devanne
- IRIMAS, University of Haute-Alsace, France
| | - J Weber
- IRIMAS, University of Haute-Alsace, France
| | - C Truntzer
- Platform of Transform in Biological Oncology, Dijon, France
| | - V Derangère
- Platform of Transform in Biological Oncology, Dijon, France
| | - F Ghiringhelli
- Platform of Transform in Biological Oncology, Dijon, France
| | | | - C Wemmert
- ICube, University of Strasbourg, France
| |
Collapse
|
15
|
Habib AR, Kajbafzadeh M, Hasan Z, Wong E, Gunasekera H, Perry C, Sacks R, Kumar A, Singh N. Artificial intelligence to classify ear disease from otoscopy: A systematic review and meta-analysis. Clin Otolaryngol 2022; 47:401-413. [PMID: 35253378 PMCID: PMC9310803 DOI: 10.1111/coa.13925] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2021] [Revised: 01/08/2022] [Accepted: 02/27/2022] [Indexed: 11/29/2022]
Abstract
Objectives To summarise the accuracy of artificial intelligence (AI) computer vision algorithms to classify ear disease from otoscopy. Design Systematic review and meta‐analysis. Methods Using the PRISMA guidelines, nine online databases were searched for articles that used AI computer vision algorithms developed from various methods (convolutional neural networks, artificial neural networks, support vector machines, decision trees and k‐nearest neighbours) to classify otoscopic images. Diagnostic classes of interest: normal tympanic membrane, acute otitis media (AOM), otitis media with effusion (OME), chronic otitis media (COM) with or without perforation, cholesteatoma and canal obstruction. Main outcome measures Accuracy to correctly classify otoscopic images compared to otolaryngologists (ground truth). The Quality Assessment of Diagnostic Accuracy Studies Version 2 tool was used to assess the quality of methodology and risk of bias. Results Thirty‐nine articles were included. Algorithms achieved 90.7% (95%CI: 90.1–91.3%) accuracy to difference between normal or abnormal otoscopy images in 14 studies. The most common multiclassification algorithm (3 or more diagnostic classes) achieved 97.6% (95%CI: 97.3–97.9%) accuracy to differentiate between normal, AOM and OME in three studies. AI algorithms outperformed human assessors to classify otoscopy images achieving 93.4% (95%CI: 90.5–96.4%) versus 73.2% (95%CI: 67.9–78.5%) accuracy in three studies. Convolutional neural networks achieved the highest accuracy compared to other classification methods. Conclusion AI can classify ear disease from otoscopy. A concerted effort is required to establish a comprehensive and reliable otoscopy database for algorithm training. An AI‐supported otoscopy system may assist health care workers, trainees and primary care practitioners with less otology experience identify ear disease.
Collapse
Affiliation(s)
- Al-Rahim Habib
- Faculty of Medicine and Health, University of Sydney, New South Wales, Australia.,Department of Otolaryngology - Head and Neck Surgery, Princess Alexandra Hospital, Queensland, Australia.,Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, New South Wales, Australia
| | - Majid Kajbafzadeh
- Faculty of Medicine and Health, University of Sydney, New South Wales, Australia
| | - Zubair Hasan
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, New South Wales, Australia
| | - Eugene Wong
- Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, New South Wales, Australia
| | - Hasantha Gunasekera
- Faculty of Medicine and Health, University of Sydney, New South Wales, Australia.,The Children's Hospital at Westmead, New South Wales, Australia
| | - Chris Perry
- Department of Otolaryngology - Head and Neck Surgery, Princess Alexandra Hospital, Queensland, Australia.,University of Queensland Medical School, Queensland, Australia
| | - Raymond Sacks
- Faculty of Medicine and Health, University of Sydney, New South Wales, Australia
| | - Ashnil Kumar
- School of Biomedical Engineering, Faculty of Engineering, University of Sydney, New South Wales, Australia
| | - Narinder Singh
- Faculty of Medicine and Health, University of Sydney, New South Wales, Australia.,Department of Otolaryngology - Head and Neck Surgery, Westmead Hospital, New South Wales, Australia
| |
Collapse
|
16
|
Improved U-Net Remote Sensing Classification Algorithm Based on Multi-Feature Fusion Perception. REMOTE SENSING 2022. [DOI: 10.3390/rs14051118] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
The selection and representation of remote sensing image classification features play crucial roles in image classification accuracy. To effectively improve the classification accuracy of features, an improved U-Net network framework based on multi-feature fusion perception is proposed in this paper. This framework adds the channel attention module (CAM-UNet) to the original U-Net framework and cascades the shallow features with the deep semantic features, replaces the classification layer in the original U-Net network with a support vector machine, and finally uses the majority voting game theory algorithm to fuse the multifeature classification results and obtain the final classification results. This study used the forest distribution in Xingbin District, Laibin City, Guangxi Zhuang Autonomous Region as the research object, which is based on Landsat 8 multispectral remote sensing images, and, by combining spectral features, spatial features, and advanced semantic features, overcame the influence of the reduction in spatial resolution that occurs with the deepening of the network on the classification results. The experimental results showed that the improved algorithm can improve classification accuracy. Before the improvement, the overall segmentation accuracy and segmentation accuracy of the forestland increased from 90.50% to 92.82% and from 95.66% to 97.16%, respectively. The forest cover results obtained by the algorithm proposed in this paper can be used as input data for regional ecological models, which is conducive to the development of accurate and real-time vegetation growth change models.
Collapse
|