1
|
Nan Y, Xing X, Wang S, Tang Z, Felder FN, Zhang S, Ledda RE, Ding X, Yu R, Liu W, Shi F, Sun T, Cao Z, Zhang M, Gu Y, Zhang H, Gao J, Wang P, Tang W, Yu P, Kang H, Chen J, Lu X, Zhang B, Mamalakis M, Prinzi F, Carlini G, Cuneo L, Banerjee A, Xing Z, Zhu L, Mesbah Z, Jain D, Mayet T, Yuan H, Lyu Q, Qayyum A, Mazher M, Wells A, Walsh SL, Yang G. Hunting imaging biomarkers in pulmonary fibrosis: Benchmarks of the AIIB23 challenge. Med Image Anal 2024; 97:103253. [PMID: 38968907 DOI: 10.1016/j.media.2024.103253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2023] [Revised: 04/16/2024] [Accepted: 06/22/2024] [Indexed: 07/07/2024]
Abstract
Airway-related quantitative imaging biomarkers are crucial for examination, diagnosis, and prognosis in pulmonary diseases. However, the manual delineation of airway structures remains prohibitively time-consuming. While significant efforts have been made towards enhancing automatic airway modelling, current public-available datasets predominantly concentrate on lung diseases with moderate morphological variations. The intricate honeycombing patterns present in the lung tissues of fibrotic lung disease patients exacerbate the challenges, often leading to various prediction errors. To address this issue, the 'Airway-Informed Quantitative CT Imaging Biomarker for Fibrotic Lung Disease 2023' (AIIB23) competition was organized in conjunction with the official 2023 International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI). The airway structures were meticulously annotated by three experienced radiologists. Competitors were encouraged to develop automatic airway segmentation models with high robustness and generalization abilities, followed by exploring the most correlated QIB of mortality prediction. A training set of 120 high-resolution computerised tomography (HRCT) scans were publicly released with expert annotations and mortality status. The online validation set incorporated 52 HRCT scans from patients with fibrotic lung disease and the offline test set included 140 cases from fibrosis and COVID-19 patients. The results have shown that the capacity of extracting airway trees from patients with fibrotic lung disease could be enhanced by introducing voxel-wise weighted general union loss and continuity loss. In addition to the competitive image biomarkers for mortality prediction, a strong airway-derived biomarker (Hazard ratio>1.5, p < 0.0001) was revealed for survival prognostication compared with existing clinical measurements, clinician assessment and AI-based biomarkers.
Collapse
Affiliation(s)
- Yang Nan
- Bioengineering Department and Imperial-X, Imperial College London, London, UK; Royal Brompton Hospital, London, UK.
| | - Xiaodan Xing
- Bioengineering Department and Imperial-X, Imperial College London, London, UK.
| | - Shiyi Wang
- National Heart and Lung Institute, Imperial College London, London, UK
| | - Zeyu Tang
- Bioengineering Department and Imperial-X, Imperial College London, London, UK
| | - Federico N Felder
- Royal Brompton Hospital, London, UK; National Heart and Lung Institute, Imperial College London, London, UK
| | - Sheng Zhang
- National Heart and Lung Institute, Imperial College London, London, UK
| | | | - Xiaoliu Ding
- Shanghai MicroPort MedBot (Group) Co., Ltd., China
| | - Ruiqi Yu
- Shanghai MicroPort MedBot (Group) Co., Ltd., China
| | - Weiping Liu
- Shanghai MicroPort MedBot (Group) Co., Ltd., China
| | - Feng Shi
- Shanghai United Imaging Intelligence Co., Ltd., China
| | - Tianyang Sun
- Shanghai United Imaging Intelligence Co., Ltd., China
| | - Zehong Cao
- Shanghai United Imaging Intelligence Co., Ltd., China
| | - Minghui Zhang
- Institute of Medical Robotics, Shanghai Jiao Tong University, China
| | - Yun Gu
- Institute of Medical Robotics, Shanghai Jiao Tong University, China
| | - Hanxiao Zhang
- Institute of Medical Robotics, Shanghai Jiao Tong University, China
| | - Jian Gao
- Department Computational Biology, School of Life Sciences, Fudan University, Shanghai, China
| | - Pingyu Wang
- Cambridge International Exam Centre in Shanghai Experimental School, China
| | - Wen Tang
- InferVision Medical Technology Co., Ltd., China
| | - Pengxin Yu
- InferVision Medical Technology Co., Ltd., China
| | - Han Kang
- InferVision Medical Technology Co., Ltd., China
| | - Junqiang Chen
- Shanghai MediWorks Precision Instruments Co., Ltd, China
| | - Xing Lu
- Sanmed Biotech Ltd., Zhuhai, China
| | | | | | - Francesco Prinzi
- Department of Biomedicine, University of Palermo, Palermo, Italy
| | - Gianluca Carlini
- IRCCS Istituto delle Scienze Neurologiche di Bologna, Bologna, Italy
| | - Lisa Cuneo
- Istituto Italiano di Tecnologia, Nanoscopy, Genova, Italy
| | - Abhirup Banerjee
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, UK
| | - Zhaohu Xing
- Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China
| | - Lei Zhu
- Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China
| | - Zacharia Mesbah
- INSA Rouen Normandie, Univ Rouen Normandie, Université Le Havre Normandie, France; Nuclear Medicine Department, Henri Becquerel Cancer Center, Rouen, France
| | - Dhruv Jain
- INSA Rouen Normandie, Univ Rouen Normandie, Université Le Havre Normandie, France
| | - Tsiry Mayet
- INSA Rouen Normandie, Univ Rouen Normandie, Université Le Havre Normandie, France
| | - Hongyu Yuan
- Department of Radiology, Wake Forest University School of Medicine, USA
| | - Qing Lyu
- Department of Radiology, Wake Forest University School of Medicine, USA
| | - Abdul Qayyum
- National Heart and Lung Institute, Imperial College London, London, UK
| | - Moona Mazher
- Department of Computer Science, University College London, United Kingdom
| | - Athol Wells
- Royal Brompton Hospital, London, UK; National Heart and Lung Institute, Imperial College London, London, UK
| | - Simon Lf Walsh
- Royal Brompton Hospital, London, UK; National Heart and Lung Institute, Imperial College London, London, UK
| | - Guang Yang
- Bioengineering Department and Imperial-X, Imperial College London, London, UK; Royal Brompton Hospital, London, UK; National Heart and Lung Institute, Imperial College London, London, UK; School of Biomedical Engineering & Imaging Sciences, King's College London, UK.
| |
Collapse
|
2
|
Zhu L, Li J, Hu Y, Zhu R, Zeng S, Rong P, Zhang Y, Gu X, Wang Y, Zhang Z, Yang L, Ren Q, Lu Y. Choroidal Optical Coherence Tomography Angiography: Noninvasive Choroidal Vessel Analysis via Deep Learning. HEALTH DATA SCIENCE 2024; 4:0170. [PMID: 39257642 PMCID: PMC11383389 DOI: 10.34133/hds.0170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/10/2024] [Accepted: 06/25/2024] [Indexed: 09/12/2024]
Abstract
Background: The choroid is the most vascularized structure in the human eye, associated with numerous retinal and choroidal diseases. However, the vessel distribution of choroidal sublayers has yet to be effectively explored due to the lack of suitable tools for visualization and analysis. Methods: In this paper, we present a novel choroidal angiography strategy to more effectively evaluate vessels within choroidal sublayers in the clinic. Our approach utilizes a segmentation model to extract choroidal vessels from OCT B-scans layer by layer. Furthermore, we ensure that the model, trained on B-scans with high choroidal quality, can proficiently handle the low-quality B-scans commonly collected in clinical practice for reconstruction vessel distributions. By treating this process as a cross-domain segmentation task, we propose an ensemble discriminative mean teacher structure to address the specificities inherent in this cross-domain segmentation process. The proposed structure can select representative samples with minimal label noise for self-training and enhance the adaptation strength of adversarial training. Results: Experiments demonstrate the effectiveness of the proposed structure, achieving a dice score of 77.28 for choroidal vessel segmentation. This validates our strategy to provide satisfactory choroidal angiography noninvasively, supportting the analysis of choroidal vessel distribution for paitients with choroidal diseases. We observed that patients with central serous chorioretinopathy have evidently (P < 0.05) lower vascular indexes at all choroidal sublayers than healthy individuals, especially in the region beyond central fovea of macula (larger than 6 mm). Conclusions: We release the code and training set of the proposed method as the first noninvasive mechnism to assist clinical application for the analysis of choroidal vessels.
Collapse
Affiliation(s)
- Lei Zhu
- Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing 100191, China
- Department of Biomedical Engineering, Peking University, Beijing 100871, China
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Junmeng Li
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Yicheng Hu
- Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing 100191, China
- Department of Biomedical Engineering, Peking University, Beijing 100871, China
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Ruilin Zhu
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Shuang Zeng
- Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing 100191, China
- Department of Biomedical Engineering, Peking University, Beijing 100871, China
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Pei Rong
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Yadi Zhang
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Xiaopeng Gu
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Yuwei Wang
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Zhiyue Zhang
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Liu Yang
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| | - Qiushi Ren
- Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing 100191, China
- Department of Biomedical Engineering, Peking University, Beijing 100871, China
- National Biomedical Imaging Center, Peking University, Beijing 100871, China
| | - Yanye Lu
- Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing 100191, China
- Department of Biomedical Engineering, Peking University, Beijing 100871, China
- Department of Ophthalmology, Peking University First Hospital, Beijing 100034, China
| |
Collapse
|
3
|
De A, Das N, Saha PK, Comellas A, Hoffman E, Basu S, Chakraborti T. MSO-GP: 3-D segmentation of large and complex conjoined tree structures. Methods 2024; 229:9-16. [PMID: 38838947 DOI: 10.1016/j.ymeth.2024.05.016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2023] [Revised: 05/03/2024] [Accepted: 05/29/2024] [Indexed: 06/07/2024] Open
Abstract
Robust segmentation of large and complex conjoined tree structures in 3-D is a major challenge in computer vision. This is particularly true in computational biology, where we often encounter large data structures in size, but few in number, which poses a hard problem for learning algorithms. We show that merging multiscale opening with geodesic path propagation, can shed new light on this classic machine vision challenge, while circumventing the learning issue by developing an unsupervised visual geometry approach (digital topology/morphometry). The novelty of the proposed MSO-GP method comes from the geodesic path propagation being guided by a skeletonization of the conjoined structure that helps to achieve robust segmentation results in a particularly challenging task in this area, that of artery-vein separation from non-contrast pulmonary computed tomography angiograms. This is an important first step in measuring vascular geometry to then diagnose pulmonary diseases and to develop image-based phenotypes. We first present proof-of-concept results on synthetic data, and then verify the performance on pig lung and human lung data with less segmentation time and user intervention needs than those of the competing methods.
Collapse
Affiliation(s)
- Arijit De
- Department of Electronics & Telecommunication Engineering, Jadavpur University, Kolkata, India.
| | - Nirmal Das
- Department of Computer Science and Engineering (AIML), Institute of Engineering and Management, Kolkata, India; Department of Computer Science and Engineering, Jadavpur University, Kolkata, India.
| | - Punam K Saha
- Department of Electrical and Computer Engineering & Department of Radiology, University of Iowa, Iowa City, IA 52242, USA.
| | | | - Eric Hoffman
- Department of Internal Medicine, University of Iowa, Iowa City, USA.
| | - Subhadip Basu
- Department of Computer Science and Engineering (AIML), Institute of Engineering and Management, Kolkata, India.
| | - Tapabrata Chakraborti
- Health Sciences Programme, The Alan Turing Institute, London, UK; Department of Medical Physics and Biomedical Engineering, University College London, London, UK.
| |
Collapse
|
4
|
Wu Y, Xia S, Liang Z, Chen R, Qi S. Artificial intelligence in COPD CT images: identification, staging, and quantitation. Respir Res 2024; 25:319. [PMID: 39174978 PMCID: PMC11340084 DOI: 10.1186/s12931-024-02913-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Accepted: 07/09/2024] [Indexed: 08/24/2024] Open
Abstract
Chronic obstructive pulmonary disease (COPD) stands as a significant global health challenge, with its intricate pathophysiological manifestations often demanding advanced diagnostic strategies. The recent applications of artificial intelligence (AI) within the realm of medical imaging, especially in computed tomography, present a promising avenue for transformative changes in COPD diagnosis and management. This review delves deep into the capabilities and advancements of AI, particularly focusing on machine learning and deep learning, and their applications in COPD identification, staging, and imaging phenotypes. Emphasis is laid on the AI-powered insights into emphysema, airway dynamics, and vascular structures. The challenges linked with data intricacies and the integration of AI in the clinical landscape are discussed. Lastly, the review casts a forward-looking perspective, highlighting emerging innovations in AI for COPD imaging and the potential of interdisciplinary collaborations, hinting at a future where AI doesn't just support but pioneers breakthroughs in COPD care. Through this review, we aim to provide a comprehensive understanding of the current state and future potential of AI in shaping the landscape of COPD diagnosis and management.
Collapse
Affiliation(s)
- Yanan Wu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China
| | - Shuyue Xia
- Respiratory Department, Central Hospital Affiliated to Shenyang Medical College, Shenyang, China
- Key Laboratory of Medicine and Engineering for Chronic Obstructive Pulmonary Disease in Liaoning Province, Shenyang, China
| | - Zhenyu Liang
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, Guangzhou Institute of Respiratory Health, The National Center for Respiratory Medicine, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Rongchang Chen
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, Guangzhou Institute of Respiratory Health, The National Center for Respiratory Medicine, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
- Shenzhen Institute of Respiratory Disease, Shenzhen People's Hospital, Shenzhen, China
| | - Shouliang Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| |
Collapse
|
5
|
Min Y, Li J, Jia S, Li Y, Nie S. Automated Cerebrovascular Segmentation and Visualization of Intracranial Time-of-Flight Magnetic Resonance Angiography Based on Deep Learning. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01215-6. [PMID: 39133457 DOI: 10.1007/s10278-024-01215-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/02/2024] [Revised: 07/15/2024] [Accepted: 07/29/2024] [Indexed: 08/13/2024]
Abstract
Time-of-flight magnetic resonance angiography (TOF-MRA) is a non-contrast technique used to visualize neurovascular. However, manual reconstruction of the volume render (VR) by radiologists is time-consuming and labor-intensive. Deep learning-based (DL-based) vessel segmentation technology may provide intelligent automation workflow. To evaluate the image quality of DL vessel segmentation for automatically acquiring intracranial arteries in TOF-MRA. A total of 394 TOF-MRA scans were selected, which included cerebral vascular health, aneurysms, or stenoses. Both our proposed method and two state-of-the-art DL methods are evaluated on external datasets for generalization ability. For qualitative assessment, two experienced clinical radiologists evaluated the image quality of cerebrovascular diagnostic and visualization (scoring 0-5 as unacceptable to excellent) obtained by manual VR reconstruction or automatic convolutional neural network (CNN) segmentation. The proposed CNN outperforms the other two DL-based methods in clinical scoring on external datasets, and its visualization was evaluated by readers as having the appearance of the radiologists' manual reconstructions. Scoring of proposed CNN and VR of intracranial arteries demonstrated good to excellent agreement with no significant differences (median, 5.0 and 5.0, P ≥ 12) at healthy-type scans. All proposed CNN image quality were considered to have adequate diagnostic quality (median scores > 2). Quantitative analysis demonstrated a superior dice similarity coefficient of cerebrovascular overlap (training sets and validation sets; 0.947 and 0.927). Automatic cerebrovascular segmentation using DL is feasible and the image quality in terms of vessel integrity, collateral circulation and lesion morphology is comparable to expert manual VR without significant differences.
Collapse
Affiliation(s)
- Yuqin Min
- Institute for Medical Imaging Technology, Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, No.889, Shuang Ding Road, Shanghai, 201801, China
- Institute of Medical Imaging Engineering, School of Health Science and Engineering, University of Shanghai for Science and Technology, No.334, Jun Gong Road, Shanghai, 200093, China
| | - Jing Li
- Institute of Diagnostic and Interventional Radiology, Shanghai Jiao Tong University Affiliated Sixth People's Hospital, No.600, Yi Shan Road, Shanghai, 200233, China
| | - Shouqiang Jia
- Department of Imaging, Jinan People's Hospital affiliated to Shandong First Medical University, Shandong, 271100, China
| | - Yuehua Li
- Institute of Diagnostic and Interventional Radiology, Shanghai Jiao Tong University Affiliated Sixth People's Hospital, No.600, Yi Shan Road, Shanghai, 200233, China
| | - Shengdong Nie
- Institute of Medical Imaging Engineering, School of Health Science and Engineering, University of Shanghai for Science and Technology, No.334, Jun Gong Road, Shanghai, 200093, China.
| |
Collapse
|
6
|
Wang S, Nan Y, Zhang S, Felder F, Xing X, Fang Y, Del Ser J, Walsh SLF, Yang G. Probing perfection: The relentless art of meddling for pulmonary airway segmentation from HRCT via a human-AI collaboration based active learning method. Artif Intell Med 2024; 154:102930. [PMID: 39047631 DOI: 10.1016/j.artmed.2024.102930] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2023] [Revised: 06/27/2024] [Accepted: 07/03/2024] [Indexed: 07/27/2024]
Abstract
In the realm of pulmonary tracheal segmentation, the scarcity of annotated data stands as a prevalent pain point in most medical segmentation endeavors. Concurrently, most Deep Learning (DL) methodologies employed in this domain invariably grapple with other dual challenges: the inherent opacity of 'black box' models and the ongoing pursuit of performance enhancement. In response to these intertwined challenges, the core concept of our Human-Computer Interaction (HCI) based learning models (RS_UNet, LC_UNet, UUNet and WD_UNet) hinge on the versatile combination of diverse query strategies and an array of deep learning models. We train four HCI models based on the initial training dataset and sequentially repeat the following steps 1-4: (1) Query Strategy: Our proposed HCI models selects those samples which contribute the most additional representative information when labeled in each iteration of the query strategy (showing the names and sequence numbers of the samples to be annotated). Additionally, in this phase, the model selects the unlabeled samples with the greatest predictive disparity by calculating the Wasserstein Distance, Least Confidence, Entropy Sampling, and Random Sampling. (2) Central line correction: The selected samples in previous stage are then used for domain expert correction of the system-generated tracheal central lines in each training round. (3) Update training dataset: When domain experts are involved in each epoch of the DL model's training iterations, they update the training dataset with greater precision after each epoch, thereby enhancing the trustworthiness of the 'black box' DL model and improving the performance of models. (4) Model training: Proposed HCI model is trained using the updated training dataset and an enhanced version of existing UNet. Experimental results validate the effectiveness of this Human-Computer Interaction-based approaches, demonstrating that our proposed WD-UNet, LC-UNet, UUNet, RS-UNet achieve comparable or even superior performance than the state-of-the-art DL models, such as WD-UNet with only 15 %-35 % of the training data, leading to substantial reductions (65 %-85 % reduction of annotation effort) in physician annotation time.
Collapse
Affiliation(s)
- Shiyi Wang
- National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK
| | - Yang Nan
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, UK
| | - Sheng Zhang
- National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK
| | - Federico Felder
- Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, UK
| | - Xiaodan Xing
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, UK
| | - Yingying Fang
- National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, UK
| | - Javier Del Ser
- TECNALIA, Basque Research and Technology Alliance (BRTA), Derio 48160, Spain; Department of Communications Engineering, University of the Basque Country UPV/EHU, Bilbao 48013, Spain
| | - Simon L F Walsh
- National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, UK
| | - Guang Yang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, UK; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, UK; School of Biomedical Engineering & Imaging Sciences, King's College London, London WC2R 2LS, UK.
| |
Collapse
|
7
|
Nan Y, Ser JD, Tang Z, Tang P, Xing X, Fang Y, Herrera F, Pedrycz W, Walsh S, Yang G. Fuzzy Attention Neural Network to Tackle Discontinuity in Airway Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:7391-7404. [PMID: 37204954 DOI: 10.1109/tnnls.2023.3269223] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/21/2023]
Abstract
Airway segmentation is crucial for the examination, diagnosis, and prognosis of lung diseases, while its manual delineation is unduly burdensome. To alleviate this time-consuming and potentially subjective manual procedure, researchers have proposed methods to automatically segment airways from computerized tomography (CT) images. However, some small-sized airway branches (e.g., bronchus and terminal bronchioles) significantly aggravate the difficulty of automatic segmentation by machine learning models. In particular, the variance of voxel values and the severe data imbalance in airway branches make the computational module prone to discontinuous and false-negative predictions, especially for cohorts with different lung diseases. The attention mechanism has shown the capacity to segment complex structures, while fuzzy logic can reduce the uncertainty in feature representations. Therefore, the integration of deep attention networks and fuzzy theory, given by the fuzzy attention layer, should be an escalated solution for better generalization and robustness. This article presents an efficient method for airway segmentation, comprising a novel fuzzy attention neural network (FANN) and a comprehensive loss function to enhance the spatial continuity of airway segmentation. The deep fuzzy set is formulated by a set of voxels in the feature map and a learnable Gaussian membership function. Different from the existing attention mechanism, the proposed channel-specific fuzzy attention addresses the issue of heterogeneous features in different channels. Furthermore, a novel evaluation metric is proposed to assess both the continuity and completeness of airway structures. The efficiency, generalization, and robustness of the proposed method have been proved by training on normal lung disease while testing on datasets of lung cancer, COVID-19, and pulmonary fibrosis.
Collapse
|
8
|
Chang GH, Wu MY, Yen LH, Huang DY, Lin YH, Luo YR, Liu YD, Xu B, Leong KW, Lai WS, Chiang AS, Wang KC, Lin CH, Wang SL, Chu LA. Isotropic multi-scale neuronal reconstruction from high-ratio expansion microscopy with contrastive unsupervised deep generative models. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 244:107991. [PMID: 38185040 DOI: 10.1016/j.cmpb.2023.107991] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Revised: 12/10/2023] [Accepted: 12/19/2023] [Indexed: 01/09/2024]
Abstract
BACKGROUND AND OBJECTIVE Current methods for imaging reconstruction from high-ratio expansion microscopy (ExM) data are limited by anisotropic optical resolution and the requirement for extensive manual annotation, creating a significant bottleneck in the analysis of complex neuronal structures. METHODS We devised an innovative approach called the IsoGAN model, which utilizes a contrastive unsupervised generative adversarial network to sidestep these constraints. This model leverages multi-scale and isotropic neuron/protein/blood vessel morphology data to generate high-fidelity 3D representations of these structures, eliminating the need for rigorous manual annotation and supervision. The IsoGAN model introduces simplified structures with idealized morphologies as shape priors to ensure high consistency in the generated neuronal profiles across all points in space and scalability for arbitrarily large volumes. RESULTS The efficacy of the IsoGAN model in accurately reconstructing complex neuronal structures was quantitatively assessed by examining the consistency between the axial and lateral views and identifying a reduction in erroneous imaging artifacts. The IsoGAN model accurately reconstructed complex neuronal structures, as evidenced by the consistency between the axial and lateral views and a reduction in erroneous imaging artifacts, and can be further applied to various biological samples. CONCLUSION With its ability to generate detailed 3D neurons/proteins/blood vessel structures using significantly fewer axial view images, IsoGAN can streamline the process of imaging reconstruction while maintaining the necessary detail, offering a transformative solution to the existing limitations in high-throughput morphology analysis across different structures.
Collapse
Affiliation(s)
- Gary Han Chang
- Institute of Medical Device and Imaging, College of Medicine, National Taiwan University, Taipei, Taiwan, ROC; Graduate School of Advanced Technology, National Taiwan University, Taipei, Taiwan, ROC.
| | - Meng-Yun Wu
- Institute of Medical Device and Imaging, College of Medicine, National Taiwan University, Taipei, Taiwan, ROC
| | - Ling-Hui Yen
- Department of Biomedical Engineering and Environmental Sciences, National Tsing Hua University, Hsinchu, Taiwan, ROC; Brain Research Center, National Tsing Hua University, Hsinchu, Taiwan, ROC
| | - Da-Yu Huang
- Institute of Medical Device and Imaging, College of Medicine, National Taiwan University, Taipei, Taiwan, ROC
| | - Ya-Hui Lin
- Department of Biomedical Engineering and Environmental Sciences, National Tsing Hua University, Hsinchu, Taiwan, ROC; Brain Research Center, National Tsing Hua University, Hsinchu, Taiwan, ROC
| | - Yi-Ru Luo
- Department of Biomedical Engineering and Environmental Sciences, National Tsing Hua University, Hsinchu, Taiwan, ROC; Brain Research Center, National Tsing Hua University, Hsinchu, Taiwan, ROC
| | - Ya-Ding Liu
- Department of Biomedical Engineering and Environmental Sciences, National Tsing Hua University, Hsinchu, Taiwan, ROC; Brain Research Center, National Tsing Hua University, Hsinchu, Taiwan, ROC
| | - Bin Xu
- Department of Psychiatry, Columbia University, New York, NY 10032, USA
| | - Kam W Leong
- Department of Biomedical Engineering, Columbia University, New York, NY 10032, USA
| | - Wen-Sung Lai
- Department of Psychology, National Taiwan University, Taipei, Taiwan, ROC
| | - Ann-Shyn Chiang
- Brain Research Center, National Tsing Hua University, Hsinchu, Taiwan, ROC; Institute of System Neuroscience, National Tsing Hua University, Hsinchu, Taiwan, ROC
| | - Kuo-Chuan Wang
- Department of Neurosurgery, National Taiwan University Hospital, Taipei, Taiwan, ROC
| | - Chin-Hsien Lin
- Department of Neurosurgery, National Taiwan University Hospital, Taipei, Taiwan, ROC
| | - Shih-Luen Wang
- Department of Physics and Center for Interdisciplinary Research on Complex Systems, Northeastern University, Boston, MA 02115, USA
| | - Li-An Chu
- Department of Biomedical Engineering and Environmental Sciences, National Tsing Hua University, Hsinchu, Taiwan, ROC; Brain Research Center, National Tsing Hua University, Hsinchu, Taiwan, ROC.
| |
Collapse
|
9
|
Su H, Zhao D, Heidari AA, Cai Z, Chen H, Zhu J. Kernel extreme learning with harmonized bat algorithm for prediction of pyrene toxicity in rats. Basic Clin Pharmacol Toxicol 2024; 134:250-271. [PMID: 37945549 DOI: 10.1111/bcpt.13959] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2022] [Revised: 10/12/2023] [Accepted: 11/07/2023] [Indexed: 11/12/2023]
Abstract
Polycyclic aromatic hydrocarbons (PAHs) are organic pollutants and manufactured substances conferring toxicity to human health. The present study investigated whether pyrene, a type of PAH, harms rats. Our research provides an effective feature selection strategy for the animal dataset from Wenzhou Medical University's Experimental Animal Center to thoroughly examine the impacts of PAH toxicity on rat features. Initially, we devised a high-performance optimization method (SCBA) and added the Sobol sequence, vertical crossover and horizontal crossover mechanisms to the bat algorithm (BA). The SCBA-KELM model, which combines SCBA with the kernel extreme learning machine model (KELM), has excellent accuracy and high stability for selecting features. Benchmark function tests are then used in this research to verify the overall optimization performance of SCBA. In this paper, the feature selection performance of SCBA-KELM is verified using various comparative experiments. According to the results, the features of the genes PXR, CAR, CYP2B1/2 and CYP1A1/2 have the most impact on rats. The SCBA-KELM model's classification performance for the gene dataset was 100%, and the model's precision value for the public dataset was around 96%, as determined by the classification index. In conclusion, the model utilized in this research is anticipated to be a reliable and valuable approach for toxicological classification and assessment.
Collapse
Affiliation(s)
- Hang Su
- College of Computer Science and Technology, Changchun Normal University, Changchun, China
| | - Dong Zhao
- College of Computer Science and Technology, Changchun Normal University, Changchun, China
| | - Ali Asghar Heidari
- School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran, Tehran, Iran
| | - Zhennao Cai
- Key Laboratory of Intelligent Informatics for Safety & Emergency of Zhejiang Province, Wenzhou University, Wenzhou, China
| | - Huiling Chen
- Key Laboratory of Intelligent Informatics for Safety & Emergency of Zhejiang Province, Wenzhou University, Wenzhou, China
| | - Jiayin Zhu
- Laboratory Animal Center, Wenzhou Medical University, Wenzhou, China
| |
Collapse
|
10
|
Materka A, Jurek J. Using Deep Learning and B-Splines to Model Blood Vessel Lumen from 3D Images. SENSORS (BASEL, SWITZERLAND) 2024; 24:846. [PMID: 38339562 PMCID: PMC10857344 DOI: 10.3390/s24030846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/23/2023] [Revised: 01/20/2024] [Accepted: 01/26/2024] [Indexed: 02/12/2024]
Abstract
Accurate geometric modeling of blood vessel lumen from 3D images is crucial for vessel quantification as part of the diagnosis, treatment, and monitoring of vascular diseases. Our method, unlike other approaches which assume a circular or elliptical vessel cross-section, employs parametric B-splines combined with image formation system equations to accurately localize the highly curved lumen boundaries. This approach avoids the need for image segmentation, which may reduce the localization accuracy due to spatial discretization. We demonstrate that the model parameters can be reliably identified by a feedforward neural network which, driven by the cross-section images, predicts the parameter values many times faster than a reference least-squares (LS) model fitting algorithm. We present and discuss two example applications, modeling the lower extremities of artery-vein complexes visualized in steady-state contrast-enhanced magnetic resonance images (MRI) and the coronary arteries pictured in computed tomography angiograms (CTA). Beyond applications in medical diagnosis, blood-flow simulation and vessel-phantom design, the method can serve as a tool for automated annotation of image datasets to train machine-learning algorithms.
Collapse
Affiliation(s)
- Andrzej Materka
- Institute of Electronics, Lodz University of Technology, 90-924 Lodz, Poland;
| | | |
Collapse
|
11
|
Zhang M, Wu Y, Zhang H, Qin Y, Zheng H, Tang W, Arnold C, Pei C, Yu P, Nan Y, Yang G, Walsh S, Marshall DC, Komorowski M, Wang P, Guo D, Jin D, Wu Y, Zhao S, Chang R, Zhang B, Lu X, Qayyum A, Mazher M, Su Q, Wu Y, Liu Y, Zhu Y, Yang J, Pakzad A, Rangelov B, Estepar RSJ, Espinosa CC, Sun J, Yang GZ, Gu Y. Multi-site, Multi-domain Airway Tree Modeling. Med Image Anal 2023; 90:102957. [PMID: 37716199 DOI: 10.1016/j.media.2023.102957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Revised: 06/07/2023] [Accepted: 09/04/2023] [Indexed: 09/18/2023]
Abstract
Open international challenges are becoming the de facto standard for assessing computer vision and image analysis algorithms. In recent years, new methods have extended the reach of pulmonary airway segmentation that is closer to the limit of image resolution. Since EXACT'09 pulmonary airway segmentation, limited effort has been directed to the quantitative comparison of newly emerged algorithms driven by the maturity of deep learning based approaches and extensive clinical efforts for resolving finer details of distal airways for early intervention of pulmonary diseases. Thus far, public annotated datasets are extremely limited, hindering the development of data-driven methods and detailed performance evaluation of new algorithms. To provide a benchmark for the medical imaging community, we organized the Multi-site, Multi-domain Airway Tree Modeling (ATM'22), which was held as an official challenge event during the MICCAI 2022 conference. ATM'22 provides large-scale CT scans with detailed pulmonary airway annotation, including 500 CT scans (300 for training, 50 for validation, and 150 for testing). The dataset was collected from different sites and it further included a portion of noisy COVID-19 CTs with ground-glass opacity and consolidation. Twenty-three teams participated in the entire phase of the challenge and the algorithms for the top ten teams are reviewed in this paper. Both quantitative and qualitative results revealed that deep learning models embedded with the topological continuity enhancement achieved superior performance in general. ATM'22 challenge holds as an open-call design, the training data and the gold standard evaluation are available upon successful registration via its homepage (https://atm22.grand-challenge.org/).
Collapse
Affiliation(s)
- Minghui Zhang
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China; Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, 200240, China; Department of Automation, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yangqian Wu
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China; Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, 200240, China; Department of Automation, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Hanxiao Zhang
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yulei Qin
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Hao Zheng
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Wen Tang
- InferVision Medical Technology Co., Ltd., Beijing, China
| | | | - Chenhao Pei
- InferVision Medical Technology Co., Ltd., Beijing, China
| | - Pengxin Yu
- InferVision Medical Technology Co., Ltd., Beijing, China
| | - Yang Nan
- Imperial College London, London, UK
| | | | | | | | | | - Puyang Wang
- Alibaba DAMO Academy, 969 West Wen Yi Road, Hangzhou, Zhejiang, China
| | - Dazhou Guo
- Alibaba DAMO Academy USA, 860 Washington Street, 8F, NY, USA
| | - Dakai Jin
- Alibaba DAMO Academy USA, 860 Washington Street, 8F, NY, USA
| | - Ya'nan Wu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Shuiqing Zhao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Runsheng Chang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Boyu Zhang
- A.I R&D Center, Sanmed Biotech Inc., No. 266 Tongchang Road, Xiangzhou District, Zhuhai, Guangdong, China
| | - Xing Lu
- A.I R&D Center, Sanmed Biotech Inc., T220 Trade st. SanDiego, CA, USA
| | - Abdul Qayyum
- ENIB, UMR CNRS 6285 LabSTICC, Brest, 29238, France
| | - Moona Mazher
- Department of Computer Engineering and Mathematics, University Rovira I Virgili, Tarragona, Spain
| | - Qi Su
- Shanghai Jiao Tong University, Shanghai, China
| | - Yonghuang Wu
- School of Information Science and Technology, Fudan University, Shanghai, China
| | - Ying'ao Liu
- University of Science and Technology of China, Hefei, Anhui, China
| | | | - Jiancheng Yang
- Dianei Technology, Shanghai, China; EPFL, Lausanne, Switzerland
| | - Ashkan Pakzad
- Medical Physics and Biomedical Engineering Department, University College London, London, UK
| | - Bojidar Rangelov
- Center for Medical Image Computing, University College London, London, UK
| | | | | | - Jiayuan Sun
- Department of Respiratory and Critical Care Medicine, Department of Respiratory Endoscopy, Shanghai Chest Hospital, Shanghai, China.
| | - Guang-Zhong Yang
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China.
| | - Yun Gu
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, 200240, China; Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, 200240, China; Department of Automation, Shanghai Jiao Tong University, Shanghai, 200240, China.
| |
Collapse
|
12
|
Cui C, Yang H, Wang Y, Zhao S, Asad Z, Coburn LA, Wilson KT, Landman BA, Huo Y. Deep multimodal fusion of image and non-image data in disease diagnosis and prognosis: a review. PROGRESS IN BIOMEDICAL ENGINEERING (BRISTOL, ENGLAND) 2023; 5:10.1088/2516-1091/acc2fe. [PMID: 37360402 PMCID: PMC10288577 DOI: 10.1088/2516-1091/acc2fe] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/28/2023]
Abstract
The rapid development of diagnostic technologies in healthcare is leading to higher requirements for physicians to handle and integrate the heterogeneous, yet complementary data that are produced during routine practice. For instance, the personalized diagnosis and treatment planning for a single cancer patient relies on various images (e.g. radiology, pathology and camera images) and non-image data (e.g. clinical data and genomic data). However, such decision-making procedures can be subjective, qualitative, and have large inter-subject variabilities. With the recent advances in multimodal deep learning technologies, an increasingly large number of efforts have been devoted to a key question: how do we extract and aggregate multimodal information to ultimately provide more objective, quantitative computer-aided clinical decision making? This paper reviews the recent studies on dealing with such a question. Briefly, this review will include the (a) overview of current multimodal learning workflows, (b) summarization of multimodal fusion methods, (c) discussion of the performance, (d) applications in disease diagnosis and prognosis, and (e) challenges and future directions.
Collapse
Affiliation(s)
- Can Cui
- Department of Computer Science, Vanderbilt University, Nashville, TN 37235, United States of America
| | - Haichun Yang
- Department of Pathology, Microbiology and Immunology, Vanderbilt University Medical Center, Nashville, TN 37215, United States of America
| | - Yaohong Wang
- Department of Pathology, Microbiology and Immunology, Vanderbilt University Medical Center, Nashville, TN 37215, United States of America
| | - Shilin Zhao
- Department of Biostatistics, Vanderbilt University Medical Center, Nashville, TN 37215, United States of America
| | - Zuhayr Asad
- Department of Computer Science, Vanderbilt University, Nashville, TN 37235, United States of America
| | - Lori A Coburn
- Division of Gastroenterology Hepatology, and Nutrition, Department of Medicine, Vanderbilt University Medical Center, Nashville, TN 37232, United States of America
- Veterans Affairs Tennessee Valley Healthcare System, Nashville, TN 37212, United States of America
| | - Keith T Wilson
- Department of Pathology, Microbiology and Immunology, Vanderbilt University Medical Center, Nashville, TN 37215, United States of America
- Division of Gastroenterology Hepatology, and Nutrition, Department of Medicine, Vanderbilt University Medical Center, Nashville, TN 37232, United States of America
- Veterans Affairs Tennessee Valley Healthcare System, Nashville, TN 37212, United States of America
| | - Bennett A Landman
- Department of Computer Science, Vanderbilt University, Nashville, TN 37235, United States of America
- Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN 37235, United States of America
| | - Yuankai Huo
- Department of Computer Science, Vanderbilt University, Nashville, TN 37235, United States of America
- Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN 37235, United States of America
| |
Collapse
|
13
|
Wang Y, Fan J, Tong Y, Wang L, Wang L, Weng C, Lai C, Song J, Zhang W. Bioinformatics analysis of ferroptosis-related gene AKR1C3 as a potential biomarker of asthma and its identification in BEAS-2B cells. Comput Biol Med 2023; 158:106740. [PMID: 36996663 DOI: 10.1016/j.compbiomed.2023.106740] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2022] [Revised: 01/24/2023] [Accepted: 03/02/2023] [Indexed: 03/17/2023]
Abstract
Ferroptosis is a newly discovered type of cell death and has recently been shown to be associated with asthma. However, the relationship between them at the genetic level has not been elucidated via informatics analysis. In this study, bioinformatics analyses are conducted using asthma and ferroptosis datasets to identify candidate ferroptosis-related genes using the R software. Weighted gene co-expression network analysis is performed to identify co-expressed genes. Protein-protein interaction networks, the Kyoto encyclopedia of genes and genomes, and gene ontology enrichment analysis are used to identify the potential functions of the candidate genes. We experimentally validate the results of our analysis using small interfering RNAs and plasmids to silence and upregulate the expression of the candidate gene in human bronchial epithelial cells (BEAS-2B). The ferroptosis signature levels are examined. Bioinformatics analysis of the asthma dataset GDS4896 shows that the level of the aldo-keto reductase family 1 member C3 (AKR1C3) gene in the peripheral blood of patients with severe therapy-resistant asthma and controlled persistent mild asthma (MA) is significantly upregulated. The AUC values for asthma diagnosis and MA are 0.823 and 0.915, respectively. The diagnostic value of AKR1C3 is verified using the GSE64913 dataset. The gene module of AKR1C3 is evident in MA and functions through redox reactions and metabolic processes. Ferroptosis indicators are downregulated by the overexpression of AKR1C3 and upregulated by silencing AKR1C3. The ferroptosis-related gene AKR1C3 can be used as a diagnostic biomarker for asthma, particularly for MA, and regulates ferroptosis in BEAS-2B cells.
Collapse
Affiliation(s)
- Yufei Wang
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Junwen Fan
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Yu Tong
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Lei Wang
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Lingya Wang
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Cuiye Weng
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China; Department of Neonatology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Chuqiao Lai
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China
| | - Jingjing Song
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China.
| | - Weixi Zhang
- Department of Pediatric Allergy and Immunology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, 325027, China.
| |
Collapse
|
14
|
Li H, Nan Y, Del Ser J, Yang G. Large-Kernel Attention for 3D Medical Image Segmentation. Cognit Comput 2023; 16:2063-2077. [PMID: 38974012 PMCID: PMC11226511 DOI: 10.1007/s12559-023-10126-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Accepted: 02/09/2023] [Indexed: 03/03/2023]
Abstract
Automated segmentation of multiple organs and tumors from 3D medical images such as magnetic resonance imaging (MRI) and computed tomography (CT) scans using deep learning methods can aid in diagnosing and treating cancer. However, organs often overlap and are complexly connected, characterized by extensive anatomical variation and low contrast. In addition, the diversity of tumor shape, location, and appearance, coupled with the dominance of background voxels, makes accurate 3D medical image segmentation difficult. In this paper, a novel 3D large-kernel (LK) attention module is proposed to address these problems to achieve accurate multi-organ segmentation and tumor segmentation. The advantages of biologically inspired self-attention and convolution are combined in the proposed LK attention module, including local contextual information, long-range dependencies, and channel adaptation. The module also decomposes the LK convolution to optimize the computational cost and can be easily incorporated into CNNs such as U-Net. Comprehensive ablation experiments demonstrated the feasibility of convolutional decomposition and explored the most efficient and effective network design. Among them, the best Mid-type 3D LK attention-based U-Net network was evaluated on CT-ORG and BraTS 2020 datasets, achieving state-of-the-art segmentation performance when compared to avant-garde CNN and Transformer-based methods for medical image segmentation. The performance improvement due to the proposed 3D LK attention module was statistically validated.
Collapse
Affiliation(s)
- Hao Li
- National Heart and Lung Institute, Faculty of Medicine, Imperial College London, London, UK
- Department of Bioengineering, Faculty of Engineering, Imperial College London, London, UK
| | - Yang Nan
- National Heart and Lung Institute, Faculty of Medicine, Imperial College London, London, UK
| | - Javier Del Ser
- TECNALIA, Basque Research & Technology Alliance (BRTA), Derio, Spain
- University of the Basque Country (UPV/EHU), Bilbao, Spain
| | - Guang Yang
- National Heart and Lung Institute, Faculty of Medicine, Imperial College London, London, UK
- Royal Brompton Hospital, London, UK
| |
Collapse
|