1
|
Hu Y, Mu N, Liu L, Zhang L, Jiang J, Li X. Slimmable transformer with hybrid axial-attention for medical image segmentation. Comput Biol Med 2024; 173:108370. [PMID: 38564854 DOI: 10.1016/j.compbiomed.2024.108370] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2023] [Revised: 03/14/2024] [Accepted: 03/24/2024] [Indexed: 04/04/2024]
Abstract
The transformer architecture has achieved remarkable success in medical image analysis owing to its powerful capability for capturing long-range dependencies. However, due to the lack of intrinsic inductive bias in modeling visual structural information, the transformer generally requires a large-scale pre-training schedule, limiting the clinical applications over expensive small-scale medical data. To this end, we propose a slimmable transformer to explore intrinsic inductive bias via position information for medical image segmentation. Specifically, we empirically investigate how different position encoding strategies affect the prediction quality of the region of interest (ROI) and observe that ROIs are sensitive to different position encoding strategies. Motivated by this, we present a novel Hybrid Axial-Attention (HAA) that can be equipped with pixel-level spatial structure and relative position information as inductive bias. Moreover, we introduce a gating mechanism to achieve efficient feature selection and further improve the representation quality over small-scale datasets. Experiments on LGG and COVID-19 datasets prove the superiority of our method over the baseline and previous works. Internal workflow visualization with interpretability is conducted to validate our success better; the proposed slimmable transformer has the potential to be further developed into a visual software tool for improving computer-aided lesion diagnosis and treatment planning.
Collapse
Affiliation(s)
- Yiyue Hu
- College of Computer Science, Sichuan Normal University, Chengdu, 610101, China
| | - Nan Mu
- College of Computer Science, Sichuan Normal University, Chengdu, 610101, China; Department of Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA; Visual Computing and Virtual Reality Key Laboratory of Sichuan, Sichuan Normal University, Chengdu, 610068, China.
| | - Lei Liu
- School of Science and Engineering, The Chinese University of Hong Kong, Shenzhen, Shenzhen, 518172, China
| | - Lei Zhang
- College of Computer Science, Sichuan Normal University, Chengdu, 610101, China
| | - Jingfeng Jiang
- Department of Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA
| | - Xiaoning Li
- College of Computer Science, Sichuan Normal University, Chengdu, 610101, China; Education Big Data Collaborative Innovation Center of Sichuan 2011, Chengdu, 610101, China
| |
Collapse
|
2
|
Liang J, Feng J, Lin Z, Wei J, Luo X, Wang QM, He B, Chen H, Ye Y. Research on prognostic risk assessment model for acute ischemic stroke based on imaging and multidimensional data. Front Neurol 2023; 14:1294723. [PMID: 38192576 PMCID: PMC10773779 DOI: 10.3389/fneur.2023.1294723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 11/30/2023] [Indexed: 01/10/2024] Open
Abstract
Accurately assessing the prognostic outcomes of patients with acute ischemic stroke and adjusting treatment plans in a timely manner for those with poor prognosis is crucial for intervening in modifiable risk factors. However, there is still controversy regarding the correlation between imaging-based predictions of complications in acute ischemic stroke. To address this, we developed a cross-modal attention module for integrating multidimensional data, including clinical information, imaging features, treatment plans, prognosis, and complications, to achieve complementary advantages. The fused features preserve magnetic resonance imaging (MRI) characteristics while supplementing clinical relevant information, providing a more comprehensive and informative basis for clinical diagnosis and treatment. The proposed framework based on multidimensional data for activity of daily living (ADL) scoring in patients with acute ischemic stroke demonstrates higher accuracy compared to other state-of-the-art network models, and ablation experiments confirm the effectiveness of each module in the framework.
Collapse
Affiliation(s)
- Jiabin Liang
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
- Graduate School, Guangzhou University of Chinese Medicine, Guangzhou, China
- Medical Imaging Institute of Panyu, Guangzhou, China
| | - Jie Feng
- Radiology Department of Sun Yat-sen Memorial Hospital, Sun Yat-sen University, Guangzhou, China
| | - Zhijie Lin
- Laboratory for Intelligent Information Processing, Guangdong University of Technology, Guangzhou, China
| | - Jinbo Wei
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
| | - Xun Luo
- Kerry Rehabilitation Medicine Research Institute, Shenzhen, China
| | - Qing Mei Wang
- Stroke Biological Recovery Laboratory, Spaulding Rehabilitation Hospital, Teaching Affiliate of Harvard Medical School, Charlestown, MA, United States
| | - Bingjie He
- Panyu Health Management Center, Guangzhou, China
| | - Hanwei Chen
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
- Medical Imaging Institute of Panyu, Guangzhou, China
- Panyu Health Management Center, Guangzhou, China
| | - Yufeng Ye
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
- Medical Imaging Institute of Panyu, Guangzhou, China
| |
Collapse
|
3
|
Mu N, Lyu Z, Zhang X, McBane R, Pandey AS, Jiang J. Exploring a frequency-domain attention-guided cascade U-Net: Towards spatially tunable segmentation of vasculature. Comput Biol Med 2023; 167:107648. [PMID: 37931523 PMCID: PMC10841687 DOI: 10.1016/j.compbiomed.2023.107648] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Revised: 10/14/2023] [Accepted: 10/27/2023] [Indexed: 11/08/2023]
Abstract
Developing fully automatic and highly accurate medical image segmentation methods is critically important for vascular disease diagnosis and treatment planning. Although advances in convolutional neural networks (CNNs) have spawned an array of automatic segmentation models converging to saturated high performance, none have explored whether CNNs can achieve (spatially) tunable segmentation. As a result, we propose multiple attention modules from a frequency-domain perspective to construct a unified CNN architecture for segmenting vasculature with desired (spatial) scales. The proposed CNN architecture is named frequency-domain attention-guided cascaded U-Net (FACU-Net). Specifically, FACU-Net contains two innovative components: (1) a frequency-domain-based channel attention module that adaptively tunes channel-wise feature responses and (2) a frequency-domain-based spatial attention module that enables the deep network to concentrate on foreground regions of interest (ROIs) effectively. Furthermore, we devised a novel frequency-domain-based content attention module to enhance or weaken the high (spatial) frequency information, allowing us to strengthen or eliminate vessels of interest. Extensive experiments using clinical data from patients with intracranial aneurysms (IA) and abdominal aortic aneurysms (AAA) demonstrated that the proposed FACU-Net met its design goal. In addition, we further investigated the association between varying (spatial) frequency components and the desirable vessel size/scale attributes. In summary, our preliminary findings are encouraging, and further developments may lead to deployable image segmentation models that are spatially tunable for clinical applications.
Collapse
Affiliation(s)
- Nan Mu
- Department of Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA; School of Computer Science, Sichuan Normal University, Chengdu, 610101, China
| | - Zonghan Lyu
- Department of Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA
| | | | | | - Aditya S Pandey
- Department of Neurosurgery, University of Michigan, Ann Arbor, MI, 48105, USA
| | - Jingfeng Jiang
- Department of Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA; Center for Biocomputing and Digital Health, Health Research Institute and Institute of Computing and Cybernetics, Michigan Technological University, Houghton, MI, 49931, USA.
| |
Collapse
|
4
|
Mu N, Guo J, Wang R. Automated polyp segmentation based on a multi-distance feature dissimilarity-guided fully convolutional network. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:20116-20134. [PMID: 38052639 DOI: 10.3934/mbe.2023891] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
Colorectal malignancies often arise from adenomatous polyps, which typically begin as solitary, asymptomatic growths before progressing to malignancy. Colonoscopy is widely recognized as a highly efficacious clinical polyp detection method, offering valuable visual data that facilitates precise identification and subsequent removal of these tumors. Nevertheless, accurately segmenting individual polyps poses a considerable difficulty because polyps exhibit intricate and changeable characteristics, including shape, size, color, quantity and growth context during different stages. The presence of similar contextual structures around polyps significantly hampers the performance of commonly used convolutional neural network (CNN)-based automatic detection models to accurately capture valid polyp features, and these large receptive field CNN models often overlook the details of small polyps, which leads to the occurrence of false detections and missed detections. To tackle these challenges, we introduce a novel approach for automatic polyp segmentation, known as the multi-distance feature dissimilarity-guided fully convolutional network. This approach comprises three essential components, i.e., an encoder-decoder, a multi-distance difference (MDD) module and a hybrid loss (HL) module. Specifically, the MDD module primarily employs a multi-layer feature subtraction (MLFS) strategy to propagate features from the encoder to the decoder, which focuses on extracting information differences between neighboring layers' features at short distances, and both short and long-distance feature differences across layers. Drawing inspiration from pyramids, the MDD module effectively acquires discriminative features from neighboring layers or across layers in a continuous manner, which helps to strengthen feature complementary across different layers. The HL module is responsible for supervising the feature maps extracted at each layer of the network to improve prediction accuracy. Our experimental results on four challenge datasets demonstrate that the proposed approach exhibits superior automatic polyp performance in terms of the six evaluation criteria compared to five current state-of-the-art approaches.
Collapse
Affiliation(s)
- Nan Mu
- College of Computer Science, Sichuan Normal University, Chengdu 610101, China
- Visual Computing and Virtual Reality Key Laboratory of Sichuan, Sichuan Normal University, Chengdu 610068, China
- Education Big Data Collaborative Innovation Center of Sichuan 2011, Chengdu 610101, China
| | - Jinjia Guo
- Chongqing University-University of Cincinnati Joint Co-op Institution, Chongqing University, Chongqing 400044, China
| | - Rong Wang
- College of Computer Science, Sichuan Normal University, Chengdu 610101, China
- Visual Computing and Virtual Reality Key Laboratory of Sichuan, Sichuan Normal University, Chengdu 610068, China
- Education Big Data Collaborative Innovation Center of Sichuan 2011, Chengdu 610101, China
| |
Collapse
|
5
|
Chen M, Yi S, Yang M, Yang Z, Zhang X. UNet segmentation network of COVID-19 CT images with multi-scale attention. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:16762-16785. [PMID: 37920033 DOI: 10.3934/mbe.2023747] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/04/2023]
Abstract
In recent years, the global outbreak of COVID-19 has posed an extremely serious life-safety risk to humans, and in order to maximize the diagnostic efficiency of physicians, it is extremely valuable to investigate the methods of lesion segmentation in images of COVID-19. Aiming at the problems of existing deep learning models, such as low segmentation accuracy, poor model generalization performance, large model parameters and difficult deployment, we propose an UNet segmentation network integrating multi-scale attention for COVID-19 CT images. Specifically, the UNet network model is utilized as the base network, and the structure of multi-scale convolutional attention is proposed in the encoder stage to enhance the network's ability to capture multi-scale information. Second, a local channel attention module is proposed to extract spatial information by modeling local relationships to generate channel domain weights, to supplement detailed information about the target region to reduce information redundancy and to enhance important information. Moreover, the network model encoder segment uses the Meta-ACON activation function to avoid the overfitting phenomenon of the model and to improve the model's representational ability. A large number of experimental results on publicly available mixed data sets show that compared with the current mainstream image segmentation algorithms, the pro-posed method can more effectively improve the accuracy and generalization performance of COVID-19 lesions segmentation and provide help for medical diagnosis and analysis.
Collapse
Affiliation(s)
- Mingju Chen
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| | - Sihang Yi
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| | - Mei Yang
- Zigong Third People's Hospital, Zigong 643000, China
| | - Zhiwen Yang
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| | - Xingyue Zhang
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| |
Collapse
|
6
|
Mu N, Lyu Z, Rezaeitaleshmahalleh M, Zhang X, Rasmussen T, McBane R, Jiang J. Automatic segmentation of abdominal aortic aneurysms from CT angiography using a context-aware cascaded U-Net. Comput Biol Med 2023; 158:106569. [PMID: 36989747 PMCID: PMC10625464 DOI: 10.1016/j.compbiomed.2023.106569] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2022] [Revised: 11/22/2022] [Accepted: 01/22/2023] [Indexed: 01/24/2023]
Abstract
We delineate abdominal aortic aneurysms, including lumen and intraluminal thrombosis (ILT), from contrast-enhanced computed tomography angiography (CTA) data in 70 patients with complete automation. A novel context-aware cascaded U-Net configuration enables automated image segmentation. Notably, auto-context structure, in conjunction with dilated convolutions, anisotropic context module, hierarchical supervision, and a multi-class loss function, are proposed to improve the delineation of ILT in an unbalanced, low-contrast multi-class labeling problem. A quantitative analysis shows that the automated image segmentation produces comparable results with trained human users (e.g., DICE scores of 0.945 and 0.804 for lumen and ILT, respectively). Resultant morphological metrics (e.g., volume, surface area, etc.) are highly correlated to those parameters generated by trained human users. In conclusion, the proposed automated multi-class image segmentation tool has the potential to be further developed as a translational software tool that can be used to improve the clinical management of AAAs.
Collapse
Affiliation(s)
- Nan Mu
- Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA
| | - Zonghan Lyu
- Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA
| | | | | | | | | | - Jingfeng Jiang
- Biomedical Engineering, Michigan Technological University, Houghton, MI, 49931, USA; Center for Biocomputing and Digital Health, Health Research Institute, Institute of Computing and Cybernetics, Michigan Technological University, Houghton, MI, 49931, USA.
| |
Collapse
|
7
|
Sun Y, Li X, Liu Y, Yuan Z, Wang J, Shi C. A lightweight dual-path cascaded network for vessel segmentation in fundus image. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:10790-10814. [PMID: 37322961 DOI: 10.3934/mbe.2023479] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
Automatic and fast segmentation of retinal vessels in fundus images is a prerequisite in clinical ophthalmic diseases; however, the high model complexity and low segmentation accuracy still limit its application. This paper proposes a lightweight dual-path cascaded network (LDPC-Net) for automatic and fast vessel segmentation. We designed a dual-path cascaded network via two U-shaped structures. Firstly, we employed a structured discarding (SD) convolution module to alleviate the over-fitting problem in both codec parts. Secondly, we introduced the depthwise separable convolution (DSC) technique to reduce the parameter amount of the model. Thirdly, a residual atrous spatial pyramid pooling (ResASPP) model is constructed in the connection layer to aggregate multi-scale information effectively. Finally, we performed comparative experiments on three public datasets. Experimental results show that the proposed method achieved superior performance on the accuracy, connectivity, and parameter quantity, thus proving that it can be a promising lightweight assisted tool for ophthalmic diseases.
Collapse
Affiliation(s)
- Yanxia Sun
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Xiang Li
- Shenzhen Institute for Advanced Study, University of Electronic Science and Technology of China, Shenzhen 518110, China
| | - Yuechang Liu
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Zhongzheng Yuan
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Jinke Wang
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Changfa Shi
- Mobile E-business Collaborative Innovation Center of Hunan Province, Hunan University of Technology and Business, Changsha 410205, China
| |
Collapse
|
8
|
Li Z, Huang J, Tong X, Zhang C, Lu J, Zhang W, Song A, Ji S. GL-FusionNet: Fusing global and local features to classify deep and superficial partial thickness burn. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:10153-10173. [PMID: 37322927 DOI: 10.3934/mbe.2023445] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
Burns constitute one of the most common injuries in the world, and they can be very painful for the patient. Especially in the judgment of superficial partial thickness burns and deep partial thickness burns, many inexperienced clinicians are easily confused. Therefore, in order to make burn depth classification automated as well as accurate, we have introduced the deep learning method. This methodology uses a U-Net to segment burn wounds. On this basis, a new thickness burn classification model that fuses global and local features (GL-FusionNet) is proposed. For the thickness burn classification model, we use a ResNet50 to extract local features, use a ResNet101 to extract global features, and finally implement the add method to perform feature fusion and obtain the deep partial or superficial partial thickness burn classification results. Burns images are collected clinically, and they are segmented and labeled by professional physicians. Among the segmentation methods, the U-Net used achieved a Dice score of 85.352 and IoU score of 83.916, which are the best results among all of the comparative experiments. In the classification model, different existing classification networks are mainly used, as well as a fusion strategy and feature extraction method that are adjusted to conduct experiments; the proposed fusion network model also achieved the best results. Our method yielded the following: accuracy of 93.523, recall of 93.67, precision of 93.51, and F1-score of 93.513. In addition, the proposed method can quickly complete the auxiliary diagnosis of the wound in the clinic, which can greatly improve the efficiency of the initial diagnosis of burns and the nursing care of clinical medical staff.
Collapse
Affiliation(s)
- Zhiwei Li
- School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
| | - Jie Huang
- Department of Burn Surgery, the First Affiliated Hospital of Naval Medical University, Shanghai 200444, China
| | - Xirui Tong
- Department of Burn Surgery, the First Affiliated Hospital of Naval Medical University, Shanghai 200444, China
| | - Chenbei Zhang
- School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
| | - Jianyu Lu
- Department of Burn Surgery, the First Affiliated Hospital of Naval Medical University, Shanghai 200444, China
| | - Wei Zhang
- Department of Burn Surgery, the First Affiliated Hospital of Naval Medical University, Shanghai 200444, China
| | - Anping Song
- School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
| | - Shizhao Ji
- Department of Burn Surgery, the First Affiliated Hospital of Naval Medical University, Shanghai 200444, China
| |
Collapse
|
9
|
PDAtt-Unet: Pyramid Dual-Decoder Attention Unet for Covid-19 infection segmentation from CT-scans. Med Image Anal 2023; 86:102797. [PMID: 36966605 PMCID: PMC10027962 DOI: 10.1016/j.media.2023.102797] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Revised: 01/10/2023] [Accepted: 03/08/2023] [Indexed: 03/23/2023]
Abstract
Since the emergence of the Covid-19 pandemic in late 2019, medical imaging has been widely used to analyse this disease. Indeed, CT-scans of the lungs can help diagnose, detect, and quantify Covid-19 infection. In this paper, we address the segmentation of Covid-19 infection from CT-scans. To improve the performance of the Att-Unet architecture and maximize the use of the Attention Gate, we propose the PAtt-Unet and DAtt-Unet architectures. PAtt-Unet aims to exploit the input pyramids to preserve the spatial awareness in all of the encoder layers. On the other hand, DAtt-Unet is designed to guide the segmentation of Covid-19 infection inside the lung lobes. We also propose to combine these two architectures into a single one, which we refer to as PDAtt-Unet. To overcome the blurry boundary pixels segmentation of Covid-19 infection, we propose a hybrid loss function. The proposed architectures were tested on four datasets with two evaluation scenarios (intra and cross datasets). Experimental results showed that both PAtt-Unet and DAtt-Unet improve the performance of Att-Unet in segmenting Covid-19 infections. Moreover, the combination architecture PDAtt-Unet led to further improvement. To Compare with other methods, three baseline segmentation architectures (Unet, Unet++, and Att-Unet) and three state-of-the-art architectures (InfNet, SCOATNet, and nCoVSegNet) were tested. The comparison showed the superiority of the proposed PDAtt-Unet trained with the proposed hybrid loss (PDEAtt-Unet) over all other methods. Moreover, PDEAtt-Unet is able to overcome various challenges in segmenting Covid-19 infections in four datasets and two evaluation scenarios.
Collapse
|
10
|
Mu N, Lyu Z, Rezaeitaleshmahalleh M, Tang J, Jiang J. An attention residual u-net with differential preprocessing and geometric postprocessing: Learning how to segment vasculature including intracranial aneurysms. Med Image Anal 2023; 84:102697. [PMID: 36462374 PMCID: PMC9830590 DOI: 10.1016/j.media.2022.102697] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2022] [Revised: 10/05/2022] [Accepted: 11/17/2022] [Indexed: 11/20/2022]
Abstract
OBJECTIVE Intracranial aneurysms (IA) are lethal, with high morbidity and mortality rates. Reliable, rapid, and accurate segmentation of IAs and their adjacent vasculature from medical imaging data is important to improve the clinical management of patients with IAs. However, due to the blurred boundaries and complex structure of IAs and overlapping with brain tissue or other cerebral arteries, image segmentation of IAs remains challenging. This study aimed to develop an attention residual U-Net (ARU-Net) architecture with differential preprocessing and geometric postprocessing for automatic segmentation of IAs and their adjacent arteries in conjunction with 3D rotational angiography (3DRA) images. METHODS The proposed ARU-Net followed the classic U-Net framework with the following key enhancements. First, we preprocessed the 3DRA images based on boundary enhancement to capture more contour information and enhance the presence of small vessels. Second, we introduced the long skip connections of the attention gate at each layer of the fully convolutional decoder-encoder structure to emphasize the field of view (FOV) for IAs. Third, residual-based short skip connections were also embedded in each layer to implement in-depth supervision to help the network converge. Fourth, we devised a multiscale supervision strategy for independent prediction at different levels of the decoding path, integrating multiscale semantic information to facilitate the segmentation of small vessels. Fifth, the 3D conditional random field (3DCRF) and 3D connected component optimization (3DCCO) were exploited as postprocessing to optimize the segmentation results. RESULTS Comprehensive experimental assessments validated the effectiveness of our ARU-Net. The proposed ARU-Net model achieved comparable or superior performance to the state-of-the-art methods through quantitative and qualitative evaluations. Notably, we found that ARU-Net improved the identification of arteries connecting to an IA, including small arteries that were hard to recognize by other methods. Consequently, IA geometries segmented by the proposed ARU-Net model yielded superior performance during subsequent computational hemodynamic studies (also known as "patient-specific" computational fluid dynamics [CFD] simulations). Furthermore, in an ablation study, the five key enhancements mentioned above were confirmed. CONCLUSIONS The proposed ARU-Net model can automatically segment the IAs in 3DRA images with relatively high accuracy and potentially has significant value for clinical computational hemodynamic analysis.
Collapse
Affiliation(s)
- Nan Mu
- Biomedical Engineering, Michigan Technological University, Houghton, MI United States; Center for Biocomputing and Digital Health, Health Research Institute and Institute of Computing and Cybernetics, Michigan Technological University, Houghton, Michigan, United States
| | - Zonghan Lyu
- Biomedical Engineering, Michigan Technological University, Houghton, MI United States; Center for Biocomputing and Digital Health, Health Research Institute and Institute of Computing and Cybernetics, Michigan Technological University, Houghton, Michigan, United States
| | - Mostafa Rezaeitaleshmahalleh
- Biomedical Engineering, Michigan Technological University, Houghton, MI United States; Center for Biocomputing and Digital Health, Health Research Institute and Institute of Computing and Cybernetics, Michigan Technological University, Houghton, Michigan, United States
| | - Jinshan Tang
- Department of Health Administration and Policy, George Mason University, Fairfax, Virginia, United States
| | - Jingfeng Jiang
- Biomedical Engineering, Michigan Technological University, Houghton, MI United States; Center for Biocomputing and Digital Health, Health Research Institute and Institute of Computing and Cybernetics, Michigan Technological University, Houghton, Michigan, United States.
| |
Collapse
|
11
|
Ilhan A, Alpan K, Sekeroglu B, Abiyev R. COVID-19 Lung CT image segmentation using localization and enhancement methods with U-Net. PROCEDIA COMPUTER SCIENCE 2023; 218:1660-1667. [PMID: 36743788 PMCID: PMC9886330 DOI: 10.1016/j.procs.2023.01.144] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/02/2023]
Abstract
Segmentation of pneumonia lesions from Lung CT images has become vital for diagnosing the disease and evaluating the severity of the patients during the COVID-19 pandemic. Several AI-based systems have been proposed for this task. However, some low-contrast abnormal zones in CT images make the task challenging. The researchers investigated image preprocessing techniques to accomplish this problem and to enable more accurate segmentation by the AI-based systems. This study proposes a COVID-19 Lung-CT segmentation system based on histogram-based non-parametric region localization and enhancement (LE) methods prior to the U-Net architecture. The COVID-19-infected lung CT images were initially processed by the LE method, and the infected regions were detected and enhanced to provide more discriminative features to the deep learning segmentation methods. The U-Net is trained using the enhanced images to segment the regions affected by COVID-19. The proposed system achieved 97.75%, 0.85, and 0.74 accuracy, dice score, and Jaccard index, respectively. The comparison results suggested that the use of LE methods as a preprocessing step in CT Lung images significantly improved the feature extraction and segmentation abilities of the U-Net model by a 0.21 dice score. The results might lead to implementing the LE method in segmenting varied medical images.
Collapse
Affiliation(s)
- Ahmet Ilhan
- Department of Computer Engineering, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
- Applied Artificial Intelligence Research Center, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
| | - Kezban Alpan
- Department of Information Systems Engineering, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
- Applied Artificial Intelligence Research Center, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
| | - Boran Sekeroglu
- Department of Information Systems Engineering, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
- Applied Artificial Intelligence Research Center, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
| | - Rahib Abiyev
- Department of Computer Engineering, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
- Applied Artificial Intelligence Research Center, Near East University, Nicosia, 99138, Cyprus, Mersin 10 Turkey
| |
Collapse
|
12
|
Li M, Li P, Liu Y. IDEFE algorithm: IDE algorithm optimizes the fuzzy entropy for the gland segmentation. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:4896-4911. [PMID: 36896528 DOI: 10.3934/mbe.2023227] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Breast cancer occurs in the epithelial tissue of the gland, so the accuracy of gland segmentation is crucial to the physician's diagnosis. An innovative technique for breast mammography image gland segmentation is put forth in this paper. In the first step, the algorithm designed the gland segmentation evaluation function. Then a new mutation strategy is established, and the adaptive controlled variables are used to balance the ability of improved differential evolution (IDE) in terms of investigation and convergence. To evaluate its performance, The proposed method is validated on a number of benchmark breast images, including four types of glands from the Quanzhou First Hospital, Fujian, China. Furthermore, the proposed algorithm is been systematically compared to five state-of-the-art algorithms. From the average MSSIM and boxplot, the evidence suggests that the mutation strategy may be effective in searching the topography of the segmented gland problem. The experiment results demonstrated that the proposed method has the best gland segmentation results compared to other algorithms.
Collapse
Affiliation(s)
- Mingzhu Li
- Department of Thyroid and Breast Surgery, East Branch of Quanzhou First Hospital, Fujian 362000, China
| | - Ping Li
- Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou, Fujian 362000, China
| | - Yao Liu
- College of Medicine, Huaqiao University, Quanzhou, Fujian 362021, China
| |
Collapse
|
13
|
Jiang X, Xiao J, Zhang Q, Wang L, Jiang J, Lan K. Improved U-Net based on cross-layer connection for pituitary adenoma MRI image segmentation. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:34-51. [PMID: 36650756 DOI: 10.3934/mbe.2023003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
Pituitary adenoma is a common neuroendocrine neoplasm, and most of its MR images are characterized by blurred edges, high noise and similar to surrounding normal tissues. Therefore, it is extremely difficult to accurately locate and outline the lesion of pituitary adenoma. To sovle these limitations, we design a novel deep learning framework for pituitary adenoma MRI image segmentation. Under the framework of U-Net, a newly cross-layer connection is introduced to capture richer multi-scale features and contextual information. At the same time, full-scale skip structure can reasonably utilize the above information obtained by different layers. In addition, an improved inception-dense block is designed to replace the classical convolution layer, which can enlarge the effectiveness of the receiving field and increase the depth of our network. Finally, a novel loss function based on binary cross-entropy and Jaccard losses is utilized to eliminate the problem of small samples and unbalanced data. The sample data were collected from 30 patients in Quzhou People's Hospital, with a total of 500 lesion images. Experimental results show that although the amount of patient sample is small, the proposed method has better performance in pituitary adenoma image compared with existing algorithms, and its Dice, Intersection over Union (IoU), Matthews correlation coefficient (Mcc) and precision reach 88.87, 80.67, 88.91 and 97.63%, respectively.
Collapse
Affiliation(s)
- Xiaoliang Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou 324000, China
| | - Junjian Xiao
- College of Mechanical Engineering, Quzhou University, Quzhou 324000, China
| | - Qile Zhang
- Department of Rehabilitation, The Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou People's Hospital, Quzhou 324000, China
| | - Lihui Wang
- Department of Science and Education, The Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou People's Hospital, Quzhou 324000, China
| | - Jinyun Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou 324000, China
| | - Kun Lan
- College of Mechanical Engineering, Quzhou University, Quzhou 324000, China
| |
Collapse
|
14
|
Fan C, Zeng Z, Xiao L, Qu X. GFNet: Automatic segmentation of COVID-19 lung infection regions using CT images based on boundary features. PATTERN RECOGNITION 2022; 132:108963. [PMID: 35966970 PMCID: PMC9359771 DOI: 10.1016/j.patcog.2022.108963] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/25/2021] [Revised: 07/31/2022] [Accepted: 08/07/2022] [Indexed: 05/03/2023]
Abstract
In early 2020, the global spread of the COVID-19 has presented the world with a serious health crisis. Due to the large number of infected patients, automatic segmentation of lung infections using computed tomography (CT) images has great potential to enhance traditional medical strategies. However, the segmentation of infected regions in CT slices still faces many challenges. Specially, the most core problem is the high variability of infection characteristics and the low contrast between the infected and the normal regions. This problem leads to fuzzy regions in lung CT segmentation. To address this problem, we have designed a novel global feature network(GFNet) for COVID-19 lung infections: VGG16 as backbone, we design a Edge-guidance module(Eg) that fuses the features of each layer. First, features are extracted by reverse attention module and Eg is combined with it. This series of steps enables each layer to fully extract boundary details that are difficult to be noticed by previous models, thus solving the fuzzy problem of infected regions. The multi-layer output features are fused into the final output to finally achieve automatic and accurate segmentation of infected areas. We compared the traditional medical segmentation networks, UNet, UNet++, the latest model Inf-Net, and methods of few shot learning field. Experiments show that our model is superior to the above models in Dice, Sensitivity, Specificity and other evaluation metrics, and our segmentation results are clear and accurate from the visual effect, which proves the effectiveness of GFNet. In addition, we verify the generalization ability of GFNet on another "never seen" dataset, and the results prove that our model still has better generalization ability than the above model. Our code has been shared at https://github.com/zengzhenhuan/GFNet.
Collapse
Affiliation(s)
- Chaodong Fan
- School of Computer Science and Technology, Hainan University, Haikou 570228, China
- School of Computer Science, Xiangtan University, Xiangtan 411100, China
- Foshan Green Intelligent Manufacturing Research Institute of Xiangtan University, Foshan 528000, China
- School of Information Technology and Management, Hunan University of Finance and Economics, Changsha 410205, China
| | - Zhenhuan Zeng
- School of Computer Science, Xiangtan University, Xiangtan 411100, China
| | - Leyi Xiao
- School of Computer Science and Technology, Hainan University, Haikou 570228, China
- School of Information Technology and Management, Hunan University of Finance and Economics, Changsha 410205, China
- AnHui Key Laboratory of Detection Technology and Energy Saving Devices, AnHui Polytechnic University, Wuhu 241000, China
- Fujian Provincial Key Laboratory of Data Intensive Computing, Quanzhou Normal University, Quanzhou 362000 China
- Vehicle Measurement, Control and Safety Key Laboratory of Sichuan Province, Xihua University, Chengdu 610039, China
| | - Xilong Qu
- School of Information Technology and Management, Hunan University of Finance and Economics, Changsha 410205, China
| |
Collapse
|
15
|
Estimation of Caenorhabditis Elegans Lifespan Stages Using a Dual-Path Network Combining Biomarkers and Physiological Changes. Bioengineering (Basel) 2022; 9:bioengineering9110689. [PMID: 36421090 PMCID: PMC9687340 DOI: 10.3390/bioengineering9110689] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Revised: 11/08/2022] [Accepted: 11/09/2022] [Indexed: 11/16/2022] Open
Abstract
Assessing individual aging has always been an important topic in aging research. Caenorhabditis elegans (C. elegans) has a short lifespan and is a popular model organism widely utilized in aging research. Studying the differences in C. elegans life stages is of great significance for human health and aging. In order to study the differences in C. elegans lifespan stages, the classification of lifespan stages is the first task to be performed. In the past, biomarkers and physiological changes captured with imaging were commonly used to assess aging in isogenic C. elegans individuals. However, all of the current research has focused only on physiological changes or biomarkers for the assessment of aging, which affects the accuracy of assessment. In this paper, we combine two types of features for the assessment of lifespan stages to improve assessment accuracy. To fuse the two types of features, an improved high-efficiency network (Att-EfficientNet) is proposed. In the new EfficientNet, attention mechanisms are introduced so that accuracy can be further improved. In addition, in contrast to previous research, which divided the lifespan into three stages, we divide the lifespan into six stages. We compared the classification method with other CNN-based methods as well as other classic machine learning methods. The results indicate that the classification method has a higher accuracy rate (72%) than other CNN-based methods and some machine learning methods.
Collapse
|
16
|
Peng Y, Zhang T, Guo Y. Cov-TransNet: Dual branch fusion network with transformer for COVID-19 infection segmentation. Biomed Signal Process Control 2022; 80:104366. [PMCID: PMC9671472 DOI: 10.1016/j.bspc.2022.104366] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2022] [Revised: 09/06/2022] [Accepted: 10/30/2022] [Indexed: 11/09/2022]
Abstract
Segmentation of COVID-19 infection is a challenging task due to the blurred boundaries and low contrast between the infected and the non-infected areas in COVID-19 CT images, especially for small infection regions. COV-TransNet is presented to achieve high-precision segmentation of COVID-19 infection regions in this paper. The proposed segmentation network is composed of the auxiliary branch and the backbone branch. The auxiliary branch network adopts transformer to provide global information, helping the convolution layers in backbone branch to learn specific local features better. A multi-scale feature attention module is introduced to capture contextual information and adaptively enhance feature representations. Specially, a high internal resolution is maintained during the attention calculation process. Moreover, feature activation module can effectively reduce the loss of valid information during sampling. The proposed network can take full advantage of different depth and multi-scale features to achieve high sensitivity for identifying lesions of varied sizes and locations. We experiment on several datasets of the COVID-19 lesion segmentation task, including COVID-19-CT-Seg, UESTC-COVID-19, MosMedData and COVID-19-MedSeg. Comprehensive results demonstrate that COV-TransNet outperforms the existing state-of-the-art segmentation methods and achieves better segmentation performance for multi-scale lesions.
Collapse
|
17
|
Kumar S, Mallik A. COVID-19 Detection from Chest X-rays Using Trained Output Based Transfer Learning Approach. Neural Process Lett 2022; 55:1-24. [PMID: 36339644 PMCID: PMC9616430 DOI: 10.1007/s11063-022-11060-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/16/2022] [Indexed: 10/31/2022]
Abstract
The recent Coronavirus disease (COVID-19), which started in 2019, has spread across the globe and become a global pandemic. The efficient and effective COVID-19 detection using chest X-rays helps in early detection and curtailing the spread of the disease. In this paper, we propose a novel Trained Output-based Transfer Learning (TOTL) approach for COVID-19 detection from chest X-rays. We start by preprocessing the Chest X-rays of the patients with techniques like denoising, contrasting, segmentation. These processed images are then fed to several pre-trained transfer learning models like InceptionV3, InceptionResNetV2, Xception, MobileNet, ResNet50, ResNet50V2, VGG16, and VGG19. We fine-tune these models on the processed chest X-rays. Then we further train the outputs of these models using a deep neural network architecture to achieve enhanced performance and aggregate the capabilities of each of them. The proposed model has been tested on four recent COVID-19 chest X-rays datasets by computing several popular evaluation metrics. The performance of our model has also been compared with various deep transfer learning models and several contemporary COVID-19 detection methods. The obtained results demonstrate the efficiency and efficacy of our proposed model.
Collapse
Affiliation(s)
- Sanjay Kumar
- Department of Computer Science and Engineering, Delhi Technological University, New Delhi, 110042 India
| | - Abhishek Mallik
- Department of Computer Science and Engineering, Delhi Technological University, New Delhi, 110042 India
| |
Collapse
|
18
|
Connell M, Xin Y, Gerard SE, Herrmann J, Shah PK, Martin KT, Rezoagli E, Ippolito D, Rajaei J, Baron R, Delvecchio P, Humayun S, Rizi RR, Bellani G, Cereda M. Unsupervised segmentation and quantification of COVID-19 lesions on computed Tomography scans using CycleGAN. Methods 2022; 205:200-209. [PMID: 35817338 PMCID: PMC9288584 DOI: 10.1016/j.ymeth.2022.07.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Revised: 05/18/2022] [Accepted: 07/06/2022] [Indexed: 01/04/2023] Open
Abstract
BACKGROUND Lesion segmentation is a critical step in medical image analysis, and methods to identify pathology without time-intensive manual labeling of data are of utmost importance during a pandemic and in resource-constrained healthcare settings. Here, we describe a method for fully automated segmentation and quantification of pathological COVID-19 lung tissue on chest Computed Tomography (CT) scans without the need for manually segmented training data. METHODS We trained a cycle-consistent generative adversarial network (CycleGAN) to convert images of COVID-19 scans into their generated healthy equivalents. Subtraction of the generated healthy images from their corresponding original CT scans yielded maps of pathological tissue, without background lung parenchyma, fissures, airways, or vessels. We then used these maps to construct three-dimensional lesion segmentations. Using a validation dataset, Dice scores were computed for our lesion segmentations and other published segmentation networks using ground truth segmentations reviewed by radiologists. RESULTS The COVID-to-Healthy generator eliminated high Hounsfield unit (HU) voxels within pulmonary lesions and replaced them with lower HU voxels. The generator did not distort normal anatomy such as vessels, airways, or fissures. The generated healthy images had higher gas content (2.45 ± 0.93 vs 3.01 ± 0.84 L, P < 0.001) and lower tissue density (1.27 ± 0.40 vs 0.73 ± 0.29 Kg, P < 0.001) than their corresponding original COVID-19 images, and they were not significantly different from those of the healthy images (P < 0.001). Using the validation dataset, lesion segmentations scored an average Dice score of 55.9, comparable to other weakly supervised networks that do require manual segmentations. CONCLUSION Our CycleGAN model successfully segmented pulmonary lesions in mild and severe COVID-19 cases. Our model's performance was comparable to other published models; however, our model is unique in its ability to segment lesions without the need for manual segmentations.
Collapse
Affiliation(s)
- Marc Connell
- Department of Anesthesiology and Critical Care, University of Pennsylvania, Philadelphia, PA, USA
| | - Yi Xin
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, USA
| | - Sarah E Gerard
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - Jacob Herrmann
- Department of Biomedical Engineering, Boston University, Boston, MA, USA
| | - Parth K Shah
- Department of Medicine, University of Pennsylvania, Philadelphia, PA, USA
| | - Kevin T Martin
- Department of Medicine, University of Pennsylvania, Philadelphia, PA, USA
| | - Emanuele Rezoagli
- Department of Emergency and Intensive Care, San Gerardo Hospital, Monza, Italy; Department of Medicine and Surgery, University of Milano-Bicocca, Monza, Italy
| | - Davide Ippolito
- Department of Diagnostic and Interventional Radiology, San Gerardo Hospital, Monza, Italy
| | - Jennia Rajaei
- Department of Medicine, Stanford University, Stanford, CA, USA
| | - Ryan Baron
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, USA
| | - Paolo Delvecchio
- Department of Anesthesiology and Critical Care, University of Pennsylvania, Philadelphia, PA, USA
| | - Shiraz Humayun
- Department of Anesthesiology and Critical Care, University of Pennsylvania, Philadelphia, PA, USA
| | - Rahim R Rizi
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, USA
| | - Giacomo Bellani
- Department of Medicine and Surgery, University of Milano-Bicocca, Monza, Italy
| | - Maurizio Cereda
- Department of Anesthesiology and Critical Care, University of Pennsylvania, Philadelphia, PA, USA.
| |
Collapse
|
19
|
Jiang S, Li J, Hua Z. Transformer with progressive sampling for medical cellular image segmentation. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:12104-12126. [PMID: 36653988 DOI: 10.3934/mbe.2022563] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
The convolutional neural network, as the backbone network for medical image segmentation, has shown good performance in the past years. However, its drawbacks cannot be ignored, namely, convolutional neural networks focus on local regions and are difficult to model global contextual information. For this reason, transformer, which is used for text processing, was introduced into the field of medical segmentation, and thanks to its expertise in modelling global relationships, the accuracy of medical segmentation was further improved. However, the transformer-based network structure requires a certain training set size to achieve satisfactory segmentation results, and most medical segmentation datasets are small in size. Therefore, in this paper we introduce a gated position-sensitive axial attention mechanism in the self-attention module, so that the transformer-based network structure can also be adapted to the case of small datasets. The common operation of the visual transformer introduced to visual processing when dealing with segmentation tasks is to divide the input image into equal patches of the same size and then perform visual processing on each patch, but this simple division may lead to the destruction of the structure of the original image, and there may be large unimportant regions in the divided grid, causing attention to stay on the uninteresting regions, affecting the segmentation performance. Therefore, in this paper, we add iterative sampling to update the sampling positions, so that the attention stays on the region to be segmented, reducing the interference of irrelevant regions and further improving the segmentation performance. In addition, we introduce the strip convolution module (SCM) and pyramid pooling module (PPM) to capture the global contextual information. The proposed network is evaluated on several datasets and shows some improvement in segmentation accuracy compared to networks of recent years.
Collapse
Affiliation(s)
- Shen Jiang
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai 264005, China
| | - Jinjiang Li
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai 264005, China
| | - Zhen Hua
- School of Information and Electronic Engineering, Shandong Technology and Business University, Yantai 264005, China
| |
Collapse
|
20
|
Li H, Liu X, Jia D, Chen Y, Hou P, Li H. Research on chest radiography recognition model based on deep learning. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:11768-11781. [PMID: 36124613 DOI: 10.3934/mbe.2022548] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
With the development of medical informatization and against the background of the spread of global epidemic, the demand for automated chest X-ray detection by medical personnel and patients continues to increase. Although the rapid development of deep learning technology has made it possible to automatically generate a single conclusive sentence, the results produced by existing methods are not reliable enough due to the complexity of medical images. To solve this problem, this paper proposes an improved RCLN (Recurrent Learning Network) model as a solution. The model can generate high-level conclusive impressions and detailed descriptive findings sentence-by-sentence and realize the imitation of the doctoros standard tone by combining a convolutional neural network (CNN) with a long short-term memory (LSTM) network through a recurrent structure, and adding a multi-head attention mechanism. The proposed algorithm has been experimentally verified on publicly available chest X-ray images from the Open-i image set. The results show that it can effectively solve the problem of automatic generation of colloquial medical reports.
Collapse
Affiliation(s)
- Hui Li
- School of Computer Engineering, Jiangsu Ocean University, China
| | - Xintang Liu
- School of Computer Engineering, Jiangsu Ocean University, China
| | - Dongbao Jia
- School of Computer Engineering, Jiangsu Ocean University, China
| | - Yanyan Chen
- School of Computer Engineering, Jiangsu Ocean University, China
| | - Pengfei Hou
- School of Computer Engineering, Jiangsu Ocean University, China
| | - Haining Li
- Department of Neurology, General Hospital of Ningxia Medical University, China
| |
Collapse
|
21
|
COVID-19 Infection Segmentation and Severity Assessment Using a Self-Supervised Learning Approach. Diagnostics (Basel) 2022; 12:diagnostics12081805. [PMID: 35892518 PMCID: PMC9332359 DOI: 10.3390/diagnostics12081805] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Revised: 07/14/2022] [Accepted: 07/23/2022] [Indexed: 11/16/2022] Open
Abstract
Background: Automated segmentation of COVID-19 infection lesions and the assessment of the severity of the infections are critical in COVID-19 diagnosis and treatment. Based on a large amount of annotated data, deep learning approaches have been widely used in COVID-19 medical image analysis. However, the number of medical image samples is generally huge, and it is challenging to obtain enough annotated medical images for training a deep CNN model. Methods: To address these challenges, we propose a novel self-supervised deep learning method for automated segmentation of COVID-19 infection lesions and assessing the severity of infection, which can reduce the dependence on the annotation of the training samples. In the proposed method, first, many unlabeled data are used to pre-train an encoder-decoder model to learn rotation-dependent and rotation-invariant features. Then, a small amount of labeled data is used to fine-tune the pre-trained encoder-decoder for COVID-19 severity classification and lesion segmentation. Results: The proposed methods were tested on two public COVID-19 CT datasets and one self-built dataset. Accuracy, precision, recall, and F1-score were used to measure classification performance and Dice coefficient was used to measure segmentation performance. For COVID-19 severity classification, the proposed method outperformed other unsupervised feature learning methods by about 7.16% in accuracy. For segmentation, when the amount of labeled data was 100%, the Dice value of the proposed method was 5.58% higher than that of U-Net.; in 70% of the cases, our method was 8.02% higher than U-Net; in 30% of the cases, our method was 11.88% higher than U-Net; and in 10% of the cases, our method was 16.88% higher than U-Net. Conclusions: The proposed method provides better classification and segmentation performance under limited labeled data than other methods.
Collapse
|
22
|
Sun J, Pi P, Tang C, Wang SH, Zhang YD. TSRNet: Diagnosis of COVID-19 based on self-supervised learning and hybrid ensemble model. Comput Biol Med 2022; 146:105531. [PMID: 35489140 PMCID: PMC9013277 DOI: 10.1016/j.compbiomed.2022.105531] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2021] [Revised: 03/12/2022] [Accepted: 04/13/2022] [Indexed: 12/16/2022]
Abstract
BACKGROUND As of Feb 27, 2022, coronavirus (COVID-19) has caused 434,888,591 infections and 5,958,849 deaths worldwide, dealing a severe blow to the economies and cultures of most countries around the world. As the virus has mutated, its infectious capacity has further increased. Effective diagnosis of suspected cases is an important tool to stop the spread of the pandemic. Therefore, we intended to develop a computer-aided diagnosis system for the diagnosis of suspected cases. METHODS To address the shortcomings of commonly used pre-training methods and exploit the information in unlabeled images, we proposed a new pre-training method based on transfer learning with self-supervised learning (TS). After that, a new convolutional neural network based on attention mechanism and deep residual network (RANet) was proposed to extract features. Based on this, a hybrid ensemble model (TSRNet) was proposed for classifying lung CT images of suspected patients as COVID-19 and normal. RESULTS Compared with the existing five models in terms of accuracy (DarkCOVIDNet: 98.08%; Deep-COVID: 97.58%; NAGNN: 97.86%; COVID-ResNet: 97.78%; Patch-based CNN: 88.90%), TSRNet has the highest accuracy of 99.80%. In addition, the recall, f1-score, and AUC of the model reached 99.59%, 99.78%, and 1, respectively. CONCLUSION TSRNet can effectively diagnose suspected COVID-19 cases with the help of the information in unlabeled and labeled images, thus helping physicians to adopt early treatment plans for confirmed cases.
Collapse
Affiliation(s)
- Junding Sun
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan, 454000, PR China.
| | - Pengpeng Pi
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan, 454000, PR China.
| | - Chaosheng Tang
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan, 454000, PR China.
| | - Shui-Hua Wang
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan, 454000, PR China; School of Computing and Mathematical Sciences, University of Leicester, Leicester, LE1 7RH, UK.
| | - Yu-Dong Zhang
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan, 454000, PR China; School of Computing and Mathematical Sciences, University of Leicester, Leicester, LE1 7RH, UK.
| |
Collapse
|
23
|
A Comprehensive Performance Analysis of Transfer Learning Optimization in Visual Field Defect Classification. Diagnostics (Basel) 2022; 12:diagnostics12051258. [PMID: 35626413 PMCID: PMC9140208 DOI: 10.3390/diagnostics12051258] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2022] [Revised: 05/16/2022] [Accepted: 05/17/2022] [Indexed: 02/05/2023] Open
Abstract
Numerous research have demonstrated that Convolutional Neural Network (CNN) models are capable of classifying visual field (VF) defects with great accuracy. In this study, we evaluated the performance of different pre-trained models (VGG-Net, MobileNet, ResNet, and DenseNet) in classifying VF defects and produced a comprehensive comparative analysis to compare the performance of different CNN models before and after hyperparameter tuning and fine-tuning. Using 32 batch sizes, 50 epochs, and ADAM as the optimizer to optimize weight, bias, and learning rate, VGG-16 obtained the highest accuracy of 97.63 percent, according to experimental findings. Subsequently, Bayesian optimization was utilized to execute automated hyperparameter tuning and automated fine-tuning layers of the pre-trained models to determine the optimal hyperparameter and fine-tuning layer for classifying many VF defect with the highest accuracy. We found that the combination of different hyperparameters and fine-tuning of the pre-trained models significantly impact the performance of deep learning models for this classification task. In addition, we also discovered that the automated selection of optimal hyperparameters and fine-tuning by Bayesian has significantly enhanced the performance of the pre-trained models. The results observed the best performance for the DenseNet-121 model with a validation accuracy of 98.46% and a test accuracy of 99.57% for the tested datasets.
Collapse
|
24
|
Automatic COVID-19 Lung Infection Segmentation through Modified Unet Model. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:6566982. [PMID: 35422980 PMCID: PMC9002904 DOI: 10.1155/2022/6566982] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/12/2022] [Revised: 02/23/2022] [Accepted: 02/28/2022] [Indexed: 11/23/2022]
Abstract
The coronavirus (COVID-19) pandemic has had a terrible impact on human lives globally, with far-reaching consequences for the health and well-being of many people around the world. Statistically, 305.9 million people worldwide tested positive for COVID-19, and 5.48 million people died due to COVID-19 up to 10 January 2022. CT scans can be used as an alternative to time-consuming RT-PCR testing for COVID-19. This research work proposes a segmentation approach to identifying ground glass opacity or ROI in CT images developed by coronavirus, with a modified structure of the Unet model having been used to classify the region of interest at the pixel level. The problem with segmentation is that the GGO often appears indistinguishable from a healthy lung in the initial stages of COVID-19, and so, to cope with this, the increased set of weights in contracting and expanding the Unet path and an improved convolutional module is added in order to establish the connection between the encoder and decoder pipeline. This has a major capacity to segment the GGO in the case of COVID-19, with the proposed model being referred to as “convUnet.” The experiment was performed on the Medseg1 dataset, and the addition of a set of weights at each layer of the model and modification in the connected module in Unet led to an improvement in overall segmentation results. The quantitative results obtained using accuracy, recall, precision, dice-coefficient, F1score, and IOU were 93.29%, 93.01%, 93.67%, 92.46%, 93.34%, 86.96%, respectively, which is better than that obtained using Unet and other state-of-the-art models. Therefore, this segmentation approach proved to be more accurate, fast, and reliable in helping doctors to diagnose COVID-19 quickly and efficiently.
Collapse
|
25
|
Punn NS, Agarwal S. CHS-Net: A Deep Learning Approach for Hierarchical Segmentation of COVID-19 via CT Images. Neural Process Lett 2022; 54:3771-3792. [PMID: 35310011 PMCID: PMC8924740 DOI: 10.1007/s11063-022-10785-x] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/24/2022] [Indexed: 01/19/2023]
Abstract
The pandemic of novel severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) also known as COVID-19 has been spreading worldwide, causing rampant loss of lives. Medical imaging such as computed tomography (CT), X-ray, etc., plays a significant role in diagnosing the patients by presenting the visual representation of the functioning of the organs. However, for any radiologist analyzing such scans is a tedious and time-consuming task. The emerging deep learning technologies have displayed its strength in analyzing such scans to aid in the faster diagnosis of the diseases and viruses such as COVID-19. In the present article, an automated deep learning based model, COVID-19 hierarchical segmentation network (CHS-Net) is proposed that functions as a semantic hierarchical segmenter to identify the COVID-19 infected regions from lungs contour via CT medical imaging using two cascaded residual attention inception U-Net (RAIU-Net) models. RAIU-Net comprises of a residual inception U-Net model with spectral spatial and depth attention network (SSD) that is developed with the contraction and expansion phases of depthwise separable convolutions and hybrid pooling (max and spectral pooling) to efficiently encode and decode the semantic and varying resolution information. The CHS-Net is trained with the segmentation loss function that is the defined as the average of binary cross entropy loss and dice loss to penalize false negative and false positive predictions. The approach is compared with the recently proposed approaches and evaluated using the standard metrics like accuracy, precision, specificity, recall, dice coefficient and Jaccard similarity along with the visualized interpretation of the model prediction with GradCam++ and uncertainty maps. With extensive trials, it is observed that the proposed approach outperformed the recently proposed approaches and effectively segments the COVID-19 infected regions in the lungs.
Collapse
|
26
|
Zhang X, Wang G, Zhao SG. CapsNet-COVID19: Lung CT image classification method based on CapsNet model. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:5055-5074. [PMID: 35430853 DOI: 10.3934/mbe.2022236] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
The outbreak of the Corona Virus Disease 2019 (COVID-19) has posed a serious threat to human health and life around the world. As the number of COVID-19 cases continues to increase, many countries are facing problems such as errors in nucleic acid testing (RT-PCR), shortage of testing reagents, and lack of testing personnel. In order to solve such problems, it is necessary to propose a more accurate and efficient method as a supplement to the detection and diagnosis of COVID-19. This research uses a deep network model to classify some of the COVID-19, general pneumonia, and normal lung CT images in the 2019 Novel Coronavirus Information Database. The first level of the model uses convolutional neural networks to locate lung regions in lung CT images. The second level of the model uses the capsule network to classify and predict the segmented images. The accuracy of our method is 84.291% on the test set and 100% on the training set. Experiment shows that our classification method is suitable for medical image classification with complex background, low recognition rate, blurred boundaries and large image noise. We believe that this classification method is of great value for monitoring and controlling the growth of patients in COVID-19 infected areas.
Collapse
Affiliation(s)
- XiaoQing Zhang
- Nanjing University of Science and Technology, Taizhou Technology Institute, Taizhou 225300, China
| | - GuangYu Wang
- Donghua University, College of Information Science and Technology, Shanghai 201620, China
| | - Shu-Guang Zhao
- Donghua University, College of Information Science and Technology, Shanghai 201620, China
| |
Collapse
|
27
|
Bi R, Ji C, Yang Z, Qiao M, Lv P, Wang H. Residual based attention-Unet combing DAC and RMP modules for automatic liver tumor segmentation in CT. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:4703-4718. [PMID: 35430836 DOI: 10.3934/mbe.2022219] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Purpose: Due to the complex distribution of liver tumors in the abdomen, the accuracy of liver tumor segmentation cannot meet the needs of clinical assistance yet. This paper aims to propose a new end-to-end network to improve the segmentation accuracy of liver tumors from CT. Method: We proposed a hybrid network, leveraging the residual block, the context encoder (CE), and the Attention-Unet, called ResCEAttUnet. The CE comprises a dense atrous convolution (DAC) module and a residual multi-kernel pooling (RMP) module. The DAC module ensures the network derives high-level semantic information and minimizes detailed information loss. The RMP module improves the ability of the network to extract multi-scale features. Moreover, a hybrid loss function based on cross-entropy and Tversky loss function is employed to distribute the weights of the two-loss parts through training iterations. Results: We evaluated the proposed method in LiTS17 and 3DIRCADb databases. It significantly improved the segmentation accuracy compared to state-of-the-art methods. Conclusions: Experimental results demonstrate the satisfying effects of the proposed method through both quantitative and qualitative analyses, thus proving a promising tool in liver tumor segmentation.
Collapse
Affiliation(s)
- Rongrong Bi
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Chunlei Ji
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Zhipeng Yang
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Meixia Qiao
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Peiqing Lv
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Haiying Wang
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| |
Collapse
|
28
|
Qi Q, Qi S, Wu Y, Li C, Tian B, Xia S, Ren J, Yang L, Wang H, Yu H. Fully automatic pipeline of convolutional neural networks and capsule networks to distinguish COVID-19 from community-acquired pneumonia via CT images. Comput Biol Med 2022; 141:105182. [PMID: 34979404 PMCID: PMC8715632 DOI: 10.1016/j.compbiomed.2021.105182] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2021] [Revised: 12/25/2021] [Accepted: 12/25/2021] [Indexed: 01/08/2023]
Abstract
Background Chest computed tomography (CT) is crucial in the diagnosis of coronavirus disease 2019 (COVID-19). However, the persistent pandemic and similar CT manifestations between COVID-19 and community-acquired pneumonia (CAP) raise methodological requirements. Methods A fully automatic pipeline of deep learning is proposed for distinguishing COVID-19 from CAP using CT images. Inspired by the diagnostic process of radiologists, the pipeline comprises four connected modules for lung segmentation, selection of slices with lesions, slice-level prediction, and patient-level prediction. The roles of the first and second modules and the effectiveness of the capsule network for slice-level prediction were investigated. A dataset of 326 CT scans was collected to train and test the pipeline. Another public dataset of 110 patients was used to evaluate the generalization capability. Results LinkNet exhibited the largest intersection over union (0.967) and Dice coefficient (0.983) for lung segmentation. For the selection of slices with lesions, the capsule network with the ResNet50 block achieved an accuracy of 92.5% and an area under the curve (AUC) of 0.933. The capsule network using the DenseNet121 block demonstrated better performance for slice-level prediction, with an accuracy of 97.1% and AUC of 0.992. For both datasets, the prediction accuracy of our pipeline was 100% at the patient level. Conclusions The proposed fully automatic deep learning pipeline of deep learning can distinguish COVID-19 from CAP via CT images rapidly and accurately, thereby accelerating diagnosis and augmenting the performance of radiologists. This pipeline is convenient for use by radiologists and provides explainable predictions.
Collapse
Affiliation(s)
- Qianqian Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| | - Shouliang Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| | - Yanan Wu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| | - Chen Li
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
| | - Bin Tian
- Department of Radiology, The Second People's Hospital of Guiyang, Guiyang, China.
| | - Shuyue Xia
- Department of Respiratory Medicine, Central Hospital Affiliated to Shenyang Medical College, Shenyang, China.
| | - Jigang Ren
- Department of Radiology, The Affiliated Hospital of Guizhou Medical University, Guiyang, China.
| | - Liming Yang
- Department of Radiology, The Affiliated Hospital of Guizhou Medical University, Guiyang, China.
| | - Hanlin Wang
- Department of Radiology, General Hospital of the Yangtze River Shipping, Wuhan, China.
| | - Hui Yu
- Department of Radiology, The Affiliated Hospital of Guizhou Medical University, Guiyang, China; Department of Radiology, The Seventh Affiliated Hospital, Southern Medical University, Guangzhou, China.
| |
Collapse
|
29
|
Deshpande G, Batliner A, Schuller BW. AI-Based human audio processing for COVID-19: A comprehensive overview. PATTERN RECOGNITION 2022; 122:108289. [PMID: 34483372 PMCID: PMC8404390 DOI: 10.1016/j.patcog.2021.108289] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/26/2021] [Revised: 08/24/2021] [Accepted: 08/29/2021] [Indexed: 06/02/2023]
Abstract
The Coronavirus (COVID-19) pandemic impelled several research efforts, from collecting COVID-19 patients' data to screening them for virus detection. Some COVID-19 symptoms are related to the functioning of the respiratory system that influences speech production; this suggests research on identifying markers of COVID-19 in speech and other human generated audio signals. In this article, we give an overview of research on human audio signals using 'Artificial Intelligence' techniques to screen, diagnose, monitor, and spread the awareness about COVID-19. This overview will be useful for developing automated systems that can help in the context of COVID-19, using non-obtrusive and easy to use bio-signals conveyed in human non-speech and speech audio productions.
Collapse
Affiliation(s)
- Gauri Deshpande
- Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
- TCS Research Pune, India
| | - Anton Batliner
- Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
| | - Björn W Schuller
- Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Germany
- GLAM - Group on Language, Audio, & Music, Imperial College London, UK
| |
Collapse
|
30
|
Multiscale U-Net with Spatial Positional Attention for Retinal Vessel Segmentation. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:5188362. [PMID: 35047151 PMCID: PMC8763561 DOI: 10.1155/2022/5188362] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/22/2021] [Revised: 12/06/2021] [Accepted: 12/09/2021] [Indexed: 11/18/2022]
Abstract
Retinal vessel segmentation is essential for the detection and diagnosis of eye diseases. However, it is difficult to accurately identify the vessel boundary due to the large variations of scale in the retinal vessels and the low contrast between the vessel and the background. Deep learning has a good effect on retinal vessel segmentation since it can capture representative and distinguishing features for retinal vessels. An improved U-Net algorithm for retinal vessel segmentation is proposed in this paper. To better identify vessel boundaries, the traditional convolutional operation CNN is replaced by a global convolutional network and boundary refinement in the coding part. To better divide the blood vessel and background, the improved position attention module and channel attention module are introduced in the jumping connection part. Multiscale input and multiscale dense feature pyramid cascade modules are used to better obtain feature information. In the decoding part, convolutional long and short memory networks and deep dilated convolution are used to extract features. In public datasets, DRIVE and CHASE_DB1, the accuracy reached 96.99% and 97.51%. The average performance of the proposed algorithm is better than that of existing algorithms.
Collapse
|