1
|
Xie X, Yang M. USCT-UNet: Rethinking the Semantic Gap in U-Net Network From U-Shaped Skip Connections With Multichannel Fusion Transformer. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3782-3793. [PMID: 39325601 DOI: 10.1109/tnsre.2024.3468339] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/28/2024]
Abstract
Medical image segmentation is a crucial component of computer-aided clinical diagnosis, with state-of-the-art models often being variants of U-Net. Despite their success, these models' skip connections introduce an unnecessary semantic gap between the encoder and decoder, which hinders their ability to achieve the high precision required for clinical applications. Awareness of this semantic gap and its detrimental influences have increased over time. However, a quantitative understanding of how this semantic gap compromises accuracy and reliability remains lacking, emphasizing the need for effective mitigation strategies. In response, we present the first quantitative evaluation of the semantic gap between corresponding layers of U-Net and identify two key characteristics: 1) The direct skip connection (DSC) exhibits a semantic gap that negatively impacts models' performance; 2) The magnitude of the semantic gap varies across different layers. Based on these findings, we re-examine this issue through the lens of skip connections. We introduce a Multichannel Fusion Transformer (MCFT) and propose a novel USCT-UNet architecture, which incorporates U-shaped skip connections (USC) to replace DSC, allocates varying numbers of MCFT blocks based on the semantic gap magnitude at different layers, and employs a spatial channel cross-attention (SCCA) module to facilitate the fusion of features between the decoder and USC. We evaluate USCT-UNet on four challenging datasets, and the results demonstrate that it effectively eliminates the semantic gap. Compared to using DSC, our USC and SCCA strategies achieve maximum improvements of 4.79% in the Dice coefficient, 5.70% in mean intersection over union (MIoU), and 3.26 in Hausdorff distance.
Collapse
|
2
|
Shamrat FMJM, Shakil R, Idris MYI, Akter B, Zhou X. FruitSeg30_Segmentation dataset & mask annotations: A novel dataset for diverse fruit segmentation and classification. Data Brief 2024; 56:110821. [PMID: 39252785 PMCID: PMC11381999 DOI: 10.1016/j.dib.2024.110821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2024] [Revised: 07/16/2024] [Accepted: 08/05/2024] [Indexed: 09/11/2024] Open
Abstract
Fruits are mature ovaries of flowering plants that are integral to human diets, providing essential nutrients such as vitamins, minerals, fiber and antioxidants that are crucial for health and disease prevention. Accurate classification and segmentation of fruits are crucial in the agricultural sector for enhancing the efficiency of sorting and quality control processes, which significantly benefit automated systems by reducing labor costs and improving product consistency. This paper introduces the "FruitSeg30_Segmentation Dataset & Mask Annotations", a novel dataset designed to advance the capability of deep learning models in fruit segmentation and classification. Comprising 1969 high-quality images across 30 distinct fruit classes, this dataset provides diverse visuals essential for a robust model. Utilizing a U-Net architecture, the model trained on this dataset achieved training accuracy of 94.72 %, validation accuracy of 92.57 %, precision of 94 %, recall of 91 %, f1-score of 92.5 %, IoU score of 86 %, and maximum dice score of 0.9472, demonstrating superior performance in segmentation tasks. The FruitSeg30 dataset fills a critical gap and sets new standards in dataset quality and diversity, enhancing agricultural technology and food industry applications.
Collapse
Affiliation(s)
| | - Rashiduzzaman Shakil
- Department of Computer Science and Engineering, Daffodil International University, Daffodil Smart City (DSC), Birulia, Savar, Dhaka 1216, Bangladesh
| | - Mohd Yamani Idna Idris
- Department of Computer System and Technology, Universiti Malaya, Kuala Lumpur 50603, Malaysia
| | - Bonna Akter
- Department of Computer Science and Engineering, Daffodil International University, Daffodil Smart City (DSC), Birulia, Savar, Dhaka 1216, Bangladesh
| | - Xujuan Zhou
- School of Business, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
3
|
Ali M, Wu T, Hu H, Mahmood T. Breast tumor segmentation using neural cellular automata and shape guided segmentation in mammography images. PLoS One 2024; 19:e0309421. [PMID: 39352900 PMCID: PMC11444406 DOI: 10.1371/journal.pone.0309421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Accepted: 08/13/2024] [Indexed: 10/04/2024] Open
Abstract
PURPOSE Using computer-aided design (CAD) systems, this research endeavors to enhance breast cancer segmentation by addressing data insufficiency and data complexity during model training. As perceived by computer vision models, the inherent symmetry and complexity of mammography images make segmentation difficult. The objective is to optimize the precision and effectiveness of medical imaging. METHODS The study introduces a hybrid strategy combining shape-guided segmentation (SGS) and M3D-neural cellular automata (M3D-NCA), resulting in improved computational efficiency and performance. The implementation of Shape-guided segmentation (SGS) during the initialization phase, coupled with the elimination of convolutional layers, enables the model to effectively reduce computation time. The research proposes a novel loss function that combines segmentation losses from both components for effective training. RESULTS The robust technique provided aims to improve the accuracy and consistency of breast tumor segmentation, leading to significant improvements in medical imaging and breast cancer detection and treatment. CONCLUSION This study enhances breast cancer segmentation in medical imaging using CAD systems. Combining shape-guided segmentation (SGS) and M3D-neural cellular automata (M3D-NCA) is a hybrid approach that improves performance and computational efficiency by dealing with complex data and not having enough training data. The approach also reduces computing time and improves training efficiency. The study aims to improve breast cancer detection and treatment methods in medical imaging technology.
Collapse
Affiliation(s)
- Mudassar Ali
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Tong Wu
- University of Illinois Urbana-Champaign Institute, Zhejiang University, Hangzhou, Zhejiang, China
| | - Haoji Hu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Tariq Mahmood
- Artificial Intelligence & Data Analytics Lab, CCIS Prince Sultan University, Riyadh, Kingdom of Saudi Arabia
- Faculty of Information Sciences, University of Education, Vehari Campus, Vehari, Pakistan
| |
Collapse
|
4
|
Li Y, Fang R, Zhang N, Liao C, Chen X, Wang X, Luo Y, Li L, Mao M, Zhang Y. An improved algorithm for salient object detection of microscope based on U 2-Net. Med Biol Eng Comput 2024:10.1007/s11517-024-03205-w. [PMID: 39322859 DOI: 10.1007/s11517-024-03205-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2024] [Accepted: 09/11/2024] [Indexed: 09/27/2024]
Abstract
With the rapid advancement of modern medical technology, microscopy imaging systems have become one of the key technologies in medical image analysis. However, manual use of microscopes presents issues such as operator dependency, inefficiency, and time consumption. To enhance the efficiency and accuracy of medical image capture and reduce the burden of subsequent quantitative analysis, this paper proposes an improved microscope salient object detection algorithm based on U2-Net, incorporating deep learning technology. The improved algorithm first enhances the network's key information extraction capability by incorporating the Convolutional Block Attention Module (CBAM) into U2-Net. It then optimizes network complexity by constructing a Simple Pyramid Pooling Module (SPPM) and uses Ghost convolution to achieve model lightweighting. Additionally, data augmentation is applied to the slides to improve the algorithm's robustness and generalization. The experimental results show that the size of the improved algorithm model is 72.5 MB, which represents a 56.85% reduction compared to the original U2-Net model size of 168.0 MB. Additionally, the model's prediction accuracy has increased from 92.24 to 97.13%, providing an efficient means for subsequent image processing and analysis tasks in microscopy imaging systems.
Collapse
Affiliation(s)
- Yunchai Li
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Run Fang
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China.
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China.
| | - Nangang Zhang
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Chengsheng Liao
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Xiaochang Chen
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Xiaoyu Wang
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Yunfei Luo
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Leheng Li
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Min Mao
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| | - Yunlong Zhang
- School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan, 430200, China
- State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan, 430200, China
| |
Collapse
|
5
|
Jiang B, Bao L, He S, Chen X, Jin Z, Ye Y. Deep learning applications in breast cancer histopathological imaging: diagnosis, treatment, and prognosis. Breast Cancer Res 2024; 26:137. [PMID: 39304962 DOI: 10.1186/s13058-024-01895-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Accepted: 09/16/2024] [Indexed: 09/22/2024] Open
Abstract
Breast cancer is the most common malignant tumor among women worldwide and remains one of the leading causes of death among women. Its incidence and mortality rates are continuously rising. In recent years, with the rapid advancement of deep learning (DL) technology, DL has demonstrated significant potential in breast cancer diagnosis, prognosis evaluation, and treatment response prediction. This paper reviews relevant research progress and applies DL models to image enhancement, segmentation, and classification based on large-scale datasets from TCGA and multiple centers. We employed foundational models such as ResNet50, Transformer, and Hover-net to investigate the performance of DL models in breast cancer diagnosis, treatment, and prognosis prediction. The results indicate that DL techniques have significantly improved diagnostic accuracy and efficiency, particularly in predicting breast cancer metastasis and clinical prognosis. Furthermore, the study emphasizes the crucial role of robust databases in developing highly generalizable models. Future research will focus on addressing challenges related to data management, model interpretability, and regulatory compliance, ultimately aiming to provide more precise clinical treatment and prognostic evaluation programs for breast cancer patients.
Collapse
Affiliation(s)
- Bitao Jiang
- Department of Hematology and Oncology, Beilun District People's Hospital, Ningbo, 315800, China.
- Department of Hematology and Oncology, Beilun Branch of the First Affiliated Hospital of Zhejiang University, Ningbo, 315800, China.
| | - Lingling Bao
- Department of Hematology and Oncology, Beilun District People's Hospital, Ningbo, 315800, China
- Department of Hematology and Oncology, Beilun Branch of the First Affiliated Hospital of Zhejiang University, Ningbo, 315800, China
| | - Songqin He
- Department of Oncology, The 906th Hospital of the Joint Logistics Force of the Chinese People's Liberation Army, Ningbo, 315100, China
| | - Xiao Chen
- Department of Oncology, The 906th Hospital of the Joint Logistics Force of the Chinese People's Liberation Army, Ningbo, 315100, China
| | - Zhihui Jin
- Department of Hematology and Oncology, Beilun District People's Hospital, Ningbo, 315800, China
- Department of Hematology and Oncology, Beilun Branch of the First Affiliated Hospital of Zhejiang University, Ningbo, 315800, China
| | - Yingquan Ye
- Department of Oncology, The 906th Hospital of the Joint Logistics Force of the Chinese People's Liberation Army, Ningbo, 315100, China.
| |
Collapse
|
6
|
Contino S, Cruciata L, Gambino O, Pirrone R. IODeep: An IOD for the introduction of deep learning in the DICOM standard. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 248:108113. [PMID: 38479148 DOI: 10.1016/j.cmpb.2024.108113] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/15/2024] [Revised: 02/22/2024] [Accepted: 03/01/2024] [Indexed: 04/02/2024]
Abstract
BACKGROUND AND OBJECTIVE In recent years, Artificial Intelligence (AI) and in particular Deep Neural Networks (DNN) became a relevant research topic in biomedical image segmentation due to the availability of more and more data sets along with the establishment of well known competitions. Despite the popularity of DNN based segmentation on the research side, these techniques are almost unused in the daily clinical practice even if they could support effectively the physician during the diagnostic process. Apart from the issues related to the explainability of the predictions of a neural model, such systems are not integrated in the diagnostic workflow, and a standardization of their use is needed to achieve this goal. METHODS This paper presents IODeep a new DICOM Information Object Definition (IOD) aimed at storing both the weights and the architecture of a DNN already trained on a particular image dataset that is labeled as regards the acquisition modality, the anatomical region, and the disease under investigation. RESULTS The IOD architecture is presented along with a DNN selection algorithm from the PACS server based on the labels outlined above, and a simple PACS viewer purposely designed for demonstrating the effectiveness of the DICOM integration, while no modifications are required on the PACS server side. Also a service based architecture in support of the entire workflow has been implemented. CONCLUSION IODeep ensures full integration of a trained AI model in a DICOM infrastructure, and it is also enables a scenario where a trained model can be either fine-tuned with hospital data or trained in a federated learning scheme shared by different hospitals. In this way AI models can be tailored to the real data produced by a Radiology ward thus improving the physician decision making process. Source code is freely available at https://github.com/CHILab1/IODeep.git.
Collapse
Affiliation(s)
- Salvatore Contino
- Department of Engineering, University of Palermo, Palermo, 90128, Sicily, Italy
| | - Luca Cruciata
- Department of Engineering, University of Palermo, Palermo, 90128, Sicily, Italy
| | - Orazio Gambino
- Department of Engineering, University of Palermo, Palermo, 90128, Sicily, Italy.
| | - Roberto Pirrone
- Department of Engineering, University of Palermo, Palermo, 90128, Sicily, Italy
| |
Collapse
|
7
|
Yeh CH, Lo C, He CH. Multibranch Wavelet-Based Network for Image Demoiréing. SENSORS (BASEL, SWITZERLAND) 2024; 24:2762. [PMID: 38732870 PMCID: PMC11086364 DOI: 10.3390/s24092762] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/20/2024] [Revised: 04/24/2024] [Accepted: 04/24/2024] [Indexed: 05/13/2024]
Abstract
Moiré patterns caused by aliasing between the camera's sensor and the monitor can severely degrade image quality. Image demoiréing is a multi-task image restoration method that includes texture and color restoration. This paper proposes a new multibranch wavelet-based image demoiréing network (MBWDN) for moiré pattern removal. Moiré images are separated into sub-band images using wavelet decomposition, and demoiréing can be achieved using the different learning strategies of two networks: moiré removal network (MRN) and detail-enhanced moiré removal network (DMRN). MRN removes moiré patterns from low-frequency images while preserving the structure of smooth areas. DMRN simultaneously removes high-frequency moiré patterns and enhances fine details in images. Wavelet decomposition is used to replace traditional upsampling, and max pooling effectively increases the receptive field of the network without losing the spatial information. Through decomposing the moiré image into different levels using wavelet transform, the feature learning results of each branch can be fully preserved and fed into the next branch; therefore, possible distortions in the recovered image are avoided. Thanks to the separation of high- and low-frequency images during feature training, the proposed two networks achieve impressive moiré removal effects. Based on extensive experiments conducted using public datasets, the proposed method shows good demoiréing validity both quantitatively and qualitatively when compared with the state-of-the-art approaches.
Collapse
Affiliation(s)
- Chia-Hung Yeh
- Department of Electrical Engineering, National Taiwan Normal University, Taipei 10610, Taiwan; (C.L.)
- Department of Electrical Engineering, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| | - Chen Lo
- Department of Electrical Engineering, National Taiwan Normal University, Taipei 10610, Taiwan; (C.L.)
| | - Cheng-Han He
- Department of Electrical Engineering, National Taiwan Normal University, Taipei 10610, Taiwan; (C.L.)
| |
Collapse
|
8
|
Jiang X, Zheng H, Yuan Z, Lan K, Wu Y. HIMS-Net: Horizontal-vertical interaction and multiple side-outputs network for cyst segmentation in jaw images. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:4036-4055. [PMID: 38549317 DOI: 10.3934/mbe.2024178] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/02/2024]
Abstract
Jaw cysts are mainly caused by abnormal tooth development, chronic oral inflammation, or jaw damage, which may lead to facial swelling, deformity, tooth loss, and other symptoms. Due to the diversity and complexity of cyst images, deep-learning algorithms still face many difficulties and challenges. In response to these problems, we present a horizontal-vertical interaction and multiple side-outputs network for cyst segmentation in jaw images. First, the horizontal-vertical interaction mechanism facilitates complex communication paths in the vertical and horizontal dimensions, and it has the ability to capture a wide range of context dependencies. Second, the feature-fused unit is introduced to adjust the network's receptive field, which enhances the ability of acquiring multi-scale context information. Third, the multiple side-outputs strategy intelligently combines feature maps to generate more accurate and detailed change maps. Finally, experiments were carried out on the self-established jaw cyst dataset and compared with different specialist physicians to evaluate its clinical usability. The research results indicate that the Matthews correlation coefficient (Mcc), Dice, and Jaccard of HIMS-Net were 93.61, 93.66 and 88.10% respectively, which may contribute to rapid and accurate diagnosis in clinical practice.
Collapse
Affiliation(s)
- Xiaoliang Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou 324000, China
| | - Huixia Zheng
- Department of Stomatology, The Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou People's Hospital, Quzhou 324000, China
| | - Zhenfei Yuan
- Department of Stomatology, The Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou People's Hospital, Quzhou 324000, China
| | - Kun Lan
- College of Mechanical Engineering, Quzhou University, Quzhou 324000, China
| | - Yaoyang Wu
- Department of Computer and Information Science, University of Macau, Macau 999078, China
| |
Collapse
|
9
|
Su C, Zhou Y, Ma J, Chi H, Jing X, Jiao J, Yan Q. JANet: A joint attention network for balancing accuracy and speed in left ventricular ultrasound video segmentation. Comput Biol Med 2024; 169:107856. [PMID: 38154159 DOI: 10.1016/j.compbiomed.2023.107856] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2023] [Revised: 11/08/2023] [Accepted: 12/11/2023] [Indexed: 12/30/2023]
Abstract
Multiple cardiac diseases are closely associated with functional parameters of the left ventricle, but functional parameter quantification still requires manual involvement, a time-consuming and less reproducible task. We develop a joint attention network (JANet) and expand it into two versions (V1 and V2) that can be used to segment the left ventricular region in echocardiograms to assist physicians in diagnosis. V1 is a smaller model with a size of 56.3 MB, and V2 has a higher accuracy. The proposed JANet V1 and V2 achieve a mean dice score (DSC) of 93.59/93.69(V1/V2), respectively, outperforming the state-of-the-art models. We grade 1264 patients with 87.24/87.50 (V1/V2) accuracy when using the 2-level classification criteria and 83.62/84.18 (V1/V2) when using the 5-level classification criteria. The results of the consistency analysis show that the proposed method is comparable to that of clinicians.
Collapse
Affiliation(s)
- Chenkai Su
- School of Integrated Circuits, Shandong University, Jinan, 250101, China
| | - Yuxiang Zhou
- School of Integrated Circuits, Shandong University, Jinan, 250101, China
| | - Jinlian Ma
- School of Integrated Circuits, Shandong University, Jinan, 250101, China; Shenzhen Research Institute of Shandong University, A301 Virtual University Park in South District of Shenzhen, China.
| | - Haoyu Chi
- School of Integrated Circuits, Shandong University, Jinan, 250101, China
| | - Xin Jing
- School of Integrated Circuits, Shandong University, Jinan, 250101, China
| | - Junyan Jiao
- School of Integrated Circuits, Shandong University, Jinan, 250101, China
| | - Qiqi Yan
- School of Integrated Circuits, Shandong University, Jinan, 250101, China
| |
Collapse
|
10
|
Ma Z, Li X. An improved supervised and attention mechanism-based U-Net algorithm for retinal vessel segmentation. Comput Biol Med 2024; 168:107770. [PMID: 38056215 DOI: 10.1016/j.compbiomed.2023.107770] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Revised: 11/08/2023] [Accepted: 11/26/2023] [Indexed: 12/08/2023]
Abstract
The segmentation results of retinal blood vessels are crucial for automatically diagnosing ophthalmic diseases such as diabetic retinopathy, hypertension, cardiovascular and cerebrovascular diseases. To improve the accuracy of vessel segmentation and better extract information about small vessels and edges, we introduce the U-Net algorithm with a supervised attention mechanism for retinal vessel segmentation. We achieve this by introducing a decoder fusion module (DFM) in the encoding part, effectively combining different convolutional blocks to extract features comprehensively. Additionally, in the decoding part of U-Net, we propose the context squeeze and excitation (CSE) decoding module to enhance important contextual feature information and the detection of tiny blood vessels. For the final output, we introduce the supervised fusion mechanism (SFM), which combines multiple branches from shallow to deep layers, effectively fusing multi-scale features and capturing information from different levels, fully integrating low-level and high-level features to improve segmentation performance. Our experimental results on the public datasets of DRIVE, STARE, and CHASED_B1 demonstrate the excellent performance of our proposed network.
Collapse
Affiliation(s)
- Zhendi Ma
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China
| | - Xiaobo Li
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China.
| |
Collapse
|
11
|
Khoshkhabar M, Meshgini S, Afrouzian R, Danishvar S. Automatic Liver Tumor Segmentation from CT Images Using Graph Convolutional Network. SENSORS (BASEL, SWITZERLAND) 2023; 23:7561. [PMID: 37688038 PMCID: PMC10490641 DOI: 10.3390/s23177561] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Revised: 08/13/2023] [Accepted: 08/16/2023] [Indexed: 09/10/2023]
Abstract
Segmenting the liver and liver tumors in computed tomography (CT) images is an important step toward quantifiable biomarkers for a computer-aided decision-making system and precise medical diagnosis. Radiologists and specialized physicians use CT images to diagnose and classify liver organs and tumors. Because these organs have similar characteristics in form, texture, and light intensity values, other internal organs such as the heart, spleen, stomach, and kidneys confuse visual recognition of the liver and tumor division. Furthermore, visual identification of liver tumors is time-consuming, complicated, and error-prone, and incorrect diagnosis and segmentation can hurt the patient's life. Many automatic and semi-automatic methods based on machine learning algorithms have recently been suggested for liver organ recognition and tumor segmentation. However, there are still difficulties due to poor recognition precision and speed and a lack of dependability. This paper presents a novel deep learning-based technique for segmenting liver tumors and identifying liver organs in computed tomography maps. Based on the LiTS17 database, the suggested technique comprises four Chebyshev graph convolution layers and a fully connected layer that can accurately segment the liver and liver tumors. Thus, the accuracy, Dice coefficient, mean IoU, sensitivity, precision, and recall obtained based on the proposed method according to the LiTS17 dataset are around 99.1%, 91.1%, 90.8%, 99.4%, 99.4%, and 91.2%, respectively. In addition, the effectiveness of the proposed method was evaluated in a noisy environment, and the proposed network could withstand a wide range of environmental signal-to-noise ratios (SNRs). Thus, at SNR = -4 dB, the accuracy of the proposed method for liver organ segmentation remained around 90%. The proposed model has obtained satisfactory and favorable results compared to previous research. According to the positive results, the proposed model is expected to be used to assist radiologists and specialist doctors in the near future.
Collapse
Affiliation(s)
- Maryam Khoshkhabar
- Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz 51666-16471, Iran
| | - Saeed Meshgini
- Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz 51666-16471, Iran
| | - Reza Afrouzian
- Miyaneh Faculty of Engineering, University of Tabriz, Miyaneh 51666-16471, Iran
| | - Sebelan Danishvar
- College of Engineering, Design, and Physical Sciences, Brunel University London, Uxbridge UB8 3PH, UK
| |
Collapse
|
12
|
Cheng Z, Wang L. Dynamic hierarchical multi-scale fusion network with axial MLP for medical image segmentation. Sci Rep 2023; 13:6342. [PMID: 37072483 PMCID: PMC10113245 DOI: 10.1038/s41598-023-32813-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Accepted: 04/03/2023] [Indexed: 05/03/2023] Open
Abstract
Medical image segmentation provides various effective methods for accuracy and robustness of organ segmentation, lesion detection, and classification. Medical images have fixed structures, simple semantics, and diverse details, and thus fusing rich multi-scale features can augment segmentation accuracy. Given that the density of diseased tissue may be comparable to that of surrounding normal tissue, both global and local information are critical for segmentation results. Therefore, considering the importance of multi-scale, global, and local information, in this paper, we propose the dynamic hierarchical multi-scale fusion network with axial mlp (multilayer perceptron) (DHMF-MLP), which integrates the proposed hierarchical multi-scale fusion (HMSF) module. Specifically, HMSF not only reduces the loss of detail information by integrating the features of each stage of the encoder, but also has different receptive fields, thereby improving the segmentation results for small lesions and multi-lesion regions. In HMSF, we not only propose the adaptive attention mechanism (ASAM) to adaptively adjust the semantic conflicts arising during the fusion process but also introduce Axial-mlp to improve the global modeling capability of the network. Extensive experiments on public datasets confirm the excellent performance of our proposed DHMF-MLP. In particular, on the BUSI, ISIC 2018, and GlaS datasets, IoU reaches 70.65%, 83.46%, and 87.04%, respectively.
Collapse
Affiliation(s)
- Zhikun Cheng
- College of Information Science and Engineering, Xinjiang University, Urumqi, 830046, China
| | - Liejun Wang
- College of Information Science and Engineering, Xinjiang University, Urumqi, 830046, China.
| |
Collapse
|