51
|
Zhang H, Zhong X, Li G, Liu W, Liu J, Ji D, Li X, Wu J. BCU-Net: Bridging ConvNeXt and U-Net for medical image segmentation. Comput Biol Med 2023; 159:106960. [PMID: 37099973 DOI: 10.1016/j.compbiomed.2023.106960] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2022] [Revised: 04/12/2023] [Accepted: 04/17/2023] [Indexed: 04/28/2023]
Abstract
Medical image segmentation enables doctors to observe lesion regions better and make accurate diagnostic decisions. Single-branch models such as U-Net have achieved great progress in this field. However, the complementary local and global pathological semantics of heterogeneous neural networks have not yet been fully explored. The class-imbalance problem remains a serious issue. To alleviate these two problems, we propose a novel model called BCU-Net, which leverages the advantages of ConvNeXt in global interaction and U-Net in local processing. We propose a new multilabel recall loss (MRL) module to relieve the class imbalance problem and facilitate deep-level fusion of local and global pathological semantics between the two heterogeneous branches. Extensive experiments were conducted on six medical image datasets including retinal vessel and polyp images. The qualitative and quantitative results demonstrate the superiority and generalizability of BCU-Net. In particular, BCU-Net can handle diverse medical images with diverse resolutions. It has a flexible structure owing to its plug-and-play characteristics, which promotes its practicality.
Collapse
Affiliation(s)
- Hongbin Zhang
- School of Software, East China Jiaotong University, China.
| | - Xiang Zhong
- School of Software, East China Jiaotong University, China.
| | - Guangli Li
- School of Information Engineering, East China Jiaotong University, China.
| | - Wei Liu
- School of Software, East China Jiaotong University, China.
| | - Jiawei Liu
- School of Software, East China Jiaotong University, China.
| | - Donghong Ji
- School of Cyber Science and Engineering, Wuhan University, China.
| | - Xiong Li
- School of Software, East China Jiaotong University, China.
| | - Jianguo Wu
- The Second Affiliated Hospital of Nanchang University, China.
| |
Collapse
|
52
|
Zheng T, Qin H, Cui Y, Wang R, Zhao W, Zhang S, Geng S, Zhao L. Segmentation of thyroid glands and nodules in ultrasound images using the improved U-Net architecture. BMC Med Imaging 2023; 23:56. [PMID: 37060061 PMCID: PMC10105426 DOI: 10.1186/s12880-023-01011-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2022] [Accepted: 04/05/2023] [Indexed: 04/16/2023] Open
Abstract
BACKGROUND Identifying thyroid nodules' boundaries is crucial for making an accurate clinical assessment. However, manual segmentation is time-consuming. This paper utilized U-Net and its improved methods to automatically segment thyroid nodules and glands. METHODS The 5822 ultrasound images used in the experiment came from two centers, 4658 images were used as the training dataset, and 1164 images were used as the independent mixed test dataset finally. Based on U-Net, deformable-pyramid split-attention residual U-Net (DSRU-Net) by introducing ResNeSt block, atrous spatial pyramid pooling, and deformable convolution v3 was proposed. This method combined context information and extracts features of interest better, and had advantages in segmenting nodules and glands of different shapes and sizes. RESULTS DSRU-Net obtained 85.8% mean Intersection over Union, 92.5% mean dice coefficient and 94.1% nodule dice coefficient, which were increased by 1.8%, 1.3% and 1.9% compared with U-Net. CONCLUSIONS Our method is more capable of identifying and segmenting glands and nodules than the original method, as shown by the results of correlational studies.
Collapse
Affiliation(s)
- Tianlei Zheng
- School of Information and Control Engineering, China University of Mining and Technology, Xuzhou, 221116, China
- Artificial Intelligence Unit, Department of Medical Equipment Management, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China
| | - Hang Qin
- Department of Medical Equipment Management, Nanjing First Hospital, Nanjing, 221000, China
| | - Yingying Cui
- Department of Pathology, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China
| | - Rong Wang
- Department of Ultrasound Medicine, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China
| | - Weiguo Zhao
- Artificial Intelligence Unit, Department of Medical Equipment Management, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China
| | - Shijin Zhang
- Artificial Intelligence Unit, Department of Medical Equipment Management, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China
| | - Shi Geng
- Artificial Intelligence Unit, Department of Medical Equipment Management, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China
| | - Lei Zhao
- Artificial Intelligence Unit, Department of Medical Equipment Management, Affiliated Hospital of Xuzhou Medical University, Xuzhou, 221004, China.
| |
Collapse
|
53
|
Zhou Z, Bian Y, Pan S, Meng Q, Zhu W, Shi F, Chen X, Shao C, Xiang D. A dual branch and fine-grained enhancement network for pancreatic tumor segmentation in contrast enhanced CT images. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104516] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/25/2022]
|
54
|
Diao Z, Jiang H, Zhou Y. Leverage prior texture information in deep learning-based liver tumor segmentation: A plug-and-play Texture-Based Auto Pseudo Label module. Comput Med Imaging Graph 2023; 106:102217. [PMID: 36958076 DOI: 10.1016/j.compmedimag.2023.102217] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2022] [Revised: 01/16/2023] [Accepted: 03/04/2023] [Indexed: 03/17/2023]
Abstract
Segmenting the liver and tumor regions using CT scans is crucial for the subsequent treatment in clinical practice and radiotherapy. Recently, liver and tumor segmentation techniques based on U-Net have gained popularity. However, there are numerous varieties of liver tumors, and they differ greatly in terms of their shapes and textures. It is unreasonable to regard all liver tumors as one class for learning. Meanwhile, texture information is crucial for the identification of liver tumors. We propose a plug-and-play Texture-based Auto Pseudo Label (TAPL) module to take use of the texture information of tumors and enable the neural network actively learn the texture differences between various tumors to increase the segmentation accuracy, especially for small tumors. The TPAL module consists of two parts, texture enhancement and texture-based pseudo label generator. To highlight the regions where the texture varies significantly, we enhance the textured areas of the CT image. Based on their texture information, tumors are automatically divided into several classes by the texture-based pseudo label generator. The multi-class tumors produced by the neural network during the prediction step are combined into a single tumor label, which is then used as the outcome of the segmentation. Experiments on clinical dataset and public dataset Lits2017 show that the proposed algorithm outperforms single liver tumor label segmentation methods and is more friendly to small tumors.
Collapse
Affiliation(s)
- Zhaoshuo Diao
- Software College, Northeastern University, Shenyang 110819, China
| | - Huiyan Jiang
- Software College, Northeastern University, Shenyang 110819, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang 110819, China.
| | - Yang Zhou
- Software College, Northeastern University, Shenyang 110819, China
| |
Collapse
|
55
|
Deep Learning-Based Semantic Segmentation Methods for Pavement Cracks. INFORMATION 2023. [DOI: 10.3390/info14030182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/17/2023] Open
Abstract
As road mileage continues to expand, the number of disasters caused by expanding pavement cracks is increasing. Two main methods, image processing and deep learning, are used to detect these cracks to improve the efficiency and quality of pavement crack segmentation. The classical segmentation network, UNet, has a poor ability to extract target edge information and small target segmentation, and is susceptible to the influence of distracting objects in the environment, thus failing to better segment the tiny cracks on the pavement. To resolve this problem, we propose a U-shaped network, ALP-UNet, which adds an attention module to each encoding layer. In the decoding phase, we incorporated the Laplacian pyramid to make the feature map contain more boundary information. We also propose adding a PAN auxiliary head to provide an additional loss for the backbone to improve the overall network segmentation effect. The experimental results show that the proposed method can effectively reduce the interference of other factors on the pavement and effectively improve the mIou and mPA values compared to the previous methods.
Collapse
|
56
|
Xie X, Zhang W, Pan X, Xie L, Shao F, Zhao W, An J. CANet: Context aware network with dual-stream pyramid for medical image segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104437] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
57
|
Shao D, Su F, Zou X, Lu J, Wu S, Tian R, Ran D, Guo Z, Jin D. Pixel-Level Classification of Five Histologic Patterns of Lung Adenocarcinoma. Anal Chem 2023; 95:2664-2670. [PMID: 36701546 DOI: 10.1021/acs.analchem.2c03020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
Lung adenocarcinoma is the most common histologic type of lung cancer. The pixel-level labeling of histologic patterns of lung adenocarcinoma can assist pathologists in determining tumor grading with more details than normal classification. We manually annotated a dataset containing a total of 1000 patches (200 patches for each pattern) of 512 × 512 pixels and 420 patches (contains test sets) of 1024 × 1024 pixels according to the morphological features of the five histologic patterns of lung adenocarcinoma (lepidic, acinar, papillary, micropapillary, and solid). To generate an even large amount of data patches, we developed a data stitching strategy as a data augmentation for classification in model training. Stitched patches improve the Dice similarity coefficient (DSC) scores by 24.06% on the whole-slide image (WSI) with the solid pattern. We propose a WSI analysis framework for lung adenocarcinoma pathology, intelligently labeling lung adenocarcinoma histologic patterns at the pixel level. Our framework contains five branches of deep neural networks for segmenting each histologic pattern. We test our framework with 200 unclassified patches. The DSC scores of our results outpace comparing networks (U-Net, LinkNet, and FPN) by up to 10.78%. We also perform results on four WSIs with an overall accuracy of 99.6%, demonstrating that our network framework exhibits better accuracy and robustness in most cases.
Collapse
Affiliation(s)
- Dan Shao
- School of Electronic and Information, Yangtze University, Jingzhou434023, China.,UTS-SUSTech Joint Research Centre for Biomedical Materials and Devices, Department of Biomedical Engineering, Southern University of Science and Technology, Shenzhen518055, China
| | - Fei Su
- UTS-SUSTech Joint Research Centre for Biomedical Materials and Devices, Department of Biomedical Engineering, Southern University of Science and Technology, Shenzhen518055, China.,Institute for Biomedical Materials and Devices (IBMD), Faculty of Science, University of Technology Sydney, Sydney, NSW2007, Australia
| | - Xueyu Zou
- School of Electronic and Information, Yangtze University, Jingzhou434023, China
| | - Jie Lu
- UTS-SUSTech Joint Research Centre for Biomedical Materials and Devices, Department of Biomedical Engineering, Southern University of Science and Technology, Shenzhen518055, China.,Guangdong Provincial Key Laboratory of Advanced Biomaterials, Southern University of Science and Technology, Shenzhen518055, China
| | - Sitong Wu
- UTS-SUSTech Joint Research Centre for Biomedical Materials and Devices, Department of Biomedical Engineering, Southern University of Science and Technology, Shenzhen518055, China.,Institute for Biomedical Materials and Devices (IBMD), Faculty of Science, University of Technology Sydney, Sydney, NSW2007, Australia
| | - Ruijun Tian
- Department of Chemistry, Southern University of Science and Technology, Shenzhen518055, China
| | - Dongmei Ran
- Department of Pathology, Southern University of Science and Technology Hospital, Shenzhen518055, China
| | - Zhiyong Guo
- UTS-SUSTech Joint Research Centre for Biomedical Materials and Devices, Department of Biomedical Engineering, Southern University of Science and Technology, Shenzhen518055, China.,Guangdong Provincial Key Laboratory of Advanced Biomaterials, Southern University of Science and Technology, Shenzhen518055, China
| | - Dayong Jin
- UTS-SUSTech Joint Research Centre for Biomedical Materials and Devices, Department of Biomedical Engineering, Southern University of Science and Technology, Shenzhen518055, China.,Institute for Biomedical Materials and Devices (IBMD), Faculty of Science, University of Technology Sydney, Sydney, NSW2007, Australia.,Guangdong Provincial Key Laboratory of Advanced Biomaterials, Southern University of Science and Technology, Shenzhen518055, China
| |
Collapse
|
58
|
Zhang S, Mu W, Dong D, Wei J, Fang M, Shao L, Zhou Y, He B, Zhang S, Liu Z, Liu J, Tian J. The Applications of Artificial Intelligence in Digestive System Neoplasms: A Review. HEALTH DATA SCIENCE 2023; 3:0005. [PMID: 38487199 PMCID: PMC10877701 DOI: 10.34133/hds.0005] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/05/2022] [Accepted: 12/05/2022] [Indexed: 03/17/2024]
Abstract
Importance Digestive system neoplasms (DSNs) are the leading cause of cancer-related mortality with a 5-year survival rate of less than 20%. Subjective evaluation of medical images including endoscopic images, whole slide images, computed tomography images, and magnetic resonance images plays a vital role in the clinical practice of DSNs, but with limited performance and increased workload of radiologists or pathologists. The application of artificial intelligence (AI) in medical image analysis holds promise to augment the visual interpretation of medical images, which could not only automate the complicated evaluation process but also convert medical images into quantitative imaging features that associated with tumor heterogeneity. Highlights We briefly introduce the methodology of AI for medical image analysis and then review its clinical applications including clinical auxiliary diagnosis, assessment of treatment response, and prognosis prediction on 4 typical DSNs including esophageal cancer, gastric cancer, colorectal cancer, and hepatocellular carcinoma. Conclusion AI technology has great potential in supporting the clinical diagnosis and treatment decision-making of DSNs. Several technical issues should be overcome before its application into clinical practice of DSNs.
Collapse
Affiliation(s)
- Shuaitong Zhang
- School of Engineering Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
| | - Wei Mu
- School of Engineering Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
| | - Di Dong
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Jingwei Wei
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Mengjie Fang
- School of Engineering Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
| | - Lizhi Shao
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Yu Zhou
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Bingxi He
- School of Engineering Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
| | - Song Zhang
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Zhenyu Liu
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Jianhua Liu
- Department of Oncology, Guangdong Provincial People's Hospital/Second Clinical Medical College of Southern Medical University/Guangdong Academy of Medical Sciences, Guangzhou, Guangdong, China
| | - Jie Tian
- School of Engineering Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
59
|
Kushnure DT, Tyagi S, Talbar SN. LiM-Net: Lightweight multi-level multiscale network with deep residual learning for automatic liver segmentation in CT images. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104305] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
60
|
Du J, Guan K, Liu P, Li Y, Wang T. Boundary-Sensitive Loss Function With Location Constraint for Hard Region Segmentation. IEEE J Biomed Health Inform 2023; 27:992-1003. [PMID: 36378793 DOI: 10.1109/jbhi.2022.3222390] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
In computer-aided diagnosis and treatment planning, accurate segmentation of medical images plays an essential role, especially for some hard regions including boundaries, small objects and background interference. However, existing segmentation loss functions including distribution-, region- and boundary-based losses cannot achieve satisfactory performances on these hard regions. In this paper, a boundary-sensitive loss function with location constraint is proposed for hard region segmentation in medical images, which provides three advantages: i) our Boundary-Sensitive loss (BS-loss) can automatically pay more attention to the hard-to-segment boundaries (e.g., thin structures and blurred boundaries), thus obtaining finer object boundaries; ii) BS-loss also can adjust its attention to small objects during training to segment them more accurately; and iii) our location constraint can alleviate the negative impact of the background interference, through the distribution matching of pixels between prediction and Ground Truth (GT) along each axis. By resorting to the proposed BS-loss and location constraint, the hard regions in both foreground and background are considered. Experimental results on three public datasets demonstrate the superiority of our method. Specifically, compared to the second-best method tested in this study, our method improves performance on hard regions in terms of Dice similarity coefficient (DSC) and 95% Hausdorff distance (95%HD) of up to 4.17% and 73% respectively. In addition, it also achieves the best overall segmentation performance. Hence, we can conclude that our method can accurately segment these hard regions and improve the overall segmentation performance in medical images.
Collapse
|
61
|
Chen H, Yu MA, Chen C, Zhou K, Qi S, Chen Y, Xiao R. FDE-net: Frequency-domain enhancement network using dynamic-scale dilated convolution for thyroid nodule segmentation. Comput Biol Med 2023; 153:106514. [PMID: 36628913 DOI: 10.1016/j.compbiomed.2022.106514] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2022] [Revised: 12/22/2022] [Accepted: 12/31/2022] [Indexed: 01/05/2023]
Abstract
Thyroid nodules, a common disease of endocrine system, have a probability of nearly 10% to turn into malignant nodules and thus pose a serious threat to health. Automatic segmentation of thyroid nodules is of great importance for clinicopathological diagnosis. This work proposes FDE-Net, a combined segmental frequency domain enhancement and dynamic scale cavity convolutional network for thyroid nodule segmentation. In FDE-Net, traditional image omics method is introduced to enhance the feature image in the segmented frequency domain. Such an approach reduces the influence of noise and strengthens the detail and contour information of the image. The proposed method introduces a cascade cross-scale attention module, which addresses the insensitivity of the network to the change in target scale by fusing the features of different receptive fields and improves the ability of the network to identify multiscale target regions. It repeatedly uses the high-dimensional feature image to improve segmentation accuracy in accordance with the simple structure of thyroid nodules. In this study, 1355 ultrasound images are used for training and testing. Quantitative evaluation results showed that the Dice coefficient of FDE-Net in thyroid nodule segmentation was 83.54%, which is better than other methods. Therefore, FDE-Net can enable the accurate and rapid segmentation of thyroid nodules.
Collapse
Affiliation(s)
- Hongyu Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ming-An Yu
- Department of Interventional Medicine, China-Japan Friendship Hospital, Beijing, 100029, China
| | - Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Siyu Qi
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Yunqing Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China; Shunde Innovation School, University of Science and Technology Beijing, Foshan, 100024, China.
| |
Collapse
|
62
|
Wang M, Jiang H. Memory-Net: Coupling feature maps extraction and hierarchical feature maps reuse for efficient and effective PET/CT multi-modality image-based tumor segmentation. Knowl Based Syst 2023. [DOI: 10.1016/j.knosys.2023.110399] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023]
|
63
|
Chen Y, Zheng C, Zhou T, Feng L, Liu L, Zeng Q, Wang G. A deep residual attention-based U-Net with a biplane joint method for liver segmentation from CT scans. Comput Biol Med 2023; 152:106421. [PMID: 36527780 DOI: 10.1016/j.compbiomed.2022.106421] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Revised: 11/17/2022] [Accepted: 12/11/2022] [Indexed: 12/15/2022]
Abstract
Liver tumours are diseases with high morbidity and high deterioration probabilities, and accurate liver area segmentation from computed tomography (CT) scans is a prerequisite for quick tumour diagnosis. While 2D network segmentation methods can perform segmentation with lower device performance requirements, they often discard the rich 3D spatial information contained in CT scans, limiting their segmentation accuracy. Hence, a deep residual attention-based U-shaped network (DRAUNet) with a biplane joint method for liver segmentation is proposed in this paper, where the biplane joint method introduces coronal CT slices to assist the transverse slices with segmentation, incorporating more 3D spatial information into the segmentation results to improve the segmentation performance of the network. Additionally, a novel deep residual block (DR block) and dual-effect attention module (DAM) are introduced in DRAUNet, where the DR block has deeper layers and two shortcut paths. The DAM efficiently combines the correlations of feature channels and the spatial locations of feature maps. The DRAUNet with the biplane joint method is tested on three datasets, Liver Tumour Segmentation (LiTS), 3D Image Reconstruction for Comparison of Algorithms Database (3DIRCADb), and Segmentation of the Liver Competition 2007 (Sliver07), and it achieves 97.3%, 97.4%, and 96.9% Dice similarity coefficients (DSCs) for liver segmentation, respectively, outperforming most state-of-the-art networks; this strongly demonstrates the segmentation performance of DRAUNet and the ability of the biplane joint method to obtain 3D spatial information from 3D images.
Collapse
Affiliation(s)
- Ying Chen
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Cheng Zheng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Taohui Zhou
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Longfeng Feng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Lan Liu
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China.
| | - Qiao Zeng
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China
| | - Guoqing Wang
- Zhejiang Suosi Technology Co. Ltd, Wenzhou, 325000, PR China.
| |
Collapse
|
64
|
Wang X, Wang S, Zhang Z, Yin X, Wang T, Li N. CPAD-Net: Contextual parallel attention and dilated network for liver tumor segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
65
|
Xiong M, Xu Y, Zhao Y, He S, Zhu Q, Wu Y, Hu X, Liu L. Quantitative analysis of artificial intelligence on liver cancer: A bibliometric analysis. Front Oncol 2023; 13:990306. [PMID: 36874099 PMCID: PMC9978515 DOI: 10.3389/fonc.2023.990306] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2022] [Accepted: 02/03/2023] [Indexed: 02/18/2023] Open
Abstract
Objective To provide the current research progress, hotspots, and emerging trends for AI in liver cancer, we have compiled a relative comprehensive and quantitative report on the research of liver disease using artificial intelligence by employing bibliometrics in this study. Methods In this study, the Web of Science Core Collection (WoSCC) database was used to perform systematic searches using keywords and a manual screening strategy, VOSviewer was used to analyze the degree of cooperation between countries/regions and institutions, as well as the co-occurrence of cooperation between authors and cited authors. Citespace was applied to generate a dual map to analyze the relationship of citing journals and citied journals and conduct a strong citation bursts ranking analysis of references. Online SRplot was used for in-depth keyword analysis and Microsoft Excel 2019 was used to collect the targeted variables from retrieved articles. Results 1724 papers were collected in this study, including 1547 original articles and 177 reviews. The study of AI in liver cancer mostly began from 2003 and has developed rapidly from 2017. China has the largest number of publications, and the United States has the highest H-index and total citation counts. The top three most productive institutions are the League of European Research Universities, Sun Yat Sen University, and Zhejiang University. Jasjit S. Suri and Frontiers in Oncology are the most published author and journal, respectively. Keyword analysis showed that in addition to the research on liver cancer, research on liver cirrhosis, fatty liver disease, and liver fibrosis were also common. Computed tomography was the most used diagnostic tool, followed by ultrasound and magnetic resonance imaging. The diagnosis and differential diagnosis of liver cancer are currently the most widely adopted research goals, and comprehensive analyses of multi-type data and postoperative analysis of patients with advanced liver cancer are rare. The use of convolutional neural networks is the main technical method used in studies of AI on liver cancer. Conclusion AI has undergone rapid development and has a wide application in the diagnosis and treatment of liver diseases, especially in China. Imaging is an indispensable tool in this filed. Mmulti-type data fusion analysis and development of multimodal treatment plans for liver cancer could become the major trend of future research in AI in liver cancer.
Collapse
Affiliation(s)
- Ming Xiong
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China
| | - Yaona Xu
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China
| | - Yang Zhao
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China
| | - Si He
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China
| | - Qihan Zhu
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China
| | - Yi Wu
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China
| | - Xiaofei Hu
- Department of Nuclear Medicine, Southwest Hospital, Third Military Medical University (Army Medical University), Chongqing, China
| | - Li Liu
- Department of Digital Medicine, School of Biomedical Engineering and Medical Imaging, Third Military Medical University (Army Medical University), Chongqing, China.,Department of Ultrasound, Southwest Hospital, Third Military Medical University (Army Medical University), Chongqing, China
| |
Collapse
|
66
|
Jung Y, Kim S, Kim J, Hwang B, Lee S, Kim EY, Kim JH, Hwang H. Abdominal Aortic Thrombus Segmentation in Postoperative Computed Tomography Angiography Images Using Bi-Directional Convolutional Long Short-Term Memory Architecture. SENSORS (BASEL, SWITZERLAND) 2022; 23:175. [PMID: 36616773 PMCID: PMC9823540 DOI: 10.3390/s23010175] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/09/2022] [Revised: 12/19/2022] [Accepted: 12/22/2022] [Indexed: 06/17/2023]
Abstract
Abdominal aortic aneurysm (AAA) is a fatal clinical condition with high mortality. Computed tomography angiography (CTA) imaging is the preferred minimally invasive modality for the long-term postoperative observation of AAA. Accurate segmentation of the thrombus region of interest (ROI) in a postoperative CTA image volume is essential for quantitative assessment and rapid clinical decision making by clinicians. Few investigators have proposed the adoption of convolutional neural networks (CNN). Although these methods demonstrated the potential of CNN architectures by automating the thrombus ROI segmentation, the segmentation performance can be further improved. The existing methods performed the segmentation process independently per 2D image and were incapable of using adjacent images, which could be useful for the robust segmentation of thrombus ROIs. In this work, we propose a thrombus ROI segmentation method to utilize not only the spatial features of a target image, but also the volumetric coherence available from adjacent images. We newly adopted a recurrent neural network, bi-directional convolutional long short-term memory (Bi-CLSTM) architecture, which can learn coherence between a sequence of data. This coherence learning capability can be useful for challenging situations, for example, when the target image exhibits inherent postoperative artifacts and noises, the inclusion of adjacent images would facilitate learning more robust features for thrombus ROI segmentation. We demonstrate the segmentation capability of our Bi-CLSTM-based method with a comparison of the existing 2D-based thrombus ROI segmentation counterpart as well as other established 2D- and 3D-based alternatives. Our comparison is based on a large-scale clinical dataset of 60 patient studies (i.e., 60 CTA image volumes). The results suggest the superior segmentation performance of our Bi-CLSTM-based method by achieving the highest scores of the evaluation metrics, e.g., our Bi-CLSTM results were 0.0331 higher on total overlap and 0.0331 lower on false negative when compared to 2D U-net++ as the second-best.
Collapse
Affiliation(s)
- Younhyun Jung
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea
| | - Suhyeon Kim
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea
| | - Jihu Kim
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea
| | - Byunghoon Hwang
- Department of Software Convergence, Kyung Hee University, Yongin 17104, Republic of Korea
| | - Sungmin Lee
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea
| | - Eun Young Kim
- Department of Radiology, Gil Medical Center, Gachon University, Incheon 21565, Republic of Korea
| | - Jeong Ho Kim
- Department of Radiology, Gil Medical Center, Gachon University, Incheon 21565, Republic of Korea
| | - Hyoseok Hwang
- Department of Software Convergence, Kyung Hee University, Yongin 17104, Republic of Korea
| |
Collapse
|
67
|
Feng H, Fu Z, Wang Y, Zhang P, Lai H, Zhao J. Automatic segmentation of thrombosed aortic dissection in post-operative CT-angiography images. Med Phys 2022. [PMID: 36542417 DOI: 10.1002/mp.16169] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2022] [Revised: 11/02/2022] [Accepted: 11/23/2022] [Indexed: 12/24/2022] Open
Abstract
PURPOSE The thrombus in the false lumen (FL) of aortic dissection (AD) patients is a meaningful indicator to determine aortic remodeling but difficult to measure in clinic. In this study, a novel segmentation strategy based on deep learning was proposed to automatically extract the thrombus in the FL in post-operative computed tomography angiography (CTA) images of AD patients, which provided an efficient and convenient segmentation method with high accuracy. METHODS A two-step segmentation strategy was proposed. Each step contained a convolutional neural network (CNN) to segment the aorta and the thrombus, respectively. In the first step, a CNN was used to obtain the binary segmentation mask of the whole aorta. In the second step, another CNN was introduced to segment the thrombus. The results of the first step were used as additional input to the second step to highlight the aorta in the complex background. Moreover, skip connection attention refinement (SAR) modules were designed and added in the second step to improve the segmentation accuracy of the thrombus details by efficiently using the low-level features. RESULTS The proposed method provided accurate thrombus segmentation results (0.903 ± 0.062 in dice score, 0.828 ± 0.092 in Jaccard index, and 2.209 ± 2.945 in 95% Hausdorff distance), which showed improvement compared to the methods without prior information (0.846 ± 0.085 in dice score) and the method without SAR (0.899 ± 0.060 in dice score). Moreover, the proposed method achieved 0.967 ± 0.029 and 0.948 ± 0.041 in dice score of true lumen (TL) and patent FL (PFL) segmentation, respectively, indicating the excellence of the proposed method in the segmentation task of the overall aorta. CONCLUSIONS A novel CNN-based segmentation framework was proposed to automatically obtain thrombus segmentation for thrombosed AD in post-operative CTA images, which provided a useful tool for further application of thrombus-related indicators in clinical and research application.
Collapse
Affiliation(s)
- Hanying Feng
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, People's Republic of China
| | - Zheng Fu
- Department of Cardiovascular Surgery, Zhongshan Hospital Fudan University, Shanghai, People's Republic of China
| | - Yulin Wang
- Department of Cardiovascular Surgery, Zhongshan Hospital Fudan University, Shanghai, People's Republic of China
| | - Puming Zhang
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, People's Republic of China
| | - Hao Lai
- Department of Cardiovascular Surgery, Zhongshan Hospital Fudan University, Shanghai, People's Republic of China
| | - Jun Zhao
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, People's Republic of China
| |
Collapse
|
68
|
Zhang F, Wang Q, Fan E, Lu N, Chen D, Jiang H, Wang Y. Automatic segmentation of the tumor in nonsmall-cell lung cancer by combining coarse and fine segmentation. Med Phys 2022. [PMID: 36514264 DOI: 10.1002/mp.16158] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2022] [Revised: 08/05/2022] [Accepted: 11/26/2022] [Indexed: 12/15/2022] Open
Abstract
OBJECTIVES Radiotherapy plays an important role in the treatment of nonsmall-cell lung cancer (NSCLC). Accurate delineation of tumor is the key to successful radiotherapy. Compared with the commonly used manual delineation ways, which are time-consuming and laborious, the automatic segmentation methods based on deep learning can greatly improve the treatment efficiency. METHODS In this paper, we introduce an automatic segmentation method by combining coarse and fine segmentations for NSCLC. Coarse segmentation network is the first level, identifing the rough region of the tumor. In this network, according to the tissue structure distribution of the thoracic cavity where tumor is located, we designed a competition method between tumors and organs at risk (OARs), which can increase the proportion of the identified tumor covering the ground truth and reduce false identification. Fine segmentation network is the second level, carrying out precise segmentation on the results of the coarse level. These two networks are independent of each other during training. When they are used, morphological processing of small scale corrosion and large scale expansion is used for the coarse segmentation results, and the outcomes are sent to the fine segmentation part as input, so as to achieve the complementary advantages of the two networks. RESULTS In the experiment, CT images of 200 patients with NSCLC are used to train the network, and CT images of 60 patients are used to test. Finally, our method produced the Dice similarity coefficient of 0.78 ± 0.10. CONCLUSIONS The experimental results show that the proposed method can accurately segment the tumor with NSCLC, and can also provide support for clinical diagnosis and treatment.
Collapse
Affiliation(s)
- Fuli Zhang
- Radiation Oncology Department, The Seventh Medical Center of Chinese PLA General Hospital, Beijing, China
| | - Qiusheng Wang
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, China
| | - Enyu Fan
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, China
| | - Na Lu
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Diandian Chen
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Huayong Jiang
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Yadi Wang
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| |
Collapse
|
69
|
Jin R, Wang M, Xu L, Lu J, Song E, Ma G. Automatic 3D CT liver segmentation based on fast global minimization of probabilistic active contour. Med Phys 2022; 50:2100-2120. [PMID: 36413182 DOI: 10.1002/mp.16116] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Revised: 10/27/2022] [Accepted: 11/05/2022] [Indexed: 11/23/2022] Open
Abstract
PURPOSE Automatic liver segmentation from computed tomography (CT) images is an essential preprocessing step for computer-aided diagnosis of liver diseases. However, due to the large differences in liver shapes, low-contrast to adjacent tissues, and existence of tumors or other abnormalities, liver segmentation has been very challenging. This study presents an accurate and fast liver segmentation method based on a novel probabilistic active contour (PAC) model and its fast global minimization scheme (3D-FGMPAC), which is explainable as compared with deep learning methods. METHODS The proposed method first constructs a slice-indexed-histogram to localize the volume of interest (VOI) and estimate the probability that a voxel belongs to the liver according its intensity. The probabilistic image would be used to initialize the 3D PAC model. Secondly, a new contour indicator function, which is a component of the model, is produced by combining the gradient-based edge detection and Hessian-matrix-based surface detection. Then, a fast numerical scheme derived for the 3D PAC model is performed to evolve the initial probabilistic image into the global minimizer of the model, which is a smoothed probabilistic image showing a distinctly highlighted liver. Next, a simple region-growing strategy is applied to extract the whole liver mask from the smoothed probabilistic image. Finally, a B-spline surface is constructed to fit the patch of the rib cage to prevent possible leakage into adjacent intercostal tissues. RESULTS The proposed method is evaluated on two public datasets. The average Dice score, volume overlap error, volume difference, symmetric surface distance and volume processing time are 0.96, 7.35%, 0.02%, 1.17 mm and 19.8 s for the Sliver07 dataset, and 0.95, 8.89%, - 0.02 % $-0.02\%$ , 1.45 mm and 23.08 s for the 3Dircadb dataset, respectively. CONCLUSIONS The proposed fully-automatic approach can effectively segment the liver from low-contrast and complex backgrounds. The quantitative and qualitative results demonstrate that the proposed segmentation method outperforms state-of-the-art traditional automatic liver segmentation algorithms and achieves very competitive performance compared with recent deep leaning-based methods.
Collapse
Affiliation(s)
- Renchao Jin
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Manyang Wang
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Lijun Xu
- School of Computer and Information Engineering, Hubei University, Wuhan, China
| | - Jiayi Lu
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Enmin Song
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Guangzhi Ma
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, China
| |
Collapse
|
70
|
Li J, Huang G, He J, Chen Z, Pun CM, Yu Z, Ling WK, Liu L, Zhou J, Huang J. Shift-channel attention and weighted-region loss function for liver and dense tumor segmentation. Med Phys 2022; 49:7193-7206. [PMID: 35746843 DOI: 10.1002/mp.15816] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Revised: 04/07/2022] [Accepted: 04/28/2022] [Indexed: 12/13/2022] Open
Abstract
PURPOSE To assist physicians in the diagnosis and treatment planning of tumor, a robust and automatic liver and tumor segmentation method is highly demanded in the clinical practice. Recently, numerous researchers have improved the segmentation accuracy of liver and tumor by introducing multiscale contextual information and attention mechanism. However, this tends to introduce more training parameters and suffer from a heavier computational burden. In addition, the tumor has various sizes, shapes, locations, and numbers, which is the main reason for the poor accuracy of automatic segmentation. Although current loss functions can improve the learning ability of the model for hard samples to a certain extent, these loss functions are difficult to optimize the segmentation effect of small tumor regions when the large tumor regions in the sample are in the majority. METHODS We propose a Liver and Tumor Segmentation Network (LiTS-Net) framework. First, the Shift-Channel Attention Module (S-CAM) is designed to model the feature interdependencies in adjacent channels and does not require additional training parameters. Second, the Weighted-Region (WR) loss function is proposed to emphasize the weight of small tumors in dense tumor regions and reduce the weight of easily segmented samples. Moreover, the Multiple 3D Inception Encoder Units (MEU) is adopted to capture the multiscale contextual information for better segmentation of liver and tumor. RESULTS Efficacy of the LiTS-Net is demonstrated through the public dataset of MICCAI 2017 Liver Tumor Segmentation (LiTS) challenge, with Dice per case of 96.9 % ${\bf \%}$ and 75.1 % ${\bf \%}$ , respectively. For the 3D Image Reconstruction for Comparison of Algorithm and DataBase (3Dircadb), Dices are 96.47 % ${\bf \%}$ for the liver and 74.54 % ${\bf \%}$ for tumor segmentation. The proposed LiTS-Net outperforms existing state-of-the-art networks. CONCLUSIONS We demonstrated the effectiveness of LiTS-Net and its core components for liver and tumor segmentation. The S-CAM is designed to model the feature interdependencies in the adjacent channels, which is characterized by no need to add additional training parameters. Meanwhile, we conduct an in-depth study of the feature shift proportion of adjacent channels to determine the optimal shift proportion. In addition, the WR loss function can implicitly learn the weights among regions without the need to manually specify the weights. In dense tumor segmentation tasks, WR aims to enhance the weights of small tumor regions and alleviate the problem that small tumor segmentation is difficult to optimize further when large tumor regions occupy the majority. Last but not least, our proposed method outperforms other state-of-the-art methods on both the LiTS dataset and the 3Dircadb dataset.
Collapse
Affiliation(s)
- Jiajian Li
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, China
| | - Guoheng Huang
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, China
| | - Junlin He
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, China
| | - Ziyang Chen
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, China
| | - Chi-Man Pun
- Department of Computer and Information Science, University of Macau, Macau, SAR, China
| | - Zhiwen Yu
- School of Computer Science and Engineering, South China University of Technology, Guangzhou, China
| | - Wing-Kuen Ling
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
| | - Lizhi Liu
- Department of Medical Imaging, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangzhou, China
| | - Jian Zhou
- Department of Medical Imaging, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangzhou, China
| | - Jinhua Huang
- Department of Minimal Invasive Interventional Therapy, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangzhou, China
| |
Collapse
|
71
|
Zhong X, Zhang H, Li G, Ji D. Do you need sharpened details? Asking MMDC-Net: Multi-layer multi-scale dilated convolution network for retinal vessel segmentation. Comput Biol Med 2022; 150:106198. [PMID: 37859292 DOI: 10.1016/j.compbiomed.2022.106198] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Revised: 09/19/2022] [Accepted: 10/09/2022] [Indexed: 11/24/2022]
Abstract
Convolutional neural networks (CNN), especially numerous U-shaped models, have achieved great progress in retinal vessel segmentation. However, a great quantity of global information in fundus images has not been fully explored. And the class imbalance problem of background and blood vessels is still serious. To alleviate these issues, we design a novel multi-layer multi-scale dilated convolution network (MMDC-Net) based on U-Net. We propose an MMDC module to capture sufficient global information under diverse receptive fields through a cascaded mode. Then, we place a new multi-layer fusion (MLF) module behind the decoder, which can not only fuse complementary features but filter noisy information. This enables MMDC-Net to capture the blood vessel details after continuous up-sampling. Finally, we employ a recall loss to resolve the class imbalance problem. Extensive experiments have been done on diverse fundus color image datasets, including STARE, CHASEDB1, DRIVE, and HRF. HRF has a large resolution of 3504 × 2336 whereas others have a small resolution of slightly more than 512 × 512. Qualitative and quantitative results verify the superiority of MMDC-Net. Notably, satisfactory accuracy and sensitivity are acquired by our model. Hence, some key blood vessel details are sharpened. In addition, a large number of further validations and discussions prove the effectiveness and generalization of the proposed MMDC-Net.
Collapse
Affiliation(s)
- Xiang Zhong
- School of Software, East China Jiaotong University, China
| | - Hongbin Zhang
- School of Software, East China Jiaotong University, China.
| | - Guangli Li
- School of Information Engineering, East China Jiaotong University, China
| | - Donghong Ji
- School of Cyber Science and Engineering, Wuhan University, China
| |
Collapse
|
72
|
Wu L, Zhuang J, Chen W, Tang Y, Hou C, Li C, Zhong Z, Luo S. Data augmentation based on multiple oversampling fusion for medical image segmentation. PLoS One 2022; 17:e0274522. [PMID: 36256637 PMCID: PMC9578635 DOI: 10.1371/journal.pone.0274522] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2022] [Accepted: 08/28/2022] [Indexed: 11/18/2022] Open
Abstract
A high-performance medical image segmentation model based on deep learning depends on the availability of large amounts of annotated training data. However, it is not trivial to obtain sufficient annotated medical images. Generally, the small size of most tissue lesions, e.g., pulmonary nodules and liver tumours, could worsen the class imbalance problem in medical image segmentation. In this study, we propose a multidimensional data augmentation method combining affine transform and random oversampling. The training data is first expanded by affine transformation combined with random oversampling to improve the prior data distribution of small objects and the diversity of samples. Secondly, class weight balancing is used to avoid having biased networks since the number of background pixels is much higher than the lesion pixels. The class imbalance problem is solved by utilizing weighted cross-entropy loss function during the training of the CNN model. The LUNA16 and LiTS17 datasets were introduced to evaluate the performance of our works, where four deep neural network models, Mask-RCNN, U-Net, SegNet and DeepLabv3+, were adopted for small tissue lesion segmentation in CT images. In addition, the small tissue segmentation performance of the four different deep learning architectures on both datasets could be greatly improved by incorporating the data augmentation strategy. The best pixelwise segmentation performance for both pulmonary nodules and liver tumours was obtained by the Mask-RCNN model, with DSC values of 0.829 and 0.879, respectively, which were similar to those of state-of-the-art methods.
Collapse
Affiliation(s)
- Liangsheng Wu
- Academy of Interdisciplinary Studies, Guangdong Polytechnic Normal University, Guangzhou, China
- Academy of Contemporary Agriculture Engineering Innovations, Zhongkai University of Agriculture and Engineering, Guangzhou, China
- Institute of Intelligent Manufacturing, Guangdong Academy of Sciences, Guangzhou, China
| | - Jiajun Zhuang
- Academy of Contemporary Agriculture Engineering Innovations, Zhongkai University of Agriculture and Engineering, Guangzhou, China
| | - Weizhao Chen
- Academy of Interdisciplinary Studies, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Yu Tang
- Academy of Interdisciplinary Studies, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Chaojun Hou
- Academy of Contemporary Agriculture Engineering Innovations, Zhongkai University of Agriculture and Engineering, Guangzhou, China
| | - Chentong Li
- Institute of Intelligent Manufacturing, Guangdong Academy of Sciences, Guangzhou, China
| | - Zhenyu Zhong
- Institute of Intelligent Manufacturing, Guangdong Academy of Sciences, Guangzhou, China
| | - Shaoming Luo
- Academy of Interdisciplinary Studies, Guangdong Polytechnic Normal University, Guangzhou, China
| |
Collapse
|
73
|
Lv P, Wang J, Zhang X, Shi C. Deep supervision and atrous inception-based U-Net combining CRF for automatic liver segmentation from CT. Sci Rep 2022; 12:16995. [PMID: 36216965 PMCID: PMC9550798 DOI: 10.1038/s41598-022-21562-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2022] [Accepted: 09/28/2022] [Indexed: 12/29/2022] Open
Abstract
Due to low contrast and the blurred boundary between liver tissue and neighboring organs sharing similar intensity values, the problem of liver segmentation from CT images has not yet achieved satisfactory performance and remains a challenge. To alleviate these problems, we introduce deep supervision (DS) and atrous inception (AI) technologies with conditional random field (CRF) and propose three major improvements that are experimentally shown to have substantive and practical value. First, we replace the encoder's standard convolution with the residual block. Residual blocks can increase the depth of the network. Second, we provide an AI module to connect the encoder and decoder. AI allows us to obtain multi-scale features. Third, we incorporate the DS mechanism into the decoder. This helps to make full use of information of the shallow layers. In addition, we employ the Tversky loss function to balance the segmented and non-segmented regions and perform further refinement with a dense CRF. Finally, we extensively validate the proposed method on three public databases: LiTS17, 3DIRCADb, and SLiver07. Compared to the state-of-the-art methods, the proposed method achieved increased segmentation accuracy for the livers with low contrast and the fuzzy boundary between liver tissue and neighboring organs and is, therefore, more suited for automatic segmentation of these livers.
Collapse
Affiliation(s)
- Peiqing Lv
- School of Automation, Harbin University of Science and Technology, Harbin, 150080, China
| | - Jinke Wang
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng, 264300, China.
- School of Automation, Harbin University of Science and Technology, Harbin, 150080, China.
| | - Xiangyang Zhang
- School of Automation, Harbin University of Science and Technology, Harbin, 150080, China
| | - Changfa Shi
- Mobile E-business Collaborative Innovation Center of Hunan Province, Hunan University of Technology and Business, Changsha, 410205, China
| |
Collapse
|
74
|
Hu W, Jiang H, Wang M. Flexible needle puncture path planning for liver tumors based on deep reinforcement learning. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac8fdd] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2022] [Accepted: 09/06/2022] [Indexed: 11/12/2022]
Abstract
Abstract
Objective. Minimally invasive surgery has been widely adopted in the treatment of patients with liver tumors. In liver tumor puncture surgery, an image-guided ablation needle for puncture surgery, which first reaches a target tumor along a predetermined path, and then ablates the tumor or injects drugs near the tumor, is often used to reduce patient trauma, improving the safety of surgery operations and avoiding possible damage to large blood vessels and key organs. In this paper, a path planning method for computer tomography (CT) guided ablation needle in liver tumor puncture surgery is proposed. Approach. Given a CT volume containing abdominal organs, we first classify voxels and optimize the number of voxels to reduce volume rendering pressure, then we reconstruct a multi-scale 3D model of the liver and hepatic vessels. Secondly, multiple entry points of the surgical path are selected based on the strong and weak constraints of clinical puncture surgery through multi-agent reinforcement learning. We select the optimal needle entry point based on the length measurement. Then, through the incremental training of the double deep Q-learning network (DDQN), the transmission of network parameters from the small-scale environment to the larger-scale environment is accomplished, and the optimal surgical path with more optimized details is obtained. Main results. To avoid falling into local optimum in network training, improve both the convergence speed and performance of the network, and maximize the cumulative reward, we train the path planning network on different scales 3D reconstructed organ models, and validate our method on tumor samples from public datasets. The scores of human surgeons verified the clinical relevance of the proposed method. Significance. Our method can robustly provide the optimal puncture path of flexible needle for liver tumors, which is expected to provide a reference for surgeons’ preoperative planning.
Collapse
|
75
|
Jiang X, Jiang J, Wang B, Yu J, Wang J. SEACU-Net: Attentive ConvLSTM U-Net with squeeze-and-excitation layer for skin lesion segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 225:107076. [PMID: 36027859 DOI: 10.1016/j.cmpb.2022.107076] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/20/2022] [Revised: 07/08/2022] [Accepted: 08/14/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Accurate segmentation of skin lesions is a pivotal step in dermoscopy image classification, which provides a powerful means for dermatologists to diagnose skin diseases. However, due to blurred boundaries, low contrast between the lesion and its surrounding skin, and changes in color and shape, most existing segmentation methods still face great challenges in obtaining receptive fields and extracting image feature information. To settle the above issues, we construct a new framework, named SEACU-Net, to analyze and segment skin lesion images. METHODS Inspired by the U-Net, we utilize dense convolution blocks to obtain more discriminative information. Then, at each encoding and decoding stage, a channel and spatial squeeze & excitation layer are designed after each convolution, to adaptively enhance useful information features and suppress low-value ones from different feature channels. In addition, the attention mechanism is integrated into the convolutional long short-term memory (ConvLSTM) structure, which improves sensitivity and prediction accuracy. Furthermore, this network introduces a novel loss based on binary cross-entropy and Jaccard losses, which can ensure more balanced segmentation. RESULTS The proposed method is applied to the ISIC 2017 and 2018 publicly image databases, then obtains a better performance in Dice, Jaccard, and Accuracy, with 89.11% and 87.58% Dice value, 80.50% and 78.12% Jaccard value, 95.01%, and 93.60% Accuracy value, respectively. CONCLUSION The results of quantitative and qualitative experiments show that our method reaches high-performance skin lesion segmentation, and can help radiologists make radiotherapy treatment plans in clinical practice.
Collapse
Affiliation(s)
- Xiaoliang Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou, Zhejiang, 324000, China.
| | - Jinyun Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou, Zhejiang, 324000, China
| | - Ban Wang
- School of Mechanical Engineering, Hangzhou Dianzi University, Hangzhou, Zhejiang, 310018, China
| | - Jianping Yu
- College of Mechanical Engineering, Quzhou University, Quzhou, Zhejiang, 324000, China
| | - Jun Wang
- College of Mechanical Engineering, Quzhou University, Quzhou, Zhejiang, 324000, China
| |
Collapse
|
76
|
Chen Y, Zhou T, Chen Y, Feng L, Zheng C, Liu L, Hu L, Pan B. HADCNet: Automatic segmentation of COVID-19 infection based on a hybrid attention dense connected network with dilated convolution. Comput Biol Med 2022; 149:105981. [PMID: 36029749 PMCID: PMC9391231 DOI: 10.1016/j.compbiomed.2022.105981] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2022] [Revised: 08/03/2022] [Accepted: 08/14/2022] [Indexed: 12/01/2022]
Abstract
the automatic segmentation of lung infections in CT slices provides a rapid and effective strategy for diagnosing, treating, and assessing COVID-19 cases. However, the segmentation of the infected areas presents several difficulties, including high intraclass variability and interclass similarity among infected areas, as well as blurred edges and low contrast. Therefore, we propose HADCNet, a deep learning framework that segments lung infections based on a dual hybrid attention strategy. HADCNet uses an encoder hybrid attention module to integrate feature information at different scales across the peer hierarchy to refine the feature map. Furthermore, a decoder hybrid attention module uses an improved skip connection to embed the semantic information of higher-level features into lower-level features by integrating multi-scale contextual structures and assigning the spatial information of lower-level features to higher-level features, thereby capturing the contextual dependencies of lesion features across levels and refining the semantic structure, which reduces the semantic gap between feature maps at different levels and improves the model segmentation performance. We conducted fivefold cross-validations of our model on four publicly available datasets, with final mean Dice scores of 0.792, 0.796, 0.785, and 0.723. These results show that the proposed model outperforms popular state-of-the-art semantic segmentation methods and indicate its potential use in the diagnosis and treatment of COVID-19.
Collapse
Affiliation(s)
- Ying Chen
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Taohui Zhou
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Yi Chen
- Department of Computer Science and Artificial Intelligence, Wenzhou University, Wenzhou, 325035, PR China.
| | - Longfeng Feng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Cheng Zheng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Lan Liu
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China.
| | - Liping Hu
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China.
| | - Bujian Pan
- Department of Hepatobiliary Surgery, Wenzhou Central Hospital, The Dingli Clinical Institute of Wenzhou Medical University, Wenzhou, Zhejiang, 325000, PR China.
| |
Collapse
|
77
|
Ma J, Zhang Y, Gu S, Zhu C, Ge C, Zhang Y, An X, Wang C, Wang Q, Liu X, Cao S, Zhang Q, Liu S, Wang Y, Li Y, He J, Yang X. AbdomenCT-1K: Is Abdominal Organ Segmentation a Solved Problem? IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:6695-6714. [PMID: 34314356 DOI: 10.1109/tpami.2021.3100536] [Citation(s) in RCA: 49] [Impact Index Per Article: 24.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
With the unprecedented developments in deep learning, automatic segmentation of main abdominal organs seems to be a solved problem as state-of-the-art (SOTA) methods have achieved comparable results with inter-rater variability on many benchmark datasets. However, most of the existing abdominal datasets only contain single-center, single-phase, single-vendor, or single-disease cases, and it is unclear whether the excellent performance can generalize on diverse datasets. This paper presents a large and diverse abdominal CT organ segmentation dataset, termed AbdomenCT-1K, with more than 1000 (1K) CT scans from 12 medical centers, including multi-phase, multi-vendor, and multi-disease cases. Furthermore, we conduct a large-scale study for liver, kidney, spleen, and pancreas segmentation and reveal the unsolved segmentation problems of the SOTA methods, such as the limited generalization ability on distinct medical centers, phases, and unseen diseases. To advance the unsolved problems, we further build four organ segmentation benchmarks for fully supervised, semi-supervised, weakly supervised, and continual learning, which are currently challenging and active research topics. Accordingly, we develop a simple and effective method for each benchmark, which can be used as out-of-the-box methods and strong baselines. We believe the AbdomenCT-1K dataset will promote future in-depth research towards clinical applicable abdominal organ segmentation methods.
Collapse
|
78
|
Senthilvelan J, Jamshidi N. A pipeline for automated deep learning liver segmentation (PADLLS) from contrast enhanced CT exams. Sci Rep 2022; 12:15794. [PMID: 36138084 PMCID: PMC9500060 DOI: 10.1038/s41598-022-20108-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2021] [Accepted: 09/08/2022] [Indexed: 11/23/2022] Open
Abstract
Multiple studies have created state-of-the-art liver segmentation models using Deep Convolutional Neural Networks (DCNNs) such as the V-net and H-DenseUnet. Oversegmentation however continues to be a problem. We set forth to address these limitations by developing a an automated workflow that leverages the strengths of different DCNN architectures, resulting in a pipeline that enables fully automated liver segmentation. A Pipeline for Automated Deep Learning Liver Segmentation (PADLLS) was developed and implemented that cascades multiple DCNNs that were trained on more than 200 CT scans. First, a V-net is used to create a rough liver, spleen, and stomach mask. After stomach and spleen pixels are removed using their respective masks and ascites is removed using a morphological algorithm, the scan is passed to a H-DenseUnet to yield the final segmentation. The segmentation accuracy of the pipleline was compared to the H-DenseUnet and the V-net using the SLIVER07 and 3DIRCADb datasets as benchmarks. The PADLLS Dice score for the SLIVER07 dataset was calculated to be 0.957 ± 0.033 and was significantly better than the H-DenseUnet's score of 0.927 ± 0.044 (p = 0.0219) and the V-net's score of 0.872 ± 0.121 (p = 0.0067). The PADLLS Dice score for the 3DIRCADb dataset was 0.965 ± 0.016 and was significantly better than the H-DenseUnet's score of 0.930 ± 0.041 (p = 0.0014) the V-net's score of 0.874 ± 0.060 (p < 0.001). In conclusion, our pipeline (PADLLS) outperforms existing liver segmentation models, serves as a valuable tool for image-based analysis, and is freely available for download and use.
Collapse
Affiliation(s)
- Jayasuriya Senthilvelan
- Department of Radiological Sciences, David Geffen School of Medicine, University of California, Los Angeles, 757 Westwood Ave, Suite 2125, Los Angeles, CA, 90095, USA
| | - Neema Jamshidi
- Department of Radiological Sciences, David Geffen School of Medicine, University of California, Los Angeles, 757 Westwood Ave, Suite 2125, Los Angeles, CA, 90095, USA.
| |
Collapse
|
79
|
Yu Q, Qi L, Gao Y, Wang W, Shi Y. Crosslink-Net: Double-Branch Encoder Network via Fusing Vertical and Horizontal Convolutions for Medical Image Segmentation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:5893-5908. [PMID: 36074869 DOI: 10.1109/tip.2022.3203223] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Accurate image segmentation plays a crucial role in medical image analysis, yet it faces great challenges caused by various shapes, diverse sizes, and blurry boundaries. To address these difficulties, square kernel-based encoder-decoder architectures have been proposed and widely used, but their performance remains unsatisfactory. To further address these challenges, we present a novel double-branch encoder architecture. Our architecture is inspired by two observations. (1) Since the discrimination of the features learned via square convolutional kernels needs to be further improved, we propose utilizing nonsquare vertical and horizontal convolutional kernels in a double-branch encoder so that the features learned by both branches can be expected to complement each other. (2) Considering that spatial attention can help models to better focus on the target region in a large-sized image, we develop an attention loss to further emphasize the segmentation of small-sized targets. With the above two schemes, we develop a novel double-branch encoder-based segmentation framework for medical image segmentation, namely, Crosslink-Net, and validate its effectiveness on five datasets with experiments. The code is released at https://github.com/Qianyu1226/Crosslink-Net.
Collapse
|
80
|
Czipczer V, Manno-Kovacs A. Adaptable volumetric liver segmentation model for CT images using region-based features and convolutional neural network. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.07.024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/17/2022]
|
81
|
Hao W, Zhang J, Su J, Song Y, Liu Z, Liu Y, Qiu C, Han K. HPM-Net: Hierarchical progressive multiscale network for liver vessel segmentation in CT images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 224:107003. [PMID: 35868034 DOI: 10.1016/j.cmpb.2022.107003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/30/2022] [Revised: 06/29/2022] [Accepted: 07/03/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE The segmentation and visualization of liver vessels in 3D CT images are essential for computer-aided diagnosis and preoperative planning of liver diseases. Due to the irregular structure of liver vessels and image noise, accurate extraction of liver vessels is difficult. In particular, accurate segmentation of small vessels is always a challenge, as multiple single down-sampling usually results in a loss of information. METHODS In this paper, we propose a hierarchical progressive multiscale learning network (HPM-Net) framework for liver vessel segmentation. Firstly, the hierarchical progressive multiscale learning network combines internal and external progressive learning methods to learn semantic information about liver vessels at different scales by acquiring receptive fields of different sizes. Secondly, to better capture vessel features, we propose a dual-branch progressive 3D Unet, which uses a dual-branch progressive (DBP) down-sampling strategy to reduce the loss of detailed information in the process of network down-sampling. Finally, a deep supervision mechanism is introduced into the framework and backbone network to speed up the network convergence and achieve better training of the network. RESULTS We conducted experiments on the public dataset 3Dircadb for liver vessel segmentation. The average dice coefficient and sensitivity of the proposed method reached 75.18% and 78.84%, respectively, both higher than the original network. CONCLUSION Experimental results show that the proposed hierarchical progressive multiscale network can accurately segment the labeled liver vessels from the CT images.
Collapse
Affiliation(s)
- Wen Hao
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
| | - Jing Zhang
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
| | - Jun Su
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
| | - Yuqing Song
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
| | - Zhe Liu
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China.
| | - Yi Liu
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
| | - Chengjian Qiu
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
| | - Kai Han
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
| |
Collapse
|
82
|
Dual attention-guided and learnable spatial transformation data augmentation multi-modal unsupervised medical image segmentation. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103849] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|
83
|
|
84
|
Ryselis K, Blažauskas T, Damaševičius R, Maskeliūnas R. Agrast-6: Abridged VGG-Based Reflected Lightweight Architecture for Binary Segmentation of Depth Images Captured by Kinect. SENSORS (BASEL, SWITZERLAND) 2022; 22:6354. [PMID: 36080813 PMCID: PMC9460068 DOI: 10.3390/s22176354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/03/2022] [Revised: 08/15/2022] [Accepted: 08/19/2022] [Indexed: 06/15/2023]
Abstract
Binary object segmentation is a sub-area of semantic segmentation that could be used for a variety of applications. Semantic segmentation models could be applied to solve binary segmentation problems by introducing only two classes, but the models to solve this problem are more complex than actually required. This leads to very long training times, since there are usually tens of millions of parameters to learn in this category of convolutional neural networks (CNNs). This article introduces a novel abridged VGG-16 and SegNet-inspired reflected architecture adapted for binary segmentation tasks. The architecture has 27 times fewer parameters than SegNet but yields 86% segmentation cross-intersection accuracy and 93% binary accuracy. The proposed architecture is evaluated on a large dataset of depth images collected using the Kinect device, achieving an accuracy of 99.25% in human body shape segmentation and 87% in gender recognition tasks.
Collapse
|
85
|
Ansari MY, Yang Y, Balakrishnan S, Abinahed J, Al-Ansari A, Warfa M, Almokdad O, Barah A, Omer A, Singh AV, Meher PK, Bhadra J, Halabi O, Azampour MF, Navab N, Wendler T, Dakua SP. A lightweight neural network with multiscale feature enhancement for liver CT segmentation. Sci Rep 2022; 12:14153. [PMID: 35986015 PMCID: PMC9391485 DOI: 10.1038/s41598-022-16828-6] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Accepted: 07/18/2022] [Indexed: 11/18/2022] Open
Abstract
Segmentation of abdominal Computed Tomography (CT) scan is essential for analyzing, diagnosing, and treating visceral organ diseases (e.g., hepatocellular carcinoma). This paper proposes a novel neural network (Res-PAC-UNet) that employs a fixed-width residual UNet backbone and Pyramid Atrous Convolutions, providing a low disk utilization method for precise liver CT segmentation. The proposed network is trained on medical segmentation decathlon dataset using a modified surface loss function. Additionally, we evaluate its quantitative and qualitative performance; the Res16-PAC-UNet achieves a Dice coefficient of 0.950 ± 0.019 with less than half a million parameters. Alternatively, the Res32-PAC-UNet obtains a Dice coefficient of 0.958 ± 0.015 with an acceptable parameter count of approximately 1.2 million.
Collapse
Affiliation(s)
| | - Yin Yang
- Hamad Bin Khalifa University, Doha, Qatar
| | | | | | | | - Mohamed Warfa
- Wake Forest Baptist Medical Center, Winston-Salem, USA
| | | | - Ali Barah
- Hamad Medical Corporation, Doha, Qatar
| | | | | | | | | | | | | | | | | | | |
Collapse
|
86
|
Deep Learning-Based CT Imaging in the Diagnosis of Treatment Effect of Pulmonary Nodules and Radiofrequency Ablation. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:7326537. [PMID: 35996649 PMCID: PMC9392615 DOI: 10.1155/2022/7326537] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/08/2022] [Accepted: 05/26/2022] [Indexed: 11/17/2022]
Abstract
To study the effect of computerized tomography (CT) images based on deep learning algorithms on the diagnosis of pulmonary nodules and the effect of radiofrequency ablation (RFA), the U-shaped fully convolutional neural network (FCNN) (U-Net) was enhanced. The convolutional neural network (CNN) algorithm was compared with the U-Net algorithm, and segmentation performances were analyzed. Then, it was applied to the CT image diagnosis of 110 lung cancer patients admitted to hospital. The patients in the observation group (55 cases) were diagnosed based on the improved U-Net algorithm, while those in the control group (55 cases) were diagnosed by traditional methods and then treated with RFA. The Dice coefficient (0.8753) and intersection over union (IOU) (0.8788) obtained by the proposed algorithm were remarkably higher than the Dice coefficient (0.7212) and IOU (0.7231) obtained by the CNN algorithm, and the differences were considerable (P < 0.05). The boundary of the pulmonary nodule can be segmented more accurately by the proposed algorithm, which had the segmentation result closest to the gold standard among the three algorithms. The diagnostic accuracy of the pulmonary nodule in the observation group (95.3%) was superior to that of the control group (90.7%). The long diameter, volume, and maximum area of the pulmonary nodule of the observation group were significantly higher than those of the control group, with substantial differences (P < 0.05). Patients were reexamined after one, three, and six months of treatment, and 71 patients (64.55%) had complete remission, 32 patients (29.10%) had partial remission, 6 patients (5.45%) had stable disease, and 1 patient (0.90%) had disease progression. The remission rate (complete remission + partial remission) was 93.65%. The improved U-NET algorithm had good image segmentation performance and ideal segmentation effect. It can clearly display the shape of pulmonary nodules, locate the lesions, and accurately evaluate the therapeutic effect of RFA, which had clinical application value.
Collapse
|
87
|
Anderson BM, Rigaud B, Lin YM, Jones AK, Kang HC, Odisio BC, Brock KK. Automated segmentation of colorectal liver metastasis and liver ablation on contrast-enhanced CT images. Front Oncol 2022; 12:886517. [PMID: 36033508 PMCID: PMC9403767 DOI: 10.3389/fonc.2022.886517] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2022] [Accepted: 07/01/2022] [Indexed: 11/13/2022] Open
Abstract
Objectives Colorectal cancer (CRC), the third most common cancer in the USA, is a leading cause of cancer-related death worldwide. Up to 60% of patients develop liver metastasis (CRLM). Treatments like radiation and ablation therapies require disease segmentation for planning and therapy delivery. For ablation, ablation-zone segmentation is required to evaluate disease coverage. We hypothesize that fully convolutional (FC) neural networks, trained using novel methods, will provide rapid and accurate identification and segmentation of CRLM and ablation zones. Methods Four FC model styles were investigated: Standard 3D-UNet, Residual 3D-UNet, Dense 3D-UNet, and Hybrid-WNet. Models were trained on 92 patients from the liver tumor segmentation (LiTS) challenge. For the evaluation, we acquired 15 patients from the 3D-IRCADb database, 18 patients from our institution (CRLM = 24, ablation-zone = 19), and those submitted to the LiTS challenge (n = 70). Qualitative evaluations of our institutional data were performed by two board-certified radiologists (interventional and diagnostic) and a radiology-trained physician fellow, using a Likert scale of 1-5. Results The most accurate model was the Hybrid-WNet. On a patient-by-patient basis in the 3D-IRCADb dataset, the median (min-max) Dice similarity coefficient (DSC) was 0.73 (0.41-0.88), the median surface distance was 1.75 mm (0.57-7.63 mm), and the number of false positives was 1 (0-4). In the LiTS challenge (n = 70), the global DSC was 0.810. The model sensitivity was 98% (47/48) for sites ≥15 mm in diameter. Qualitatively, 100% (24/24; minority vote) of the CRLM and 84% (16/19; majority vote) of the ablation zones had Likert scores ≥4. Conclusion The Hybrid-WNet model provided fast (<30 s) and accurate segmentations of CRLM and ablation zones on contrast-enhanced CT scans, with positive physician reviews.
Collapse
Affiliation(s)
- Brian M. Anderson
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
- UTHealth Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Bastien Rigaud
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Yuan-Mao Lin
- Department of Interventional Radiology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - A. Kyle Jones
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - HynSeon Christine Kang
- Department of Abdominal Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Bruno C. Odisio
- Department of Interventional Radiology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Kristy K. Brock
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| |
Collapse
|
88
|
Li Y, Zhang Y, Cui W, Lei B, Kuang X, Zhang T. Dual Encoder-Based Dynamic-Channel Graph Convolutional Network With Edge Enhancement for Retinal Vessel Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1975-1989. [PMID: 35167444 DOI: 10.1109/tmi.2022.3151666] [Citation(s) in RCA: 34] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Retinal vessel segmentation with deep learning technology is a crucial auxiliary method for clinicians to diagnose fundus diseases. However, the deep learning approaches inevitably lose the edge information, which contains spatial features of vessels while performing down-sampling, leading to the limited segmentation performance of fine blood vessels. Furthermore, the existing methods ignore the dynamic topological correlations among feature maps in the deep learning framework, resulting in the inefficient capture of the channel characterization. To address these limitations, we propose a novel dual encoder-based dynamic-channel graph convolutional network with edge enhancement (DE-DCGCN-EE) for retinal vessel segmentation. Specifically, we first design an edge detection-based dual encoder to preserve the edge of vessels in down-sampling. Secondly, we investigate a dynamic-channel graph convolutional network to map the image channels to the topological space and synthesize the features of each channel on the topological map, which solves the limitation of insufficient channel information utilization. Finally, we study an edge enhancement block, aiming to fuse the edge and spatial features in the dual encoder, which is beneficial to improve the accuracy of fine blood vessel segmentation. Competitive experimental results on five retinal image datasets validate the efficacy of the proposed DE-DCGCN-EE, which achieves more remarkable segmentation results against the other state-of-the-art methods, indicating its potential clinical application.
Collapse
|
89
|
Gul S, Khan MS, Bibi A, Khandakar A, Ayari MA, Chowdhury ME. Deep learning techniques for liver and liver tumor segmentation: A review. Comput Biol Med 2022; 147:105620. [DOI: 10.1016/j.compbiomed.2022.105620] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2021] [Revised: 02/18/2022] [Accepted: 03/19/2022] [Indexed: 12/29/2022]
|
90
|
CEDRNN: A Convolutional Encoder-Decoder Residual Neural Network for Liver Tumour Segmentation. Neural Process Lett 2022. [DOI: 10.1007/s11063-022-10953-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
|
91
|
Zhang H, Zhong X, Li Z, Chen Y, Zhu Z, Lv J, Li C, Zhou Y, Li G. TiM-Net: Transformer in M-Net for Retinal Vessel Segmentation. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:9016401. [PMID: 35859930 PMCID: PMC9293566 DOI: 10.1155/2022/9016401] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/30/2022] [Revised: 06/04/2022] [Accepted: 06/21/2022] [Indexed: 11/17/2022]
Abstract
retinal image is a crucial window for the clinical observation of cardiovascular, cerebrovascular, or other correlated diseases. Retinal vessel segmentation is of great benefit to the clinical diagnosis. Recently, the convolutional neural network (CNN) has become a dominant method in the retinal vessel segmentation field, especially the U-shaped CNN models. However, the conventional encoder in CNN is vulnerable to noisy interference, and the long-rang relationship in fundus images has not been fully utilized. In this paper, we propose a novel model called Transformer in M-Net (TiM-Net) based on M-Net, diverse attention mechanisms, and weighted side output layers to efficaciously perform retinal vessel segmentation. First, to alleviate the effects of noise, a dual-attention mechanism based on channel and spatial is designed. Then the self-attention mechanism in Transformer is introduced into skip connection to re-encode features and model the long-range relationship explicitly. Finally, a weighted SideOut layer is proposed for better utilization of the features from each side layer. Extensive experiments are conducted on three public data sets to show the effectiveness and robustness of our TiM-Net compared with the state-of-the-art baselines. Both quantitative and qualitative results prove its clinical practicality. Moreover, variants of TiM-Net also achieve competitive performance, demonstrating its scalability and generalization ability. The code of our model is available at https://github.com/ZX-ECJTU/TiM-Net.
Collapse
Affiliation(s)
- Hongbin Zhang
- School of Software, East China Jiaotong University, Nanchang, China
| | - Xiang Zhong
- School of Software, East China Jiaotong University, Nanchang, China
| | - Zhijie Li
- School of Software, East China Jiaotong University, Nanchang, China
| | - Yanan Chen
- School of International, East China Jiaotong University, Nanchang, China
| | - Zhiliang Zhu
- School of Software, East China Jiaotong University, Nanchang, China
| | - Jingqin Lv
- School of Software, East China Jiaotong University, Nanchang, China
| | - Chuanxiu Li
- School of Information Engineering, East China Jiaotong University, Nanchang, China
| | - Ying Zhou
- Medical School, Nanchang University, Nanchang, China
| | - Guangli Li
- School of Information Engineering, East China Jiaotong University, Nanchang, China
| |
Collapse
|
92
|
Maiti S, Maji D, Kumar Dhara A, Sarkar G. Automatic detection and segmentation of optic disc using a modified convolution network. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103633] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
93
|
Lee Z, Qi S, Fan C, Xie Z, Meng J. RA V-Net: deep learning network for automated liver segmentation. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac7193] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2021] [Accepted: 05/19/2022] [Indexed: 11/12/2022]
Abstract
Abstract
Objective. Segmenting liver from CT images is the first step for doctors to diagnose a patient’s disease. Processing medical images with deep learning models has become a current research trend. Although it can automate segmenting region of interest of medical images, the inability to achieve the required segmentation accuracy is an urgent problem to be solved. Approach. Residual Attention V-Net (RA V-Net) based on U-Net is proposed to improve the performance of medical image segmentation. Composite Original Feature Residual Module is proposed to achieve a higher level of image feature extraction capability and prevent gradient disappearance or explosion. Attention Recovery Module is proposed to add spatial attention to the model. Channel Attention Module is introduced to extract relevant channels with dependencies and strengthen them by matrix dot product. Main results. Through test, evaluation index has improved significantly. Lits2017 and 3Dircadb are chosen as our experimental datasets. On the Dice Similarity Coefficient, RA V-Net exceeds U-Net 0.1107 in Lits2017, and 0.0754 in 3Dircadb. On the Jaccard Similarity Coefficient, RA V-Net exceeds U-Net 0.1214 in Lits2017, and 0.13 in 3Dircadb. Significance. Combined with all the innovations, the model performs brightly in liver segmentation without clear over-segmentation and under-segmentation. The edges of organs are sharpened considerably with high precision. The model we proposed provides a reliable basis for the surgeon to design the surgical plans.
Collapse
|
94
|
Karmakar R, Nooshabadi S. Mobile-PolypNet: Lightweight Colon Polyp Segmentation Network for Low-Resource Settings. J Imaging 2022; 8:jimaging8060169. [PMID: 35735968 PMCID: PMC9225047 DOI: 10.3390/jimaging8060169] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Revised: 06/08/2022] [Accepted: 06/10/2022] [Indexed: 11/16/2022] Open
Abstract
Colon polyps, small clump of cells on the lining of the colon, can lead to colorectal cancer (CRC), one of the leading types of cancer globally. Hence, early detection of these polyps automatically is crucial in the prevention of CRC. The deep learning models proposed for the detection and segmentation of colorectal polyps are resource-consuming. This paper proposes a lightweight deep learning model for colorectal polyp segmentation that achieved state-of-the-art accuracy while significantly reducing the model size and complexity. The proposed deep learning autoencoder model employs a set of state-of-the-art architectural blocks and optimization objective functions to achieve the desired efficiency. The model is trained and tested on five publicly available colorectal polyp segmentation datasets (CVC-ClinicDB, CVC-ColonDB, EndoScene, Kvasir, and ETIS). We also performed ablation testing on the model to test various aspects of the autoencoder architecture. We performed the model evaluation by using most of the common image-segmentation metrics. The backbone model achieved a DICE score of 0.935 on the Kvasir dataset and 0.945 on the CVC-ClinicDB dataset, improving the accuracy by 4.12% and 5.12%, respectively, over the current state-of-the-art network, while using 88 times fewer parameters, 40 times less storage space, and being computationally 17 times more efficient. Our ablation study showed that the addition of ConvSkip in the autoencoder slightly improves the model's performance but it was not significant (p-value = 0.815).
Collapse
|
95
|
Wang X, Wei Y. Optimization algorithm of CT image edge segmentation using improved convolution neural network. PLoS One 2022; 17:e0265338. [PMID: 35657969 PMCID: PMC9165790 DOI: 10.1371/journal.pone.0265338] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2021] [Accepted: 03/01/2022] [Indexed: 11/30/2022] Open
Abstract
To address the problem of high failure rate and low accuracy in computed tomography (CT) image edge segmentation, we proposed a CT sequence image edge segmentation optimization algorithm using improved convolution neural network. Firstly, the pattern clustering algorithm is applied to cluster the pixels with relationship in the CT sequence image space to extract the edge information of the real CT image; secondly, Euclidean distance is used to calculate similarity and measure similarity, according to the measurement results, convolution neural network (CNN) hierarchical optimization is carried out to improve the convergence ability of CNN; finally, the pixel classification of CT sequence images is carried out, and the edge segmentation of CT sequence images is optimized according to the classification results. The results show that the overall recognition rate of this method is at a high level. The training time is obviously reduced when the training times exceed 12 times, the recall rate is always about 90%, and the accuracy of image segmentation is high, which solves the problem of large failure rate and low accuracy.
Collapse
Affiliation(s)
- Xiaojuan Wang
- College of Electronic Information Technology, Jiamusi University, Jiamusi, China
| | - Yuntao Wei
- College of Electronic Information Technology, Jiamusi University, Jiamusi, China
- * E-mail:
| |
Collapse
|
96
|
Li C, Mao Y, Guo Y, Li J, Wang Y. Multi-Dimensional Cascaded Net with Uncertain Probability Reduction for Abdominal Multi-Organ Segmentation in CT Sequences. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 221:106887. [PMID: 35597204 DOI: 10.1016/j.cmpb.2022.106887] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/27/2022] [Revised: 05/03/2022] [Accepted: 05/11/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Deep learning abdominal multi-organ segmentation provides preoperative guidance for abdominal surgery. However, due to the large volume of 3D CT sequences, the existing methods cannot balance complete semantic features and high-resolution detail information, which leads to uncertain, rough, and inaccurate segmentation, especially in small and irregular organs. In this paper, we propose a two-stage algorithm named multi-dimensional cascaded net (MDCNet) to solve the above problems and segment multi-organs in CT images, including the spleen, kidney, gallbladder, esophagus, liver, stomach, pancreas, and duodenum. METHODS MDCNet combines the powerful semantic encoder ability of a 3D net and the rich high-resolution information of a 2.5D net. In stage1, a prior-guided shallow-layer-enhanced 3D location net extracts entire semantic features from a downsampled CT volume to perform rough segmentation. Additionally, we use circular inference and parameter Dice loss to alleviate uncertain boundary. The inputs of stage2 are high-resolution slices, which are obtained by the original image and coarse segmentation of stage1. Stage2 offsets the details lost during downsampling, resulting in smooth and accurate refined contours. The 2.5D net from the axial, coronal, and sagittal views also compensates for the missing spatial information of a single view. RESULTS The experiments on the two datasets both obtained the best performance, particularly a higher Dice on small gallbladders and irregular duodenums, which reached 0.85±0.12 and 0.77±0.07 respectively, increasing by 0.02 and 0.03 compared to the state-of-the-art method. CONCLUSION Our method can extract all semantic and high-resolution detail information from a large-volume CT image. It reduces the boundary uncertainty while yielding smoother segmentation edges, indicating good clinical application prospects.
Collapse
Affiliation(s)
- Chengkang Li
- School of Information Science and Technology of Fudan University, Shanghai 200433, China; Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention (MICCAI) of Shanghai, Shanghai 200032, China
| | - Yishen Mao
- Department of Pancreatic Surgery, Pancreatic Disease Institute, Huashan Hospital, Shanghai Medical College, Fudan University, Shanghai 200040, China
| | - Yi Guo
- School of Information Science and Technology of Fudan University, Shanghai 200433, China; Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention (MICCAI) of Shanghai, Shanghai 200032, China.
| | - Ji Li
- Department of Pancreatic Surgery, Pancreatic Disease Institute, Huashan Hospital, Shanghai Medical College, Fudan University, Shanghai 200040, China.
| | - Yuanyuan Wang
- School of Information Science and Technology of Fudan University, Shanghai 200433, China; Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention (MICCAI) of Shanghai, Shanghai 200032, China.
| |
Collapse
|
97
|
Ansari MY, Abdalla A, Ansari MY, Ansari MI, Malluhi B, Mohanty S, Mishra S, Singh SS, Abinahed J, Al-Ansari A, Balakrishnan S, Dakua SP. Practical utility of liver segmentation methods in clinical surgeries and interventions. BMC Med Imaging 2022; 22:97. [PMID: 35610600 PMCID: PMC9128093 DOI: 10.1186/s12880-022-00825-2] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2022] [Accepted: 05/09/2022] [Indexed: 12/15/2022] Open
Abstract
Clinical imaging (e.g., magnetic resonance imaging and computed tomography) is a crucial adjunct for clinicians, aiding in the diagnosis of diseases and planning of appropriate interventions. This is especially true in malignant conditions such as hepatocellular carcinoma (HCC), where image segmentation (such as accurate delineation of liver and tumor) is the preliminary step taken by the clinicians to optimize diagnosis, staging, and treatment planning and intervention (e.g., transplantation, surgical resection, radiotherapy, PVE, embolization, etc). Thus, segmentation methods could potentially impact the diagnosis and treatment outcomes. This paper comprehensively reviews the literature (during the year 2012-2021) for relevant segmentation methods and proposes a broad categorization based on their clinical utility (i.e., surgical and radiological interventions) in HCC. The categorization is based on the parameters such as precision, accuracy, and automation.
Collapse
|
98
|
2.5D lightweight RIU-Net for automatic liver and tumor segmentation from CT. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103567] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022]
|
99
|
Ashkani Chenarlogh V, Shabanzadeh A, Ghelich Oghli M, Sirjani N, Farzin Moghadam S, Akhavan A, Arabi H, Shiri I, Shabanzadeh Z, Sanei Taheri M, Kazem Tarzamni M. Clinical target segmentation using a novel deep neural network: double attention Res-U-Net. Sci Rep 2022; 12:6717. [PMID: 35468984 PMCID: PMC9038725 DOI: 10.1038/s41598-022-10429-z] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2021] [Accepted: 03/24/2022] [Indexed: 01/10/2023] Open
Abstract
We introduced Double Attention Res-U-Net architecture to address medical image segmentation problem in different medical imaging system. Accurate medical image segmentation suffers from some challenges including, difficulty of different interest object modeling, presence of noise, and signal dropout throughout the measurement. The base line image segmentation approaches are not sufficient for complex target segmentation throughout the various medical image types. To overcome the issues, a novel U-Net-based model proposed that consists of two consecutive networks with five and four encoding and decoding levels respectively. In each of networks, there are four residual blocks between the encoder-decoder path and skip connections that help the networks to tackle the vanishing gradient problem, followed by the multi-scale attention gates to generate richer contextual information. To evaluate our architecture, we investigated three distinct data-sets, (i.e., CVC-ClinicDB dataset, Multi-site MRI dataset, and a collected ultrasound dataset). The proposed algorithm achieved Dice and Jaccard coefficients of 95.79%, 91.62%, respectively for CRL, and 93.84% and 89.08% for fetal foot segmentation. Moreover, the proposed model outperformed the state-of-the-art U-Net based model on the external CVC-ClinicDB, and multi-site MRI datasets with Dice and Jaccard coefficients of 83%, 75.31% for CVC-ClinicDB, and 92.07% and 87.14% for multi-site MRI dataset, respectively.
Collapse
Affiliation(s)
- Vahid Ashkani Chenarlogh
- Research and Development Department, Med Fanavaran Plus Co., Karaj, Iran
- Department of Electrical and Computer Engineering, National Center for Audiology, Western University, London, Canada
| | - Ali Shabanzadeh
- Research and Development Department, Med Fanavaran Plus Co., Karaj, Iran
| | - Mostafa Ghelich Oghli
- Research and Development Department, Med Fanavaran Plus Co., Karaj, Iran.
- Department of Cardiovascular Sciences, KU Leuven, Leuven, Belgium.
| | - Nasim Sirjani
- Research and Development Department, Med Fanavaran Plus Co., Karaj, Iran
| | | | - Ardavan Akhavan
- Research and Development Department, Med Fanavaran Plus Co., Karaj, Iran
| | - Hossein Arabi
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, 1211, Geneva 4, Switzerland
| | - Isaac Shiri
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, 1211, Geneva 4, Switzerland
| | - Zahra Shabanzadeh
- School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Morteza Sanei Taheri
- Department of Radiology, Shohada-e-Tajrish Hospital, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Mohammad Kazem Tarzamni
- Department of Radiology, Imam Reza Hospital, Tabriz University of Medical Sciences, Tabriz, Iran
| |
Collapse
|
100
|
Liu Q, Wang J, Zuo M, Cao W, Zheng J, Zhao H, Xie J. NCRNet: Neighborhood Context Refinement Network for skin lesion segmentation. Comput Biol Med 2022; 146:105545. [PMID: 35477048 DOI: 10.1016/j.compbiomed.2022.105545] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2022] [Revised: 04/11/2022] [Accepted: 04/17/2022] [Indexed: 12/24/2022]
Abstract
Accurate skin lesion segmentation plays a fundamental role in computer-aided melanoma analysis. Recently, some FCN-based methods have been proposed and achieved promising results in lesion segmentation tasks. However, due to the variable shapes, different scales, noise interference, and ambiguous boundaries of skin lesions, the capabilities of lesion location and boundary delineation of these works are still insufficient. To overcome the above challenges, in this paper, we propose a novel Neighborhood Context Refinement Network (NCRNet) by using the coarse-to-fine strategy to achieve accurate skin lesion segmentation. The proposed NCRNet contains a shared encoder and two different but closely related decoders for locating the skin lesions and refining the lesion boundaries. Specifically, we first design the Parallel Attention Decoder (PAD), which can effectively extract and fuse the local detail information and global semantic information on multiple levels to locate skin lesions of different sizes and shapes. Then, based on the initial lesion location, we further design the Neighborhood Context Refinement Decoder (NCRD), aiming at leveraging the fine-grained multi-stage neighborhood context cues to refine the lesion boundaries continuously. Furthermore, the neighborhood-based deep supervision used in the NCRD can make the shared encoder pay more attention to the lesion boundary areas and promote convergence of the segmentation network. The public skin lesion segmentation dataset ISIC2017 is adopted to validate the effectiveness of the proposed NCRNet. Comprehensive experiments prove that the proposed NCRNet reaches the state-of-the-art performance than the other nine competitive methods and gets 78.62%, 86.55%, and 94.01% on Jaccard, Dice, and Accuracy, respectively.
Collapse
Affiliation(s)
- Qi Liu
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Jingkun Wang
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Mengying Zuo
- Department of Cardiology, Children's Hospital of Soochow University, Suzhou, 215003, China
| | - Weiwei Cao
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China.
| | - Jian Zheng
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China; Jinan Guoke Medical Technology Development Co., Ltd, Jinan, 250101, China
| | - Hui Zhao
- The Wenzhou Third Clinical Institute Affiliated to Wenzhou Medical University, (Wenzhou People's Hospital), Wenzhou, 325000, China
| | - Jing Xie
- The Wenzhou Third Clinical Institute Affiliated to Wenzhou Medical University, (Wenzhou People's Hospital), Wenzhou, 325000, China.
| |
Collapse
|