1
|
Guo X, Xu L, Liu Z, Hao Y, Wang P, Zhu H, Du Y. Automated classification of ulcerative lesions in small intestine using densenet with channel attention and residual dilated blocks. Phys Med Biol 2024; 69:055017. [PMID: 38316034 DOI: 10.1088/1361-6560/ad2637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Accepted: 02/05/2024] [Indexed: 02/07/2024]
Abstract
Objective. Ulceration of the small intestine, which has a high incidence, includes Crohn's disease (CD), intestinal tuberculosis (ITB), primary small intestinal lymphoma (PSIL), cryptogenic multifocal ulcerous stenosing enteritis (CMUSE), and non-specific ulcer (NSU). However, the ulceration morphology can easily be misdiagnosed through enteroscopy.Approach. In this study, DRCA-DenseNet169, which is based on DenseNet169, with residual dilated blocks and a channel attention block, is proposed to identify CD, ITB, PSIL, CMUSE, and NSU intelligently. In addition, a novel loss function that incorporates dynamic weights is designed to enhance the precision of imbalanced datasets with limited samples. DRCA-Densenet169 was evaluated using 10883 enteroscopy images, including 5375 ulcer images and 5508 normal images, which were obtained from the Shanghai Changhai Hospital.Main results. DRCA-Densenet169 achieved an overall accuracy of 85.27% ± 0.32%, a weighted-precision of 83.99% ± 2.47%, a weighted-recall of 84.36% ± 0.88% and a weighted-F1-score of 84.07% ± 2.14%.Significance. The results demonstrate that DRCA-Densenet169 has high recognition accuracy and strong robustness in identifying different types of ulcers when obtaining immediate and preliminary diagnoses.
Collapse
Affiliation(s)
- Xudong Guo
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Lei Xu
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Zhang Liu
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Youguo Hao
- Department of Rehabilitation, Shanghai Putuo People's Hospital, Shanghai 200060, People's Republic of China
| | - Peng Wang
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Huiyun Zhu
- Department of Gastroenterology, Changhai Hospital, Naval Medical University, Shanghai 200433, People's Republic of China
| | - Yiqi Du
- Department of Gastroenterology, Changhai Hospital, Naval Medical University, Shanghai 200433, People's Republic of China
| |
Collapse
|
2
|
You H, Yu L, Tian S, Cai W. A stereo spatial decoupling network for medical image classification. COMPLEX INTELL SYST 2023; 9:1-10. [PMID: 37361963 PMCID: PMC10107597 DOI: 10.1007/s40747-023-01049-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2022] [Accepted: 03/09/2023] [Indexed: 06/28/2023]
Abstract
Deep convolutional neural network (CNN) has made great progress in medical image classification. However, it is difficult to establish effective spatial associations, and always extracts similar low-level features, resulting in redundancy of information. To solve these limitations, we propose a stereo spatial discoupling network (TSDNets), which can leverage the multi-dimensional spatial details of medical images. Then, we use an attention mechanism to progressively extract the most discriminative features from three directions: horizontal, vertical, and depth. Moreover, a cross feature screening strategy is used to divide the original feature maps into three levels: important, secondary and redundant. Specifically, we design a cross feature screening module (CFSM) and a semantic guided decoupling module (SGDM) to model multi-dimension spatial relationships, thereby enhancing the feature representation capabilities. The extensive experiments conducted on multiple open source baseline datasets demonstrate that our TSDNets outperforms previous state-of-the-art models.
Collapse
Affiliation(s)
- Hongfeng You
- School of Information Science and Engineering, Xinjiang University, Urumqi, 830000 China
| | - Long Yu
- Network Center, Xinjiang University, Urumqi, 830000 China
| | - Shengwei Tian
- Software College, Xinjiang University, Urumqi, 830000 China
| | - Weiwei Cai
- School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi, 214122 China
| |
Collapse
|
3
|
Li S, Yao J, Cao J, Kong X, Zhu J. Effective high-to-low-level feature aggregation network for endoscopic image classification. Int J Comput Assist Radiol Surg 2022; 17:1225-1233. [PMID: 35568744 DOI: 10.1007/s11548-022-02591-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2021] [Accepted: 03/04/2022] [Indexed: 11/28/2022]
Abstract
PURPOSE The accuracy improvement in endoscopic image classification matters to the endoscopists in diagnosing and choosing suitable treatment for patients. Existing CNN-based methods for endoscopic image classification tend to use the deepest abstract features without considering the contribution of low-level features, while the latter is of great significance in the actual diagnosis of intestinal diseases. METHODS To make full use of both high-level and low-level features, we propose a novel two-stream network for endoscopic image classification. Specifically, the backbone stream is utilized to extract high-level features. In the fusion stream, low-level features are generated by a bottom-up multi-scale gradual integration (BMGI) method, and the input of BMGI is refined by top-down attention learning modules. Besides, a novel correction loss is proposed to clarify the relationship between high-level and low-level features. RESULTS Experiments on the KVASIR dataset demonstrate that the proposed framework can obtain an overall classification accuracy of 97.33% with Kappa coefficient of 95.25%. Compared to the existing models, the two evaluation indicators have increased by 2% and 2.25%, respectively, at least. CONCLUSION In this study, we proposed a two-stream network that fuses the high-level and low-level features for endoscopic image classification. The experiment results show that the high-to-low-level feature can better represent the endoscopic image and enable our model to outperform several state-of-the-art classification approaches. In addition, the proposed correction loss could regularize the consistency between backbone stream and fusion stream. Thus, the fused feature can reduce the intra-class distances and make accurate label prediction.
Collapse
Affiliation(s)
- Sheng Li
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, 310023, Zhejiang, People's Republic of China
| | - Jiafeng Yao
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, 310023, Zhejiang, People's Republic of China
| | - Jing Cao
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, 310023, Zhejiang, People's Republic of China
| | - Xueting Kong
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, 310023, Zhejiang, People's Republic of China
| | - Jinhui Zhu
- The Second Affiliated Hospital of Hospital of Zhejiang University School of Medicine, Hangzhou, 310023, Zhejiang, People's Republic of China.
| |
Collapse
|
4
|
Reuss J, Pascual G, Wenzek H, Seguí S. Sequential Models for Endoluminal Image Classification. Diagnostics (Basel) 2022; 12:501. [PMID: 35204591 PMCID: PMC8871077 DOI: 10.3390/diagnostics12020501] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2021] [Revised: 02/04/2022] [Accepted: 02/11/2022] [Indexed: 11/30/2022] Open
Abstract
Wireless Capsule Endoscopy (WCE) is a procedure to examine the human digestive system for potential mucosal polyps, tumours, or bleedings using an encapsulated camera. This work focuses on polyp detection within WCE videos through Machine Learning. When using Machine Learning in the medical field, scarce and unbalanced datasets often make it hard to receive a satisfying performance. We claim that using Sequential Models in order to take the temporal nature of the data into account improves the performance of previous approaches. Thus, we present a bidirectional Long Short-Term Memory Network (BLSTM), a sequential network that is particularly designed for temporal data. We find the BLSTM Network outperforms non-sequential architectures and other previous models, receiving a final Area under the Curve of 93.83%. Experiments show that our method of extracting spatial and temporal features yields better performance and could be a possible method to decrease the time needed by physicians to analyse the video material.
Collapse
Affiliation(s)
- Joana Reuss
- Department of Mathematics and Computer Science, Universitat de Barcelona, 08007 Barcelona, Spain; (G.P.); (S.S.)
- Chair of Remote Sensing Technology, Technical University of Munich, 80333 Munich, Germany
| | - Guillem Pascual
- Department of Mathematics and Computer Science, Universitat de Barcelona, 08007 Barcelona, Spain; (G.P.); (S.S.)
| | - Hagen Wenzek
- CorporateHealth International ApS, 5230 Odense, Denmark;
| | - Santi Seguí
- Department of Mathematics and Computer Science, Universitat de Barcelona, 08007 Barcelona, Spain; (G.P.); (S.S.)
| |
Collapse
|
5
|
Zhong Y, Piao Y, Zhang G. Dilated and soft attention-guided convolutional neural network for breast cancer histology images classification. Microsc Res Tech 2021; 85:1248-1257. [PMID: 34859543 DOI: 10.1002/jemt.23991] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2021] [Revised: 10/03/2021] [Accepted: 10/18/2021] [Indexed: 01/22/2023]
Abstract
Breast cancer is one of the most common types of cancer in women, and histopathological imaging is considered the gold standard for its diagnosis. However, the great complexity of histopathological images and the considerable workload make this work extremely time-consuming, and the results may be affected by the subjectivity of the pathologist. Therefore, the development of an accurate, automated method for analysis of histopathological images is critical to this field. In this article, we propose a deep learning method guided by the attention mechanism for fast and effective classification of haematoxylin and eosin-stained breast biopsy images. First, this method takes advantage of DenseNet and uses the feature map's information. Second, we introduce dilated convolution to produce a larger receptive field. Finally, spatial attention and channel attention are used to guide the extraction of the most useful visual features. With the use of fivefold cross-validation, the best model obtained an accuracy of 96.47% on the BACH2018 dataset. We also evaluated our method on other datasets, and the experimental results demonstrated that our model has reliable performance. This study indicates that our histopathological image classifier with a soft attention-guided deep learning model for breast cancer shows significantly better results than the latest methods. It has great potential as an effective tool for automatic evaluation of digital histopathological microscopic images for computer-aided diagnosis.
Collapse
Affiliation(s)
- Yutong Zhong
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, China
| | - Yan Piao
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, China
| | - Guohui Zhang
- Pneumoconiosis Diagnosis and Treatment Center, Occupational Preventive and Treatment Hospital in Jilin Province, Changchun, China
| |
Collapse
|
6
|
Zhu M, Chen Z, Yuan Y. DSI-Net: Deep Synergistic Interaction Network for Joint Classification and Segmentation With Endoscope Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3315-3325. [PMID: 34033538 DOI: 10.1109/tmi.2021.3083586] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Automatic classification and segmentation of wireless capsule endoscope (WCE) images are two clinically significant and relevant tasks in a computer-aided diagnosis system for gastrointestinal diseases. Most of existing approaches, however, considered these two tasks individually and ignored their complementary information, leading to limited performance. To overcome this bottleneck, we propose a deep synergistic interaction network (DSI-Net) for joint classification and segmentation with WCE images, which mainly consists of the classification branch (C-Branch), the coarse segmentation (CS-Branch) and the fine segmentation branches (FS-Branch). In order to facilitate the classification task with the segmentation knowledge, a lesion location mining (LLM) module is devised in C-Branch to accurately highlight lesion regions through mining neglected lesion areas and erasing misclassified background areas. To assist the segmentation task with the classification prior, we propose a category-guided feature generation (CFG) module in FS-Branch to improve pixel representation by leveraging the category prototypes of C-Branch to obtain the category-aware features. In such way, these modules enable the deep synergistic interaction between these two tasks. In addition, we introduce a task interaction loss to enhance the mutual supervision between the classification and segmentation tasks and guarantee the consistency of their predictions. Relying on the proposed deep synergistic interaction mechanism, DSI-Net achieves superior classification and segmentation performance on public dataset in comparison with state-of-the-art methods. The source code is available at https://github.com/CityU-AIM-Group/DSI-Net.
Collapse
|
7
|
Chen Z, Guo X, Woo PYM, Yuan Y. Super-Resolution Enhanced Medical Image Diagnosis With Sample Affinity Interaction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1377-1389. [PMID: 33507866 DOI: 10.1109/tmi.2021.3055290] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The degradation in image resolution harms the performance of medical image diagnosis. By inferring high-frequency details from low-resolution (LR) images, super-resolution (SR) techniques can introduce additional knowledge and assist high-level tasks. In this paper, we propose a SR enhanced diagnosis framework, consisting of an efficient SR network and a diagnosis network. Specifically, a Multi-scale Refined Context Network (MRC-Net) with Refined Context Fusion (RCF) is devised to leverage global and local features for SR tasks. Instead of learning from scratch, we first develop a recursive MRC-Net with temporal context, and then propose a recursion distillation scheme to enhance the performance of MRC-Net from the knowledge of the recursive one and reduce the computational cost. The diagnosis network jointly utilizes the reliable original images and more informative SR images by two branches, with the proposed Sample Affinity Interaction (SAI) blocks at different stages to effectively extract and integrate discriminative features towards diagnosis. Moreover, two novel constraints, sample affinity consistency and sample affinity regularization, are devised to refine the features and achieve the mutual promotion of these two branches. Extensive experiments of synthetic and real LR cases are conducted on wireless capsule endoscopy and histopathology images, verifying that our proposed method is significantly effective for medical image diagnosis.
Collapse
|
8
|
Guo X, Yang C, Liu Y, Yuan Y. Learn to Threshold: ThresholdNet With Confidence-Guided Manifold Mixup for Polyp Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1134-1146. [PMID: 33360986 DOI: 10.1109/tmi.2020.3046843] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The automatic segmentation of polyp in endoscopy images is crucial for early diagnosis and cure of colorectal cancer. Existing deep learning-based methods for polyp segmentation, however, are inadequate due to the limited annotated dataset and the class imbalance problems. Moreover, these methods obtained the final polyp segmentation results by simply thresholding the likelihood maps at an eclectic and equivalent value (often set to 0.5). In this paper, we propose a novel ThresholdNet with a confidence-guided manifold mixup (CGMMix) data augmentation method, mainly for addressing the aforementioned issues in polyp segmentation. The CGMMix conducts manifold mixup at the image and feature levels, and adaptively lures the decision boundary away from the under-represented polyp class with the confidence guidance to alleviate the limited training dataset and the class imbalance problems. Two consistency regularizations, mixup feature map consistency (MFMC) loss and mixup confidence map consistency (MCMC) loss, are devised to exploit the consistent constraints in the training of the augmented mixup data. We then propose a two-branch approach, termed ThresholdNet, to collaborate the segmentation and threshold learning in an alternative training strategy. The threshold map supervision generator (TMSG) is embedded to provide supervision for the threshold map, thereby inducing better optimization of the threshold branch. As a consequence, ThresholdNet is able to calibrate the segmentation result with the learned threshold map. We illustrate the effectiveness of the proposed method on two polyp segmentation datasets, and our methods achieved the state-of-the-art result with 87.307% and 87.879% dice score on the EndoScene dataset and the WCE polyp dataset. The source code is available at https://github.com/Guo-Xiaoqing/ThresholdNet.
Collapse
|
9
|
Wu Z, Ge R, Shi G, Zhang L, Chen Y, Luo L, Cao Y, Yu H. MD-NDNet: a multi-dimensional convolutional neural network for false-positive reduction in pulmonary nodule detection. Phys Med Biol 2020; 65:235053. [PMID: 32698172 DOI: 10.1088/1361-6560/aba87c] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/29/2023]
Abstract
Pulmonary nodule false-positive reduction is of great significance for automated nodule detection in clinical diagnosis of low-dose computed tomography (LDCT) lung cancer screening. Due to individual intra-nodule variations and visual similarities between true nodules and false positives as soft tissues in LDCT images, the current clinical practices remain subject to shortcomings of potential high-risk and time-consumption issues. In this paper, we propose a multi-dimensional nodule detection network (MD-NDNet) for automatic nodule false-positive reduction using deep convolutional neural network (DCNNs). The underlying method collaboratively integrates multi-dimensional nodule information to complementarily and comprehensively extract nodule inter-plane volumetric correlation features using three-dimensional CNNs (3D CNNs) and spatial nodule correlation features from sagittal, coronal, and axial planes using two-dimensional CNNs (2D CNNs) with attention module. To incorporate different sizes and shapes of nodule candidates, a multi-scale ensemble strategy is employed for probability aggregation with weights. The proposed method is evaluated on the LUNA16 challenge dataset in ISBI 2016 with ten-fold cross-validation. Experiment results show that the proposed framework achieves classification performance with a CPM score of 0.9008. All of these indicate that our method enables an efficient, accurate and reliable pulmonary nodule detection for clinical diagnosis.
Collapse
Affiliation(s)
- Zhan Wu
- School of Cyberspace Security, Southeast University, Nanjing, Jiangsu, China. Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA, United States of America
| | | | | | | | | | | | | | | |
Collapse
|
10
|
Xing X, Yuan Y, Meng MQH. Zoom in Lesions for Better Diagnosis: Attention Guided Deformation Network for WCE Image Classification. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:4047-4059. [PMID: 32746146 DOI: 10.1109/tmi.2020.3010102] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Wireless capsule endoscopy (WCE) is a novel imaging tool that allows noninvasive visualization of the entire gastrointestinal (GI) tract without causing discomfort to patients. Convolutional neural networks (CNNs), though perform favorably against traditional machine learning methods, show limited capacity in WCE image classification due to the small lesions and background interference. To overcome these limits, we propose a two-branch Attention Guided Deformation Network (AGDN) for WCE image classification. Specifically, the attention maps of branch1 are utilized to guide the amplification of lesion regions on the input images of branch2, thus leading to better representation and inspection of the small lesions. What's more, we devise and insert Third-order Long-range Feature Aggregation (TLFA) modules into the network. By capturing long-range dependencies and aggregating contextual features, TLFAs endow the network with a global contextual view and stronger feature representation and discrimination capability. Furthermore, we propose a novel Deformation based Attention Consistency (DAC) loss to refine the attention maps and achieve the mutual promotion of the two branches. Finally, the global feature embeddings from the two branches are fused to make image label predictions. Extensive experiments show that the proposed AGDN outperforms state-of-the-art methods with an overall classification accuracy of 91.29% on two public WCE datasets. The source code is available at https://github.com/hathawayxxh/WCE-AGDN.
Collapse
|