51
|
Zhao L, Asis-Cruz JD, Feng X, Wu Y, Kapse K, Largent A, Quistorff J, Lopez C, Wu D, Qing K, Meyer C, Limperopoulos C. Automated 3D Fetal Brain Segmentation Using an Optimized Deep Learning Approach. AJNR Am J Neuroradiol 2022; 43:448-454. [PMID: 35177547 PMCID: PMC8910820 DOI: 10.3174/ajnr.a7419] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2021] [Accepted: 12/06/2021] [Indexed: 01/01/2023]
Abstract
BACKGROUND AND PURPOSE MR imaging provides critical information about fetal brain growth and development. Currently, morphologic analysis primarily relies on manual segmentation, which is time-intensive and has limited repeatability. This work aimed to develop a deep learning-based automatic fetal brain segmentation method that provides improved accuracy and robustness compared with atlas-based methods. MATERIALS AND METHODS A total of 106 fetal MR imaging studies were acquired prospectively from fetuses between 23 and 39 weeks of gestation. We trained a deep learning model on the MR imaging scans of 65 healthy fetuses and compared its performance with a 4D atlas-based segmentation method using the Wilcoxon signed-rank test. The trained model was also evaluated on data from 41 fetuses diagnosed with congenital heart disease. RESULTS The proposed method showed high consistency with the manual segmentation, with an average Dice score of 0.897. It also demonstrated significantly improved performance (P < .001) based on the Dice score and 95% Hausdorff distance in all brain regions compared with the atlas-based method. The performance of the proposed method was consistent across gestational ages. The segmentations of the brains of fetuses with high-risk congenital heart disease were also highly consistent with the manual segmentation, though the Dice score was 7% lower than that of healthy fetuses. CONCLUSIONS The proposed deep learning method provides an efficient and reliable approach for fetal brain segmentation, which outperformed segmentation based on a 4D atlas and has been used in clinical and research settings.
Collapse
Affiliation(s)
- L Zhao
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
- Department of Biomedical Engineering (L.Z., D.W.), Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering & Instrument Science, Zhejiang University, China
| | - J D Asis-Cruz
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| | - X Feng
- Department of Biomedical Engineering (X.F., C.M.), University of Virginia, Charlottesville, Virginia
| | - Y Wu
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| | - K Kapse
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| | - A Largent
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| | - J Quistorff
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| | - C Lopez
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| | - D Wu
- Department of Biomedical Engineering (L.Z., D.W.), Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering & Instrument Science, Zhejiang University, China
| | - K Qing
- Department of Radiation Oncology (K.Q.), City of Hope National Center, Duarte, California
| | - C Meyer
- Department of Biomedical Engineering (X.F., C.M.), University of Virginia, Charlottesville, Virginia
| | - C Limperopoulos
- From the Department of Diagnostic Imaging and Radiology (L.Z., J.D.A.-C., Y.W., K.K., A.L., J.Q., C. Lopez, C. Limperopoulos), Developing Brain Institute, Children's National, Washington, DC
| |
Collapse
|
52
|
A Multi-Agent Deep Reinforcement Learning Approach for Enhancement of COVID-19 CT Image Segmentation. J Pers Med 2022; 12:jpm12020309. [PMID: 35207796 PMCID: PMC8880720 DOI: 10.3390/jpm12020309] [Citation(s) in RCA: 31] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2022] [Revised: 02/14/2022] [Accepted: 02/15/2022] [Indexed: 11/18/2022] Open
Abstract
Currently, most mask extraction techniques are based on convolutional neural networks (CNNs). However, there are still numerous problems that mask extraction techniques need to solve. Thus, the most advanced methods to deploy artificial intelligence (AI) techniques are necessary. The use of cooperative agents in mask extraction increases the efficiency of automatic image segmentation. Hence, we introduce a new mask extraction method that is based on multi-agent deep reinforcement learning (DRL) to minimize the long-term manual mask extraction and to enhance medical image segmentation frameworks. A DRL-based method is introduced to deal with mask extraction issues. This new method utilizes a modified version of the Deep Q-Network to enable the mask detector to select masks from the image studied. Based on COVID-19 computed tomography (CT) images, we used DRL mask extraction-based techniques to extract visual features of COVID-19 infected areas and provide an accurate clinical diagnosis while optimizing the pathogenic diagnostic test and saving time. We collected CT images of different cases (normal chest CT, pneumonia, typical viral cases, and cases of COVID-19). Experimental validation achieved a precision of 97.12% with a Dice of 80.81%, a sensitivity of 79.97%, a specificity of 99.48%, a precision of 85.21%, an F1 score of 83.01%, a structural metric of 84.38%, and a mean absolute error of 0.86%. Additionally, the results of the visual segmentation clearly reflected the ground truth. The results reveal the proof of principle for using DRL to extract CT masks for an effective diagnosis of COVID-19.
Collapse
|
53
|
Protonotarios NE, Katsamenis I, Sykiotis S, Dikaios N, Kastis GA, Chatziioannou SN, Metaxas M, Doulamis N, Doulamis A. A few-shot U-Net deep learning model for lung cancer lesion segmentation via PET/CT imaging. Biomed Phys Eng Express 2022; 8. [PMID: 35144242 DOI: 10.1088/2057-1976/ac53bd] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2021] [Accepted: 02/10/2022] [Indexed: 11/12/2022]
Abstract
Over the past few years, positron emission tomography/computed tomography (PET/CT) imaging for computer-aided diagnosis has received increasing attention. Supervised deep learning architectures are usually employed for the detection of abnormalities, with anatomical localization, especially in the case of CT scans. However, the main limitations of the supervised learning paradigm include (i) large amounts of data required for model training, and (ii) the assumption of fixed network weights upon training completion, implying that the performance of the model cannot be further improved after training. In order to overcome these limitations, we apply a few-shot learning (FSL) scheme. Contrary to traditional deep learning practices, in FSL the model is provided with less data during training. The model then utilizes end-user feedback after training to constantly improve its performance. We integrate FSL in a U-Net architecture for lung cancer lesion segmentation on PET/CT scans, allowing for dynamic model weight fine-tuning and resulting in an online supervised learning scheme. Constant online readjustments of the model weights according to the user's feedback, increase the detection and classification accuracy, especially in cases where low detection performance is encountered. Our proposed method is validated on the Lung-PET-CT-DX TCIA database. PET/CT scans from 87 patients were included in the dataset and were acquired 60 minutes after intravenous18F-FDG injection. Experimental results indicate the superiority of our approach compared to other state of the art methods.
Collapse
Affiliation(s)
- Nicholas E Protonotarios
- Department of Applied Mathematics and Theoretical Physics (DAMTP), University of Cambridge, University of Cambridge, Cambridge, CB3 0WA, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
| | - Iason Katsamenis
- School of Rural and Surveying Engineering, National Technical University of Athens, 9, Heroon Polytechniou, Zografou, Attica, 157 73, GREECE
| | - Stavros Sykiotis
- School of Rural and Surveying Engineering, National Technical University of Athens, 9, Heroon Polytechniou, Zografou, Attica, 157 73, GREECE
| | - Nikolaos Dikaios
- Mathematics Research Center, Academy of Athens, 4, Soranou Efesiou, Athens, 115 27, GREECE
| | - George Anthony Kastis
- Mathematics Research Center, Academy of Athens, 4, Soranou Efesiou, Athens, Attica, 115 27, GREECE
| | - Sofia N Chatziioannou
- PET/CT, Biomedical Research Foundation of the Academy of Athens, 4, Soranou Efesiou, Athens, Attica, 115 27, GREECE
| | - Marinos Metaxas
- PET/CT, Biomedical Research Foundation of the Academy of Athens, 4, Soranou Efesiou, Athens, Attica, 115 27, GREECE
| | - Nikolaos Doulamis
- School of Rural and Surveying Engineering, National Technical University of Athens, 9, Heroon Polytechniou, Zografou, Attica, 157 73, GREECE
| | - Anastasios Doulamis
- School of Rural and Surveying Engineering, National Technical University of Athens, 9, Heroon Polytechniou, Zografou, Attica, 157 73, GREECE
| |
Collapse
|
54
|
Abdou MA. Literature review: efficient deep neural networks techniques for medical image analysis. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-06960-9] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2023]
|
55
|
|
56
|
Xiong H, Liu S, Sharan RV, Coiera E, Berkovsky S. Weak label based Bayesian U-Net for optic disc segmentation in fundus images. Artif Intell Med 2022; 126:102261. [DOI: 10.1016/j.artmed.2022.102261] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2021] [Revised: 01/18/2022] [Accepted: 02/20/2022] [Indexed: 01/27/2023]
|
57
|
Liu X, Yuan Q, Gao Y, He K, Wang S, Tang X, Tang J, Shen D. Weakly Supervised Segmentation of COVID19 Infection with Scribble Annotation on CT Images. PATTERN RECOGNITION 2022; 122:108341. [PMID: 34565913 PMCID: PMC8452156 DOI: 10.1016/j.patcog.2021.108341] [Citation(s) in RCA: 34] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Revised: 08/23/2021] [Accepted: 09/18/2021] [Indexed: 05/19/2023]
Abstract
Segmentation of infections from CT scans is important for accurate diagnosis and follow-up in tackling the COVID-19. Although the convolutional neural network has great potential to automate the segmentation task, most existing deep learning-based infection segmentation methods require fully annotated ground-truth labels for training, which is time-consuming and labor-intensive. This paper proposed a novel weakly supervised segmentation method for COVID-19 infections in CT slices, which only requires scribble supervision and is enhanced with the uncertainty-aware self-ensembling and transformation-consistent techniques. Specifically, to deal with the difficulty caused by the shortage of supervision, an uncertainty-aware mean teacher is incorporated into the scribble-based segmentation method, encouraging the segmentation predictions to be consistent under different perturbations for an input image. This mean teacher model can guide the student model to be trained using information in images without requiring manual annotations. On the other hand, considering the output of the mean teacher contains both correct and unreliable predictions, equally treating each prediction in the teacher model may degrade the performance of the student network. To alleviate this problem, the pixel level uncertainty measure on the predictions of the teacher model is calculated, and then the student model is only guided by reliable predictions from the teacher model. To further regularize the network, a transformation-consistent strategy is also incorporated, which requires the prediction to follow the same transformation if a transform is performed on an input image of the network. The proposed method has been evaluated on two public datasets and one local dataset. The experimental results demonstrate that the proposed method is more effective than other weakly supervised methods and achieves similar performance as those fully supervised.
Collapse
Affiliation(s)
- Xiaoming Liu
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan, China
- Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, Wuhan, China
| | - Quan Yuan
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan, China
- Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, Wuhan, China
| | - Yaozong Gao
- Department of Research and Development, Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
| | - Kelei He
- Medical School, Nanjing University, Nanjing, China
- National Institute of Healthcare Data Science at Nanjing University, Nanjing, China
| | - Shuo Wang
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan, China
- Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, Wuhan, China
| | - Xiao Tang
- Department of Medical Imaging, Tianyou Hospital Affiliated to Wuhan University of Science and Technology, Wuhan, China
| | - Jinshan Tang
- Department of Health Administration and Policy, George Mason University, Fairfax, VA, 22030, USA
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
- Department of Research and Development, Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
- Department of Artificial Intelligence, Korea University, Seoul 02841, Republic of Korea
| |
Collapse
|
58
|
Ma T, Wang Q, Zhang H, Zuo W. Delving Deeper Into Pixel Prior for Box-Supervised Semantic Segmentation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:1406-1417. [PMID: 35038294 DOI: 10.1109/tip.2022.3141878] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Weakly supervised semantic segmentation (WSSS) based on bounding box annotations has attracted considerable recent attention and has achieved promising performance. However, most of existing methods focus on generation of high-quality pseudo labels for segmented objects using box indicators, but they fail to fully explore and exploit prior from bounding box annotations, which limits performance of WSSS methods, especially for fine parts and boundaries. To overcome above issues, this paper proposes a novel Pixel-as-Instance Prior (PIP) for WSSS methods by delving deeper into pixel prior from bounding box annotations. Specifically, the proposed PIP is built on two important observations on pixels around bounding boxes. First, since objects are usually irregularity and tightly close to bounding boxes (dubbed irregular-filling prior), so each row or column of bounding boxes basically have at least one pixel belonging to foreground objects and background, respectively. Second, pixels near the bounding boxes tend to be highly ambiguous and more difficult to classify (dubbed label-ambiguity prior). To implement our PIP, a constrained loss alike multiple instance learning (MIL) and a labeling-balance loss are developed to jointly train WSSS models, which regards each pixel as a weighted positive or negative instance while considering more effective prior (i.e., irregular-filling and label-ambiguity priors) from bounding box annotations in an efficient way. Note that our PIP can be flexibly integrated with various WSSS methods, while clearly improving their performance with negligible computational overload in training stage. The experiments are conducted on most widely used PASCAL VOC 2012 and Cityscapes benchmarks, and the results show that our PIP has a good ability to improve performance of various WSSS methods, while achieving very competitive results.
Collapse
|
59
|
Li R, Chen X. An efficient interactive multi-label segmentation tool for 2D and 3D medical images using fully connected conditional random field. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 213:106534. [PMID: 34839271 DOI: 10.1016/j.cmpb.2021.106534] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/10/2021] [Revised: 09/18/2021] [Accepted: 11/12/2021] [Indexed: 06/13/2023]
Abstract
OBJECTIVE Image segmentation is a crucial and fundamental step in many medical image analysis tasks, such as tumor measurement, surgery planning, disease diagnosis, etc. To ensure the quality of image segmentation, most of the current solutions require labor-intensive manual processes by tracing the boundaries of the objects. The workload increases tremendously for the case of three dimensional (3D) image with multiple objects to be segmented. METHOD In this paper, we introduce our developed interactive image segmentation tool that provides efficient segmentation of multiple labels for both 2D and 3D medical images. The core segmentation method is based on a fast implementation of the fully connected conditional random field. The software also enables automatic recommendation of the next slice to be annotated in 3D, leading to a higher efficiency. RESULTS We have evaluated the tool on many 2D and 3D medical image modalities (e.g. CT, MRI, ultrasound, X-ray, etc.) and different objects of interest (abdominal organs, tumor, bones, etc.), in terms of segmentation accuracy, repeatability and computational time. CONCLUSION In contrast to other interactive image segmentation tools, our software produces high quality image segmentation results without the requirement of parameter tuning for each application. Both the software and source code are freely available for research purpose1. 1Software and source code download: https://drive.google.com/file/d/1JIzWkT3M-X7jeB8tTwVcEw240TGbJAvj/view?usp=sharing.
Collapse
Affiliation(s)
- Ruizhe Li
- Intelligent Modelling and Analysis Group, School of Computer Science, University of Nottingham, UK.
| | - Xin Chen
- Intelligent Modelling and Analysis Group, School of Computer Science, University of Nottingham, UK.
| |
Collapse
|
60
|
Faheem Saleem M, Muhammad Adnan Shah S, Nazir T, Mehmood A, Nawaz M, Attique Khan M, Kadry S, Majumdar A, Thinnukool O. Signet Ring Cell Detection from Histological Images Using Deep Learning. COMPUTERS, MATERIALS & CONTINUA 2022; 72:5985-5997. [DOI: 10.32604/cmc.2022.023101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/27/2021] [Accepted: 11/29/2021] [Indexed: 08/25/2024]
|
61
|
Liu T, Wang Y, Yang Y, Sun M, Fan W, Bunger C, Wu C. A multi-scale keypoint estimation network with self-supervision for spinal curvature assessment of idiopathic scoliosis from the imperfect dataset. Artif Intell Med 2022; 125:102235. [DOI: 10.1016/j.artmed.2021.102235] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2020] [Revised: 12/05/2021] [Accepted: 12/29/2021] [Indexed: 11/24/2022]
|
62
|
Patel G, Dolz J. Weakly supervised segmentation with cross-modality equivariant constraints. Med Image Anal 2022; 77:102374. [DOI: 10.1016/j.media.2022.102374] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2021] [Revised: 01/12/2022] [Accepted: 01/18/2022] [Indexed: 10/19/2022]
|
63
|
A Suspicious Multi-Object Detection and Recognition Method for Millimeter Wave SAR Security Inspection Images Based on Multi-Path Extraction Network. REMOTE SENSING 2021. [DOI: 10.3390/rs13244978] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
There are several major challenges in detecting and recognizing multiple hidden objects from millimeter wave SAR security inspection images: inconsistent clarity of objects, similar objects, and complex background interference. To address these problems, a suspicious multi-object detection and recognition method based on the Multi-Path Extraction Network (MPEN) is proposed. In MPEN, You Only Look Once (YOLO) v3 is used as the base network, and then the Multi-Path Feature Pyramid (MPFP) module and modified residual block distribution are proposed. MPFP is designed to output the deep network feature layers separately. Then, to distinguish similar objects more easily, the residual block distribution is modified to improve the ability of the shallow network to capture details. To verify the effectiveness of the proposed method, the millimeter wave SAR images from the laboratory’s self-developed security inspection system are utilized in conducting research on multi-object detection and recognition. The detection rate (probability of detecting a target) and average false alarm (probability of error detection) rate of our method on the target are 94.6% and 14.6%, respectively. The mean Average Precision (mAP) of recognizing multi-object is 82.39%. Compared with YOLOv3, our method shows a better performance in detecting and recognizing similar targets.
Collapse
|
64
|
Thanellas A, Peura H, Wennervirta J, Korja M. Foundations of Brain Image Segmentation: Pearls and Pitfalls in Segmenting Intracranial Blood on Computed Tomography Images. ACTA NEUROCHIRURGICA. SUPPLEMENT 2021; 134:153-159. [PMID: 34862539 DOI: 10.1007/978-3-030-85292-4_19] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
Abstract
Not only the time-dependent varying of signal intensity (i.e. haematoma evolution) characteristics of the intracranial blood in computed tomography images, but also the fluctuating image quality, the distortions introduced after medical interventions, and the brain deformations and intensity profile variations due to underlying pathologies make the segmentation of intracranial blood a challenging task. In addition to describing various challenges with blood segmentation, this chapter also reviews the following: (1) the general concept of segmentation-explaining why a proper segmentation is a critical step when creating machine learning algorithms for image detection purposes, (2) the different segmentation types and how different medical conditions and technical issues can further complicate this task, (3) how to choose a proper software to facilitate the segmentation task, and (4) useful tips that may be applied before launching a similar segmentation project.
Collapse
Affiliation(s)
- Antonios Thanellas
- Department of Information Management, Helsinki University Hospital, Helsinki, Finland
| | - Heikki Peura
- Department of Neurosurgery, Helsinki University Hospital and University of Helsinki, Helsinki, Finland
| | - Jenni Wennervirta
- Department of Neurosurgery, Helsinki University Hospital and University of Helsinki, Helsinki, Finland
| | - Miikka Korja
- Department of Neurosurgery, Helsinki University Hospital and University of Helsinki, Helsinki, Finland.
| |
Collapse
|
65
|
Dang VN, Galati F, Cortese R, Di Giacomo G, Marconetto V, Mathur P, Lekadir K, Lorenzi M, Prados F, Zuluaga MA. Vessel-CAPTCHA: An efficient learning framework for vessel annotation and segmentation. Med Image Anal 2021; 75:102263. [PMID: 34731770 DOI: 10.1016/j.media.2021.102263] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2021] [Revised: 10/04/2021] [Accepted: 10/06/2021] [Indexed: 12/22/2022]
Abstract
Deep learning techniques for 3D brain vessel image segmentation have not been as successful as in the segmentation of other organs and tissues. This can be explained by two factors. First, deep learning techniques tend to show poor performances at the segmentation of relatively small objects compared to the size of the full image. Second, due to the complexity of vascular trees and the small size of vessels, it is challenging to obtain the amount of annotated training data typically needed by deep learning methods. To address these problems, we propose a novel annotation-efficient deep learning vessel segmentation framework. The framework avoids pixel-wise annotations, only requiring weak patch-level labels to discriminate between vessel and non-vessel 2D patches in the training set, in a setup similar to the CAPTCHAs used to differentiate humans from bots in web applications. The user-provided weak annotations are used for two tasks: (1) to synthesize pixel-wise pseudo-labels for vessels and background in each patch, which are used to train a segmentation network, and (2) to train a classifier network. The classifier network allows to generate additional weak patch labels, further reducing the annotation burden, and it acts as a second opinion for poor quality images. We use this framework for the segmentation of the cerebrovascular tree in Time-of-Flight angiography (TOF) and Susceptibility-Weighted Images (SWI). The results show that the framework achieves state-of-the-art accuracy, while reducing the annotation time by ∼77% w.r.t. learning-based segmentation methods using pixel-wise labels for training.
Collapse
Affiliation(s)
- Vien Ngoc Dang
- Data Science Department, EURECOM, Sophia Antipolis, France; Artificial Intelligence in Medicine Lab, Facultat de Matemátiques I Informática, Universitat de Barcelona, Spain
| | | | - Rosa Cortese
- Queen Square MS Centre, Department of Neuroinflammation, UCL Queen Square Institute of Neurology, Faculty of Brain Sciences, University College London, UK; Department of Medicine, Surgery and Neuroscience, University of Siena, Italy
| | - Giuseppe Di Giacomo
- Data Science Department, EURECOM, Sophia Antipolis, France; Politecnico di Torino, Turin, Italy
| | - Viola Marconetto
- Data Science Department, EURECOM, Sophia Antipolis, France; Politecnico di Torino, Turin, Italy
| | - Prateek Mathur
- Data Science Department, EURECOM, Sophia Antipolis, France
| | - Karim Lekadir
- Artificial Intelligence in Medicine Lab, Facultat de Matemátiques I Informática, Universitat de Barcelona, Spain
| | - Marco Lorenzi
- Université Côte d'Azur, Inria Sophia Antipolis, Epione Research Group, Valbonne, France
| | - Ferran Prados
- Centre for Medical Image Computing, Department of Medical Physics and Bioengineering, University College London, UK; Queen Square MS Centre, Department of Neuroinflammation, UCL Queen Square Institute of Neurology, Faculty of Brain Sciences, University College London, UK; National Institute for Health Research, University College London Hospitals, Biomedical Research Centre, London, UK; e-health Center, Universitat Oberta de Catalunya, Barcelona, Spain
| | | |
Collapse
|
66
|
Ma C, Xu Q, Wang X, Jin B, Zhang X, Wang Y, Zhang Y. Boundary-Aware Supervoxel-Level Iteratively Refined Interactive 3D Image Segmentation With Multi-Agent Reinforcement Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2563-2574. [PMID: 33382649 DOI: 10.1109/tmi.2020.3048477] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Interactive segmentation has recently been explored to effectively and efficiently harvest high-quality segmentation masks by iteratively incorporating user hints. While iterative in nature, most existing interactive segmentation methods tend to ignore the dynamics of successive interactions and take each interaction independently. We here propose to model iterative interactive image segmentation with a Markov decision process (MDP) and solve it with reinforcement learning (RL) where each voxel is treated as an agent. Considering the large exploration space for voxel-wise prediction and the dependence among neighboring voxels for the segmentation tasks, multi-agent reinforcement learning is adopted, where the voxel-level policy is shared among agents. Considering that boundary voxels are more important for segmentation, we further introduce a boundary-aware reward, which consists of a global reward in the form of relative cross-entropy gain, to update the policy in a constrained direction, and a boundary reward in the form of relative weight, to emphasize the correctness of boundary predictions. To combine the advantages of different types of interactions, i. e., simple and efficient for point-clicking, and stable and robust for scribbles, we propose a supervoxel-clicking based interaction design. Experimental results on four benchmark datasets have shown that the proposed method significantly outperforms the state-of-the-arts, with the advantage of fewer interactions, higher accuracy, and enhanced robustness.
Collapse
|
67
|
Single-Input Multi-Output U-Net for Automated 2D Foetal Brain Segmentation of MR Images. J Imaging 2021; 7:jimaging7100200. [PMID: 34677286 PMCID: PMC8536962 DOI: 10.3390/jimaging7100200] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2021] [Revised: 09/14/2021] [Accepted: 09/26/2021] [Indexed: 11/16/2022] Open
Abstract
In this work, we develop the Single-Input Multi-Output U-Net (SIMOU-Net), a hybrid network for foetal brain segmentation inspired by the original U-Net fused with the holistically nested edge detection (HED) network. The SIMOU-Net is similar to the original U-Net but it has a deeper architecture and takes account of the features extracted from each side output. It acts similar to an ensemble neural network, however, instead of averaging the outputs from several independently trained models, which is computationally expensive, our approach combines outputs from a single network to reduce the variance of predications and generalization errors. Experimental results using 200 normal foetal brains consisting of over 11,500 2D images produced Dice and Jaccard coefficients of 94.2 ± 5.9% and 88.7 ± 6.9%, respectively. We further tested the proposed network on 54 abnormal cases (over 3500 images) and achieved Dice and Jaccard coefficients of 91.2 ± 6.8% and 85.7 ± 6.6%, respectively.
Collapse
|
68
|
Wang L, Guo D, Wang G, Zhang S. Annotation-Efficient Learning for Medical Image Segmentation Based on Noisy Pseudo Labels and Adversarial Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2795-2807. [PMID: 33370237 DOI: 10.1109/tmi.2020.3047807] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Despite that deep learning has achieved state-of-the-art performance for medical image segmentation, its success relies on a large set of manually annotated images for training that are expensive to acquire. In this paper, we propose an annotation-efficient learning framework for segmentation tasks that avoids annotations of training images, where we use an improved Cycle-Consistent Generative Adversarial Network (GAN) to learn from a set of unpaired medical images and auxiliary masks obtained either from a shape model or public datasets. We first use the GAN to generate pseudo labels for our training images under the implicit high-level shape constraint represented by a Variational Auto-encoder (VAE)-based discriminator with the help of the auxiliary masks, and build a Discriminator-guided Generator Channel Calibration (DGCC) module which employs our discriminator's feedback to calibrate the generator for better pseudo labels. To learn from the pseudo labels that are noisy, we further introduce a noise-robust iterative learning method using noise-weighted Dice loss. We validated our framework with two situations: objects with a simple shape model like optic disc in fundus images and fetal head in ultrasound images, and complex structures like lung in X-Ray images and liver in CT images. Experimental results demonstrated that 1) Our VAE-based discriminator and DGCC module help to obtain high-quality pseudo labels. 2) Our proposed noise-robust learning method can effectively overcome the effect of noisy pseudo labels. 3) The segmentation performance of our method without using annotations of training images is close or even comparable to that of learning from human annotations.
Collapse
|
69
|
Zhao T, Yin Z. Weakly Supervised Cell Segmentation by Point Annotation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2736-2747. [PMID: 33347404 DOI: 10.1109/tmi.2020.3046292] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
We propose weakly supervised training schemes to train end-to-end cell segmentation networks that only require a single point annotation per cell as the training label and generate a high-quality segmentation mask close to those fully supervised methods using mask annotation on cells. Three training schemes are investigated to train cell segmentation networks, using the point annotation. First, self-training is performed to learn additional information near the annotated points. Next, co-training is applied to learn more cell regions using multiple networks that supervise each other. Finally, a hybrid-training scheme is proposed to leverage the advantages of both self-training and co-training. During the training process, we propose a divergence loss to avoid the overfitting and a consistency loss to enforce the consensus among multiple co-trained networks. Furthermore, we propose weakly supervised learning with human in the loop, aiming at achieving high segmentation accuracy and annotation efficiency simultaneously. Evaluated on two benchmark datasets, our proposal achieves high-quality cell segmentation results comparable to the fully supervised methods, but with much less amount of human annotation effort.
Collapse
|
70
|
Hoffmann M, Abaci Turk E, Gagoski B, Morgan L, Wighton P, Tisdall MD, Reuter M, Adalsteinsson E, Grant PE, Wald LL, van der Kouwe AJW. Rapid head-pose detection for automated slice prescription of fetal-brain MRI. INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY 2021; 31:1136-1154. [PMID: 34421216 PMCID: PMC8372849 DOI: 10.1002/ima.22563] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/28/2020] [Revised: 01/29/2021] [Accepted: 02/09/2021] [Indexed: 06/13/2023]
Abstract
In fetal-brain MRI, head-pose changes between prescription and acquisition present a challenge to obtaining the standard sagittal, coronal and axial views essential to clinical assessment. As motion limits acquisitions to thick slices that preclude retrospective resampling, technologists repeat ~55-second stack-of-slices scans (HASTE) with incrementally reoriented field of view numerous times, deducing the head pose from previous stacks. To address this inefficient workflow, we propose a robust head-pose detection algorithm using full-uterus scout scans (EPI) which take ~5 seconds to acquire. Our ~2-second procedure automatically locates the fetal brain and eyes, which we derive from maximally stable extremal regions (MSERs). The success rate of the method exceeds 94% in the third trimester, outperforming a trained technologist by up to 20%. The pipeline may be used to automatically orient the anatomical sequence, removing the need to estimate the head pose from 2D views and reducing delays during which motion can occur.
Collapse
Affiliation(s)
- Malte Hoffmann
- Department of Radiology, Massachusetts General HospitalBostonMassachusettsUSA
- Department of RadiologyHarvard Medical SchoolBostonMassachusettsUSA
| | - Esra Abaci Turk
- Fetal‐Neonatal Neuroimaging and Developmental Science Center, Boston Children's HospitalBostonMassachusettsUSA
- Electrical Engineering and Computer ScienceMassachusetts Institute of TechnologyCambridgeMassachusettsUSA
| | - Borjan Gagoski
- Department of RadiologyHarvard Medical SchoolBostonMassachusettsUSA
- Fetal‐Neonatal Neuroimaging and Developmental Science Center, Boston Children's HospitalBostonMassachusettsUSA
| | - Leah Morgan
- Department of Radiology, Massachusetts General HospitalBostonMassachusettsUSA
| | - Paul Wighton
- Department of Radiology, Massachusetts General HospitalBostonMassachusettsUSA
| | - Matthew Dylan Tisdall
- Radiology, Perelman School of MedicineUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Martin Reuter
- Department of Radiology, Massachusetts General HospitalBostonMassachusettsUSA
- Department of RadiologyHarvard Medical SchoolBostonMassachusettsUSA
- German Center for Neurodegenerative DiseasesBonnGermany
| | - Elfar Adalsteinsson
- Electrical Engineering and Computer ScienceMassachusetts Institute of TechnologyCambridgeMassachusettsUSA
- Institute for Medical Engineering and ScienceMassachusetts Institute of TechnologyCambridgeMassachusettsUSA
| | - Patricia Ellen Grant
- Department of RadiologyHarvard Medical SchoolBostonMassachusettsUSA
- Fetal‐Neonatal Neuroimaging and Developmental Science Center, Boston Children's HospitalBostonMassachusettsUSA
| | - Lawrence L. Wald
- Department of Radiology, Massachusetts General HospitalBostonMassachusettsUSA
- Department of RadiologyHarvard Medical SchoolBostonMassachusettsUSA
| | - André J. W. van der Kouwe
- Department of Radiology, Massachusetts General HospitalBostonMassachusettsUSA
- Department of RadiologyHarvard Medical SchoolBostonMassachusettsUSA
| |
Collapse
|
71
|
Zou H, Gong X, Luo J, Li T. A Robust Breast ultrasound segmentation method under noisy annotations. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2021; 209:106327. [PMID: 34428680 DOI: 10.1016/j.cmpb.2021.106327] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/17/2020] [Accepted: 07/30/2021] [Indexed: 06/13/2023]
Abstract
BACKGROUND AND OBJECTIVE A large-scale training data and accurate annotations are fundamental for current segmentation networks. However, the characteristic artifacts of ultrasound images always make the annotation task complicated, such as attenuation, speckle, shadows and signal dropout. Further complications arise as the contrast between the region of interest and background is often low. Without double-check from professionals, it is hard to guarantee that there is no noisy annotation in segmentation datasets. However, among the deep learning methods applied to ultrasound segmentation so far, no one can solve this problem. METHOD Given a dataset with poorly labeled masks, including a certain amount of noises, we propose an end-to-end noisy annotation tolerance network (NAT-Net). NAT-Net can detect noise by the proposed noise index (NI) and dynamically correct noisy annotations in the training stage. Simultaneously, noise index is used to correct the noise along with the output of the learning model. This method does not need any auxiliary clean datasets or prior knowledge of noise distributions, so it is more general, robust and easier to apply than the existing methods. RESULTS NAT-Net outperforms previous state-of-the-art methods on synthesized data with different noise ratio. For real-world dataset with more complex noise types, the IoU of NAT-Net is higher than that of state-of-art approaches by nearly 6%. Experimental results show that our method can also achieve good results compared with the existing methods for clean dataset. CONCLUSION The NAT-Net reduces manual interaction of data annotation, reduces dependence on medical personnel. After tumor segmentation, disease diagnosis efficiency is improved, which provides an auxiliary strategies for subsequent medical diagnosis systems based on ultrasound.
Collapse
Affiliation(s)
- Haipeng Zou
- School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu, Sichuan, China.
| | - Xun Gong
- School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu, Sichuan, China.
| | - Jun Luo
- Sichuan Academy of Medical Sciences Sichuan Provincial Peoples Hospital, Chengdu, Sichuan, China.
| | - Tianrui Li
- School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu, Sichuan, China.
| |
Collapse
|
72
|
Chong Y, Chen X, Tao Y, Pan S. Erase then grow: Generating correct class activation maps for weakly-supervised semantic segmentation. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.04.103] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
73
|
|
74
|
Bitarafan A, Nikdan M, Baghshah MS. 3D Image Segmentation With Sparse Annotation by Self-Training and Internal Registration. IEEE J Biomed Health Inform 2021; 25:2665-2672. [PMID: 33211667 DOI: 10.1109/jbhi.2020.3038847] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Anatomical image segmentation is one of the foundations for medical planning. Recently, convolutional neural networks (CNN) have achieved much success in segmenting volumetric (3D) images when a large number of fully annotated 3D samples are available. However, rarely a volumetric medical image dataset containing a sufficient number of segmented 3D images is accessible since providing manual segmentation masks is monotonous and time-consuming. Thus, to alleviate the burden of manual annotation, we attempt to effectively train a 3D CNN using a sparse annotation where ground truth on just one 2D slice of the axial axis of each training 3D image is available. To tackle this problem, we propose a self-training framework that alternates between two steps consisting of assigning pseudo annotations to unlabeled voxels and updating the 3D segmentation network by employing both the labeled and pseudo labeled voxels. To produce pseudo labels more accurately, we benefit from both propagation of labels (or pseudo-labels) between adjacent slices and 3D processing of voxels. More precisely, a 2D registration-based method is proposed to gradually propagate labels between consecutive 2D slices and a 3D U-Net is employed to utilize volumetric information. Ablation studies on benchmarks show that cooperation between the 2D registration and the 3D segmentation provides accurate pseudo-labels that enable the segmentation network to be trained effectively when for each training sample only even one segmented slice by an expert is available. Our method is assessed on the CHAOS and Visceral datasets to segment abdominal organs. Results demonstrate that despite utilizing just one segmented slice for each 3D image (that is weaker supervision in comparison with the compared weakly supervised methods) can result in higher performance and also achieve closer results to the fully supervised manner.
Collapse
|
75
|
Bateson M, Dolz J, Kervadec H, Lombaert H, Ayed IB. Constrained Domain Adaptation for Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1875-1887. [PMID: 33750688 DOI: 10.1109/tmi.2021.3067688] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Domain Adaption tasks have recently attracted substantial attention in computer vision as they improve the transferability of deep network models from a source to a target domain with different characteristics. A large body of state-of-the-art domain-adaptation methods was developed for image classification purposes, which may be inadequate for segmentation tasks. We propose to adapt segmentation networks with a constrained formulation, which embeds domain-invariant prior knowledge about the segmentation regions. Such knowledge may take the form of anatomical information, for instance, structure size or shape, which can be known a priori or learned from the source samples via an auxiliary task. Our general formulation imposes inequality constraints on the network predictions of unlabeled or weakly labeled target samples, thereby matching implicitly the prediction statistics of the target and source domains, with permitted uncertainty of prior knowledge. Furthermore, our inequality constraints easily integrate weak annotations of the target data, such as image-level tags. We address the ensuing constrained optimization problem with differentiable penalties, fully suited for conventional stochastic gradient descent approaches. Unlike common two-step adversarial training, our formulation is based on a single segmentation network, which simplifies adaptation, while improving training quality. Comparison with state-of-the-art adaptation methods reveals considerably better performance of our model on two challenging tasks. Particularly, it consistently yields a performance gain of 1-4% Dice across architectures and datasets. Our results also show robustness to imprecision in the prior knowledge. The versatility of our novel approach can be readily used in various segmentation problems, with code available publicly.
Collapse
|
76
|
Going to Extremes: Weakly Supervised Medical Image Segmentation. MACHINE LEARNING AND KNOWLEDGE EXTRACTION 2021. [DOI: 10.3390/make3020026] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Medical image annotation is a major hurdle for developing precise and robust machine-learning models. Annotation is expensive, time-consuming, and often requires expert knowledge, particularly in the medical field. Here, we suggest using minimal user interaction in the form of extreme point clicks to train a segmentation model which, in effect, can be used to speed up medical image annotation. An initial segmentation is generated based on the extreme points using the random walker algorithm. This initial segmentation is then used as a noisy supervision signal to train a fully convolutional network that can segment the organ of interest, based on the provided user clicks. Through experimentation on several medical imaging datasets, we show that the predictions of the network can be refined using several rounds of training with the prediction from the same weakly annotated data. Further improvements are shown using the clicked points within a custom-designed loss and attention mechanism. Our approach has the potential to speed up the process of generating new training datasets for the development of new machine-learning and deep-learning-based models for, but not exclusively, medical image analysis.
Collapse
|
77
|
Shaari H, Kevrić J, Jukić S, Bešić L, Jokić D, Ahmed N, Rajs V. Deep Learning-Based Studies on Pediatric Brain Tumors Imaging: Narrative Review of Techniques and Challenges. Brain Sci 2021; 11:brainsci11060716. [PMID: 34071202 PMCID: PMC8230188 DOI: 10.3390/brainsci11060716] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2021] [Revised: 05/10/2021] [Accepted: 05/17/2021] [Indexed: 11/16/2022] Open
Abstract
Brain tumors diagnosis in children is a scientific concern due to rapid anatomical, metabolic, and functional changes arising in the brain and non-specific or conflicting imaging results. Pediatric brain tumors diagnosis is typically centralized in clinical practice on the basis of diagnostic clues such as, child age, tumor location and incidence, clinical history, and imaging (Magnetic resonance imaging MRI / computed tomography CT) findings. The implementation of deep learning has rapidly propagated in almost every field in recent years, particularly in the medical images’ evaluation. This review would only address critical deep learning issues specific to pediatric brain tumor imaging research in view of the vast spectrum of other applications of deep learning. The purpose of this review paper is to include a detailed summary by first providing a succinct guide to the types of pediatric brain tumors and pediatric brain tumor imaging techniques. Then, we will present the research carried out by summarizing the scientific contributions to the field of pediatric brain tumor imaging processing and analysis. Finally, to establish open research issues and guidance for potential study in this emerging area, the medical and technical limitations of the deep learning-based approach were included.
Collapse
Affiliation(s)
- Hala Shaari
- Department of Information Technologies, Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina;
| | - Jasmin Kevrić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Samed Jukić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Larisa Bešić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Dejan Jokić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Nuredin Ahmed
- Control Department, Technical Computer College Tripoli, Tripoli 00218, Libya;
| | - Vladimir Rajs
- Department of Power, Electronics and Telecommunication Engineering, Faculty of Technical Science, University of Novi Sad, 21000 Novi Sad, Serbia
- Correspondence:
| |
Collapse
|
78
|
Bandyk MG, Gopireddy DR, Lall C, Balaji KC, Dolz J. MRI and CT bladder segmentation from classical to deep learning based approaches: Current limitations and lessons. Comput Biol Med 2021; 134:104472. [PMID: 34023696 DOI: 10.1016/j.compbiomed.2021.104472] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2021] [Revised: 04/29/2021] [Accepted: 05/02/2021] [Indexed: 10/21/2022]
Abstract
Precise determination and assessment of bladder cancer (BC) extent of muscle invasion involvement guides proper risk stratification and personalized therapy selection. In this context, segmentation of both bladder walls and cancer are of pivotal importance, as it provides invaluable information to stage the primary tumor. Hence, multiregion segmentation on patients presenting with symptoms of bladder tumors using deep learning heralds a new level of staging accuracy and prediction of the biologic behavior of the tumor. Nevertheless, despite the success of these models in other medical problems, progress in multiregion bladder segmentation, particularly in MRI and CT modalities, is still at a nascent stage, with just a handful of works tackling a multiregion scenario. Furthermore, most existing approaches systematically follow prior literature in other clinical problems, without casting a doubt on the validity of these methods on bladder segmentation, which may present different challenges. Inspired by this, we provide an in-depth look at bladder cancer segmentation using deep learning models. The critical determinants for accurate differentiation of muscle invasive disease, current status of deep learning based bladder segmentation, lessons and limitations of prior work are highlighted.
Collapse
Affiliation(s)
- Mark G Bandyk
- Department of Urology, University of Florida, Jacksonville, FL, USA.
| | | | - Chandana Lall
- Department of Radiology, University of Florida, Jacksonville, FL, USA
| | - K C Balaji
- Department of Urology, University of Florida, Jacksonville, FL, USA
| | | |
Collapse
|
79
|
Luo X, Wang G, Song T, Zhang J, Aertsen M, Deprest J, Ourselin S, Vercauteren T, Zhang S. MIDeepSeg: Minimally interactive segmentation of unseen objects from medical images using deep learning. Med Image Anal 2021; 72:102102. [PMID: 34118654 DOI: 10.1016/j.media.2021.102102] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2020] [Revised: 03/14/2021] [Accepted: 05/05/2021] [Indexed: 12/21/2022]
Abstract
Segmentation of organs or lesions from medical images plays an essential role in many clinical applications such as diagnosis and treatment planning. Though Convolutional Neural Networks (CNN) have achieved the state-of-the-art performance for automatic segmentation, they are often limited by the lack of clinically acceptable accuracy and robustness in complex cases. Therefore, interactive segmentation is a practical alternative to these methods. However, traditional interactive segmentation methods require a large number of user interactions, and recently proposed CNN-based interactive segmentation methods are limited by poor performance on previously unseen objects. To solve these problems, we propose a novel deep learning-based interactive segmentation method that not only has high efficiency due to only requiring clicks as user inputs but also generalizes well to a range of previously unseen objects. Specifically, we first encode user-provided interior margin points via our proposed exponentialized geodesic distance that enables a CNN to achieve a good initial segmentation result of both previously seen and unseen objects, then we use a novel information fusion method that combines the initial segmentation with only a few additional user clicks to efficiently obtain a refined segmentation. We validated our proposed framework through extensive experiments on 2D and 3D medical image segmentation tasks with a wide range of previously unseen objects that were not present in the training set. Experimental results showed that our proposed framework 1) achieves accurate results with fewer user interactions and less time compared with state-of-the-art interactive frameworks and 2) generalizes well to previously unseen objects.
Collapse
Affiliation(s)
- Xiangde Luo
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Guotai Wang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China.
| | - Tao Song
- SenseTime Research, Shanghai, China
| | - Jingyang Zhang
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Michael Aertsen
- Department of Radiology, University Hospitals Leuven, Leuven, Belgium
| | - Jan Deprest
- Department of Radiology, University Hospitals Leuven, Leuven, Belgium; Department of Obstetrics and Gynaecology, University Hospitals Leuven, Leuven, Belgium; Institute for Women's Health, University College London, London, UK
| | - Sebastien Ourselin
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, UK
| | - Tom Vercauteren
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, UK
| | - Shaoting Zhang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China; SenseTime Research, Shanghai, China
| |
Collapse
|
80
|
Tian Z, Li X, Chen Z, Zheng Y, Fan H, Li Z, Li C, Du S. Interactive prostate MR image segmentation based on ConvLSTMs and GGNN. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2020.05.121] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
|
81
|
Budd S, Robinson EC, Kainz B. A survey on active learning and human-in-the-loop deep learning for medical image analysis. Med Image Anal 2021; 71:102062. [PMID: 33901992 DOI: 10.1016/j.media.2021.102062] [Citation(s) in RCA: 100] [Impact Index Per Article: 33.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2019] [Revised: 03/26/2021] [Accepted: 03/30/2021] [Indexed: 12/21/2022]
Abstract
Fully automatic deep learning has become the state-of-the-art technique for many tasks including image acquisition, analysis and interpretation, and for the extraction of clinically useful information for computer-aided detection, diagnosis, treatment planning, intervention and therapy. However, the unique challenges posed by medical image analysis suggest that retaining a human end-user in any deep learning enabled system will be beneficial. In this review we investigate the role that humans might play in the development and deployment of deep learning enabled diagnostic applications and focus on techniques that will retain a significant input from a human end user. Human-in-the-Loop computing is an area that we see as increasingly important in future research due to the safety-critical nature of working in the medical domain. We evaluate four key areas that we consider vital for deep learning in the clinical practice: (1) Active Learning to choose the best data to annotate for optimal model performance; (2) Interaction with model outputs - using iterative feedback to steer models to optima for a given prediction and offering meaningful ways to interpret and respond to predictions; (3) Practical considerations - developing full scale applications and the key considerations that need to be made before deployment; (4) Future Prospective and Unanswered Questions - knowledge gaps and related research fields that will benefit human-in-the-loop computing as they evolve. We offer our opinions on the most promising directions of research and how various aspects of each area might be unified towards common goals.
Collapse
Affiliation(s)
- Samuel Budd
- Department of Computing, Imperial College London, UK.
| | | | | |
Collapse
|
82
|
Wang D, Li M, Ben-Shlomo N, Corrales CE, Cheng Y, Zhang T, Jayender J. A novel dual-network architecture for mixed-supervised medical image segmentation. Comput Med Imaging Graph 2021; 89:101841. [PMID: 33756304 PMCID: PMC8084108 DOI: 10.1016/j.compmedimag.2020.101841] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2020] [Revised: 11/01/2020] [Accepted: 12/07/2020] [Indexed: 11/15/2022]
Abstract
In medical image segmentation tasks, deep learning-based models usually require densely and precisely annotated datasets to train, which are time-consuming and expensive to prepare. One possible solution is to train with the mixed-supervised dataset, where only a part of data is densely annotated with segmentation map and the rest is annotated with some weak form, such as bounding box. In this paper, we propose a novel network architecture called Mixed-Supervised Dual-Network (MSDN), which consists of two separate networks for the segmentation and detection tasks respectively, and a series of connection modules between the layers of the two networks. These connection modules are used to extract and transfer useful information from the detection task to help the segmentation task. We exploit a variant of a recently designed technique called 'Squeeze and Excitation' in the connection module to boost the information transfer between the two tasks. Compared with existing model with shared backbone and multiple branches, our model has flexible and trainable feature sharing fashion and thus is more effective and stable. We conduct experiments on 4 medical image segmentation datasets, and experiment results show that the proposed MSDN model outperforms multiple baselines.
Collapse
Affiliation(s)
- Duo Wang
- Department of Automation, Tsinghua University, Beijing 100084, China; Department of Radiology, Brigham and Women's Hospital, Boston 02115, USA.
| | - Ming Li
- Department of Radiology, Huadong Hospital affiliated to Fudan University, Shanghai 200040, China.
| | - Nir Ben-Shlomo
- Department of Surgery, Brigham and Women's Hospital, Boston 02115, USA.
| | - C Eduardo Corrales
- Department of Surgery, Brigham and Women's Hospital, Boston 02115, USA; Harvard Medical School, Boston 02115, USA.
| | - Yu Cheng
- Microsoft AI & Research, Redmond, WA, USA.
| | - Tao Zhang
- Department of Automation, Tsinghua University, Beijing 100084, China; Beijing National Research Center for Information Science and Technology, Tsinghua University, Beijing 100084, China.
| | - Jagadeesan Jayender
- Department of Radiology, Brigham and Women's Hospital, Boston 02115, USA; Harvard Medical School, Boston 02115, USA.
| |
Collapse
|
83
|
Li J, Udupa JK, Tong Y, Wang L, Torigian DA. Segmentation evaluation with sparse ground truth data: Simulating true segmentations as perfect/imperfect as those generated by humans. Med Image Anal 2021; 69:101980. [PMID: 33588116 PMCID: PMC7933105 DOI: 10.1016/j.media.2021.101980] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2020] [Revised: 01/19/2021] [Accepted: 01/20/2021] [Indexed: 10/22/2022]
Abstract
Fully annotated data sets play important roles in medical image segmentation and evaluation. Expense and imprecision are the two main issues in generating ground truth (GT) segmentations. In this paper, in an attempt to overcome these two issues jointly, we propose a method, named SparseGT, which exploit variability among human segmenters to maximally save manual workload in GT generation for evaluating actual segmentations by algorithms. Pseudo ground truth (p-GT) segmentations are created by only a small fraction of workload and with human-level perfection/imperfection, and they can be used in practice as a substitute for fully manual GT in evaluating segmentation algorithms at the same precision. p-GT segmentations are generated by first selecting slices sparsely, where manual contouring is conducted only on these sparse slices, and subsequently filling segmentations on other slices automatically. By creating p-GT with different levels of sparseness, we determine the largest workload reduction achievable for each considered object, where the variability of the generated p-GT is statistically indistinguishable from inter-segmenter differences in full manual GT segmentations for that object. Furthermore, we investigate the segmentation evaluation errors introduced by variability in manual GT by applying p-GT in evaluation of actual segmentations by an algorithm. Experiments are conducted on ∼500 computed tomography (CT) studies involving six objects in two body regions, Head & Neck and Thorax, where optimal sparseness and corresponding evaluation errors are determined for each object and each strategy. Our results indicate that creating p-GT by the concatenated strategy of uniformly selecting sparse slices and filling segmentations via deep-learning (DL) network show highest manual workload reduction by ∼80-96% without sacrificing evaluation accuracy compared to fully manual GT. Nevertheless, other strategies also have obvious contributions in different situations. A non-uniform strategy for slice selection shows its advantage for objects with irregular shape change from slice to slice. An interpolation strategy for filling segmentations can achieve ∼60-90% of workload reduction in simulating human-level GT without the need of an actual training stage and shows potential in enlarging data sets for training p-GT generation networks. We conclude that not only over 90% reduction in workload is feasible without sacrificing evaluation accuracy but also the suitable strategy and the optimal sparseness level achievable for creating p-GT are object- and application-specific.
Collapse
Affiliation(s)
- Jieyu Li
- Institute of Image Processing and Pattern Recognition, Department of Automation, Shanghai Jiao Tong University, 800 Dongchuan RD, Shanghai, 200240, China; Medical Image Processing Group, Department of Radiology, University of Pennsylvania, 602 Goddard building, 3710 Hamilton Walk, Philadelphia, PA, 19104, United States
| | - Jayaram K Udupa
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, 602 Goddard building, 3710 Hamilton Walk, Philadelphia, PA, 19104, United States.
| | - Yubing Tong
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, 602 Goddard building, 3710 Hamilton Walk, Philadelphia, PA, 19104, United States
| | - Lisheng Wang
- Institute of Image Processing and Pattern Recognition, Department of Automation, Shanghai Jiao Tong University, 800 Dongchuan RD, Shanghai, 200240, China
| | - Drew A Torigian
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, 602 Goddard building, 3710 Hamilton Walk, Philadelphia, PA, 19104, United States
| |
Collapse
|
84
|
Dou Y, Tsai YH, Liu CC, Hobson BA, Lein PJ. Co-localization of fluorescent signals using deep learning with Manders overlapping coefficient. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2021; 11596:115963C. [PMID: 34305257 PMCID: PMC8301216 DOI: 10.1117/12.2580650] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/14/2022]
Abstract
Object-based co-localization of fluorescent signals allows the assessment of interactions between two (or more) biological entities using spatial information. It relies on object identification with high accuracy to separate fluorescent signals from the background. Object detectors using convolutional neural networks (CNN) with annotated training samples could facilitate the process by detecting and counting fluorescent-labeled cells from fluorescence photomicrographs. However, datasets containing segmented annotations of colocalized cells are generally not available, and creating a new dataset with delineated masks is label-intensive. Also, the co-localization coefficient is often not used as a component during training with the CNN model. Yet, it may aid with localizing and detecting objects during training and testing. In this work, we propose to address these issues by using a quantification coefficient for co-localization called Manders overlapping coefficient (MOC)1 as a single-layer branch in a CNN. Fully convolutional one-state (FCOS)2 with a Resnet101 backbone served as the network to evaluate the effectiveness of the novel branch to assist with bounding box prediction. Training data were sourced from lab curated fluorescence images of neurons from the rat hippocampus, piriform cortex, somatosensory cortex, and amygdala. Results suggest that using modified FCOS with MOC outperformed the original FCOS model for accuracy in detecting fluorescence signals by 1.1% in mean average precision (mAP). The model could be downloaded from https://github.com/Alphafrey946/Colocalization-MOC.
Collapse
Affiliation(s)
- Yimeng Dou
- UW-Madison, Department of Biostatistics and Medical Informatics, Madison, Wisconsin, United States
- UC Davis School of Veterinary Medicine, Department of Molecular Biosciences, Davis, California, United States
| | - Yi-Hua Tsai
- UC Davis School of Veterinary Medicine, Department of Molecular Biosciences, Davis, California, United States
| | - Chih-Chieh Liu
- UC Davis, Department of Biomedical Engineering, Davis, California, United States
| | - Brad A. Hobson
- UC Davis, Center for Molecular and Genomic Imaging, Davis, California, United States
| | - Pamela J. Lein
- UC Davis School of Veterinary Medicine, Department of Molecular Biosciences, Davis, California, United States
| |
Collapse
|
85
|
Torrents-Barrena J, Monill N, Piella G, Gratacós E, Eixarch E, Ceresa M, González Ballester MA. Assessment of Radiomics and Deep Learning for the Segmentation of Fetal and Maternal Anatomy in Magnetic Resonance Imaging and Ultrasound. Acad Radiol 2021; 28:173-188. [PMID: 31879159 DOI: 10.1016/j.acra.2019.11.006] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2019] [Revised: 11/08/2019] [Accepted: 11/18/2019] [Indexed: 11/18/2022]
Abstract
Recent advances in fetal imaging open the door to enhanced detection of fetal disorders and computer-assisted surgical planning. However, precise segmentation of womb's tissues is challenging due to motion artifacts caused by fetal movements and maternal respiration during acquisition. This work aims to efficiently segment different intrauterine tissues in fetal magnetic resonance imaging (MRI) and 3D ultrasound (US). First, a large set of ninety-four radiomic features are extracted to characterize the mother uterus, placenta, umbilical cord, fetal lungs, and brain. The optimal features for each anatomy are identified using both K-best and Sequential Forward Feature Selection techniques. These features are then fed to a Support Vector Machine with instance balancing to accurately segment the intrauterine anatomies. To the best of our knowledge, this is the first time that "Radiomics" is expanded from classification tasks to segmentation purposes to deal with challenging fetal images. In addition, we evaluate several state-of-the-art deep learning-based segmentation approaches. Validation is extensively performed on a set of 60 axial MRI and 3D US images from pathological and clinical cases. Our results suggest that combining the selected 10 radiomic features per anatomy along with DeepLabV3+ or BiSeNet architectures for MRI, and PSPNet or Tiramisu for 3D US, can lead to the highest fetal / maternal tissue segmentation performance, robustness, informativeness, and heterogeneity. Therefore, this work opens new avenues for advancement of segmentation techniques and, in particular, for improved fetal surgical planning.
Collapse
Affiliation(s)
- Jordina Torrents-Barrena
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain.
| | - Núria Monill
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Gemma Piella
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Eduard Gratacós
- BCNatal | Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain; Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain and Centre for Biomedical Research on Rare Diseases (CIBER-ER), Barcelona, Spain
| | - Elisenda Eixarch
- BCNatal | Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain; Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain and Centre for Biomedical Research on Rare Diseases (CIBER-ER), Barcelona, Spain
| | - Mario Ceresa
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Miguel A González Ballester
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain; ICREA, Barcelona, Spain
| |
Collapse
|
86
|
Wang J, Xia B. Bounding Box Tightness Prior for Weakly Supervised Image Segmentation. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION – MICCAI 2021 2021:526-536. [DOI: 10.1007/978-3-030-87196-3_49] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/01/2023]
|
87
|
Jiang Y, Chen W, Liu M, Wang Y, Meijering E. 3D Neuron Microscopy Image Segmentation via the Ray-Shooting Model and a DC-BLSTM Network. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:26-37. [PMID: 32881683 DOI: 10.1109/tmi.2020.3021493] [Citation(s) in RCA: 26] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
The morphology reconstruction (tracing) of neurons in 3D microscopy images is important to neuroscience research. However, this task remains very challenging because of the low signal-to-noise ratio (SNR) and the discontinued segments of neurite patterns in the images. In this paper, we present a neuronal structure segmentation method based on the ray-shooting model and the Long Short-Term Memory (LSTM)-based network to enhance the weak-signal neuronal structures and remove background noise in 3D neuron microscopy images. Specifically, the ray-shooting model is used to extract the intensity distribution features within a local region of the image. And we design a neural network based on the dual channel bidirectional LSTM (DC-BLSTM) to detect the foreground voxels according to the voxel-intensity features and boundary-response features extracted by multiple ray-shooting models that are generated in the whole image. This way, we transform the 3D image segmentation task into multiple 1D ray/sequence segmentation tasks, which makes it much easier to label the training samples than many existing Convolutional Neural Network (CNN) based 3D neuron image segmentation methods. In the experiments, we evaluate the performance of our method on the challenging 3D neuron images from two datasets, the BigNeuron dataset and the Whole Mouse Brain Sub-image (WMBS) dataset. Compared with the neuron tracing results on the segmented images produced by other state-of-the-art neuron segmentation methods, our method improves the distance scores by about 32% and 27% in the BigNeuron dataset, and about 38% and 27% in the WMBS dataset.
Collapse
|
88
|
Chen J, Fang Z, Zhang G, Ling L, Li G, Zhang H, Wang L. Automatic brain extraction from 3D fetal MR image with deep learning-based multi-step framework. Comput Med Imaging Graph 2020; 88:101848. [PMID: 33385932 DOI: 10.1016/j.compmedimag.2020.101848] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2020] [Revised: 11/15/2020] [Accepted: 12/07/2020] [Indexed: 12/21/2022]
Abstract
Brain extraction is a fundamental prerequisite step in neuroimage analysis for fetus. Due to surrounding maternal tissues and unpredictable movement, brain extraction from fetal Magnetic Resonance (MR) images is a challenging task. In this paper, we propose a novel deep learning-based multi-step framework for brain extraction from 3D fetal MR images. In the first step, a global localization network is applied to estimate probability maps for brain candidates. Connected-component labeling algorithm is applied to eliminate small erroneous components and accurately locate the candidate brain area. In the second step, a local refinement network is implemented in the brain candidate area to obtain fine-grained probability maps. Final extraction results are derived by a fusion network with the two cascaded probability maps obtained from previous two steps. Experimental results demonstrate that our proposed method has superior performance compared with existing deep learning-based methods.
Collapse
Affiliation(s)
- Jian Chen
- School of Electronic, Electrical Engineering and Physics, Fujian University of Technology, Fuzhou, Fujian, 350118, China.
| | - Zhenghan Fang
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, 27517, USA
| | - Guofu Zhang
- Department of Radiology, Obstetrics and Gynecology Hospital, Fudan University, Shanghai, 200011, China
| | - Lei Ling
- Department of Radiology, Obstetrics and Gynecology Hospital, Fudan University, Shanghai, 200011, China
| | - Gang Li
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, 27517, USA
| | - He Zhang
- Department of Radiology, Obstetrics and Gynecology Hospital, Fudan University, Shanghai, 200011, China.
| | - Li Wang
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, 27517, USA.
| |
Collapse
|
89
|
Gal A, Saragosti J, Kronauer DJC. anTraX, a software package for high-throughput video tracking of color-tagged insects. eLife 2020; 9:e58145. [PMID: 33211008 PMCID: PMC7676868 DOI: 10.7554/elife.58145] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2020] [Accepted: 10/29/2020] [Indexed: 12/14/2022] Open
Abstract
Recent years have seen a surge in methods to track and analyze animal behavior. Nevertheless, tracking individuals in closely interacting, group-living organisms remains a challenge. Here, we present anTraX, an algorithm and software package for high-throughput video tracking of color-tagged insects. anTraX combines neural network classification of animals with a novel approach for representing tracking data as a graph, enabling individual tracking even in cases where it is difficult to segment animals from one another, or where tags are obscured. The use of color tags, a well-established and robust method for marking individual insects in groups, relaxes requirements for image size and quality, and makes the software broadly applicable. anTraX is readily integrated into existing tools and methods for automated image analysis of behavior to further augment its output. anTraX can handle large-scale experiments with minimal human involvement, allowing researchers to simultaneously monitor many social groups over long time periods.
Collapse
Affiliation(s)
- Asaf Gal
- Laboratory of Social Evolution and Behavior, The Rockefeller UniversityNew YorkUnited States
| | - Jonathan Saragosti
- Laboratory of Social Evolution and Behavior, The Rockefeller UniversityNew YorkUnited States
| | - Daniel JC Kronauer
- Laboratory of Social Evolution and Behavior, The Rockefeller UniversityNew YorkUnited States
| |
Collapse
|
90
|
Introducing Biomedisa as an open-source online platform for biomedical image segmentation. Nat Commun 2020; 11:5577. [PMID: 33149150 PMCID: PMC7642381 DOI: 10.1038/s41467-020-19303-w] [Citation(s) in RCA: 62] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2019] [Accepted: 10/02/2020] [Indexed: 01/07/2023] Open
Abstract
We present Biomedisa, a free and easy-to-use open-source online platform developed for semi-automatic segmentation of large volumetric images. The segmentation is based on a smart interpolation of sparsely pre-segmented slices taking into account the complete underlying image data. Biomedisa is particularly valuable when little a priori knowledge is available, e.g. for the dense annotation of the training data for a deep neural network. The platform is accessible through a web browser and requires no complex and tedious configuration of software and model parameters, thus addressing the needs of scientists without substantial computational expertise. We demonstrate that Biomedisa can drastically reduce both the time and human effort required to segment large images. It achieves a significant improvement over the conventional approach of densely pre-segmented slices with subsequent morphological interpolation as well as compared to segmentation tools that also consider the underlying image data. Biomedisa can be used for different 3D imaging modalities and various biomedical applications.
Collapse
|
91
|
Qu H, Wu P, Huang Q, Yi J, Yan Z, Li K, Riedlinger GM, De S, Zhang S, Metaxas DN. Weakly Supervised Deep Nuclei Segmentation Using Partial Points Annotation in Histopathology Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3655-3666. [PMID: 32746112 DOI: 10.1109/tmi.2020.3002244] [Citation(s) in RCA: 49] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/06/2023]
Abstract
Nuclei segmentation is a fundamental task in histopathology image analysis. Typically, such segmentation tasks require significant effort to manually generate accurate pixel-wise annotations for fully supervised training. To alleviate such tedious and manual effort, in this paper we propose a novel weakly supervised segmentation framework based on partial points annotation, i.e., only a small portion of nuclei locations in each image are labeled. The framework consists of two learning stages. In the first stage, we design a semi-supervised strategy to learn a detection model from partially labeled nuclei locations. Specifically, an extended Gaussian mask is designed to train an initial model with partially labeled data. Then, self-training with background propagation is proposed to make use of the unlabeled regions to boost nuclei detection and suppress false positives. In the second stage, a segmentation model is trained from the detected nuclei locations in a weakly-supervised fashion. Two types of coarse labels with complementary information are derived from the detected points and are then utilized to train a deep neural network. The fully-connected conditional random field loss is utilized in training to further refine the model without introducing extra computational complexity during inference. The proposed method is extensively evaluated on two nuclei segmentation datasets. The experimental results demonstrate that our method can achieve competitive performance compared to the fully supervised counterpart and the state-of-the-art methods while requiring significantly less annotation effort.
Collapse
|
92
|
Rasse TM, Hollandi R, Horvath P. OpSeF: Open Source Python Framework for Collaborative Instance Segmentation of Bioimages. Front Bioeng Biotechnol 2020; 8:558880. [PMID: 33117778 PMCID: PMC7576117 DOI: 10.3389/fbioe.2020.558880] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2020] [Accepted: 09/15/2020] [Indexed: 11/13/2022] Open
Abstract
Various pre-trained deep learning models for the segmentation of bioimages have been made available as developer-to-end-user solutions. They are optimized for ease of use and usually require neither knowledge of machine learning nor coding skills. However, individually testing these tools is tedious and success is uncertain. Here, we present the Open Segmentation Framework (OpSeF), a Python framework for deep learning-based instance segmentation. OpSeF aims at facilitating the collaboration of biomedical users with experienced image analysts. It builds on the analysts' knowledge in Python, machine learning, and workflow design to solve complex analysis tasks at any scale in a reproducible, well-documented way. OpSeF defines standard inputs and outputs, thereby facilitating modular workflow design and interoperability with other software. Users play an important role in problem definition, quality control, and manual refinement of results. OpSeF semi-automates preprocessing, convolutional neural network (CNN)-based segmentation in 2D or 3D, and postprocessing. It facilitates benchmarking of multiple models in parallel. OpSeF streamlines the optimization of parameters for pre- and postprocessing such, that an available model may frequently be used without retraining. Even if sufficiently good results are not achievable with this approach, intermediate results can inform the analysts in the selection of the most promising CNN-architecture in which the biomedical user might invest the effort of manually labeling training data. We provide Jupyter notebooks that document sample workflows based on various image collections. Analysts may find these notebooks useful to illustrate common segmentation challenges, as they prepare the advanced user for gradually taking over some of their tasks and completing their projects independently. The notebooks may also be used to explore the analysis options available within OpSeF in an interactive way and to document and share final workflows. Currently, three mechanistically distinct CNN-based segmentation methods, the U-Net implementation used in Cellprofiler 3.0, StarDist, and Cellpose have been integrated within OpSeF. The addition of new networks requires little; the addition of new models requires no coding skills. Thus, OpSeF might soon become both an interactive model repository, in which pre-trained models might be shared, evaluated, and reused with ease.
Collapse
Affiliation(s)
- Tobias M. Rasse
- Scientific Service Group Microscopy, Max Planck Institute for Heart and Lung Research, Bad Nauheim, Germany
| | - Réka Hollandi
- Synthetic and Systems Biology Unit, Biological Research Center (BRC), Szeged, Hungary
| | - Peter Horvath
- Synthetic and Systems Biology Unit, Biological Research Center (BRC), Szeged, Hungary
- Institute for Molecular Medicine Finland (FIMM), University of Helsinki, Helsinki, Finland
| |
Collapse
|
93
|
Discretely-constrained deep network for weakly supervised segmentation. Neural Netw 2020; 130:297-308. [DOI: 10.1016/j.neunet.2020.07.011] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2019] [Revised: 06/17/2020] [Accepted: 07/11/2020] [Indexed: 11/18/2022]
|
94
|
Mecheter I, Alic L, Abbod M, Amira A, Ji J. MR Image-Based Attenuation Correction of Brain PET Imaging: Review of Literature on Machine Learning Approaches for Segmentation. J Digit Imaging 2020; 33:1224-1241. [PMID: 32607906 PMCID: PMC7573060 DOI: 10.1007/s10278-020-00361-x] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2022] Open
Abstract
Recent emerging hybrid technology of positron emission tomography/magnetic resonance (PET/MR) imaging has generated a great need for an accurate MR image-based PET attenuation correction. MR image segmentation, as a robust and simple method for PET attenuation correction, has been clinically adopted in commercial PET/MR scanners. The general approach in this method is to segment the MR image into different tissue types, each assigned an attenuation constant as in an X-ray CT image. Machine learning techniques such as clustering, classification and deep networks are extensively used for brain MR image segmentation. However, only limited work has been reported on using deep learning in brain PET attenuation correction. In addition, there is a lack of clinical evaluation of machine learning methods in this application. The aim of this review is to study the use of machine learning methods for MR image segmentation and its application in attenuation correction for PET brain imaging. Furthermore, challenges and future opportunities in MR image-based PET attenuation correction are discussed.
Collapse
Affiliation(s)
- Imene Mecheter
- Department of Electronic and Computer Engineering, Brunel University London, Uxbridge, UK.
- Department of Electrical and Computer Engineering, Texas A & M University at Qatar, Doha, Qatar.
| | - Lejla Alic
- Magnetic Detection and Imaging Group, Faculty of Science and Technology, University of Twente, Enschede, Netherlands
| | - Maysam Abbod
- Department of Electronic and Computer Engineering, Brunel University London, Uxbridge, UK
| | - Abbes Amira
- Institute of Artificial Intelligence, De Montfort University, Leicester, UK
| | - Jim Ji
- Department of Electrical and Computer Engineering, Texas A & M University at Qatar, Doha, Qatar
- Department of Electrical and Computer Engineering, Texas A & M University, College Station, TX, USA
| |
Collapse
|
95
|
Torrents-Barrena J, Piella G, Gratacos E, Eixarch E, Ceresa M, Gonalez Ballester MA. Deep Q-CapsNet Reinforcement Learning Framework for Intrauterine Cavity Segmentation in TTTS Fetal Surgery Planning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3113-3124. [PMID: 32305906 DOI: 10.1109/tmi.2020.2987981] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Fetoscopic laser photocoagulation is the most effective treatment for Twin-to-Twin Transfusion Syndrome, a condition affecting twin pregnancies in which there is a deregulation of blood circulation through the placenta, that can be fatal to both babies. For the purposes of surgical planning, we design the first automatic approach to detect and segment the intrauterine cavity from axial, sagittal and coronal MRI stacks. Our methodology relies on the ability of capsule networks to successfully capture the part-whole interdependency of objects in the scene, particularly for unique class instances (i.e., intrauterine cavity). The presented deep Q-CapsNet reinforcement learning framework is built upon a context-adaptive detection policy to generate a bounding box of the womb. A capsule architecture is subsequently designed to segment (or refine) the whole intrauterine cavity. This network is coupled with a strided nnU-Net feature extractor, which encodes discriminative feature maps to construct strong primary capsules. The method is robustly evaluated with and without the localization stage using 13 performance measures, and directly compared with 15 state-of-the-art deep neural networks trained on 71 singleton and monochorionic twin pregnancies. An average Dice score above 0.91 is achieved for all ablations, revealing the potential of our approach to be used in clinical practice.
Collapse
|
96
|
Girum KB, Créhange G, Hussain R, Lalande A. Fast interactive medical image segmentation with weakly supervised deep learning method. Int J Comput Assist Radiol Surg 2020; 15:1437-1444. [DOI: 10.1007/s11548-020-02223-x] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2020] [Accepted: 06/20/2020] [Indexed: 12/27/2022]
|
97
|
Mi S, Wei Z, Xu J, Yu Z, Yang W, Liao Q. Detecting Carotid Intima-Media From Small-Sample Ultrasound Images. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2020:2129-2132. [PMID: 33018427 DOI: 10.1109/embc44109.2020.9176282] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Cardiovascular diseases are the biggest threat to human being's health all over the world, and carotid atherosclerotic plaque is the leading cause of ischemic cardiovascular diseases. To determine the location and shape of the plaque, it is of great significance to detect the intima-media (IM). In this paper, a new IM detection method based on convolution neural network (IMD-CNN) is proposed for the detection of IM of blood vessels in longitudinal ultrasonic images. In IMD-CNN, firstly the region of interest (ROI) is automatically extracted by morphological processing, then the patch-wise training data are constructed, and finally a simple CNN is trained to detect the IM. The experimental results obtained on 23 images show that the test accuracy of IMD-CNN is over 86% and the performance of IMD-CNN is also visually proved to be effective.
Collapse
|
98
|
Rundo L, Beer L, Ursprung S, Martin-Gonzalez P, Markowetz F, Brenton JD, Crispin-Ortuzar M, Sala E, Woitek R. Tissue-specific and interpretable sub-segmentation of whole tumour burden on CT images by unsupervised fuzzy clustering. Comput Biol Med 2020; 120:103751. [PMID: 32421652 PMCID: PMC7248575 DOI: 10.1016/j.compbiomed.2020.103751] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2020] [Revised: 04/03/2020] [Accepted: 04/05/2020] [Indexed: 12/18/2022]
Abstract
BACKGROUND Cancer typically exhibits genotypic and phenotypic heterogeneity, which can have prognostic significance and influence therapy response. Computed Tomography (CT)-based radiomic approaches calculate quantitative features of tumour heterogeneity at a mesoscopic level, regardless of macroscopic areas of hypo-dense (i.e., cystic/necrotic), hyper-dense (i.e., calcified), or intermediately dense (i.e., soft tissue) portions. METHOD With the goal of achieving the automated sub-segmentation of these three tissue types, we present here a two-stage computational framework based on unsupervised Fuzzy C-Means Clustering (FCM) techniques. No existing approach has specifically addressed this task so far. Our tissue-specific image sub-segmentation was tested on ovarian cancer (pelvic/ovarian and omental disease) and renal cell carcinoma CT datasets using both overlap-based and distance-based metrics for evaluation. RESULTS On all tested sub-segmentation tasks, our two-stage segmentation approach outperformed conventional segmentation techniques: fixed multi-thresholding, the Otsu method, and automatic cluster number selection heuristics for the K-means clustering algorithm. In addition, experiments showed that the integration of the spatial information into the FCM algorithm generally achieves more accurate segmentation results, whilst the kernelised FCM versions are not beneficial. The best spatial FCM configuration achieved average Dice similarity coefficient values starting from 81.94±4.76 and 83.43±3.81 for hyper-dense and hypo-dense components, respectively, for the investigated sub-segmentation tasks. CONCLUSIONS The proposed intelligent framework could be readily integrated into clinical research environments and provides robust tools for future radiomic biomarker validation.
Collapse
Affiliation(s)
- Leonardo Rundo
- Department of Radiology, University of Cambridge, Cambridge CB2 0QQ, UK; Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK.
| | - Lucian Beer
- Department of Radiology, University of Cambridge, Cambridge CB2 0QQ, UK; Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK; Department of Biomedical Imaging and Image-guided Therapy, Medical University Vienna, Vienna 1090, Austria.
| | - Stephan Ursprung
- Department of Radiology, University of Cambridge, Cambridge CB2 0QQ, UK; Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK.
| | - Paula Martin-Gonzalez
- Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK; Cancer Research UK Cambridge Institute, University of Cambridge, Cambridge CB2 0RE, UK.
| | - Florian Markowetz
- Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK; Cancer Research UK Cambridge Institute, University of Cambridge, Cambridge CB2 0RE, UK.
| | - James D Brenton
- Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK; Cancer Research UK Cambridge Institute, University of Cambridge, Cambridge CB2 0RE, UK.
| | - Mireia Crispin-Ortuzar
- Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK; Cancer Research UK Cambridge Institute, University of Cambridge, Cambridge CB2 0RE, UK.
| | - Evis Sala
- Department of Radiology, University of Cambridge, Cambridge CB2 0QQ, UK; Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK.
| | - Ramona Woitek
- Department of Radiology, University of Cambridge, Cambridge CB2 0QQ, UK; Cancer Research UK Cambridge Centre, University of Cambridge, Cambridge CB2 0RE, UK; Department of Biomedical Imaging and Image-guided Therapy, Medical University Vienna, Vienna 1090, Austria.
| |
Collapse
|
99
|
Yang G, Wang C, Yang J, Chen Y, Tang L, Shao P, Dillenseger JL, Shu H, Luo L. Weakly-supervised convolutional neural networks of renal tumor segmentation in abdominal CTA images. BMC Med Imaging 2020; 20:37. [PMID: 32293303 PMCID: PMC7161012 DOI: 10.1186/s12880-020-00435-w] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2019] [Accepted: 03/20/2020] [Indexed: 11/23/2022] Open
Abstract
Background Renal cancer is one of the 10 most common cancers in human beings. The laparoscopic partial nephrectomy (LPN) is an effective way to treat renal cancer. Localization and delineation of the renal tumor from pre-operative CT Angiography (CTA) is an important step for LPN surgery planning. Recently, with the development of the technique of deep learning, deep neural networks can be trained to provide accurate pixel-wise renal tumor segmentation in CTA images. However, constructing the training dataset with a large amount of pixel-wise annotations is a time-consuming task for the radiologists. Therefore, weakly-supervised approaches attract more interest in research. Methods In this paper, we proposed a novel weakly-supervised convolutional neural network (CNN) for renal tumor segmentation. A three-stage framework was introduced to train the CNN with the weak annotations of renal tumors, i.e. the bounding boxes of renal tumors. The framework includes pseudo masks generation, group and weighted training phases. Clinical abdominal CT angiographic images of 200 patients were applied to perform the evaluation. Results Extensive experimental results show that the proposed method achieves a higher dice coefficient (DSC) of 0.826 than the other two existing weakly-supervised deep neural networks. Furthermore, the segmentation performance is close to the fully supervised deep CNN. Conclusions The proposed strategy improves not only the efficiency of network training but also the precision of the segmentation.
Collapse
Affiliation(s)
- Guanyu Yang
- LIST, Key Laboratory of Computer Network and Information Integration, Southeast University, Ministry of Education, Nanjing, China. .,Centre de Recherche en Information Biomédicale Sino-Français (CRIBs), Rennes, France.
| | - Chuanxia Wang
- LIST, Key Laboratory of Computer Network and Information Integration, Southeast University, Ministry of Education, Nanjing, China
| | - Jian Yang
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing, 100081, China
| | - Yang Chen
- LIST, Key Laboratory of Computer Network and Information Integration, Southeast University, Ministry of Education, Nanjing, China.,Centre de Recherche en Information Biomédicale Sino-Français (CRIBs), Rennes, France
| | - Lijun Tang
- Department of Radiology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, China
| | - Pengfei Shao
- Department of Urology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, China
| | - Jean-Louis Dillenseger
- Centre de Recherche en Information Biomédicale Sino-Français (CRIBs), Rennes, France.,University Rennes, Inserm, LTSI - UMR1099, F-35000, Rennes, France
| | - Huazhong Shu
- LIST, Key Laboratory of Computer Network and Information Integration, Southeast University, Ministry of Education, Nanjing, China.,Centre de Recherche en Information Biomédicale Sino-Français (CRIBs), Rennes, France
| | - Limin Luo
- LIST, Key Laboratory of Computer Network and Information Integration, Southeast University, Ministry of Education, Nanjing, China.,Centre de Recherche en Information Biomédicale Sino-Français (CRIBs), Rennes, France
| |
Collapse
|
100
|
Attention-Guided Multi-Scale Segmentation Neural Network for Interactive Extraction of Region Objects from High-Resolution Satellite Imagery. REMOTE SENSING 2020. [DOI: 10.3390/rs12050789] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Automatic extraction of region objects from high-resolution satellite imagery presents a great challenge, because there may be very large variations of the objects in terms of their size, texture, shape, and contextual complexity in the image. To handle these issues, we present a novel, deep-learning-based approach to interactively extract non-artificial region objects, such as water bodies, woodland, farmland, etc., from high-resolution satellite imagery. First, our algorithm transforms user-provided positive and negative clicks or scribbles into guidance maps, which consist of a relevance map modified from Euclidean distance maps, two geodesic distance maps (for positive and negative, respectively), and a sampling map. Then, feature maps are extracted by applying a VGG convolutional neural network pre-trained on the ImageNet dataset to the image X, and they are then upsampled to the resolution of X. Image X, guidance maps, and feature maps are integrated as the input tensor. We feed the proposed attention-guided, multi-scale segmentation neural network (AGMSSeg-Net) with the input tensor above to obtain the mask that assigns a binary label to each pixel. After a post-processing operation based on a fully connected Conditional Random Field (CRF), we extract the selected object boundary from the segmentation result. Experiments were conducted on two typical datasets with diverse region object types from complex scenes. The results demonstrate the effectiveness of the proposed method, and our approach outperforms existing methods for interactive image segmentation.
Collapse
|