1
|
Lew CO, Harouni M, Kirksey ER, Kang EJ, Dong H, Gu H, Grimm LJ, Walsh R, Lowell DA, Mazurowski MA. A publicly available deep learning model and dataset for segmentation of breast, fibroglandular tissue, and vessels in breast MRI. Sci Rep 2024; 14:5383. [PMID: 38443410 PMCID: PMC10915139 DOI: 10.1038/s41598-024-54048-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2022] [Accepted: 02/08/2024] [Indexed: 03/07/2024] Open
Abstract
Breast density, or the amount of fibroglandular tissue (FGT) relative to the overall breast volume, increases the risk of developing breast cancer. Although previous studies have utilized deep learning to assess breast density, the limited public availability of data and quantitative tools hinders the development of better assessment tools. Our objective was to (1) create and share a large dataset of pixel-wise annotations according to well-defined criteria, and (2) develop, evaluate, and share an automated segmentation method for breast, FGT, and blood vessels using convolutional neural networks. We used the Duke Breast Cancer MRI dataset to randomly select 100 MRI studies and manually annotated the breast, FGT, and blood vessels for each study. Model performance was evaluated using the Dice similarity coefficient (DSC). The model achieved DSC values of 0.92 for breast, 0.86 for FGT, and 0.65 for blood vessels on the test set. The correlation between our model's predicted breast density and the manually generated masks was 0.95. The correlation between the predicted breast density and qualitative radiologist assessment was 0.75. Our automated models can accurately segment breast, FGT, and blood vessels using pre-contrast breast MRI data. The data and the models were made publicly available.
Collapse
Affiliation(s)
- Christopher O Lew
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA.
| | - Majid Harouni
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Ella R Kirksey
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Elianne J Kang
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Haoyu Dong
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Hanxue Gu
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Lars J Grimm
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Ruth Walsh
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Dorothy A Lowell
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University Medical Center, Box 2731, Durham, NC, 27710, USA
| |
Collapse
|
2
|
Prabhu NK, Wong MK, Klapper JA, Haney JC, Mazurowski MA, Mammarappallil JG, Hartwig MG. Computed Tomography Volumetrics for Size Matching in Lung Transplantation for Restrictive Disease. Ann Thorac Surg 2024; 117:413-421. [PMID: 37031770 DOI: 10.1016/j.athoracsur.2023.03.033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/20/2022] [Revised: 03/08/2023] [Accepted: 03/26/2023] [Indexed: 04/11/2023]
Abstract
BACKGROUND There is no consensus on the optimal allograft sizing strategy for lung transplantation in restrictive lung disease. Current methods that are based on predicted total lung capacity (pTLC) ratios do not account for the diminutive recipient chest size. The study investigators hypothesized that a new sizing ratio incorporating preoperative recipient computed tomographic lung volumes (CTVol) would be associated with postoperative outcomes. METHODS A retrospective single-institution study was conducted of adults undergoing primary bilateral lung transplantation between January 2016 and July 2020 for restrictive lung disease. CTVol was computed for recipients by using advanced segmentation software. Two sizing ratios were calculated: pTLC ratio (pTLCdonor/pTLCrecipient) and a new volumetric ratio (pTLCdonor/CTVolrecipient). Patients were divided into reference, oversized, and undersized groups on the basis of ratio quintiles, and multivariable models were used to assess the effect of the ratios on primary graft dysfunction and survival. RESULTS CTVol was successfully acquired in 218 of 220 (99.1%) patients. In adjusted analysis, undersizing on the basis of the volumetric ratio was independently associated with decreased primary graft dysfunction grade 2 or 3 within 72 hours (odds ratio, 0.42; 95% CI, 0.20-0.87; P =.02). The pTLC ratio was not significantly associated with primary graft dysfunction. Oversizing on the basis of the volumetric ratio was independently associated with an increased risk of death (hazard ratio, 2.27; 95% CI, 1.04-4.99; P =.04], whereas the pTLC ratio did not have a significant survival association. CONCLUSIONS Using computed tomography-acquired lung volumes for donor-recipient size matching in lung transplantation is feasible with advanced segmentation software. This method may be more predictive of outcome compared with current sizing methods, which use gender and height only.
Collapse
Affiliation(s)
- Neel K Prabhu
- Duke University School of Medicine, Durham, North Carolina.
| | - Megan K Wong
- Duke University School of Medicine, Durham, North Carolina
| | - Jacob A Klapper
- Duke University School of Medicine, Durham, North Carolina; Division of Thoracic and Cardiovascular Surgery, Department of Surgery, Duke University Medical Center, Durham, North Carolina
| | - John C Haney
- Duke University School of Medicine, Durham, North Carolina; Division of Thoracic and Cardiovascular Surgery, Department of Surgery, Duke University Medical Center, Durham, North Carolina
| | - Maciej A Mazurowski
- Duke University School of Medicine, Durham, North Carolina; Department of Computer Science, Duke University, Durham, North Carolina; Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina; Department of Biostatistics and Bioinformatics, Duke University, Durham, North Carolina; Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Joseph G Mammarappallil
- Duke University School of Medicine, Durham, North Carolina; Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Matthew G Hartwig
- Duke University School of Medicine, Durham, North Carolina; Division of Thoracic and Cardiovascular Surgery, Department of Surgery, Duke University Medical Center, Durham, North Carolina
| |
Collapse
|
3
|
Dong H, Zhang Y, Gu H, Konz N, Zhang Y, Mazurowski MA. SWSSL: Sliding Window-Based Self-Supervised Learning for Anomaly Detection in High-Resolution Images. IEEE Trans Med Imaging 2023; 42:3860-3870. [PMID: 37695965 PMCID: PMC10766076 DOI: 10.1109/tmi.2023.3314318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/13/2023]
Abstract
Anomaly detection (AD) aims to determine if an instance has properties different from those seen in normal cases. The success of this technique depends on how well a neural network learns from normal instances. We observe that the learning difficulty scales exponentially with the input resolution, making it infeasible to apply AD to high-resolution images. Resizing them to a lower resolution is a compromising solution and does not align with clinical practice where the diagnosis could depend on image details. In this work, we propose to train the network and perform inference at the patch level, through the sliding window algorithm. This simple operation allows the network to receive high-resolution images but introduces additional training difficulties, including inconsistent image structure and higher variance. We address these concerns by setting the network's objective to learn augmentation-invariant features. We further study the augmentation function in the context of medical imaging. In particular, we observe that the resizing operation, a key augmentation in general computer vision literature, is detrimental to detection accuracy, and the inverting operation can be beneficial. We also propose a new module that encourages the network to learn from adjacent patches to boost detection performance. Extensive experiments are conducted on breast tomosynthesis and chest X-ray datasets and our method improves 8.03% and 5.66% AUC on image-level classification respectively over the current leading techniques. The experimental results demonstrate the effectiveness of our approach.
Collapse
|
4
|
Zhang J, Santos C, Park C, Mazurowski MA, Colglazier R. Improving Image Classification of Knee Radiographs: An Automated Image Labeling Approach. J Digit Imaging 2023; 36:2402-2410. [PMID: 37620710 PMCID: PMC10584746 DOI: 10.1007/s10278-023-00894-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2023] [Revised: 07/28/2023] [Accepted: 07/31/2023] [Indexed: 08/26/2023] Open
Abstract
Large numbers of radiographic images are available in musculoskeletal radiology practices which could be used for training of deep learning models for diagnosis of knee abnormalities. However, those images do not typically contain readily available labels due to limitations of human annotations. The purpose of our study was to develop an automated labeling approach that improves the image classification model to distinguish normal knee images from those with abnormalities or prior arthroplasty. The automated labeler was trained on a small set of labeled data to automatically label a much larger set of unlabeled data, further improving the image classification performance for knee radiographic diagnosis. We used BioBERT and EfficientNet as the feature extraction backbone of the labeler and imaging model, respectively. We developed our approach using 7382 patients and validated it on a separate set of 637 patients. The final image classification model, trained using both manually labeled and pseudo-labeled data, had the higher weighted average AUC (WA-AUC 0.903) value and higher AUC values among all classes (normal AUC 0.894; abnormal AUC 0.896, arthroplasty AUC 0.990) compared to the baseline model (WA-AUC = 0.857; normal AUC 0.842; abnormal AUC 0.848, arthroplasty AUC 0.987), trained using only manually labeled data. Statistical tests show that the improvement is significant on normal (p value < 0.002), abnormal (p value < 0.001), and WA-AUC (p value = 0.001). Our findings demonstrated that the proposed automated labeling approach significantly improves the performance of image classification for radiographic knee diagnosis, allowing for facilitating patient care and curation of large knee datasets.
Collapse
Affiliation(s)
- Jikai Zhang
- Department of Electrical and Computer Engineering, Duke University, Room 10070, 2424 Erwin Road, Durham, NC, 27705, USA.
| | - Carlos Santos
- Wake Forest University, Winston-Salem, NC, 27109, USA
| | - Christine Park
- Department of Radiology, Duke University Medical Center, Durham, NC, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University Medical Center, Durham, NC, USA
- Department of Electrical and Computer Engineering, Department of Biostatistics and Bioinformatics, Department of Computer Science, Duke University, Durham, NC, USA
| | - Roy Colglazier
- Department of Radiology, Duke University Medical Center, Durham, NC, USA
| |
Collapse
|
5
|
Mazurowski MA, Dong H, Gu H, Yang J, Konz N, Zhang Y. Segment anything model for medical image analysis: An experimental study. Med Image Anal 2023; 89:102918. [PMID: 37595404 PMCID: PMC10528428 DOI: 10.1016/j.media.2023.102918] [Citation(s) in RCA: 19] [Impact Index Per Article: 19.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2023] [Revised: 07/03/2023] [Accepted: 07/31/2023] [Indexed: 08/20/2023]
Abstract
Training segmentation models for medical images continues to be challenging due to the limited availability of data annotations. Segment Anything Model (SAM) is a foundation model trained on over 1 billion annotations, predominantly for natural images, that is intended to segment user-defined objects of interest in an interactive manner. While the model performance on natural images is impressive, medical image domains pose their own set of challenges. Here, we perform an extensive evaluation of SAM's ability to segment medical images on a collection of 19 medical imaging datasets from various modalities and anatomies. In our experiments, we generated point and box prompts for SAM using a standard method that simulates interactive segmentation. We report the following findings: (1) SAM's performance based on single prompts highly varies depending on the dataset and the task, from IoU=0.1135 for spine MRI to IoU=0.8650 for hip X-ray. (2) Segmentation performance appears to be better for well-circumscribed objects with prompts with less ambiguity such as the segmentation of organs in computed tomography and poorer in various other scenarios such as the segmentation of brain tumors. (3) SAM performs notably better with box prompts than with point prompts. (4) SAM outperforms similar methods RITM, SimpleClick, and FocalClick in almost all single-point prompt settings. (5) When multiple-point prompts are provided iteratively, SAM's performance generally improves only slightly while other methods' performance improves to the level that surpasses SAM's point-based performance. We also provide several illustrations for SAM's performance on all tested datasets, iterative segmentation, and SAM's behavior given prompt ambiguity. We conclude that SAM shows impressive zero-shot segmentation performance for certain medical imaging datasets, but moderate to poor performance for others. SAM has the potential to make a significant impact in automated medical image segmentation in medical imaging, but appropriate care needs to be applied when using it. Code for evaluation SAM is made publicly available at https://github.com/mazurowski-lab/segment-anything-medical-evaluation.
Collapse
Affiliation(s)
- Maciej A Mazurowski
- Department of Radiology, Duke University, Durham, NC, 27708, USA; Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA; Department of Computer Science, Duke University, Durham, NC, 27708, USA; Department of Biostatistics & Bioinformatics, Duke University, Durham, NC, 27708, USA
| | - Haoyu Dong
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA.
| | - Hanxue Gu
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA
| | - Jichen Yang
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA
| | - Nicholas Konz
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA
| | - Yixin Zhang
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA
| |
Collapse
|
6
|
Lew CO, Zhou L, Mazurowski MA, Doraiswamy PM, Petrella JR. MRI-based Deep Learning Assessment of Amyloid, Tau, and Neurodegeneration Biomarker Status across the Alzheimer Disease Spectrum. Radiology 2023; 309:e222441. [PMID: 37815445 PMCID: PMC10623183 DOI: 10.1148/radiol.222441] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2022] [Revised: 09/03/2023] [Accepted: 09/11/2023] [Indexed: 10/11/2023]
Abstract
Background PET can be used for amyloid-tau-neurodegeneration (ATN) classification in Alzheimer disease, but incurs considerable cost and exposure to ionizing radiation. MRI currently has limited use in characterizing ATN status. Deep learning techniques can detect complex patterns in MRI data and have potential for noninvasive characterization of ATN status. Purpose To use deep learning to predict PET-determined ATN biomarker status using MRI and readily available diagnostic data. Materials and Methods MRI and PET data were retrospectively collected from the Alzheimer's Disease Imaging Initiative. PET scans were paired with MRI scans acquired within 30 days, from August 2005 to September 2020. Pairs were randomly split into subsets as follows: 70% for training, 10% for validation, and 20% for final testing. A bimodal Gaussian mixture model was used to threshold PET scans into positive and negative labels. MRI data were fed into a convolutional neural network to generate imaging features. These features were combined in a logistic regression model with patient demographics, APOE gene status, cognitive scores, hippocampal volumes, and clinical diagnoses to classify each ATN biomarker component as positive or negative. Area under the receiver operating characteristic curve (AUC) analysis was used for model evaluation. Feature importance was derived from model coefficients and gradients. Results There were 2099 amyloid (mean patient age, 75 years ± 10 [SD]; 1110 male), 557 tau (mean patient age, 75 years ± 7; 280 male), and 2768 FDG PET (mean patient age, 75 years ± 7; 1645 male) and MRI pairs. Model AUCs for the test set were as follows: amyloid, 0.79 (95% CI: 0.74, 0.83); tau, 0.73 (95% CI: 0.58, 0.86); and neurodegeneration, 0.86 (95% CI: 0.83, 0.89). Within the networks, high gradients were present in key temporal, parietal, frontal, and occipital cortical regions. Model coefficients for cognitive scores, hippocampal volumes, and APOE status were highest. Conclusion A deep learning algorithm predicted each component of PET-determined ATN status with acceptable to excellent efficacy using MRI and other available diagnostic data. © RSNA, 2023 Supplemental material is available for this article.
Collapse
Affiliation(s)
- Christopher O. Lew
- From the Department of Radiology, Division of Neuroradiology,
Alzheimer Disease Imaging Research Laboratory (C.O.L., J.R.P.), and
Neurocognitive Disorders Program, Departments of Psychiatry and Medicine
(P.M.D.), Duke University Medical Center, DUMC-Box 3808, Durham, NC 27710-3808;
and Duke Institute for Brain Sciences (P.M.D.) and Department of Electrical and
Computer Engineering, Department of Computer Science, Department of
Biostatistics and Bioinformatics (L.Z., M.A.M.), Duke University, Durham,
NC
| | - Longfei Zhou
- From the Department of Radiology, Division of Neuroradiology,
Alzheimer Disease Imaging Research Laboratory (C.O.L., J.R.P.), and
Neurocognitive Disorders Program, Departments of Psychiatry and Medicine
(P.M.D.), Duke University Medical Center, DUMC-Box 3808, Durham, NC 27710-3808;
and Duke Institute for Brain Sciences (P.M.D.) and Department of Electrical and
Computer Engineering, Department of Computer Science, Department of
Biostatistics and Bioinformatics (L.Z., M.A.M.), Duke University, Durham,
NC
| | - Maciej A. Mazurowski
- From the Department of Radiology, Division of Neuroradiology,
Alzheimer Disease Imaging Research Laboratory (C.O.L., J.R.P.), and
Neurocognitive Disorders Program, Departments of Psychiatry and Medicine
(P.M.D.), Duke University Medical Center, DUMC-Box 3808, Durham, NC 27710-3808;
and Duke Institute for Brain Sciences (P.M.D.) and Department of Electrical and
Computer Engineering, Department of Computer Science, Department of
Biostatistics and Bioinformatics (L.Z., M.A.M.), Duke University, Durham,
NC
| | - P. Murali Doraiswamy
- From the Department of Radiology, Division of Neuroradiology,
Alzheimer Disease Imaging Research Laboratory (C.O.L., J.R.P.), and
Neurocognitive Disorders Program, Departments of Psychiatry and Medicine
(P.M.D.), Duke University Medical Center, DUMC-Box 3808, Durham, NC 27710-3808;
and Duke Institute for Brain Sciences (P.M.D.) and Department of Electrical and
Computer Engineering, Department of Computer Science, Department of
Biostatistics and Bioinformatics (L.Z., M.A.M.), Duke University, Durham,
NC
| | - Jeffrey R. Petrella
- From the Department of Radiology, Division of Neuroradiology,
Alzheimer Disease Imaging Research Laboratory (C.O.L., J.R.P.), and
Neurocognitive Disorders Program, Departments of Psychiatry and Medicine
(P.M.D.), Duke University Medical Center, DUMC-Box 3808, Durham, NC 27710-3808;
and Duke Institute for Brain Sciences (P.M.D.) and Department of Electrical and
Computer Engineering, Department of Computer Science, Department of
Biostatistics and Bioinformatics (L.Z., M.A.M.), Duke University, Durham,
NC
| | | |
Collapse
|
7
|
Macdonald JA, Zhu Z, Konkel B, Mazurowski MA, Wiggins WF, Bashir MR. Duke Liver Dataset: A Publicly Available Liver MRI Dataset with Liver Segmentation Masks and Series Labels. Radiol Artif Intell 2023; 5:e220275. [PMID: 37795141 PMCID: PMC10546360 DOI: 10.1148/ryai.220275] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2022] [Revised: 06/26/2023] [Accepted: 07/03/2023] [Indexed: 10/06/2023]
Abstract
The Duke Liver Dataset contains 2146 abdominal MRI series from 105 patients, including a majority with cirrhotic features, and 310 image series with corresponding manually segmented liver masks.
Collapse
Affiliation(s)
- Jacob A. Macdonald
- From the Department of Radiology (J.A.M., Z.Z., B.K., M.A.M., W.F.W., M.R.B.), Department of Electrical and Computer Engineering (M.A.M.), Department of Computer Science (M.A.M.), Center for Advanced Magnetic Resonance Development (M.R.B.), and Division of Gastroenterology, Department of Medicine (M.R.B.), Duke University, 2301 Erwin Rd, Durham, NC 27710
| | - Zhe Zhu
- From the Department of Radiology (J.A.M., Z.Z., B.K., M.A.M., W.F.W., M.R.B.), Department of Electrical and Computer Engineering (M.A.M.), Department of Computer Science (M.A.M.), Center for Advanced Magnetic Resonance Development (M.R.B.), and Division of Gastroenterology, Department of Medicine (M.R.B.), Duke University, 2301 Erwin Rd, Durham, NC 27710
| | - Brandon Konkel
- From the Department of Radiology (J.A.M., Z.Z., B.K., M.A.M., W.F.W., M.R.B.), Department of Electrical and Computer Engineering (M.A.M.), Department of Computer Science (M.A.M.), Center for Advanced Magnetic Resonance Development (M.R.B.), and Division of Gastroenterology, Department of Medicine (M.R.B.), Duke University, 2301 Erwin Rd, Durham, NC 27710
| | - Maciej A. Mazurowski
- From the Department of Radiology (J.A.M., Z.Z., B.K., M.A.M., W.F.W., M.R.B.), Department of Electrical and Computer Engineering (M.A.M.), Department of Computer Science (M.A.M.), Center for Advanced Magnetic Resonance Development (M.R.B.), and Division of Gastroenterology, Department of Medicine (M.R.B.), Duke University, 2301 Erwin Rd, Durham, NC 27710
| | - Walter F. Wiggins
- From the Department of Radiology (J.A.M., Z.Z., B.K., M.A.M., W.F.W., M.R.B.), Department of Electrical and Computer Engineering (M.A.M.), Department of Computer Science (M.A.M.), Center for Advanced Magnetic Resonance Development (M.R.B.), and Division of Gastroenterology, Department of Medicine (M.R.B.), Duke University, 2301 Erwin Rd, Durham, NC 27710
| | - Mustafa R. Bashir
- From the Department of Radiology (J.A.M., Z.Z., B.K., M.A.M., W.F.W., M.R.B.), Department of Electrical and Computer Engineering (M.A.M.), Department of Computer Science (M.A.M.), Center for Advanced Magnetic Resonance Development (M.R.B.), and Division of Gastroenterology, Department of Medicine (M.R.B.), Duke University, 2301 Erwin Rd, Durham, NC 27710
| |
Collapse
|
8
|
Zhang J, Mazurowski MA, Grimm LJ. Feasibility of predicting a screening digital breast tomosynthesis recall using features extracted from the electronic medical record. Eur J Radiol 2023; 166:110979. [PMID: 37473618 DOI: 10.1016/j.ejrad.2023.110979] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2023] [Revised: 07/05/2023] [Accepted: 07/12/2023] [Indexed: 07/22/2023]
Abstract
PURPOSE Tools to predict a screening mammogram recall at the time of scheduling could improve patient care. We extracted patient demographic and breast care history information within the electronic medical record (EMR) for women undergoing digital breast tomosynthesis (DBT) to identify which factors were associated with a screening recall recommendation. METHOD In 2018, 21,543 women aged 40 years or greater who underwent screening DBT at our institution were identified. Demographic information and breast care factors were extracted automatically from the EMR. The primary outcome was a screening recall recommendation of BI-RADS 0. A multivariable logistic regression model was built and included age, race, ethnicity groups, family breast cancer history, personal breast cancer history, surgical breast cancer history, recall history, and days since last available screening mammogram. RESULTS Multiple factors were associated with a recall on the multivariable model: history of breast cancer surgery (OR: 2.298, 95% CI: 1.854, 2.836); prior recall within the last five years (vs no prior, OR: 0.768, 95% CI: 0.687, 0.858); prior screening mammogram within 0-18 months (vs no prior, OR: 0.601, 95% CI: 0.520, 0.691), prior screening mammogram within 18-30 months (vs no prior, OR: 0.676, 95% CI: 0.520, 0.691); and age (normalized OR: 0.723, 95% CI: 0.690, 0.758). CONCLUSIONS It is feasible to predict a DBT screening recall recommendation using patient demographics and breast care factors that can be extracted automatically from the EMR.
Collapse
Affiliation(s)
- Jikai Zhang
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, United States Room 10070, 2424 Erwin Road, Durham, NC 27705, United States.
| | - Maciej A Mazurowski
- Department of Radiology, Duke University Medical Center, Durham, NC, United States; Department of Electrical and Computer Engineering, Department of Biostatistics and Bioinformatics, Department of Computer Science, Duke University, Durham, NC, United States
| | - Lars J Grimm
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, United States Room 10070, 2424 Erwin Road, Durham, NC 27705, United States
| |
Collapse
|
9
|
Zhang J, Mazurowski MA, Allen BC, Wildman-Tobriner B. Multistep Automated Data Labelling Procedure (MADLaP) for thyroid nodules on ultrasound: An artificial intelligence approach for automating image annotation. Artif Intell Med 2023; 141:102553. [PMID: 37295897 DOI: 10.1016/j.artmed.2023.102553] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Revised: 02/14/2023] [Accepted: 04/11/2023] [Indexed: 06/12/2023]
Abstract
Machine learning (ML) for diagnosis of thyroid nodules on ultrasound is an active area of research. However, ML tools require large, well-labeled datasets, the curation of which is time-consuming and labor-intensive. The purpose of our study was to develop and test a deep-learning-based tool to facilitate and automate the data annotation process for thyroid nodules; we named our tool Multistep Automated Data Labelling Procedure (MADLaP). MADLaP was designed to take multiple inputs including pathology reports, ultrasound images, and radiology reports. Using multiple step-wise 'modules' including rule-based natural language processing, deep-learning-based imaging segmentation, and optical character recognition, MADLaP automatically identified images of a specific thyroid nodule and correctly assigned a pathology label. The model was developed using a training set of 378 patients across our health system and tested on a separate set of 93 patients. Ground truths for both sets were selected by an experienced radiologist. Performance metrics including yield (how many labeled images the model produced) and accuracy (percentage correct) were measured using the test set. MADLaP achieved a yield of 63 % and an accuracy of 83 %. The yield progressively increased as the input data moved through each module, while accuracy peaked part way through. Error analysis showed that inputs from certain examination sites had lower accuracy (40 %) than the other sites (90 %, 100 %). MADLaP successfully created curated datasets of labeled ultrasound images of thyroid nodules. While accurate, the relatively suboptimal yield of MADLaP exposed some challenges when trying to automatically label radiology images from heterogeneous sources. The complex task of image curation and annotation could be automated, allowing for enrichment of larger datasets for use in machine learning development.
Collapse
Affiliation(s)
- Jikai Zhang
- Department of Electrical and Computer Engineering, Duke University, Room 10070, 2424 Erwin Rd, Durham, NC 27705, United States.
| | - Maciej A Mazurowski
- Department of Radiology, Duke University Medical Center, Durham, NC, United States; Department of Electrical and Computer Engineering, Department of Biostatistics and Bioinformatics, Department of Computer Science, Duke University, Room 9044, 2424 Erwin Rd, Durham, NC 27705, United States
| | - Brian C Allen
- Department of Radiology, Duke University Medical Center, Duke University, Dept of Radiology, Box 3808, Durham, NC 27710, United States
| | - Benjamin Wildman-Tobriner
- Department of Radiology, Duke University Medical Center, Duke University, Dept of Radiology, Box 3808, Durham, NC 27710, United States
| |
Collapse
|
10
|
Konz N, Dong H, Mazurowski MA. Unsupervised anomaly localization in high-resolution breast scans using deep pluralistic image completion. Med Image Anal 2023; 87:102836. [PMID: 37201220 DOI: 10.1016/j.media.2023.102836] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2022] [Revised: 04/19/2023] [Accepted: 05/03/2023] [Indexed: 05/20/2023]
Abstract
Automated tumor detection in Digital Breast Tomosynthesis (DBT) is a difficult task due to natural tumor rarity, breast tissue variability, and high resolution. Given the scarcity of abnormal images and the abundance of normal images for this problem, an anomaly detection/localization approach could be well-suited. However, most anomaly localization research in machine learning focuses on non-medical datasets, and we find that these methods fall short when adapted to medical imaging datasets. The problem is alleviated when we solve the task from the image completion perspective, in which the presence of anomalies can be indicated by a discrepancy between the original appearance and its auto-completion conditioned on the surroundings. However, there are often many valid normal completions given the same surroundings, especially in the DBT dataset, making this evaluation criterion less precise. To address such an issue, we consider pluralistic image completion by exploring the distribution of possible completions instead of generating fixed predictions. This is achieved through our novel application of spatial dropout on the completion network during inference time only, which requires no additional training cost and is effective at generating diverse completions. We further propose minimum completion distance (MCD), a new metric for detecting anomalies, thanks to these stochastic completions. We provide theoretical as well as empirical support for the superiority over existing methods of using the proposed method for anomaly localization. On the DBT dataset, our model outperforms other state-of-the-art methods by at least 10% AUROC for pixel-level detection.
Collapse
Affiliation(s)
- Nicholas Konz
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA.
| | - Haoyu Dong
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA
| | - Maciej A Mazurowski
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27708, USA; Department of Computer Science, Duke University, Durham, NC, 27708, USA; Department of Radiology, Duke University, Durham, NC, 27708, USA; Department of Biostatistics & Bioinformatics, Duke University, Durham, NC, 27708, USA
| |
Collapse
|
11
|
Weng J, Wildman-Tobriner B, Buda M, Yang J, Ho LM, Allen BC, Ehieli WL, Miller CM, Zhang J, Mazurowski MA. Deep learning for classification of thyroid nodules on ultrasound: validation on an independent dataset. Clin Imaging 2023; 99:60-66. [PMID: 37116263 DOI: 10.1016/j.clinimag.2023.04.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2022] [Revised: 04/13/2023] [Accepted: 04/13/2023] [Indexed: 04/30/2023]
Abstract
OBJECTIVES The purpose is to apply a previously validated deep learning algorithm to a new thyroid nodule ultrasound image dataset and compare its performances with radiologists. METHODS Prior study presented an algorithm which is able to detect thyroid nodules and then make malignancy classifications with two ultrasound images. A multi-task deep convolutional neural network was trained from 1278 nodules and originally tested with 99 separate nodules. The results were comparable with that of radiologists. The algorithm was further tested with 378 nodules imaged with ultrasound machines from different manufacturers and product types than the training cases. Four experienced radiologists were requested to evaluate the nodules for comparison with deep learning. RESULTS The Area Under Curve (AUC) of the deep learning algorithm and four radiologists were calculated with parametric, binormal estimation. For the deep learning algorithm, the AUC was 0.69 (95% CI: 0.64-0.75). The AUC of radiologists were 0.63 (95% CI: 0.59-0.67), 0.66 (95% CI:0.61-0.71), 0.65 (95% CI: 0.60-0.70), and 0.63 (95%CI: 0.58-0.67). CONCLUSION In the new testing dataset, the deep learning algorithm achieved similar performances with all four radiologists. The relative performance difference between the algorithm and the radiologists is not significantly affected by the difference of ultrasound scanner.
Collapse
Affiliation(s)
- Jingxi Weng
- Department of Radiation Oncology, University of Florida, Gainesville, FL, USA
| | | | - Mateusz Buda
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA
| | - Jichen Yang
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA.
| | - Lisa M Ho
- Department of Radiology, Duke University Medical Center, USA
| | - Brian C Allen
- Department of Radiology, Duke University Medical Center, USA
| | - Wendy L Ehieli
- Department of Radiology, Duke University Medical Center, USA
| | - Chad M Miller
- Department of Radiology, Duke University Medical Center, USA
| | - Jikai Zhang
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA
| | | |
Collapse
|
12
|
Cao S, Konz N, Duncan J, Mazurowski MA. Deep Learning for Breast MRI Style Transfer with Limited Training Data. J Digit Imaging 2023; 36:666-678. [PMID: 36544066 PMCID: PMC10039216 DOI: 10.1007/s10278-022-00755-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2022] [Revised: 11/08/2022] [Accepted: 12/07/2022] [Indexed: 12/24/2022] Open
Abstract
In this work we introduce a novel medical image style transfer method, StyleMapper, that can transfer medical scans to an unseen style with access to limited training data. This is made possible by training our model on unlimited possibilities of simulated random medical imaging styles on the training set, making our work more computationally efficient when compared with other style transfer methods. Moreover, our method enables arbitrary style transfer: transferring images to styles unseen in training. This is useful for medical imaging, where images are acquired using different protocols and different scanner models, resulting in a variety of styles that data may need to be transferred between. Our model disentangles image content from style and can modify an image's style by simply replacing the style encoding with one extracted from a single image of the target style, with no additional optimization required. This also allows the model to distinguish between different styles of images, including among those that were unseen in training. We propose a formal description of the proposed model. Experimental results on breast magnetic resonance images indicate the effectiveness of our method for style transfer. Our style transfer method allows for the alignment of medical images taken with different scanners into a single unified style dataset, allowing for the training of other downstream tasks on such a dataset for tasks such as classification, object detection and others.
Collapse
Affiliation(s)
- Shixing Cao
- Department of Electrical and Computer Engineering, Duke University, Durham, 27704 NC USA
| | - Nicholas Konz
- Department of Electrical and Computer Engineering, Duke University, Durham, 27704 NC USA
| | - James Duncan
- Department of Biomedical Engineering, Yale University, New Haven, 06520 CT USA
| | - Maciej A. Mazurowski
- Department of Electrical and Computer Engineering, Duke University, Durham, 27704 NC USA
- Departments of Computer Science, Radiology and Biostatistics & Bioinformatics, Duke University, Durham, 27704 NC USA
| |
Collapse
|
13
|
Konz N, Buda M, Gu H, Saha A, Yang J, Chłędowski J, Park J, Witowski J, Geras KJ, Shoshan Y, Gilboa-Solomon F, Khapun D, Ratner V, Barkan E, Ozery-Flato M, Martí R, Omigbodun A, Marasinou C, Nakhaei N, Hsu W, Sahu P, Hossain MB, Lee J, Santos C, Przelaskowski A, Kalpathy-Cramer J, Bearce B, Cha K, Farahani K, Petrick N, Hadjiiski L, Drukker K, Armato SG, Mazurowski MA. A Competition, Benchmark, Code, and Data for Using Artificial Intelligence to Detect Lesions in Digital Breast Tomosynthesis. JAMA Netw Open 2023; 6:e230524. [PMID: 36821110 PMCID: PMC9951043 DOI: 10.1001/jamanetworkopen.2023.0524] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 02/24/2023] Open
Abstract
IMPORTANCE An accurate and robust artificial intelligence (AI) algorithm for detecting cancer in digital breast tomosynthesis (DBT) could significantly improve detection accuracy and reduce health care costs worldwide. OBJECTIVES To make training and evaluation data for the development of AI algorithms for DBT analysis available, to develop well-defined benchmarks, and to create publicly available code for existing methods. DESIGN, SETTING, AND PARTICIPANTS This diagnostic study is based on a multi-institutional international grand challenge in which research teams developed algorithms to detect lesions in DBT. A data set of 22 032 reconstructed DBT volumes was made available to research teams. Phase 1, in which teams were provided 700 scans from the training set, 120 from the validation set, and 180 from the test set, took place from December 2020 to January 2021, and phase 2, in which teams were given the full data set, took place from May to July 2021. MAIN OUTCOMES AND MEASURES The overall performance was evaluated by mean sensitivity for biopsied lesions using only DBT volumes with biopsied lesions; ties were broken by including all DBT volumes. RESULTS A total of 8 teams participated in the challenge. The team with the highest mean sensitivity for biopsied lesions was the NYU B-Team, with 0.957 (95% CI, 0.924-0.984), and the second-place team, ZeDuS, had a mean sensitivity of 0.926 (95% CI, 0.881-0.964). When the results were aggregated, the mean sensitivity for all submitted algorithms was 0.879; for only those who participated in phase 2, it was 0.926. CONCLUSIONS AND RELEVANCE In this diagnostic study, an international competition produced algorithms with high sensitivity for using AI to detect lesions on DBT images. A standardized performance benchmark for the detection task using publicly available clinical imaging data was released, with detailed descriptions and analyses of submitted algorithms accompanied by a public release of their predictions and code for selected methods. These resources will serve as a foundation for future research on computer-assisted diagnosis methods for DBT, significantly lowering the barrier of entry for new researchers.
Collapse
Affiliation(s)
- Nicholas Konz
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
| | - Mateusz Buda
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Faculty of Mathematics and Information Science, Warsaw University of Technology, Warsaw, Poland
| | - Hanxue Gu
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
| | - Ashirbani Saha
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Department of Oncology, McMaster University, Hamilton, Ontario, Canada
| | | | - Jakub Chłędowski
- Jagiellonian University, Kraków, Poland
- Department of Radiology, NYU Grossman School of Medicine, New York, New York
| | - Jungkyu Park
- Department of Radiology, NYU Grossman School of Medicine, New York, New York
| | - Jan Witowski
- Department of Radiology, NYU Grossman School of Medicine, New York, New York
| | - Krzysztof J. Geras
- Department of Radiology, NYU Grossman School of Medicine, New York, New York
| | - Yoel Shoshan
- Medical Image Analytics, IBM Research, Haifa, Israel
| | | | - Daniel Khapun
- Medical Image Analytics, IBM Research, Haifa, Israel
| | - Vadim Ratner
- Medical Image Analytics, IBM Research, Haifa, Israel
| | - Ella Barkan
- Medical Image Analytics, IBM Research, Haifa, Israel
| | | | - Robert Martí
- Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
| | - Akinyinka Omigbodun
- Medical and Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine, University of California Los Angeles
| | - Chrysostomos Marasinou
- Medical and Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine, University of California Los Angeles
| | - Noor Nakhaei
- Medical and Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine, University of California Los Angeles
| | - William Hsu
- Medical and Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine, University of California Los Angeles
- Department of Radiological Sciences, David Geffen School of Medicine, University of California Los Angeles
- Department of Bioengineering, University of California Los Angeles Samueli School of Engineering
| | - Pranjal Sahu
- Department of Computer Science, Stony Brook University, Stony Brook, New York
| | - Md Belayat Hossain
- Department of Radiology, University of Pittsburgh, Pittsburgh, Pennsylvania
| | - Juhun Lee
- Department of Radiology, University of Pittsburgh, Pittsburgh, Pennsylvania
| | - Carlos Santos
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Artur Przelaskowski
- Faculty of Mathematics and Information Science, Warsaw University of Technology, Warsaw, Poland
| | - Jayashree Kalpathy-Cramer
- Department of Radiology, Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown
| | - Benjamin Bearce
- Department of Radiology, Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown
| | - Kenny Cha
- US Food and Drug Administration, Silver Spring, Maryland
| | - Keyvan Farahani
- Center for Biomedical Informatics and Information Technology, National Cancer Institute, Bethesda, Maryland
| | | | | | - Karen Drukker
- Department of Radiology, University of Chicago, Chicago, Illinois
| | - Samuel G. Armato
- Department of Radiology, University of Chicago, Chicago, Illinois
| | - Maciej A. Mazurowski
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Department of Computer Science, Duke University, Durham, North Carolina
- Department of Biostatistics and Bioinformatics, Duke University Medical Center, Durham, North Carolina
| |
Collapse
|
14
|
Goldstein BA, Mazurowski MA, Li C. The Need for Targeted Labeling of Machine Learning-Based Software as a Medical Device. JAMA Netw Open 2022; 5:e2242351. [PMID: 36409502 DOI: 10.1001/jamanetworkopen.2022.42351] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Affiliation(s)
- Benjamin A Goldstein
- Department of Biostatistics and Bioinformatics, Duke University School of Medicine, Durham, North Carolina
- Duke AI Health, Duke University School of Medicine, Durham, North Carolina
| | - Maciej A Mazurowski
- Department of Biostatistics and Bioinformatics, Duke University School of Medicine, Durham, North Carolina
- Duke AI Health, Duke University School of Medicine, Durham, North Carolina
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina
| | - Cheng Li
- Independent Regulatory Consultant, Durham, North Carolina
| |
Collapse
|
15
|
Hou R, Peng Y, Grimm LJ, Ren Y, Mazurowski MA, Marks JR, King LM, Maley CC, Hwang ES, Lo JY. Anomaly Detection of Calcifications in Mammography Based on 11,000 Negative Cases. IEEE Trans Biomed Eng 2022; 69:1639-1650. [PMID: 34788216 DOI: 10.1109/tbme.2021.3126281] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
In mammography, calcifications are one of the most common signs of breast cancer. Detection of such lesions is an active area of research for computer-aided diagnosis and machine learning algorithms. Due to limited numbers of positive cases, many supervised detection models suffer from overfitting and fail to generalize. We present a one-class, semi-supervised framework using a deep convolutional autoencoder trained with over 50,000 images from 11,000 negative-only cases. Since the model learned from only normal breast parenchymal features, calcifications produced large signals when comparing the residuals between input and reconstruction output images. As a key advancement, a structural dissimilarity index was used to suppress non-structural noises. Our selected model achieved pixel-based AUROC of 0.959 and AUPRC of 0.676 during validation, where calcification masks were defined in a semi-automated process. Although not trained directly on any cancers, detection performance of calcification lesions on 1,883 testing images (645 malignant and 1238 negative) achieved 75% sensitivity at 2.5 false positives per image. Performance plateaued early when trained with only a fraction of the cases, and greater model complexity or a larger dataset did not improve performance. This study demonstrates the potential of this anomaly detection approach to detect mammographic calcifications in a semi-supervised manner with efficient use of a small number of labeled images, and may facilitate new clinical applications such as computer-aided triage and quality improvement.
Collapse
|
16
|
D’Anniballe VM, Tushar FI, Faryna K, Han S, Mazurowski MA, Rubin GD, Lo JY. Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning. BMC Med Inform Decis Mak 2022; 22:102. [PMID: 35428335 PMCID: PMC9011942 DOI: 10.1186/s12911-022-01843-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2021] [Accepted: 04/08/2022] [Indexed: 11/10/2022] Open
Abstract
Abstract
Background
There is progress to be made in building artificially intelligent systems to detect abnormalities that are not only accurate but can handle the true breadth of findings that radiologists encounter in body (chest, abdomen, and pelvis) computed tomography (CT). Currently, the major bottleneck for developing multi-disease classifiers is a lack of manually annotated data. The purpose of this work was to develop high throughput multi-label annotators for body CT reports that can be applied across a variety of abnormalities, organs, and disease states thereby mitigating the need for human annotation.
Methods
We used a dictionary approach to develop rule-based algorithms (RBA) for extraction of disease labels from radiology text reports. We targeted three organ systems (lungs/pleura, liver/gallbladder, kidneys/ureters) with four diseases per system based on their prevalence in our dataset. To expand the algorithms beyond pre-defined keywords, attention-guided recurrent neural networks (RNN) were trained using the RBA-extracted labels to classify reports as being positive for one or more diseases or normal for each organ system. Alternative effects on disease classification performance were evaluated using random initialization or pre-trained embedding as well as different sizes of training datasets. The RBA was tested on a subset of 2158 manually labeled reports and performance was reported as accuracy and F-score. The RNN was tested against a test set of 48,758 reports labeled by RBA and performance was reported as area under the receiver operating characteristic curve (AUC), with 95% CIs calculated using the DeLong method.
Results
Manual validation of the RBA confirmed 91–99% accuracy across the 15 different labels. Our models extracted disease labels from 261,229 radiology reports of 112,501 unique subjects. Pre-trained models outperformed random initialization across all diseases. As the training dataset size was reduced, performance was robust except for a few diseases with a relatively small number of cases. Pre-trained classification AUCs reached > 0.95 for all four disease outcomes and normality across all three organ systems.
Conclusions
Our label-extracting pipeline was able to encompass a variety of cases and diseases in body CT reports by generalizing beyond strict rules with exceptional accuracy. The method described can be easily adapted to enable automated labeling of hospital-scale medical data sets for training image-based disease classifiers.
Collapse
|
17
|
Zhu Z, Mittendorf A, Shropshire E, Allen B, Miller C, Bashir MR, Mazurowski MA. 3D Pyramid Pooling Network for Abdominal MRI Series Classification. IEEE Trans Pattern Anal Mach Intell 2022; 44:1688-1698. [PMID: 33112740 DOI: 10.1109/tpami.2020.3033990] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Recognizing and organizing different series in an MRI examination is important both for clinical review and research, but it is poorly addressed by the current generation of picture archiving and communication systems (PACSs) and post-processing workstations. In this paper, we study the problem of using deep convolutional neural networks for automatic classification of abdominal MRI series to one of many series types. Our contributions are three-fold. First, we created a large abdominal MRI dataset containing 3717 MRI series including 188,665 individual images, derived from liver examinations. 30 different series types are represented in this dataset. The dataset was annotated by consensus readings from two radiologists. Both the MRIs and the annotations were made publicly available. Second, we proposed a 3D pyramid pooling network, which can elegantly handle abdominal MRI series with varied sizes of each dimension, and achieved state-of-the-art classification performance. Third, we performed the first ever comparison between the algorithm and the radiologists on an additional dataset and had several meaningful findings.
Collapse
|
18
|
Hou R, Grimm LJ, Mazurowski MA, Marks JR, King LM, Maley CC, Lynch T, van Oirsouw M, Rogers K, Stone N, Wallis M, Teuwen J, Wesseling J, Hwang ES, Lo JY. Prediction of Upstaging in Ductal Carcinoma in Situ Based on Mammographic Radiomic Features. Radiology 2022; 303:54-62. [PMID: 34981975 DOI: 10.1148/radiol.210407] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2022]
Abstract
Background Improving diagnosis of ductal carcinoma in situ (DCIS) before surgery is important in choosing optimal patient management strategies. However, patients may harbor occult invasive disease not detected until definitive surgery. Purpose To assess the performance and clinical utility of mammographic radiomic features in the prediction of occult invasive cancer among women diagnosed with DCIS on the basis of core biopsy findings. Materials and Methods In this Health Insurance Portability and Accountability Act-compliant retrospective study, digital magnification mammographic images were collected from women who underwent breast core-needle biopsy for calcifications that was performed at a single institution between September 2008 and April 2017 and yielded a diagnosis of DCIS. The database query was directed at asymptomatic women with calcifications without a mass, architectural distortion, asymmetric density, or palpable disease. Logistic regression with regularization was used. Differences across training and internal test set by upstaging rate, age, lesion size, and estrogen and progesterone receptor status were assessed by using the Kruskal-Wallis or χ2 test. Results The study consisted of 700 women with DCIS (age range, 40-89 years; mean age, 59 years ± 10 [standard deviation]), including 114 with lesions (16.3%) upstaged to invasive cancer at subsequent surgery. The sample was split randomly into 400 women for the training set and 300 for the testing set (mean ages: training set, 59 years ± 10; test set, 59 years ± 10; P = .85). A total of 109 radiomic and four clinical features were extracted. The best model on the test set by using all radiomic and clinical features helped predict upstaging with an area under the receiver operating characteristic curve of 0.71 (95% CI: 0.62, 0.79). For a fixed high sensitivity (90%), the model yielded a specificity of 22%, a negative predictive value of 92%, and an odds ratio of 2.4 (95% CI: 1.8, 3.2). High specificity (90%) corresponded to a sensitivity of 37%, positive predictive value of 41%, and odds ratio of 5.0 (95% CI: 2.8, 9.0). Conclusion Machine learning models that use radiomic features applied to mammographic calcifications may help predict upstaging of ductal carcinoma in situ, which can refine clinical decision making and treatment planning. © RSNA, 2022.
Collapse
Affiliation(s)
- Rui Hou
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Lars J Grimm
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Maciej A Mazurowski
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Jeffrey R Marks
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Lorraine M King
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Carlo C Maley
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Thomas Lynch
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Marja van Oirsouw
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Keith Rogers
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Nicholas Stone
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Matthew Wallis
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Jonas Teuwen
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Jelle Wesseling
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - E Shelley Hwang
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| | - Joseph Y Lo
- From the Departments of Radiology (R.H., L.J.G., J.Y.L.) and Surgery (J.R.M., L.M.K., T.L., E.S.H.), Duke University Medical Center, Box 3513, Durham, NC 27710; Department of Electrical and Computer Engineering, Pratt School of Engineering, Duke University, Durham, NC (R.H.); School of Life Sciences, Arizona State University, Tempe, Ariz (C.C.M.); Cranfield Forensic Institute, Cranfield University, Cranfield, UK (K.R.); School of Physics and Astronomy, College of Engineering, Mathematics and Physical Sciences, Physics Building, Streatham Campus, University of Exeter, Exeter, UK (N.S.); Cambridge Breast Unit and NIHR Cambridge Biomedical Research Center, Cambridge University Hospitals NHS Trust, Cambridge Biomedical Campus, Cambridge, UK (M.W.); and Netherlands Cancer Institute, Amsterdam, the Netherlands (M.v.O., J.T., J.W.)
| |
Collapse
|
19
|
Tushar FI, D’Anniballe VM, Hou R, Mazurowski MA, Fu W, Samei E, Rubin GD, Lo JY. Classification of Multiple Diseases on Body CT Scans Using Weakly Supervised Deep Learning. Radiol Artif Intell 2022; 4:e210026. [PMID: 35146433 PMCID: PMC8823458 DOI: 10.1148/ryai.210026] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2021] [Revised: 10/04/2021] [Accepted: 11/15/2021] [Indexed: 04/14/2023]
Abstract
PURPOSE To design multidisease classifiers for body CT scans for three different organ systems using automatically extracted labels from radiology text reports. MATERIALS AND METHODS This retrospective study included a total of 12 092 patients (mean age, 57 years ± 18 [standard deviation]; 6172 women) for model development and testing. Rule-based algorithms were used to extract 19 225 disease labels from 13 667 body CT scans performed between 2012 and 2017. Using a three-dimensional DenseVNet, three organ systems were segmented: lungs and pleura, liver and gallbladder, and kidneys and ureters. For each organ system, a three-dimensional convolutional neural network classified each as no apparent disease or for the presence of four common diseases, for a total of 15 different labels across all three models. Testing was performed on a subset of 2158 CT volumes relative to 2875 manually derived reference labels from 2133 patients (mean age, 58 years ± 18; 1079 women). Performance was reported as area under the receiver operating characteristic curve (AUC), with 95% CIs calculated using the DeLong method. RESULTS Manual validation of the extracted labels confirmed 91%-99% accuracy across the 15 different labels. AUCs for lungs and pleura labels were as follows: atelectasis, 0.77 (95% CI: 0.74, 0.81); nodule, 0.65 (95% CI: 0.61, 0.69); emphysema, 0.89 (95% CI: 0.86, 0.92); effusion, 0.97 (95% CI: 0.96, 0.98); and no apparent disease, 0.89 (95% CI: 0.87, 0.91). AUCs for liver and gallbladder were as follows: hepatobiliary calcification, 0.62 (95% CI: 0.56, 0.67); lesion, 0.73 (95% CI: 0.69, 0.77); dilation, 0.87 (95% CI: 0.84, 0.90); fatty, 0.89 (95% CI: 0.86, 0.92); and no apparent disease, 0.82 (95% CI: 0.78, 0.85). AUCs for kidneys and ureters were as follows: stone, 0.83 (95% CI: 0.79, 0.87); atrophy, 0.92 (95% CI: 0.89, 0.94); lesion, 0.68 (95% CI: 0.64, 0.72); cyst, 0.70 (95% CI: 0.66, 0.73); and no apparent disease, 0.79 (95% CI: 0.75, 0.83). CONCLUSION Weakly supervised deep learning models were able to classify diverse diseases in multiple organ systems from CT scans.Keywords: CT, Diagnosis/Classification/Application Domain, Semisupervised Learning, Whole-Body Imaging© RSNA, 2022.
Collapse
|
20
|
Modanwal G, Vellal A, Mazurowski MA. Normalization of breast MRIs using cycle-consistent generative adversarial networks. Comput Methods Programs Biomed 2021; 208:106225. [PMID: 34198016 DOI: 10.1016/j.cmpb.2021.106225] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/25/2021] [Accepted: 05/29/2021] [Indexed: 06/13/2023]
Abstract
OBJECTIVES Dynamic Contrast Enhanced-Magnetic Resonance Imaging (DCE-MRI) is widely used to complement ultrasound examinations and x-ray mammography for early detection and diagnosis of breast cancer. However, images generated by various MRI scanners (e.g., GE Healthcare, and Siemens) differ both in intensity and noise distribution, preventing algorithms trained on MRIs from one scanner to generalize to data from other scanners. In this work, we propose a method to solve this problem by normalizing images between various scanners. METHODS MRI normalization is challenging because it requires normalizing intensity values and mapping noise distributions between scanners. We utilize a cycle-consistent generative adversarial network to learn a bidirectional mapping and perform normalization between MRIs produced by GE Healthcare and Siemens scanners in an unpaired setting. Initial experiments demonstrate that the traditional CycleGAN architecture struggles to preserve the anatomical structures of the breast during normalization. Thus, we propose two technical innovations in order to preserve both the shape of the breast as well as the tissue structures within the breast. First, we incorporate mutual information loss during training in order to ensure anatomical consistency. Second, we propose a modified discriminator architecture that utilizes a smaller field-of-view to ensure the preservation of finer details in the breast tissue. RESULTS Quantitative and qualitative evaluations show that the second innovation consistently preserves the breast shape and tissue structures while also performing the proper intensity normalization and noise distribution mapping. CONCLUSION Our results demonstrate that the proposed model can successfully learn a bidirectional mapping and perform normalization between MRIs produced by different vendors, potentially enabling improved diagnosis and detection of breast cancer. All the data used in this study are publicly available at https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=70226903.
Collapse
Affiliation(s)
| | - Adithya Vellal
- Department of Computer Science, Duke University, Durham, NC, USA
| | | |
Collapse
|
21
|
Buda M, Saha A, Walsh R, Ghate S, Li N, Święcicki A, Lo JY, Mazurowski MA. A Data Set and Deep Learning Algorithm for the Detection of Masses and Architectural Distortions in Digital Breast Tomosynthesis Images. JAMA Netw Open 2021; 4:e2119100. [PMID: 34398205 PMCID: PMC8369362 DOI: 10.1001/jamanetworkopen.2021.19100] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/15/2022] Open
Abstract
IMPORTANCE Breast cancer screening is among the most common radiological tasks, with more than 39 million examinations performed each year. While it has been among the most studied medical imaging applications of artificial intelligence, the development and evaluation of algorithms are hindered by the lack of well-annotated, large-scale publicly available data sets. OBJECTIVES To curate, annotate, and make publicly available a large-scale data set of digital breast tomosynthesis (DBT) images to facilitate the development and evaluation of artificial intelligence algorithms for breast cancer screening; to develop a baseline deep learning model for breast cancer detection; and to test this model using the data set to serve as a baseline for future research. DESIGN, SETTING, AND PARTICIPANTS In this diagnostic study, 16 802 DBT examinations with at least 1 reconstruction view available, performed between August 26, 2014, and January 29, 2018, were obtained from Duke Health System and analyzed. From the initial cohort, examinations were divided into 4 groups and split into training and test sets for the development and evaluation of a deep learning model. Images with foreign objects or spot compression views were excluded. Data analysis was conducted from January 2018 to October 2020. EXPOSURES Screening DBT. MAIN OUTCOMES AND MEASURES The detection algorithm was evaluated with breast-based free-response receiver operating characteristic curve and sensitivity at 2 false positives per volume. RESULTS The curated data set contained 22 032 reconstructed DBT volumes that belonged to 5610 studies from 5060 patients with a mean (SD) age of 55 (11) years and 5059 (100.0%) women. This included 4 groups of studies: (1) 5129 (91.4%) normal studies; (2) 280 (5.0%) actionable studies, for which where additional imaging was needed but no biopsy was performed; (3) 112 (2.0%) benign biopsied studies; and (4) 89 studies (1.6%) with cancer. Our data set included masses and architectural distortions that were annotated by 2 experienced radiologists. Our deep learning model reached breast-based sensitivity of 65% (39 of 60; 95% CI, 56%-74%) at 2 false positives per DBT volume on a test set of 460 examinations from 418 patients. CONCLUSIONS AND RELEVANCE The large, diverse, and curated data set presented in this study could facilitate the development and evaluation of artificial intelligence algorithms for breast cancer screening by providing data for training as well as a common set of cases for model validation. The performance of the model developed in this study showed that the task remains challenging; its performance could serve as a baseline for future model development.
Collapse
Affiliation(s)
- Mateusz Buda
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Ashirbani Saha
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Ruth Walsh
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Sujata Ghate
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Nianyi Li
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Albert Święcicki
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Joseph Y. Lo
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
| | - Maciej A. Mazurowski
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
- Department of Computer Science, Duke University, Durham, North Carolina
- Department of Biostatistics and Bioinformatics, Duke University Medical Center, Durham, North Carolina
| |
Collapse
|
22
|
Mazurowski MA. Do We Expect More from Radiology AI than from Radiologists? Radiol Artif Intell 2021; 3:e200221. [PMID: 34350411 PMCID: PMC8328102 DOI: 10.1148/ryai.2021200221] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2020] [Revised: 02/26/2021] [Accepted: 03/01/2021] [Indexed: 06/13/2023]
Abstract
The expectations of radiology artificial intelligence do not match expectations of radiologists in terms of performance and explainability.
Collapse
|
23
|
Swiecicki A, Konz N, Buda M, Mazurowski MA. A generative adversarial network-based abnormality detection using only normal images for model training with application to digital breast tomosynthesis. Sci Rep 2021; 11:10276. [PMID: 33986361 PMCID: PMC8119417 DOI: 10.1038/s41598-021-89626-1] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2020] [Accepted: 04/20/2021] [Indexed: 01/07/2023] Open
Abstract
Deep learning has shown tremendous potential in the task of object detection in images. However, a common challenge with this task is when only a limited number of images containing the object of interest are available. This is a particular issue in cancer screening, such as digital breast tomosynthesis (DBT), where less than 1% of cases contain cancer. In this study, we propose a method to train an inpainting generative adversarial network to be used for cancer detection using only images that do not contain cancer. During inference, we removed a part of the image and used the network to complete the removed part. A significant error in completing an image part was considered an indication that such location is unexpected and thus abnormal. A large dataset of DBT images used in this study was collected at Duke University. It consisted of 19,230 reconstructed volumes from 4348 patients. Cancerous masses and architectural distortions were marked with bounding boxes by radiologists. Our experiments showed that the locations containing cancer were associated with a notably higher completion error than the non-cancer locations (mean error ratio of 2.77). All data used in this study has been made publicly available by the authors.
Collapse
Affiliation(s)
- Albert Swiecicki
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA.
| | - Nicholas Konz
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA
| | - Mateusz Buda
- Department of Radiology, Duke University, Durham, NC, USA
| | - Maciej A Mazurowski
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA.,Department of Radiology, Duke University, Durham, NC, USA
| |
Collapse
|
24
|
Swiecicki A, Li N, O'Donnell J, Said N, Yang J, Mather RC, Jiranek WA, Mazurowski MA. Deep learning-based algorithm for assessment of knee osteoarthritis severity in radiographs matches performance of radiologists. Comput Biol Med 2021; 133:104334. [PMID: 33823398 DOI: 10.1016/j.compbiomed.2021.104334] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2020] [Revised: 03/09/2021] [Accepted: 03/10/2021] [Indexed: 11/26/2022]
Abstract
A fully-automated deep learning algorithm matched performance of radiologists in assessment of knee osteoarthritis severity in radiographs using the Kellgren-Lawrence grading system. PURPOSE To develop an automated deep learning-based algorithm that jointly uses Posterior-Anterior (PA) and Lateral (LAT) views of knee radiographs to assess knee osteoarthritis severity according to the Kellgren-Lawrence grading system. MATERIALS AND METHODS We used a dataset of 9739 exams from 2802 patients from Multicenter Osteoarthritis Study (MOST). The dataset was divided into a training set of 2040 patients, a validation set of 259 patients and a test set of 503 patients. A novel deep learning-based method was utilized for assessment of knee OA in two steps: (1) localization of knee joints in the images, (2) classification according to the KL grading system. Our method used both PA and LAT views as the input to the model. The scores generated by the algorithm were compared to the grades provided in the MOST dataset for the entire test set as well as grades provided by 5 radiologists at our institution for a subset of the test set. RESULTS The model obtained a multi-class accuracy of 71.90% on the entire test set when compared to the ratings provided in the MOST dataset. The quadratic weighted Kappa coefficient for this set was 0.9066. The average quadratic weighted Kappa between all pairs of radiologists from our institution who took part in the study was 0.748. The average quadratic-weighted Kappa between the algorithm and the radiologists at our institution was 0.769. CONCLUSION The proposed model performed demonstrated equivalency of KL classification to MSK radiologists, but clearly superior reproducibility. Our model also agreed with radiologists at our institution to the same extent as the radiologists with each other. The algorithm could be used to provide reproducible assessment of knee osteoarthritis severity.
Collapse
Affiliation(s)
- Albert Swiecicki
- Department of Electrical and Computer Engineering, Duke University, Durham, USA
| | - Nianyi Li
- Department of Electrical and Computer Engineering, Duke University, Durham, USA
| | | | - Nicholas Said
- Department of Radiology, Duke University, Durham, USA
| | - Jichen Yang
- Department of Electrical and Computer Engineering, Duke University, Durham, USA.
| | | | | | - Maciej A Mazurowski
- Department of Electrical and Computer Engineering, Duke University, Durham, USA; Department of Radiology, Duke University, Durham, USA
| |
Collapse
|
25
|
Draelos RL, Dov D, Mazurowski MA, Lo JY, Henao R, Rubin GD, Carin L. Machine-learning-based multiple abnormality prediction with large-scale chest computed tomography volumes. Med Image Anal 2020; 67:101857. [PMID: 33129142 DOI: 10.1016/j.media.2020.101857] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2020] [Revised: 09/15/2020] [Accepted: 09/18/2020] [Indexed: 12/11/2022]
Abstract
Machine learning models for radiology benefit from large-scale data sets with high quality labels for abnormalities. We curated and analyzed a chest computed tomography (CT) data set of 36,316 volumes from 19,993 unique patients. This is the largest multiply-annotated volumetric medical imaging data set reported. To annotate this data set, we developed a rule-based method for automatically extracting abnormality labels from free-text radiology reports with an average F-score of 0.976 (min 0.941, max 1.0). We also developed a model for multi-organ, multi-disease classification of chest CT volumes that uses a deep convolutional neural network (CNN). This model reached a classification performance of AUROC >0.90 for 18 abnormalities, with an average AUROC of 0.773 for all 83 abnormalities, demonstrating the feasibility of learning from unfiltered whole volume CT data. We show that training on more labels improves performance significantly: for a subset of 9 labels - nodule, opacity, atelectasis, pleural effusion, consolidation, mass, pericardial effusion, cardiomegaly, and pneumothorax - the model's average AUROC increased by 10% when the number of training labels was increased from 9 to all 83. All code for volume preprocessing, automated label extraction, and the volume abnormality prediction model is publicly available. The 36,316 CT volumes and labels will also be made publicly available pending institutional approval.
Collapse
Affiliation(s)
- Rachel Lea Draelos
- Computer Science Department, Duke University, LSRC Building D101, 308 Research Drive, Duke Box 90129, Durham, North Carolina 27708-0129, United States of America; School of Medicine, Duke University, DUMC 3710, Durham, North Carolina 27710, United States of America.
| | - David Dov
- Electrical and Computer Engineering Department, Edmund T. Pratt Jr. School of Engineering, Duke University, Box 90291, Durham, North Carolina 27708, United States of America
| | - Maciej A Mazurowski
- Electrical and Computer Engineering Department, Edmund T. Pratt Jr. School of Engineering, Duke University, Box 90291, Durham, North Carolina 27708, United States of America; Radiology Department, Duke University, Box 3808 DUMC, Durham, North Carolina 27710, United States of America; Biostatistics and Bioinformatics Department, Duke University, DUMC 2424 Erwin Road, Suite 1102 Hock Plaza, Box 2721 Durham, North Carolina 27710, United States of America
| | - Joseph Y Lo
- Electrical and Computer Engineering Department, Edmund T. Pratt Jr. School of Engineering, Duke University, Box 90291, Durham, North Carolina 27708, United States of America; Radiology Department, Duke University, Box 3808 DUMC, Durham, North Carolina 27710, United States of America; Biomedical Engineering Department, Edmund T. Pratt Jr. School of Engineering, Duke University, Room 1427, Fitzpatrick Center (FCIEMAS), 101 Science Drive, Campus Box 90281, Durham, North Carolina 27708-0281, United States of America
| | - Ricardo Henao
- Electrical and Computer Engineering Department, Edmund T. Pratt Jr. School of Engineering, Duke University, Box 90291, Durham, North Carolina 27708, United States of America; Biostatistics and Bioinformatics Department, Duke University, DUMC 2424 Erwin Road, Suite 1102 Hock Plaza, Box 2721 Durham, North Carolina 27710, United States of America
| | - Geoffrey D Rubin
- Radiology Department, Duke University, Box 3808 DUMC, Durham, North Carolina 27710, United States of America
| | - Lawrence Carin
- Computer Science Department, Duke University, LSRC Building D101, 308 Research Drive, Duke Box 90129, Durham, North Carolina 27708-0129, United States of America; Electrical and Computer Engineering Department, Edmund T. Pratt Jr. School of Engineering, Duke University, Box 90291, Durham, North Carolina 27708, United States of America; Statistical Science Department, Duke University, Box 90251, Durham, North Carolina 27708-0251, United States of America
| |
Collapse
|
26
|
Wildman-Tobriner B, Ahmed S, Erkanli A, Mazurowski MA, Hoang JK. Using the American College of Radiology Thyroid Imaging Reporting and Data System at the Point of Care: Sonographer Performance and Interobserver Variability. Ultrasound Med Biol 2020; 46:1928-1933. [PMID: 32507343 DOI: 10.1016/j.ultrasmedbio.2020.04.019] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/04/2020] [Revised: 04/08/2020] [Accepted: 04/16/2020] [Indexed: 06/11/2023]
Abstract
The purpose of this study was to assess inter-observer variability and performance when sonographers assign features to thyroid nodules on ultrasound using the American College of Radiology (ACR) Thyroid Imaging Reporting and Data System (TI-RADS). Fifteen sonographers retrospectively evaluated 100 thyroid nodules and assigned features to each nodule according to ACR TI-RADS lexicon. Ratings were compared with one another and to a gold standard using Fleiss' and Cohen's kappa statistics, respectively. Sonographers were also asked subjective questions regarding their comfort level assessing each feature, and opinions were compared with performance using a mixed effects model. Sonographers demonstrated only slight agreement for margin (κ = 0.18, 95% confidence interval [CI]: 0.16-0.20) and large comet tail artifact (κ = 0.08, 95% CI: 0.06-0.10) but better performance for macrocalcification (κ = 0.41, 95% CI: 0.39-0.43) and no echogenic foci (κ = 0.52, 95% CI: 0.50-0.54). Sonographer comfort level with different feature assignments did not statistically correlate with performance for a given feature. In conclusion, sonographers using ACR TI-RADS to assign thyroid nodule features on ultrasound demonstrate a range of agreement across features, with margin and large comet tail artifact showing the most variability. These results highlight potential areas of focus for sonographer education efforts as ACR TI-RADS continues to be implemented in radiology departments.
Collapse
Affiliation(s)
| | - Salmaan Ahmed
- Department of Radiology, MD Anderson Cancer Center, Neuroradiology Department, The University of Texas, Houston, Texas, USA
| | - Al Erkanli
- Department of Biostatistics and Bioinformatics, Duke University Hospital, Durham, North Carolina, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University Hospital, Durham, North Carolina, USA
| | - Jenny K Hoang
- Russell H. Morgan Department of Radiology and Radiological Science, Johns Hopkins Health System, Baltimore, Maryland, USA
| |
Collapse
|
27
|
Buda M, Wildman-Tobriner B, Castor K, Hoang JK, Mazurowski MA. Deep Learning-Based Segmentation of Nodules in Thyroid Ultrasound: Improving Performance by Utilizing Markers Present in the Images. Ultrasound Med Biol 2020; 46:415-421. [PMID: 31699547 DOI: 10.1016/j.ultrasmedbio.2019.10.003] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/17/2019] [Revised: 09/29/2019] [Accepted: 10/03/2019] [Indexed: 06/10/2023]
Abstract
Computer-aided segmentation of thyroid nodules in ultrasound imaging could assist in their accurate characterization. In this study, using data for 1278 nodules, we proposed and evaluated two methods for deep learning-based segmentation of thyroid nodules that utilize calipers present in the images. The first method used approximate nodule masks generated based on the calipers. The second method combined manual annotations with automatic guidance by the calipers. When only approximate nodule masks were used for training, the achieved Dice similarity coefficient (DSC) was 85.1%. The performance of a network trained using manual annotations was DSC = 90.4%. When the guidance by the calipers was added, the performance increased to DSC = 93.1%. An increase in the number of cases used for training resulted in increased performance for all methods. The proposed method utilizing the guidance by calipers matched the performance of the network that did not use it with a reduced number of manually annotated training cases.
Collapse
Affiliation(s)
- Mateusz Buda
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA.
| | | | - Kerry Castor
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina, USA
| | - Jenny K Hoang
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA; Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina, USA
| |
Collapse
|
28
|
Grimm LJ, Mazurowski MA. Breast Cancer Radiogenomics: Current Status and Future Directions. Acad Radiol 2020; 27:39-46. [PMID: 31818385 DOI: 10.1016/j.acra.2019.09.012] [Citation(s) in RCA: 33] [Impact Index Per Article: 8.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2019] [Revised: 07/17/2019] [Accepted: 09/08/2019] [Indexed: 12/13/2022]
Abstract
Radiogenomics is an area of research that aims to identify associations between imaging phenotypes ("radio-") and tumor genome ("-genomics"). Breast cancer radiogenomics research in particular has been an especially prolific area of investigation in recent years as evidenced by the wide number and variety of publications and conferences presentations. To date, research has primarily been focused on dynamic contrast enhanced pre-operative breast MRI and breast cancer molecular subtypes, but investigations have extended to all breast imaging modalities as well as multiple additional genetic markers including those that are commercially available. Furthermore, both human and computer-extracted features as well as deep learning techniques have been explored. This review will summarize the specific imaging modalities used in radiogenomics analysis, describe the methods of extracting imaging features, and present the types of genomics, molecular, and related information used for analysis. Finally, the limitations and future directions of breast cancer radiogenomics research will be discussed.
Collapse
|
29
|
Buda M, AlBadawy EA, Saha A, Mazurowski MA. Deep Radiogenomics of Lower-Grade Gliomas: Convolutional Neural Networks Predict Tumor Genomic Subtypes Using MR Images. Radiol Artif Intell 2020; 2:e180050. [PMID: 33937809 DOI: 10.1148/ryai.2019180050] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2018] [Revised: 08/06/2019] [Accepted: 08/30/2019] [Indexed: 11/11/2022]
Abstract
PURPOSE To employ deep learning to predict genomic subtypes of lower-grade glioma (LLG) tumors based on their appearance at MRI. MATERIALS AND METHODS Imaging data from The Cancer Imaging Archive and genomic data from The Cancer Genome Atlas from 110 patients from five institutions with lower-grade gliomas (World Health Organization grade II and III) were used in this study. A convolutional neural network was trained to predict tumor genomic subtype based on the MRI of the tumor. Two different deep learning approaches were tested: training from random initialization and transfer learning. Deep learning models were pretrained on glioblastoma MRI, instead of natural images, to determine if performance was improved for the detection of LGGs. The models were evaluated using area under the receiver operating characteristic curve (AUC) with cross-validation. Imaging data and annotations used in this study are publicly available. RESULTS The best performing model was based on transfer learning from glioblastoma MRI. It achieved AUC of 0.730 (95% confidence interval [CI]: 0.605, 0.844) for discriminating cluster-of-clusters 2 from others. For the same task, a network trained from scratch achieved an AUC of 0.680 (95% CI: 0.538, 0.811), whereas a model pretrained on natural images achieved an AUC of 0.640 (95% CI: 0.521, 0.763). CONCLUSION These findings show the potential of utilizing deep learning to identify relationships between cancer imaging and cancer genomics in LGGs. However, more accurate models are needed to justify clinical use of such tools, which might be obtained using substantially larger training datasets.Supplemental material is available for this article.© RSNA, 2020.
Collapse
Affiliation(s)
- Mateusz Buda
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Rd, Suite 302, Durham, NC 27705 (M.B., E.A.A., A.S., M.A.M.); Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.); and Department of Biostatistics and Bioinformatics, Duke University, Durham, NC (M.A.M.)
| | - Ehab A AlBadawy
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Rd, Suite 302, Durham, NC 27705 (M.B., E.A.A., A.S., M.A.M.); Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.); and Department of Biostatistics and Bioinformatics, Duke University, Durham, NC (M.A.M.)
| | - Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Rd, Suite 302, Durham, NC 27705 (M.B., E.A.A., A.S., M.A.M.); Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.); and Department of Biostatistics and Bioinformatics, Duke University, Durham, NC (M.A.M.)
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Rd, Suite 302, Durham, NC 27705 (M.B., E.A.A., A.S., M.A.M.); Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.); and Department of Biostatistics and Bioinformatics, Duke University, Durham, NC (M.A.M.)
| |
Collapse
|
30
|
Mazurowski MA. Artificial Intelligence in Radiology: Some Ethical Considerations for Radiologists and Algorithm Developers. Acad Radiol 2020; 27:127-129. [PMID: 31818378 DOI: 10.1016/j.acra.2019.04.024] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2019] [Accepted: 04/21/2019] [Indexed: 11/25/2022]
Abstract
As artificial intelligence (AI) is finding its place in radiology, it is important to consider how to guide the research and clinical implementation in a way that will be most beneficial to patients. Although there are multiple aspects of this issue, I consider a specific one: a potential misalignment of the self-interests of radiologists and AI developers with the best interests of the patients. Radiologists know that supporting research into AI and advocating for its adoption in clinical settings could diminish their employment opportunities and reduce respect for their profession. This provides an incentive to oppose AI in various ways. AI developers have an incentive to hype their discoveries to gain attention. This could provide short-term personal gains, however, it could also create a distrust toward the field if it became apparent that the state of the art was far from where it was promised to be. The future research and clinical implementation of AI in radiology will be partially determined by radiologist and AI researchers. Therefore, it is very important that we recognize our own personal motivations and biases and act responsibly to ensure the highest benefit of the AI transformation to the patients.
Collapse
|
31
|
Hou R, Mazurowski MA, Grimm LJ, Marks JR, King LM, Maley CC, Hwang ESS, Lo JY. Prediction of Upstaged Ductal Carcinoma In Situ Using Forced Labeling and Domain Adaptation. IEEE Trans Biomed Eng 2019; 67:1565-1572. [PMID: 31502960 DOI: 10.1109/tbme.2019.2940195] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
OBJECTIVE The goal of this study is to use adjunctive classes to improve a predictive model whose performance is limited by the common problems of small numbers of primary cases, high feature dimensionality, and poor class separability. Specifically, our clinical task is to use mammographic features to predict whether ductal carcinoma in situ (DCIS) identified at needle core biopsy will be later upstaged or shown to contain invasive breast cancer. METHODS To improve the prediction of pure DCIS (negative) versus upstaged DCIS (positive) cases, this study considers the adjunctive roles of two related classes: atypical ductal hyperplasia (ADH), a non-cancer type of breast abnormity, and invasive ductal carcinoma (IDC), with 113 computer vision based mammographic features extracted from each case. To improve the baseline Model A's classification of pure vs. upstaged DCIS, we designed three different strategies (Models B, C, D) with different ways of embedding features or inputs. RESULTS Based on ROC analysis, the baseline Model A performed with AUC of 0.614 (95% CI, 0.496-0.733). All three new models performed better than the baseline, with domain adaptation (Model D) performing the best with an AUC of 0.697 (95% CI, 0.595-0.797). CONCLUSION We improved the prediction performance of DCIS upstaging by embedding two related pathology classes in different training phases. SIGNIFICANCE The three new strategies of embedding related class data all outperformed the baseline model, thus demonstrating not only feature similarities among these different classes, but also the potential for improving classification by using other related classes.
Collapse
|
32
|
Buda M, Wildman-Tobriner B, Hoang JK, Thayer D, Tessler FN, Middleton WD, Mazurowski MA. Management of Thyroid Nodules Seen on US Images: Deep Learning May Match Performance of Radiologists. Radiology 2019; 292:695-701. [PMID: 31287391 DOI: 10.1148/radiol.2019181343] [Citation(s) in RCA: 107] [Impact Index Per Article: 21.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/24/2023]
Abstract
BackgroundManagement of thyroid nodules may be inconsistent between different observers and time consuming for radiologists. An artificial intelligence system that uses deep learning may improve radiology workflow for management of thyroid nodules.PurposeTo develop a deep learning algorithm that uses thyroid US images to decide whether a thyroid nodule should undergo a biopsy and to compare the performance of the algorithm with the performance of radiologists who adhere to American College of Radiology (ACR) Thyroid Imaging Reporting and Data System (TI-RADS).Materials and MethodsIn this retrospective analysis, studies in patients referred for US with subsequent fine-needle aspiration or with surgical histologic analysis used as the standard were evaluated. The study period was from August 2006 to May 2010. A multitask deep convolutional neural network was trained to provide biopsy recommendations for thyroid nodules on the basis of two orthogonal US images as the input. In the training phase, the deep learning algorithm was first evaluated by using 10-fold cross-validation. Internal validation was then performed on an independent set of 99 consecutive nodules. The sensitivity and specificity of the algorithm were compared with a consensus of three ACR TI-RADS committee experts and nine other radiologists, all of whom interpreted thyroid US images in clinical practice.ResultsIncluded were 1377 thyroid nodules in 1230 patients with complete imaging data and conclusive cytologic or histologic diagnoses. For the 99 test nodules, the proposed deep learning algorithm achieved 13 of 15 (87%: 95% confidence interval [CI]: 67%, 100%) sensitivity, the same as expert consensus (P > .99) and higher than five of nine radiologists. The specificity of the deep learning algorithm was 44 of 84 (52%; 95% CI: 42%, 62%), which was similar to expert consensus (43 of 84; 51%; 95% CI: 41%, 62%; P = .91) and higher than seven of nine other radiologists. The mean sensitivity and specificity for the nine radiologists was 83% (95% CI: 64%, 98%) and 48% (95% CI: 37%, 59%), respectively.ConclusionSensitivity and specificity of a deep learning algorithm for thyroid nodule biopsy recommendations was similar to that of expert radiologists who used American College of Radiology Thyroid Imaging and Reporting Data System guidelines.© RSNA, 2019Online supplemental material is available for this article.
Collapse
Affiliation(s)
- Mateusz Buda
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| | - Benjamin Wildman-Tobriner
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| | - Jenny K Hoang
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| | - David Thayer
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| | - Franklin N Tessler
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| | - William D Middleton
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| | - Maciej A Mazurowski
- From the Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC 27705 (M.B., B.W.T., J.K.H., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (D.T., W.D.M.); Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.); and Department of Electrical and Computer Engineering, Duke University, Durham, NC (M.A.M.)
| |
Collapse
|
33
|
Mazurowski MA. Artificial Intelligence May Cause a Significant Disruption to the Radiology Workforce. J Am Coll Radiol 2019; 16:1077-1082. [DOI: 10.1016/j.jacr.2019.01.026] [Citation(s) in RCA: 42] [Impact Index Per Article: 8.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2018] [Revised: 01/23/2019] [Accepted: 01/24/2019] [Indexed: 11/24/2022]
|
34
|
Wildman-Tobriner B, Buda M, Hoang JK, Middleton WD, Thayer D, Short RG, Tessler FN, Mazurowski MA. Using Artificial Intelligence to Revise ACR TI-RADS Risk Stratification of Thyroid Nodules: Diagnostic Accuracy and Utility. Radiology 2019; 292:112-119. [PMID: 31112088 DOI: 10.1148/radiol.2019182128] [Citation(s) in RCA: 69] [Impact Index Per Article: 13.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
Background Risk stratification systems for thyroid nodules are often complicated and affected by low specificity. Continual improvement of these systems is necessary to reduce the number of unnecessary thyroid biopsies. Purpose To use artificial intelligence (AI) to optimize the American College of Radiology (ACR) Thyroid Imaging Reporting and Data System (TI-RADS). Materials and Methods A total of 1425 biopsy-proven thyroid nodules from 1264 consecutive patients (1026 women; mean age, 52.9 years [range, 18-93 years]) were evaluated retrospectively. Expert readers assigned points based on five ACR TI-RADS categories (composition, echogenicity, shape, margin, echogenic foci), and a genetic AI algorithm was applied to a training set (1325 nodules). Point and pathologic data were used to create an optimized scoring system (hereafter, AI TI-RADS). Performance of the systems was compared by using a test set of the final 100 nodules with interpretations from the expert reader, eight nonexpert readers, and an expert panel. Initial performance of AI TI-RADS was calculated by using a test for differences between binomial proportions. Additional comparisons across readers were conducted by using bootstrapping; diagnostic performance was assessed by using area under the receiver operating curve. Results AI TI-RADS assigned new point values for eight ACR TI-RADS features. Six features were assigned zero points, which simplified categorization. By using expert reader data, the diagnostic performance of ACR TI-RADS and AI TI-RADS was area under the receiver operating curve of 0.91 and 0.93, respectively. For the same expert, specificity of AI TI-RADS (65%, 55 of 85) was higher (P < .001) than that of ACR TI-RADS (47%, 40 of 85). For the eight nonexpert radiologists, mean specificity for AI TI-RADS (55%) was also higher (P < .001) than that of ACR TI-RADS (48%). An interactive AI TI-RADS calculator can be viewed at http://deckard.duhs.duke.edu/∼ai-ti-rads . Conclusion An artificial intelligence-optimized Thyroid Imaging Reporting and Data System (TI-RADS) validates the American College of Radiology TI-RADS while slightly improving specificity and maintaining sensitivity. Additionally, it simplifies feature assignments, which may improve ease of use. © RSNA, 2019 Online supplemental material is available for this article.
Collapse
Affiliation(s)
- Benjamin Wildman-Tobriner
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - Mateusz Buda
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - Jenny K Hoang
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - William D Middleton
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - David Thayer
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - Ryan G Short
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - Franklin N Tessler
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| | - Maciej A Mazurowski
- From the Department of Radiology, Duke University Hospital, 2301 Erwin Rd, Durham, NC 27701 (B.W.T., M.B., J.K.H., R.G.S., M.A.M.); Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Mo (W.D.M., D.T.); and Department of Radiology, University of Alabama at Birmingham, Birmingham, Ala (F.N.T.)
| |
Collapse
|
35
|
Buda M, Saha A, Mazurowski MA. Association of genomic subtypes of lower-grade gliomas with shape features automatically extracted by a deep learning algorithm. Comput Biol Med 2019; 109:218-225. [PMID: 31078126 DOI: 10.1016/j.compbiomed.2019.05.002] [Citation(s) in RCA: 68] [Impact Index Per Article: 13.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2019] [Revised: 04/25/2019] [Accepted: 05/01/2019] [Indexed: 01/22/2023]
Abstract
Recent analysis identified distinct genomic subtypes of lower-grade glioma tumors which are associated with shape features. In this study, we propose a fully automatic way to quantify tumor imaging characteristics using deep learning-based segmentation and test whether these characteristics are predictive of tumor genomic subtypes. We used preoperative imaging and genomic data of 110 patients from 5 institutions with lower-grade gliomas from The Cancer Genome Atlas. Based on automatic deep learning segmentations, we extracted three features which quantify two-dimensional and three-dimensional characteristics of the tumors. Genomic data for the analyzed cohort of patients consisted of previously identified genomic clusters based on IDH mutation and 1p/19q co-deletion, DNA methylation, gene expression, DNA copy number, and microRNA expression. To analyze the relationship between the imaging features and genomic clusters, we conducted the Fisher exact test for 10 hypotheses for each pair of imaging feature and genomic subtype. To account for multiple hypothesis testing, we applied a Bonferroni correction. P-values lower than 0.005 were considered statistically significant. We found the strongest association between RNASeq clusters and the bounding ellipsoid volume ratio (p < 0.0002) and between RNASeq clusters and margin fluctuation (p < 0.005). In addition, we identified associations between bounding ellipsoid volume ratio and all tested molecular subtypes (p < 0.02) as well as between angular standard deviation and RNASeq cluster (p < 0.02). In terms of automatic tumor segmentation that was used to generate the quantitative image characteristics, our deep learning algorithm achieved a mean Dice coefficient of 82% which is comparable to human performance.
Collapse
Affiliation(s)
- Mateusz Buda
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA.
| | - Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA.
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA; Department of Electrical and Computer Engineering, Duke University, Box 90291, Durham, NC, 27708, USA.
| |
Collapse
|
36
|
Zhu Z, Albadawy E, Saha A, Zhang J, Harowicz MR, Mazurowski MA. Deep learning for identifying radiogenomic associations in breast cancer. Comput Biol Med 2019; 109:85-90. [PMID: 31048129 DOI: 10.1016/j.compbiomed.2019.04.018] [Citation(s) in RCA: 66] [Impact Index Per Article: 13.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2018] [Revised: 03/05/2019] [Accepted: 04/20/2019] [Indexed: 11/27/2022]
Abstract
RATIONALE AND OBJECTIVES To determine whether deep learning models can distinguish between breast cancer molecular subtypes based on dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI). MATERIALS AND METHODS In this institutional review board-approved single-center study, we analyzed DCE-MR images of 270 patients at our institution. Lesions of interest were identified by radiologists. The task was to automatically determine whether the tumor is of the Luminal A subtype or of another subtype based on the MR image patches representing the tumor. Three different deep learning approaches were used to classify the tumor according to their molecular subtypes: learning from scratch where only tumor patches were used for training, transfer learning where networks pre-trained on natural images were fine-tuned using tumor patches, and off-the-shelf deep features where the features extracted by neural networks trained on natural images were used for classification with a support vector machine. Network architectures utilized in our experiments were GoogleNet, VGG, and CIFAR. We used 10-fold crossvalidation method for validation and area under the receiver operating characteristic (AUC) as the measure of performance. RESULTS The best AUC performance for distinguishing molecular subtypes was 0.65 (95% CI:[0.57,0.71]) and was achieved by the off-the-shelf deep features approach. The highest AUC performance for training from scratch was 0.58 (95% CI:[0.51,0.64]) and the best AUC performance for transfer learning was 0.60 (95% CI:[0.52,0.65]) respectively. For the off-the-shelf approach, the features extracted from the fully connected layer performed the best. CONCLUSION Deep learning may play a role in discovering radiogenomic associations in breast cancer.
Collapse
Affiliation(s)
- Zhe Zhu
- Department of Radiology, Duke University, USA.
| | | | | | - Jun Zhang
- Department of Radiology, Duke University, USA.
| | | | - Maciej A Mazurowski
- Department of Radiology and Department of Electrical and Computer Engineering, Duke University, USA.
| |
Collapse
|
37
|
Mazurowski MA, Buda M, Saha A, Bashir MR. Deep learning in radiology: An overview of the concepts and a survey of the state of the art with focus on MRI. J Magn Reson Imaging 2019; 49:939-954. [PMID: 30575178 PMCID: PMC6483404 DOI: 10.1002/jmri.26534] [Citation(s) in RCA: 202] [Impact Index Per Article: 40.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2018] [Revised: 09/14/2018] [Accepted: 09/17/2018] [Indexed: 12/15/2022] Open
Abstract
Deep learning is a branch of artificial intelligence where networks of simple interconnected units are used to extract patterns from data in order to solve complex problems. Deep-learning algorithms have shown groundbreaking performance in a variety of sophisticated tasks, especially those related to images. They have often matched or exceeded human performance. Since the medical field of radiology mainly relies on extracting useful information from images, it is a very natural application area for deep learning, and research in this area has rapidly grown in recent years. In this article, we discuss the general context of radiology and opportunities for application of deep-learning algorithms. We also introduce basic concepts of deep learning, including convolutional neural networks. Then, we present a survey of the research in deep learning applied to radiology. We organize the studies by the types of specific tasks that they attempt to solve and review a broad range of deep-learning algorithms being utilized. Finally, we briefly discuss opportunities and challenges for incorporating deep learning in the radiology practice of the future. Level of Evidence: 3 Technical Efficacy: Stage 1 J. Magn. Reson. Imaging 2019;49:939-954.
Collapse
Affiliation(s)
- Maciej A. Mazurowski
- Department of Radiology, Duke University, Durham, NC
- Department of Electrical and Computer Engineering, Duke University, Durham, NC
- Duke Medical Physics Program, Duke University, Durham, NC
| | - Mateusz Buda
- Department of Radiology, Duke University, Durham, NC
| | | | - Mustafa R. Bashir
- Department of Radiology, Duke University, Durham, NC
- Center for Advanced Magnetic Resonance Development, Duke University, Durham, NC
| |
Collapse
|
38
|
Zhang J, Saha A, Zhu Z, Mazurowski MA. Hierarchical Convolutional Neural Networks for Segmentation of Breast Tumors in MRI With Application to Radiogenomics. IEEE Trans Med Imaging 2019; 38:435-447. [PMID: 30130181 DOI: 10.1109/tmi.2018.2865671] [Citation(s) in RCA: 64] [Impact Index Per Article: 12.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Breast tumor segmentation based on dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI) is a challenging problem and an active area of research. Particular challenges, similarly as in other segmentation problems, include the class-imbalance problem as well as confounding background in DCE-MR images. To address these issues, we propose a mask-guided hierarchical learning (MHL) framework for breast tumor segmentation via fully convolutional networks (FCN). Specifically, we first develop an FCN model to generate a 3D breast mask as the region of interest (ROI) for each image, to remove confounding information from input DCE-MR images. We then design a two-stage FCN model to perform coarse-to-fine segmentation for breast tumors. Particularly, we propose a Dice-Sensitivity-like loss function and a reinforcement sampling strategy to handle the class-imbalance problem. To precisely identify locations of tumors that underwent a biopsy, we further propose an FCN model to detect two landmarks located at two nipples. We finally selected the biopsied tumor based on both identified landmarks and segmentations. We validate our MHL method on 272 patients, achieving a mean Dice similarity coefficient (DSC) of 0.72 which is comparable to mutual DSC between expert radiologists. Using the segmented biopsied tumors, we also demonstrate that the automatically generated masks can be applied to radiogenomics and can identify luminal A subtype from other molecular subtypes with the similar accuracy with the analysis based on semi-manual tumor segmentation.
Collapse
|
39
|
Mazurowski MA, Saha A, Harowicz MR, Cain EH, Marks JR, Marcom PK. Association of distant recurrence-free survival with algorithmically extracted MRI characteristics in breast cancer. J Magn Reson Imaging 2019; 49:e231-e240. [PMID: 30672045 DOI: 10.1002/jmri.26648] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2018] [Revised: 12/23/2018] [Accepted: 12/26/2018] [Indexed: 01/01/2023] Open
Abstract
BACKGROUND While important in diagnosis of breast cancer, the scientific assessment of the role of imaging in prognosis of outcomes and treatment planning is limited. PURPOSE To evaluate the potential of using quantitative imaging variables for stratifying risk of distant recurrence in breast cancer patients. STUDY TYPE Retrospective. POPULATION In all, 892 female invasive breast cancer patients. SEQUENCE Dynamic contrast-enhanced MRI with field strength 1.5 T and 3 T. ASSESSMENT Computer vision algorithms were applied to extract a comprehensive set of 529 imaging features quantifying size, shape, enhancement patterns, and heterogeneity of the tumors and the surrounding tissue. Using a development set with 446 cases, we selected 20 imaging features with high prognostic value. STATISTICAL TESTS We evaluated the imaging features using an independent test set with 446 cases. The principal statistical measure was a concordance index between individual imaging features and patient distant recurrence-free survival (DRFS). RESULTS The strongest association with DRFS that persisted after controlling for known prognostic clinical and pathology variables was found for signal enhancement ratio (SER) partial tumor volume (concordance index [C] = 0.768, 95% confidence interval [CI]: 0.679-0.856), tumor major axis length (C = 0.742, 95% CI: 0.650-0.834), kurtosis of the SER map within tumor (C = 0.640, 95% CI: 0.521-0.760), tumor cluster shade (C = 0.313, 95% CI: 0.216-0.410), and washin rate information measure of correlation (C = 0.702, 95% CI: 0.601-0.803). DATA CONCLUSION Quantitative assessment of breast cancer features seen in a routine breast MRI might be able to be used for assessment of risk of distant recurrence. LEVEL OF EVIDENCE 4 Technical Efficacy: Stage 6 J. Magn. Reson. Imaging 2019.
Collapse
Affiliation(s)
- Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Michael R Harowicz
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA.,Department of Radiology, Johns Hopkins University, Baltimore, Maryland, USA
| | - Elizabeth Hope Cain
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Jeffrey R Marks
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA.,Department of Surgery, Duke University School of Medicine, Durham, North Carolina, USA
| | - P Kelly Marcom
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA.,Department of Medicine, Duke University School of Medicine, Durham, North Carolina, USA
| |
Collapse
|
40
|
Saha A, Grimm LJ, Ghate SV, Kim CE, Soo MS, Yoon SC, Mazurowski MA. Machine learning-based prediction of future breast cancer using algorithmically measured background parenchymal enhancement on high-risk screening MRI. J Magn Reson Imaging 2019; 50:456-464. [PMID: 30648316 DOI: 10.1002/jmri.26636] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2018] [Revised: 12/16/2018] [Accepted: 12/18/2018] [Indexed: 12/14/2022] Open
Abstract
BACKGROUND Preliminary work has demonstrated that background parenchymal enhancement (BPE) assessed by radiologists is predictive of future breast cancer in women undergoing high-risk screening MRI. Algorithmically assessed measures of BPE offer a more precise and reproducible means of measuring BPE than human readers and thus might improve the predictive performance of future cancer development. PURPOSE To determine if algorithmically extracted imaging features of BPE on screening breast MRI in high-risk women are associated with subsequent development of cancer. STUDY TYPE Case-control study. POPULATION In all, 133 women at high risk for developing breast cancer; 46 of these patients developed breast cancer subsequently over a follow-up period of 2 years. FIELD STRENGTH/SEQUENCE 5 T or 3.0 T T1 -weighted precontrast fat-saturated and nonfat-saturated sequences and postcontrast nonfat-saturated sequences. ASSESSMENT Automatic features of BPE were extracted with a computer algorithm. Subjective BPE scores from five breast radiologists (blinded to clinical outcomes) were also available. STATISTICAL TESTS Leave-one-out crossvalidation for a multivariate logistic regression model developed using the automatic features and receiver operating characteristic (ROC) analysis were performed to calculate the area under the curve (AUC). Comparison of automatic features and subjective features was performed using a generalized regression model and the P-value was obtained. Odds ratios for automatic and subjective features were compared. RESULTS The multivariate model discriminated patients who developed cancer from the patients who did not, with an AUC of 0.70 (95% confidence interval: 0.60-0.79, P < 0.001). The imaging features remained independently predictive of subsequent development of cancer (P < 0.003) when compared with the subjective BPE assessment of the readers. DATA CONCLUSION Automatically extracted BPE measurements may potentially be used to further stratify risk in patients undergoing high-risk screening MRI. LEVEL OF EVIDENCE 3 Technical Efficacy: Stage 5 J. Magn. Reson. Imaging 2019;50:456-464.
Collapse
Affiliation(s)
- Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Lars J Grimm
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Sujata V Ghate
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Connie E Kim
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Mary S Soo
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Sora C Yoon
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina, USA.,Department of Electrical and Computer Eng., Duke University, Durham, North Carolina, USA.,Duke University Medical Physics Program, Durham, North Carolina, USA
| |
Collapse
|
41
|
Grimm LJ, Saha A, Ghate SV, Kim C, Soo MS, Yoon SC, Mazurowski MA. Relationship between Background Parenchymal Enhancement on High-risk Screening MRI and Future Breast Cancer Risk. Acad Radiol 2019; 26:69-75. [PMID: 29602724 DOI: 10.1016/j.acra.2018.03.013] [Citation(s) in RCA: 30] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2018] [Revised: 02/24/2018] [Accepted: 03/09/2018] [Indexed: 01/10/2023]
Abstract
RATIONALE AND OBJECTIVES To determine if background parenchymal enhancement (BPE) on screening breast magnetic resonance imaging (MRI) in high-risk women correlates with future cancer. MATERIALS AND METHODS All screening breast MRIs (n = 1039) in high-risk women at our institution from August 1, 2004, to July 30, 2013, were identified. Sixty-one patients who subsequently developed breast cancer were matched 1:2 by age and high-risk indication with patients who did not develop breast cancer (n = 122). Five fellowship-trained breast radiologists independently recorded the BPE. The median reader BPE for each case was calculated and compared between the cancer and control cohorts. RESULTS Cancer cohort patients were high-risk because of a history of radiation therapy (10%, 6 of 61), high-risk lesion (18%, 11 of 61), or breast cancer (30%, 18 of 61); BRCA mutation (18%, 11 of 61); or family history (25%, 15 of 61). Subsequent malignancies were invasive ductal carcinoma (64%, 39 of 61), ductal carcinoma in situ (30%, 18 of 61) and invasive lobular carcinoma (7%, 4of 61). BPE was significantly higher in the cancer cohort than in the control cohort (P = 0.01). Women with mild, moderate, or marked BPE were 2.5 times more likely to develop breast cancer than women with minimal BPE (odds ratio = 2.5, 95% confidence interval: 1.3-4.8, P = .005). There was fair interreader agreement (κ = 0.39). CONCLUSIONS High-risk women with greater than minimal BPE at screening MRI have increased risk of future breast cancer.
Collapse
Affiliation(s)
- Lars J Grimm
- Department of Radiology, Duke University Medical Center, Box 3808, Durham, NC 27710.
| | - Ashirbani Saha
- Carl E. Ravin Advanced Imaging Laboratories, Duke University Hock Plaza, Durham, North Carolina
| | - Sujata V Ghate
- Department of Radiology, Duke University Medical Center, Box 3808, Durham, NC 27710
| | - Connie Kim
- Department of Radiology, Duke University Medical Center, Box 3808, Durham, NC 27710
| | - Mary Scott Soo
- Department of Radiology, Duke University Medical Center, Box 3808, Durham, NC 27710
| | - Sora C Yoon
- Department of Radiology, Duke University Medical Center, Box 3808, Durham, NC 27710
| | - Maciej A Mazurowski
- Carl E. Ravin Advanced Imaging Laboratories, Duke University Hock Plaza, Durham, North Carolina
| |
Collapse
|
42
|
Saha A, Harowicz MR, Cain EH, Hall AH, Hwang ESS, Marks JR, Marcom PK, Mazurowski MA. Intra-tumor molecular heterogeneity in breast cancer: definitions of measures and association with distant recurrence-free survival. Breast Cancer Res Treat 2018; 172:123-132. [PMID: 29992418 PMCID: PMC6588400 DOI: 10.1007/s10549-018-4879-7] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2017] [Accepted: 07/05/2018] [Indexed: 01/22/2023]
Abstract
PURPOSE The purpose of the study was to define quantitative measures of intra-tumor heterogeneity in breast cancer based on histopathology data gathered from multiple samples on individual patients and determine their association with distant recurrence-free survival (DRFS). METHODS We collected data from 971 invasive breast cancers, from 1st January 2000 to 23rd March 2014, that underwent repeat tumor sampling at our institution. We defined and calculated 31 measures of intra-tumor heterogeneity including ER, PR, and HER2 immunohistochemistry (IHC), proliferation, EGFR IHC, grade, and histology. For each heterogeneity measure, Cox proportional hazards models were used to determine whether patients with heterogeneous disease had different distant recurrence-free survival (DRFS) than those with homogeneous disease. RESULTS The presence of heterogeneity in ER percentage staining was prognostic of reduced DRFS with a hazard ratio of 4.26 (95% CI 2.22-8.18, p < 0.00002). It remained significant after controlling for the ER status itself (p < 0.00062) and for patients that had chemotherapy (p < 0.00032). Most of the heterogeneity measures did not show any association with DRFS despite the considerable sample size. CONCLUSIONS Intra-tumor heterogeneity of ER receptor status may be a predictor of patient DRFS. Histopathologic data from multiple tissue samples may offer a view of tumor heterogeneity and assess recurrence risk.
Collapse
Affiliation(s)
- Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, Durham, NC, 27705, USA.
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA.
| | - Michael R Harowicz
- Department of Radiology, Duke University School of Medicine, Durham, NC, 27705, USA
| | - Elizabeth Hope Cain
- Department of Radiology, Duke University School of Medicine, Durham, NC, 27705, USA
| | - Allison H Hall
- Department of Pathology, Duke University School of Medicine, Durham, NC, 27710, USA
| | - Eun-Sil Shelley Hwang
- Department of Surgical Oncology, Duke University School of Medicine, Durham, NC, 27710, USA
| | - Jeffrey R Marks
- Department of Surgery, Duke University School of Medicine, Durham, NC, 27710, USA
| | - Paul Kelly Marcom
- Department of Medicine, Duke University School of Medicine, Durham, NC, 27710, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, NC, 27705, USA
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, 27705, USA
- Duke University Medical Physics Program, Durham, NC, 27705, USA
| |
Collapse
|
43
|
Cain EH, Saha A, Harowicz MR, Marks JR, Marcom PK, Mazurowski MA. Multivariate machine learning models for prediction of pathologic response to neoadjuvant therapy in breast cancer using MRI features: a study using an independent validation set. Breast Cancer Res Treat 2018; 173:455-463. [PMID: 30328048 DOI: 10.1007/s10549-018-4990-9] [Citation(s) in RCA: 93] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2018] [Accepted: 10/01/2018] [Indexed: 02/07/2023]
Abstract
PURPOSE To determine whether a multivariate machine learning-based model using computer-extracted features of pre-treatment dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI) can predict pathologic complete response (pCR) to neoadjuvant therapy (NAT) in breast cancer patients. METHODS Institutional review board approval was obtained for this retrospective study of 288 breast cancer patients at our institution who received NAT and had a pre-treatment breast MRI. A comprehensive set of 529 radiomic features was extracted from each patient's pre-treatment MRI. The patients were divided into equal groups to form a training set and an independent test set. Two multivariate machine learning models (logistic regression and a support vector machine) based on imaging features were trained to predict pCR in (a) all patients with NAT, (b) patients with neoadjuvant chemotherapy (NACT), and (c) triple-negative or human epidermal growth factor receptor 2-positive (TN/HER2+) patients who had NAT. The multivariate models were tested using the independent test set, and the area under the receiver operating characteristics (ROC) curve (AUC) was calculated. RESULTS Out of the 288 patients, 64 achieved pCR. The AUC values for predicting pCR in TN/HER+ patients who received NAT were significant (0.707, 95% CI 0.582-0.833, p < 0.002). CONCLUSIONS The multivariate models based on pre-treatment MRI features were able to predict pCR in TN/HER2+ patients.
Collapse
Affiliation(s)
- Elizabeth Hope Cain
- Department of Radiology, Duke University School of Medicine, 2301 Erwin Road, Durham, NC, 27705, USA.
| | - Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, 2301 Erwin Road, Durham, NC, 27705, USA
| | - Michael R Harowicz
- Department of Radiology, Duke University School of Medicine, 2301 Erwin Road, Durham, NC, 27705, USA.,Department of Radiology, Johns Hopkins University, Baltimore, MD, USA
| | - Jeffrey R Marks
- Department of Surgery, Duke University School of Medicine, 2301 Erwin Road, Durham, NC, 27705, USA
| | - P Kelly Marcom
- Department of Medicine, Duke University School of Medicine, 2301 Erwin Road, Durham, NC, 27705, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, 2301 Erwin Road, Durham, NC, 27705, USA.,Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA
| |
Collapse
|
44
|
Wollin DA, Gupta RT, Young B, Cone E, Kaplan A, Marin D, Patel BN, Mazurowski MA, Scales CD, Ferrandino MN, Preminger GM, Lipkin ME. Abdominal Radiography With Digital Tomosynthesis: An Alternative to Computed Tomography for Identification of Urinary Calculi? Urology 2018; 120:56-61. [PMID: 30006268 DOI: 10.1016/j.urology.2018.06.041] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2018] [Revised: 06/28/2018] [Accepted: 06/29/2018] [Indexed: 11/15/2022]
Abstract
OBJECTIVE To compare the accuracy of plain abdominal radiography (kidneys, ureter, and bladder [KUB]) with digital tomosynthesis (DT) to noncontrast computed tomography (NCCT), the gold standard imaging modality for urinary stones. Due to radiation and cost concerns, KUB is often used for diagnosis and follow-up of nephrolithiasis. DT, a novel technique that produces high-quality radiographs with less radiation and/or cost than low-dose NCCT, has not been assessed in this situation. MATERIALS AND METHODS Seven fresh tissue cadavers were implanted with stones of known size and/or composition and imaged with KUB, DT, and NCCT. Four blinded readers (2 urologists, 2 radiologists) evaluated KUBs for presence and/or location of calculi. They then re-evaluated with addition of tomograms to assess additional value. After a memory extinction period, readers evaluated NCCT images. Accuracy of detection was determined using nearest-neighbor match with generalized linear mixed modeling. RESULTS Total of 59 stones were identified on reference read. Overall, NCCT and DT were both superior to KUB alone (P < .001) while the difference between DT and NCCT was not significant (P = .06). When evaluating uric acid stones, NCCT and DT outperformed KUB (P < .01 and P < .05, respectively) while DT and NCCT were similar (P = .16). Intrarenal stones were better evaluated on DT and NCCT (P < .001 compared to KUB), while DT and NCCT were similar (P = 1.00). Accuracy was lower than anticipated across modalities due to use of the cadaver model. CONCLUSION Our study demonstrates DT is superior to KUB for identification of intrarenal calculi and could replace routine use of KUB or NCCT for detecting renal stones, even those composed of uric acid.
Collapse
Affiliation(s)
- Daniel A Wollin
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC.
| | - Rajan T Gupta
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC; Department of Radiology, Duke University Medical Center, Durham, NC
| | - Brian Young
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC
| | - Eugene Cone
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC
| | - Adam Kaplan
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC
| | - Daniele Marin
- Department of Radiology, Duke University Medical Center, Durham, NC
| | - Bhavik N Patel
- Department of Radiology, Duke University Medical Center, Durham, NC; Department of Radiology, Stanford University, Palo Alto, CA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University Medical Center, Durham, NC; Department of Electrical and Computer Engineering, Duke University, Durham, NC
| | - Charles D Scales
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC
| | | | - Glenn M Preminger
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC
| | - Michael E Lipkin
- Division of Urologic Surgery, Duke University Medical Center, Durham, NC
| |
Collapse
|
45
|
Saha A, Harowicz MR, Mazurowski MA. Breast cancer MRI radiomics: An overview of algorithmic features and impact of inter-reader variability in annotating tumors. Med Phys 2018; 45:3076-3085. [PMID: 29663411 DOI: 10.1002/mp.12925] [Citation(s) in RCA: 44] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2017] [Revised: 03/01/2018] [Accepted: 04/04/2018] [Indexed: 12/30/2022] Open
Abstract
PURPOSE To review features used in MRI radiomics of breast cancer and study the inter-reader stability of the features. METHODS We implemented 529 algorithmic features that can be extracted from tumor and fibroglandular tissue (FGT) in breast MRIs. The features were identified based on a review of the existing literature with consideration of their usage, prognostic ability, and uniqueness. The set was then extended so that it comprehensively describes breast cancer imaging characteristics. The features were classified into 10 groups based on the type of data used to extract them and the type of calculation being performed. For the assessment of inter-reader variability, four fellowship-trained readers annotated tumors on preoperative dynamic contrast-enhanced MRIs for 50 breast cancer patients. Based on the annotations, an algorithm automatically segmented the image and extracted all features resulting in one set of features for each reader. For a given feature, the inter-reader stability was defined as the intraclass correlation coefficient (ICC) computed using the feature values obtained through all readers for all cases. RESULTS The average inter-reader stability for all features was 0.8474 (95% CI: 0.8068-0.8858). The mean inter-reader stability was lower for tumor-based features (0.6348, 95% CI: 0.5391-0.7257) than FGT-based features (0.9984, 95% CI: 0.9970-0.9992). The feature group with the highest inter-reader stability quantifies breast and FGT volume. The feature group with the lowest inter-reader stability quantifies variations in tumor enhancement. CONCLUSIONS Breast MRI radiomics features widely vary in terms of their stability in the presence of inter-reader variability. Appropriate measures need to be taken for reducing this variability in tumor-based radiomics.
Collapse
Affiliation(s)
- Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA
| | - Michael R Harowicz
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA.,Department of Electrical and Computer Engineering, Duke University, Box 90291, Durham, NC, 27708, USA.,Duke University Medical Physics Program, DUMC 2729, 2424 Erwin Road, Suite 101, Durham, NC, 27705, USA
| |
Collapse
|
46
|
Saha A, Harowicz MR, Wang W, Mazurowski MA. A study of association of Oncotype DX recurrence score with DCE-MRI characteristics using multivariate machine learning models. J Cancer Res Clin Oncol 2018; 144:799-807. [PMID: 29427210 PMCID: PMC5920720 DOI: 10.1007/s00432-018-2595-7] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2017] [Accepted: 01/23/2018] [Indexed: 01/09/2023]
Abstract
PURPOSE To determine whether multivariate machine learning models of algorithmically assessed magnetic resonance imaging (MRI) features from breast cancer patients are associated with Oncotype DX (ODX) test recurrence scores. METHODS A set of 261 female patients with invasive breast cancer, pre-operative dynamic contrast enhanced magnetic resonance (DCE-MR) images and available ODX score at our institution was identified. A computer algorithm extracted a comprehensive set of 529 features from the DCE-MR images of these patients. The set of patients was divided into a training set and a test set. Using the training set we developed two machine learning-based models to discriminate (1) high ODX scores from intermediate and low ODX scores, and (2) high and intermediate ODX scores from low ODX scores. The performance of these models was evaluated on the independent test set. RESULTS High against low and intermediate ODX scores were predicted by the multivariate model with AUC 0.77 (95% CI 0.56-0.98, p < 0.003). Low against intermediate and high ODX score was predicted with AUC 0.51 (95% CI 0.41-0.61, p = 0.75). CONCLUSION A moderate association between imaging and ODX score was identified. The evaluated models currently do not warrant replacement of ODX with imaging alone.
Collapse
Affiliation(s)
- Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA.
| | - Michael R Harowicz
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA
| | - Weiyao Wang
- Department of Mathematics, Duke University, 120 Science Drive, 117 Physics Building, Durham, NC, 27708, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, 2424 Erwin Road, Suite 302, Durham, NC, 27705, USA
- Department of Electrical and Computer Engineering, Duke University, Box 90291, Durham, NC, 27708, USA
- Duke University Medical Physics Graduate Program, 2424 Erwin Road, Suite 101, Durham, NC, 27705, USA
| |
Collapse
|
47
|
Shi B, Grimm LJ, Mazurowski MA, Baker JA, Marks JR, King LM, Maley CC, Hwang ES, Lo JY. Prediction of Occult Invasive Disease in Ductal Carcinoma in Situ Using Deep Learning Features. J Am Coll Radiol 2018; 15:527-534. [PMID: 29398498 PMCID: PMC5837927 DOI: 10.1016/j.jacr.2017.11.036] [Citation(s) in RCA: 45] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2017] [Accepted: 11/27/2017] [Indexed: 01/23/2023]
Abstract
PURPOSE The aim of this study was to determine whether deep features extracted from digital mammograms using a pretrained deep convolutional neural network are prognostic of occult invasive disease for patients with ductal carcinoma in situ (DCIS) on core needle biopsy. METHODS In this retrospective study, digital mammographic magnification views were collected for 99 subjects with DCIS at biopsy, 25 of which were subsequently upstaged to invasive cancer. A deep convolutional neural network model that was pretrained on nonmedical images (eg, animals, plants, instruments) was used as the feature extractor. Through a statistical pooling strategy, deep features were extracted at different levels of convolutional layers from the lesion areas, without sacrificing the original resolution or distorting the underlying topology. A multivariate classifier was then trained to predict which tumors contain occult invasive disease. This was compared with the performance of traditional "handcrafted" computer vision (CV) features previously developed specifically to assess mammographic calcifications. The generalization performance was assessed using Monte Carlo cross-validation and receiver operating characteristic curve analysis. RESULTS Deep features were able to distinguish DCIS with occult invasion from pure DCIS, with an area under the receiver operating characteristic curve of 0.70 (95% confidence interval, 0.68-0.73). This performance was comparable with the handcrafted CV features (area under the curve = 0.68; 95% confidence interval, 0.66-0.71) that were designed with prior domain knowledge. CONCLUSIONS Despite being pretrained on only nonmedical images, the deep features extracted from digital mammograms demonstrated comparable performance with handcrafted CV features for the challenging task of predicting DCIS upstaging.
Collapse
Affiliation(s)
- Bibo Shi
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University School of Medicine, Durham, North Carolina.
| | - Lars J Grimm
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University School of Medicine, Durham, North Carolina
| | - Maciej A Mazurowski
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University School of Medicine, Durham, North Carolina
| | - Jay A Baker
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University School of Medicine, Durham, North Carolina
| | - Jeffrey R Marks
- Department of Surgery, Duke University School of Medicine, Durham, North Carolina
| | - Lorraine M King
- Department of Surgery, Duke University School of Medicine, Durham, North Carolina
| | - Carlo C Maley
- Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, Arizona; Centre for Evolution and Cancer, Institute of Cancer Research, London, United Kingdom
| | - E Shelley Hwang
- Department of Surgery, Duke University School of Medicine, Durham, North Carolina
| | - Joseph Y Lo
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University School of Medicine, Durham, North Carolina
| |
Collapse
|
48
|
Lo JY, Grimm LJ, Mazurowski MA, Baker JA, Marks JR, King LM, Maley CC, Hwang ESS, Shi B. Abstract GS5-04: Prediction of occult invasive disease in ductal carcinoma in situ using deep learning features. Cancer Res 2018. [DOI: 10.1158/1538-7445.sabcs17-gs5-04] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Abstract
Background: Deep learning, especially deep convolutional neural network (CNN), has emerged as a promising approach for many image recognition or classification tasks, demonstrating human or even superhuman performance. Used as feature extractor, some pre-trained CNN models can match or surpass the performance of domain-specific, “handcrafted” features. In this study, we aim to determine whether deep features extracted from digital mammograms using a pre-trained deep CNN are prognostic of occult invasive disease for patients with ductal carcinoma in situ (DCIS) on core needle biopsy.
Materials and Methods: In this retrospective study, we collected digital mammography magnification views for 99 subjects with DCIS at biopsy, 25 of which were subsequently upstaged to invasive cancer. We utilized a deep CNN model that was pre-trained on non-medical images (e.g., animals, plants, instruments) as the feature extractor. Through a statistical pooling strategy, we extracted deep features at different levels of convolutional layers from the lesion areas, without sacrificing the original resolution or distorting the underlying topology. A multivariate classifier was then trained to predict which tumors contain occult invasive disease. This was compared to the performance of traditional “handcrafted” computer vision (CV) features previously developed specifically to assess mammographic calcifications. The generalization performance was assessed using Monte Carlo cross validation and receiver operating characteristic (ROC) curve analysis.
Results: Deep features were able to distinguish DCIS with occult invasion from pure DCIS, with an area under the ROC curve (AUC-ROC) equal to 0.70 (95% CI: 0.68-0.73). This performance was comparable to the "handcrafted" CV features (AUC-ROC = 0.68, 95% CI: 0.66-0.71) that were designed with prior domain knowledge.
Conclusion: In spite of being pre-trained on only non-medical images, the deep features extracted from digital mammograms demonstrated comparable performance to "handcrafted" CV features for the challenging task of predicting DCIS upstaging.
Acknowledgments: This work was supported in part by NIH/NCI R01-CQA185138 and DOD Breast Cancer Research Program W81XWH-14-1-0473.
Citation Format: Lo JY, Grimm LJ, Mazurowski MA, Baker JA, Marks JR, King LM, Maley CC, Hwang E-SS, Shi B. Prediction of occult invasive disease in ductal carcinoma in situ using deep learning features [abstract]. In: Proceedings of the 2017 San Antonio Breast Cancer Symposium; 2017 Dec 5-9; San Antonio, TX. Philadelphia (PA): AACR; Cancer Res 2018;78(4 Suppl):Abstract nr GS5-04.
Collapse
Affiliation(s)
- JY Lo
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - LJ Grimm
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - MA Mazurowski
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - JA Baker
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - JR Marks
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - LM King
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - CC Maley
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - E-SS Hwang
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| | - B Shi
- Carl E. Ravin Advanced Imaging Laboratories, Duke University School of Medicine, Durham, NC; Duke University School of Medicine, Durham, NC; Biodesign Center for Personalized Diagnostics and School of Life Sciences, Arizona State University, Tempe, AZ
| |
Collapse
|
49
|
AlBadawy EA, Saha A, Mazurowski MA. Deep learning for segmentation of brain tumors: Impact of cross-institutional training and testing. Med Phys 2018; 45:1150-1158. [PMID: 29356028 DOI: 10.1002/mp.12752] [Citation(s) in RCA: 93] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2017] [Revised: 12/13/2017] [Accepted: 12/14/2017] [Indexed: 12/31/2022] Open
Abstract
BACKGROUND AND PURPOSE Convolutional neural networks (CNNs) are commonly used for segmentation of brain tumors. In this work, we assess the effect of cross-institutional training on the performance of CNNs. METHODS We selected 44 glioblastoma (GBM) patients from two institutions in The Cancer Imaging Archive dataset. The images were manually annotated by outlining each tumor component to form ground truth. To automatically segment the tumors in each patient, we trained three CNNs: (a) one using data for patients from the same institution as the test data, (b) one using data for the patients from the other institution and (c) one using data for the patients from both of the institutions. The performance of the trained models was evaluated using Dice similarity coefficients as well as Average Hausdorff Distance between the ground truth and automatic segmentations. The 10-fold cross-validation scheme was used to compare the performance of different approaches. RESULTS Performance of the model significantly decreased (P < 0.0001) when it was trained on data from a different institution (dice coefficients: 0.68 ± 0.19 and 0.59 ± 0.19) as compared to training with data from the same institution (dice coefficients: 0.72 ± 0.17 and 0.76 ± 0.12). This trend persisted for segmentation of the entire tumor as well as its individual components. CONCLUSIONS There is a very strong effect of selecting data for training on performance of CNNs in a multi-institutional setting. Determination of the reasons behind this effect requires additional comprehensive investigation.
Collapse
Affiliation(s)
- Ehab A AlBadawy
- Department of Radiology, Duke University School of Medicine, Durham, NC, USA
| | - Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, Durham, NC, USA
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, NC, USA.,Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA.,Duke University Medical Physics Program, Durham, NC, USA
| |
Collapse
|
50
|
Abstract
PURPOSE To assess the impact of varying magnetic resonance imaging (MRI) scanner parameters on the extraction of algorithmic features in breast MRI radiomics studies. METHODS In this retrospective study, breast imaging data for 272 patients were analyzed with magnetic resonance (MR) images. From the MR images, we assembled and implemented 529 algorithmic features of breast tumors and fibrograndular tissue (FGT). We divided the features into 10 groups based on the type of data used for the feature extraction and the nature of the extracted information. Three scanner parameters were considered: scanner manufacturer, scanner magnetic field strength, and slice thickness. We assessed the impact of each of the scanner parameters on each of the feature by testing whether the feature values are systematically diverse for different values of these scanner parameters. A two-sample t-test has been used to establish whether the impact of a scanner parameter on values of a feature is significant and receiver operating characteristics have been used for to establish the extent of that effect. RESULTS On average, higher proportion (69% FGT versus 20% tumor) of FGT related features were affected by the three scanner parameters. Of all feature groups and scanner parameters, the feature group related to the variation in FGT enhancement was found to be the most sensitive to the scanner manufacturer (AUC = 0.81 ± 0.14). CONCLUSIONS Features involving calculations from FGT are particularly sensitive to the scanner parameters.
Collapse
Affiliation(s)
- Ashirbani Saha
- Department of Radiology, Duke University School of Medicine, Duke University, Durham, NC, USA
| | - Xiaozhi Yu
- Department of Radiology, Duke University School of Medicine, Duke University, Durham, NC, USA
| | - Dushyant Sahoo
- Department of Radiology, Duke University School of Medicine, Duke University, Durham, NC, USA
| | - Maciej A. Mazurowski
- Department of Radiology, Duke University School of Medicine, Duke University, Durham, NC, USA
- Department of Electrical and Computer Engineering, Duke University, Durham, NC, USA
- Duke University Medical Physics Program, Durham, NC, USA
| |
Collapse
|